Examples: query, "exact match", wildcard*, wild?ard, wild*rd
Fuzzy search: cake~ (finds cakes, bake)
Term boost: "red velvet"^4, chocolate^2
Field grouping: tags:(+work -"fun-stuff")
Escaping: Escape characters +-&|!(){}[]^"~*?:\ with \, e.g. \+
Range search: properties.timestamp:[1587729413488 TO *] (inclusive), properties.title:{A TO Z}(excluding A and Z)
Combinations: chocolate AND vanilla, chocolate OR vanilla, (chocolate OR vanilla) NOT "vanilla pudding"
Field search: properties.title:"The Title" AND text
Profile picture
NuttyLobster9
Moderator
9 Questions, 38 Answers
  Active since 07 February 2023
  Last activity one month ago

Reputation

0

Badges 1

33 × Eureka!
0 Votes
3 Answers
514 Views
0 Votes 3 Answers 514 Views
I’m wondering if someone has an example of how to use the base_task_factory option for adding a step to a PipelineController ? It’s not clear to me how we sh...
7 months ago
0 Votes
6 Answers
688 Views
0 Votes 6 Answers 688 Views
I know at least one other person has posted about this previously, but when I interact with Datasets I keep getting a bunch of repeated logging messages with...
one year ago
0 Votes
10 Answers
153 Views
0 Votes 10 Answers 153 Views
I’m trying to understand the execution flow of pipelines when translating from local to remote execution. I’ve defined a pipeline using the PipelineDecorator...
one month ago
0 Votes
8 Answers
141 Views
0 Votes 8 Answers 141 Views
one month ago
0 Votes
4 Answers
147 Views
0 Votes 4 Answers 147 Views
Hi All, I’m curious if anyone has an example of editing the configuration of a running TaskScheduler or TriggerScheduler . If I understand correctly the desc...
one month ago
0 Votes
9 Answers
624 Views
0 Votes 9 Answers 624 Views
11 months ago
0 Votes
6 Answers
205 Views
0 Votes 6 Answers 205 Views
Hi All, is there a way to clone a pipeline from the web UI like you can with a task? The goal is to be able to give the cloned version a different name so I ...
3 months ago
0 Votes
4 Answers
179 Views
0 Votes 4 Answers 179 Views
2 months ago
0 Votes
15 Answers
119 Views
0 Votes 15 Answers 119 Views
Are there any particular system dependencies needed to enable auto_resource_logging ? My understanding based on None is that this should be enabled by defaul...
one month ago
0 I Know At Least One Other Person Has Posted About This Previously, But When I Interact With

I’m using SDK version 1.10.2 and yes, it’s self-hosted. Here is the version info for the server:

 WebApp: 1.9.1-312 • Server: 1.9.1-312 • API: 2.23 

Thanks!

one year ago
one year ago
0 I’M Trying To Understand The Execution Flow Of Pipelines When Translating From Local To Remote Execution. I’Ve Defined A Pipeline Using The

The result i get in the agent is:


Traceback (most recent call last):
  File "src/clearml_pipelines_examples/pipelines/examples/train_model_on_random_data/pipeline.py", line 89, in <module>
    pipeline(**pipeline_ui_config)
TypeError: 'NoneType' object is not callable

Seems like the call to pipeline = PipelineDecorator.get_current_pipeline() returns None . Also, in the UI, I should be seeing all of the pipeline function parameters but I only see the config_file_path

one month ago
0 I’M Trying To Understand The Execution Flow Of Pipelines When Translating From Local To Remote Execution. I’Ve Defined A Pipeline Using The

Hi @<1523701435869433856:profile|SmugDolphin23> , so I need to call the pipeline function again in the remote context? I guess I thought when I start it up, my local session parses the pipeline and then transmits it to the server to run but it sounds like, it just copies the code and then i need to effectively call it again in the agent?

one month ago
0 I’M Trying To Understand The Execution Flow Of Pipelines When Translating From Local To Remote Execution. I’Ve Defined A Pipeline Using The

Gotcha. Are the parameters in @PipelineDecorator.pipeline() ignored in the remote context? Settings like the docker image and gitlab repo would already be used before the pipeline is kicked off on the agent.

one month ago
0 Hi All! Question About Pipelines Using Decorators. The First Step Of My Pipeline Uses A Specific Repo, Specified Using

@<1523701225533476864:profile|ObedientDolphin41> , I was searching for anyone having an issue like me and found this thread. I have created a simple pipeline using decorators and when I try to clone it in the UI, I get that base_task_id is empty error. It works fine when triggered programmatically from my machine. I’m wondering if you could elaborate on how you utilized the get_configuration_object and set_configuration_object methods to solve this? In my case, I’m not setting a...

one year ago
0 Hi Everyone, I'M Running Into A Weird Error When Trying To Clone And Run And Task That Has Completed Successfully. I Have A Test Task That Loads A Dummy Dataset And Trains A Toy Model With Pytorch. When Running Remotely, I Use My Own Docker Image That Has

Okay so I discovered that setting -e CLEARML_AGENT_PACKAGE_PYTORCH_RESOLVE=none solves the issue.

That said, if someone could explain to me why this error was occurring and why it only happens in the case of cloning, I'd love to understand. Thanks!

one month ago
one month ago
0 Hi Everyone, I'M Running Into A Weird Error When Trying To Clone And Run And Task That Has Completed Successfully. I Have A Test Task That Loads A Dummy Dataset And Trains A Toy Model With Pytorch. When Running Remotely, I Use My Own Docker Image That Has

Hi Martin, I see . That makes sense though I would have expected the behavior to be the same when running remotely the first time as well . In any case, this solved the issue for me . Thanks for looking at it

one month ago
0 Are There Any Particular System Dependencies Needed To Enable

Sure. I can send it on Monday. Thank you.

one month ago
0 Are There Any Particular System Dependencies Needed To Enable

Hi @<1523701205467926528:profile|AgitatedDove14> , I've actually hit on something accidentally that might be a clue. I have noticed that when running inside an agent, there is a bug wherein both Task.current_task() and Logger.current_logger() return None . If these are being used by the clearml package under the hood, this could be the reason we aren't seeing the metrics.

As a workaround, I created this utility function, which works for explicit logging (though it doesn't c...

one month ago
0 I’M Trying To Understand The Execution Flow Of Pipelines When Translating From Local To Remote Execution. I’Ve Defined A Pipeline Using The

Okay well I have to supply them again for the function to work, but the values are ignored so i can just have a hard-coded version for remote.

I am still struggling to figure out how to update the parameter defaults, though. I would like to be able to do the equivalent of the PipelineController.add_parameter() so that I can supply a local config with new defaults that are used on the remote execution. Otherwise, I’m stuck with whatever defaults are in the function signature.

one month ago
0 I’M Trying To Understand The Execution Flow Of Pipelines When Translating From Local To Remote Execution. I’Ve Defined A Pipeline Using The

Are those fixed from the local environment or do i need to also supply those again in the remote context?

one month ago
0 Are There Any Particular System Dependencies Needed To Enable

Hi @<1523701205467926528:profile|AgitatedDove14> , CLEARML_TASK_ID is set inside the agent's process, which is how I was able to get the task by running Task.get_task(environ["CLEARML_TASK_ID") . However I believe I've sorted out how to make both the resource logging and the tensorboard logging work in the agent. It seems that using Task.current_task() to get the task object does not work when running remotely, but calling Task.init() again does work. And after having called ...

one month ago
0 Hi All! Is There Any Simple Way To Use

Okay, I take it back. os.getenv("CLEARML_TASK_ID") does work. I forgot to rebuild my container after making the change. Thanks for bringing this option to my attention!

one year ago
0 Hello, All. I’Ve Recently Started Experiencing A Weird Issue With Arg Parsing Where Any String Values Are Being Repeated As Lists Of Strings When The Values Are Sent To The Clearml Server (See Attached Screenshot). I Believe This Issue Started Around The

Hi @<1523701070390366208:profile|CostlyOstrich36> , I would expect the loss_func parameter to be FocalLoss instead of ['FocalLoss', 'FocalLoss', 'FocalLoss', 'FocalLoss'] (and same for the validation_split_name parameter. I will try to put together an example, though it might take a little time before I can do it.

11 months ago
0 I’M Trying To Understand The Execution Flow Of Pipelines When Translating From Local To Remote Execution. I’Ve Defined A Pipeline Using The

Sorry, i meant the arguments that are supplied to the decorator method, itself @PipelineDecorator.pipeline() and @PipelineDecorator.component() , things like name , project , docker_args , etc.

one month ago
0 Hi Everyone, I'M Running Into A Weird Error When Trying To Clone And Run And Task That Has Completed Successfully. I Have A Test Task That Loads A Dummy Dataset And Trains A Toy Model With Pytorch. When Running Remotely, I Use My Own Docker Image That Has

Unfortunately, it's turning out to be quite time consuming to manually remove all of the private info in here. Is there a particular section of the log that would be useful to see? I can try to focus on just sharing that part.

one month ago
0 Hi All! Question About Pipelines Using Decorators. The First Step Of My Pipeline Uses A Specific Repo, Specified Using

Hi Max, thanks very much for your message! I understand what you’re saying now, though I suppose this is not my issue since I’m not setting any of the decorator values with variables. I’ll post a query in the main channel with code snippets to see if anyone has ideas. Thank you!

one year ago
0 Hi All, I’M Curious If Anyone Has An Example Of Editing The Configuration Of A Running

Ahhh okay, thank you. Perhaps in the future, it would be great to allow this from the UI as well?

one month ago
0 Are There Any Particular System Dependencies Needed To Enable

Here's my example script:

from random import randint

from clearml import Task

if __name__ == "__main__":
    task: Task = Task.init(
        project_name="clearml-examples", task_name="try-to-make-logging-work"
    )
    task.execute_remotely(queue_name="5da90f42dd4c40edab972a4bef8eab04")
    logger = task.get_logger()
    for i in range(10):
        logger.report_scalar("example plot", series="random", value=randint(0, 100), iteration=i)
one month ago
0 Are There Any Particular System Dependencies Needed To Enable

Ah interesting, okay. I'll try adding a sleep in here for testing it out. Thanks

one month ago
0 Hi All, I’M Curious If Anyone Has An Example Of Editing The Configuration Of A Running

That could happen with any task when it’s cloned. To be honest, the cron and trigger schedulers probably deserve their own UI panel since they operate differently than other tasks. Ideally, a user would be able to add and remove jobs from the schedulers purely through the UI.

one month ago
0 Are There Any Particular System Dependencies Needed To Enable

Yes, that did make it work in this case, thank you.

one month ago
0 Hi All, Is There A Way To Clone A Pipeline From The Web Ui Like You Can With A Task? The Goal Is To Be Able To Give The Cloned Version A Different Name So I Can Organize Pipeline Runs By Project.

Hi @<1523701205467926528:profile|AgitatedDove14> , sorry for the delayed reply. So what you’re saying is to first kick off a new run and then rename the underlying Pipeline Task, which will cause that particular run to become a new pipeline name? But you have to do this only after you’ve started the run.

What would be most ideal would be to be able to right-click on a pipeline run and have a “clone” option, like you can with a task, where you can start a new run with a new name in a single ...

2 months ago
0 I’M Wondering If Someone Has An Example Of How To Use The

Hi @<1523701205467926528:profile|AgitatedDove14> , thanks so the code to be executed by the task needs to be provided to the Task.create() method as script=some/path.py or does it work to have something like

def my_node_task_factory(node: PipelineController.Node) -> Task:
  task = Task.create(...)
  my_function()
  return task 
7 months ago
0 I’M Trying To Understand The Execution Flow Of Pipelines When Translating From Local To Remote Execution. I’Ve Defined A Pipeline Using The
import json
import os
import sys
from argparse import ArgumentParser
from logging import getLogger
from pathlib import Path
from typing import Callable

from clearml import PipelineDecorator, Task

from clearml_pipelines_examples.base.pipeline_settings import ExecutionMode
from clearml_pipelines_examples.pipelines.examples.train_model_on_random_data import (
    TrainModelPipelineKwargs,
    TrainModelPipelineSettings,
)
from clearml_pipelines_examples.tasks.examples import generate_dat...
one month ago
Show more results compactanswers