Examples: query, "exact match", wildcard*, wild?ard, wild*rd
Fuzzy search: cake~ (finds cakes, bake)
Term boost: "red velvet"^4, chocolate^2
Field grouping: tags:(+work -"fun-stuff")
Escaping: Escape characters +-&|!(){}[]^"~*?:\ with \, e.g. \+
Range search: properties.timestamp:[1587729413488 TO *] (inclusive), properties.title:{A TO Z}(excluding A and Z)
Combinations: chocolate AND vanilla, chocolate OR vanilla, (chocolate OR vanilla) NOT "vanilla pudding"
Field search: properties.title:"The Title" AND text
Profile picture
DangerousDragonfly8
Moderator
13 Questions, 40 Answers
  Active since 10 January 2023
  Last activity one year ago

Reputation

0

Badges 1

38 × Eureka!
0 Votes
7 Answers
1K Views
0 Votes 7 Answers 1K Views
Hello. Is it possible to show hidden projects in the UI? Like the ".pipelines" created by pipelines? Alternatively is it possible to place the pipeline tasks...
2 years ago
0 Votes
5 Answers
1K Views
0 Votes 5 Answers 1K Views
Hello. I have a question regarding pipeline parameters. Is it possible to reference pipeline parameters in other fields of the https://clear.ml/docs/latest/d...
one year ago
0 Votes
0 Answers
1K Views
0 Votes 0 Answers 1K Views
one year ago
0 Votes
7 Answers
1K Views
0 Votes 7 Answers 1K Views
one year ago
0 Votes
6 Answers
1K Views
0 Votes 6 Answers 1K Views
one year ago
0 Votes
3 Answers
942 Views
0 Votes 3 Answers 942 Views
Hello! When I use the TriggerScheduler class with the add_task_trigger function configured to watch for trigger_on_status=['published'] and a specific trigge...
2 years ago
0 Votes
1 Answers
977 Views
0 Votes 1 Answers 977 Views
Hello. When I use https://clear.ml/docs/latest/docs/references/sdk/model_outputmodel#update_weights_package with an S3 upload_uri, ClearML uploads a file cal...
one year ago
0 Votes
19 Answers
1K Views
0 Votes 19 Answers 1K Views
2 years ago
0 Votes
10 Answers
1K Views
0 Votes 10 Answers 1K Views
Can I use automation.TriggerScheduler() with add_task_trigger to trigger when a task is archived? I know that when a task is archived it gets "system_tags": ...
2 years ago
0 Votes
4 Answers
986 Views
0 Votes 4 Answers 986 Views
When using a TriggerScheduler with a add_task_trigger and schedule_function , how would I go about updating the trigger and function without a new task/exper...
2 years ago
0 Votes
1 Answers
1K Views
0 Votes 1 Answers 1K Views
2 years ago
0 Votes
4 Answers
1K Views
0 Votes 4 Answers 1K Views
Does anyone know if it is possible to add a plot from a stylized pandas data frame? I can easily log a pandas data frame with logger.report_table but can I l...
2 years ago
0 Votes
5 Answers
1K Views
0 Votes 5 Answers 1K Views
Hi. I will try again since I got no answers previously 😁
2 years ago
0 Hello. Just Re-Posting This Maybe I Can Get Come Additional Help. For Some Reason The Pipeline Node Object From My Pre_Execute_Callback Has Not Attributes Set Other Than The Name. Any Idea Why That Is The Case? The Docs Say That I Should Be Able To Modify

SmugDolphin23 ok so pipe.start with step_task_completed_callback does indeed work because step_task_completed_callback runs before the task is executed. step_task_created_callback seems to run after the task is executed however so the naming seems to be reversed.

one year ago
0 When Using A

AgitatedDove14
I do believe triggers should be unique somehow because I find them way too easy to mishandle. Especially if used with schedule_function which is defined in the same script. Updating that function requires deleting the existing trigger task first and recreating it. If not done like this you just end up with 2 trigger tasks with the same name which I assume will respond to the same event(s) but do something slightly different in response. I assume it might work like this...

2 years ago
0 Hello. I Have An Issue In Regards To A Task That I Run As A Service ( Should Always Run). I Run The Clearml Server And Agents In Kubernetes. I Think This Is A Design Problem With The Way Clearml Agents Run On Kubernetes. The K8S Glue Will Launch A Worker

Here is what I see as the ideal scenario:
If a worker pod running a task dies for any reason, clearml should mark the task as failed / aborted asap. Basically improve the feedback loop. Tasks running as services should be re-enqueued automatically if a the pod it runs on dies because of OOM, node eviction, node replacement, pod replacement because of autoscaling etc. You could argue the same for tasks which are not services. Restart them if their pod dies for the above reasons.

one year ago
0 Hello. I Have An Issue In Regards To A Task That I Run As A Service ( Should Always Run). I Run The Clearml Server And Agents In Kubernetes. I Think This Is A Design Problem With The Way Clearml Agents Run On Kubernetes. The K8S Glue Will Launch A Worker

Now for example the pod was killed because I had to replace the node. The task is stuck in "Running". Aborting from the UI says "experiment aborted successfully" but the state does not change.

one year ago
0 Hello. I Have An Issue In Regards To A Task That I Run As A Service ( Should Always Run). I Run The Clearml Server And Agents In Kubernetes. I Think This Is A Design Problem With The Way Clearml Agents Run On Kubernetes. The K8S Glue Will Launch A Worker

With pipelines is even more complicated because what I experienced is that the pod for step 2 was evicted because it was eating too much memory. So the pod has been terminated but the task was not marked as failed / aborted. Because of that, the pipeline controller pod was still running and the pipeline itself was also not marked as aborted / failed.

one year ago
0 Hello. I Have An Issue In Regards To A Task That I Run As A Service ( Should Always Run). I Run The Clearml Server And Agents In Kubernetes. I Think This Is A Design Problem With The Way Clearml Agents Run On Kubernetes. The K8S Glue Will Launch A Worker

I am trying to run with scale from zero k8s nodes for maximum cost savings. So a node should only be online if clearml actually runs a task. Waiting for the 2 hours timeout when running on expensive gpu instances for example is quite wasteful because the pipeline controller pod will keep the node online.

one year ago
Show more results compactanswers