Examples: query, "exact match", wildcard*, wild?ard, wild*rd
Fuzzy search: cake~ (finds cakes, bake)
Term boost: "red velvet"^4, chocolate^2
Field grouping: tags:(+work -"fun-stuff")
Escaping: Escape characters +-&|!(){}[]^"~*?:\ with \, e.g. \+
Range search: properties.timestamp:[1587729413488 TO *] (inclusive), properties.title:{A TO Z}(excluding A and Z)
Combinations: chocolate AND vanilla, chocolate OR vanilla, (chocolate OR vanilla) NOT "vanilla pudding"
Field search: properties.title:"The Title" AND text
Profile picture
VexedCat68
Moderator
60 Questions, 381 Answers
  Active since 10 January 2023
  Last activity 7 months ago

Reputation

0

Badges 1

371 × Eureka!
0 Votes
14 Answers
1K Views
0 Votes 14 Answers 1K Views
2 years ago
0 Votes
9 Answers
1K Views
0 Votes 9 Answers 1K Views
Pipeline_Controller.py doesn't exist in repo, or atleast the link to it in the docs is dead in the simple pipeline example. Any help would be appreciated.
3 years ago
0 Votes
10 Answers
1K Views
0 Votes 10 Answers 1K Views
3 years ago
0 Votes
13 Answers
1K Views
0 Votes 13 Answers 1K Views
Another simple query guys. I've installed clearml on ubuntu. However, it says command not found when i run any command with clearml . I feel like it might be...
3 years ago
0 Votes
3 Answers
953 Views
0 Votes 3 Answers 953 Views
In your docs for Dataset at https://clear.ml/docs/latest/docs/references/sdk/dataset#class-dataset , I think you might have duplicate explanations for list_m...
2 years ago
0 Votes
10 Answers
1K Views
0 Votes 10 Answers 1K Views
3 years ago
0 Votes
15 Answers
991 Views
0 Votes 15 Answers 991 Views
How do I get args like epochs to show up in the UI configuration panel under hyperparameters? I want to be able to change number of epochs and learning rate ...
3 years ago
0 Votes
10 Answers
1K Views
0 Votes 10 Answers 1K Views
3 years ago
0 Votes
2 Answers
1K Views
0 Votes 2 Answers 1K Views
Given this pipeline step, is there any way to get the return value outside of the pipeline? Like put split_dataset_id in a variable in the main pipeline file.
2 years ago
0 Votes
5 Answers
1K Views
0 Votes 5 Answers 1K Views
Would appreciate some help. Getting this Error. ValueError: Node train_model, parameter '${split_dataset.split_dataset_id}', input type 'split_dataset_id' is...
2 years ago
0 Votes
1 Answers
1K Views
0 Votes 1 Answers 1K Views
In the configuration section of an experiment, I can see the args I passed via argparse. Is there any way to create other sections other than args and tf_define
3 years ago
0 Votes
16 Answers
1K Views
0 Votes 16 Answers 1K Views
2 years ago
0 Votes
27 Answers
1K Views
0 Votes 27 Answers 1K Views
I wanted to ask, how to run pipeline steps conditionally? E.g if step returns a specific value, exit the pipeline or run another step instead of the sequenti...
2 years ago
0 Votes
17 Answers
1K Views
0 Votes 17 Answers 1K Views
Should Dataset Triggers also be activated if there is no trigger condition except dataset_project and a new task starts in that project? Is this expected beh...
3 years ago
0 Votes
1 Answers
905 Views
0 Votes 1 Answers 905 Views
3 years ago
0 Votes
5 Answers
1K Views
0 Votes 5 Answers 1K Views
3 years ago
0 Votes
9 Answers
893 Views
0 Votes 9 Answers 893 Views
I have a clearml server deployed. I can see in the docs that one way of doing things is through api calls. The docs mention the end point url but does not me...
2 years ago
0 Votes
7 Answers
923 Views
0 Votes 7 Answers 923 Views
Hi guys, needed a bit of clarification. Every 15 minutes, the scheduler prints, "Syncing scheduler, sleeping for 15 minutes until next sync". Can you guide m...
2 years ago
0 Votes
12 Answers
1K Views
0 Votes 12 Answers 1K Views
I'm on the machine with ClearML Server hosted. Is there any way to see datasets uploaded to ClearML Data without downloading them using ClearML Data?
3 years ago
0 Votes
19 Answers
1K Views
0 Votes 19 Answers 1K Views
Is this the write way to add a tag to an output model artifact of a task? torch.save(model, ' http://best.pt ') output_model = task.models['output'][-1] outp...
2 years ago
0 Votes
8 Answers
1K Views
0 Votes 8 Answers 1K Views
Is there a direct way to get a model using its id like it works with Dataset.get?
2 years ago
0 Votes
11 Answers
1K Views
0 Votes 11 Answers 1K Views
3 years ago
0 Votes
13 Answers
1K Views
0 Votes 13 Answers 1K Views
3 years ago
0 Votes
2 Answers
994 Views
0 Votes 2 Answers 994 Views
Is there a ClearML way to write a scheduler to watch a folder and publish a dataset when there are X number of files in that folder, or do I have to write a ...
3 years ago
0 Votes
5 Answers
913 Views
0 Votes 5 Answers 913 Views
Is it possible to add just a string or some other object as an artifact? if yes, then how?
2 years ago
0 Votes
25 Answers
941 Views
0 Votes 25 Answers 941 Views
Um, is there a way to delete an artifact from a task that is running?
2 years ago
0 Votes
5 Answers
979 Views
0 Votes 5 Answers 979 Views
3 years ago
0 Votes
10 Answers
1K Views
0 Votes 10 Answers 1K Views
Given I want to run a task in a pipeline using a base task id. One of my steps just finds the latest model to use. I want the task to output the id, and the ...
2 years ago
0 Votes
9 Answers
971 Views
0 Votes 9 Answers 971 Views
Trying to create a data pipeline on my own. Wanted to ask, for each batch of data, do I have to create a new Dataset Object or do I just create one Dataset O...
3 years ago
0 Votes
12 Answers
911 Views
0 Votes 12 Answers 911 Views
Is there a quicker way to abort all running experiments in a project? I have over a thousand running anonymous data tasks in a specific project and I want to...
2 years ago
Show more results questions
0 I'Ll Just Ask This Question Again To Get Some Fresh Attention To This. Is There Any Way To Run A Pipeline Step Conditionally? E.G, Under Certain Condition, Execute The Step Otherwise Don'T?

I checked the value is being returned, but I'm having issues accessing merged_dataset_id in the preexecute_callback like the way you showed me.

2 years ago
0 I'Ll Just Ask This Question Again To Get Some Fresh Attention To This. Is There Any Way To Run A Pipeline Step Conditionally? E.G, Under Certain Condition, Execute The Step Otherwise Don'T?

I'm not using decorators. I have a bunch of function_steps followed by a normal task step, where I've passed a base_task_id.

I want to check the value of one of the functional steps, and if it holds true, I want to execute the task step otherwise I want the pipeline to end there, since the task step is the last one.

2 years ago
0 I'Ll Just Ask This Question Again To Get Some Fresh Attention To This. Is There Any Way To Run A Pipeline Step Conditionally? E.G, Under Certain Condition, Execute The Step Otherwise Don'T?

AnxiousSeal95 I'm trying to access the specific value. I checked the type of task.artifacts and it's a ReadOnlyDict. Given that the return value I'm looking for is called merged_dataset_id, how would I go about doing that?

2 years ago
0 Hi Everyone, Is It Possible To Not Create A Copy Of A Dataset When Adding To Clearml? My Data Is Already In A Directory On The Clearml-Server Machine And I Do Not Want To Copy It, Just Add It To Clearml As Dataset.

I normally just upload the data to the ClearML server and then remove it locally from my machine but I understand that isn't what you want. A quick hack was the only thing I could come up with at the moment xd. Anyway you're welcome. Hope you find a solution.

2 years ago
0 Is There A Direct Way To Get A Model Using Its Id Like It Works With Dataset.Get?

Basically when I'm loading the model in InputModel, it loads it fine but I can't seem to get a local copy.

2 years ago
0 I'Ll Just Ask This Question Again To Get Some Fresh Attention To This. Is There Any Way To Run A Pipeline Step Conditionally? E.G, Under Certain Condition, Execute The Step Otherwise Don'T?

AnxiousSeal95 Basically its a function step return. if I do, artifacts.keys(), there are no keys, even though the step prior to it does return the output

2 years ago
0 Should Dataset Triggers Also Be Activated If There Is No Trigger Condition Except Dataset_Project And A New Task Starts In That Project? Is This Expected Behavior?

trigger_scheduler.add_dataset_trigger(schedule_task_id=TASK_ID, schedule_queue='default',
trigger_project='Cassava Leaf Disease Classification', name='start task on data - End Game')

3 years ago
0 I'Ll Just Ask This Question Again To Get Some Fresh Attention To This. Is There Any Way To Run A Pipeline Step Conditionally? E.G, Under Certain Condition, Execute The Step Otherwise Don'T?

AnxiousSeal95 I just have a question, can you share an example of accessing an artifact of a previous step in the pre execute callback?

2 years ago
0 I Keep Facing This Issue. I'M Trying To Set Up My Own Clearml Server Using This Tutorial.

There's a whole task bar on the left in the server. I only get this page when i use the ip 0.0.0.0

3 years ago
0 Hi Guys, Needed A Bit Of Clarification. Every 15 Minutes, The Scheduler Prints, "Syncing Scheduler, Sleeping For 15 Minutes Until Next Sync". Can You Guide Me As To It Is Syncing With What?

Alright. Can you guide me on how to edit the task configuration object? Is it done via the UI or programatically? Is there a config file and can it work with any config file I create or is it a specific config file? Sorry for the barrage of questions.

2 years ago
0 When I Create Clearml-Dataset From The Cli, I Get An Id. The Same Doesn'T Happen When I Use The Python Api. Is There Any Way To Get The Id In Python?

My use case is basically if I want to now access this dataset from somewhere else, shouldn't I be able to do so using its id?

3 years ago
3 years ago
0 I Keep Facing This Issue. I'M Trying To Set Up My Own Clearml Server Using This Tutorial.

When I try to access the server with the IP I set as CLEARML_HOST_IP, it looks like this. I set that IP to the ip assigned to me by the network

3 years ago
0 Is There A Quicker Way To Abort All Running Experiments In A Project? I Have Over A Thousand Running Anonymous Data Tasks In A Specific Project And I Want To Abort Them Before Debugging Them.

In the case of api call,

given that i have id of the task I want to stop, I would make a post request to [CLEARML_SERVER_URL]:8080/tasks.stop

with the request body set up like the one mentioned in the api?

2 years ago
0 When It Comes To Continuous Training, I Wanted To Know How You Train Or Would Train If You Have Annotated Data Incoming? Do You Train Completely Online Where You Train As Soon As You Have A Training Example Available? Do You Instead Train When You Have A

I get what you're saying. I was considering training on just the new data to see how it works. To me it felt like that was the fastest way to deal with data drift. I understand that it may introduce instability however. I was curious how other developers who have successfully managed to set up continuous training deal with it. 100% new data, or a ratio between new and old data. And if it is the latter, what should be the case, which should be the majority, old data or new data?

3 years ago
0 When It Comes To Continuous Training, I Wanted To Know How You Train Or Would Train If You Have Annotated Data Incoming? Do You Train Completely Online Where You Train As Soon As You Have A Training Example Available? Do You Instead Train When You Have A

Understandable. I mainly have regular image data, not video sequences so I can do the train test splits like you mentioned normally. What about the epochs though? Is there a recommended number of epochs when you train on that new batch?

3 years ago
Show more results compactanswers