Examples: query, "exact match", wildcard*, wild?ard, wild*rd
Fuzzy search: cake~ (finds cakes, bake)
Term boost: "red velvet"^4, chocolate^2
Field grouping: tags:(+work -"fun-stuff")
Escaping: Escape characters +-&|!(){}[]^"~*?:\ with \, e.g. \+
Range search: properties.timestamp:[1587729413488 TO *] (inclusive), properties.title:{A TO Z}(excluding A and Z)
Combinations: chocolate AND vanilla, chocolate OR vanilla, (chocolate OR vanilla) NOT "vanilla pudding"
Field search: properties.title:"The Title" AND text
Profile picture
VexedCat68
Moderator
60 Questions, 381 Answers
  Active since 10 January 2023
  Last activity one year ago

Reputation

0

Badges 1

371 × Eureka!
0 I Wanted To Ask, How To Run Pipeline Steps Conditionally? E.G If Step Returns A Specific Value, Exit The Pipeline Or Run Another Step Instead Of The Sequential Step

Wait, so the pipeline step only runs if the pre execute callback returns True? It'll stop if it doesn't run?

3 years ago
0 Hey Guys, Sorry For The Rapid Fire Questions In The Past Few Days. I Have Another Issue Though. I Initially Ran A Task, Directly From A Repo. It Succesfully Installed The Requirements From The Requirements File In The Repo And Ran The Task Without Any Iss

when i pass the repo in clearml-task with the parameters, it runs fine and finishes. Basically when I clone and attempt the task again, I get the above assert error I don't know why.

3 years ago
0 When Saving The Model, There'S A Label Tab But It'S Empty.

Thanks, I went through it and this seems easy

3 years ago
0 Sorry For The Barrage Of Questions. I Can'T Seem To Figure Out How Best To Get A Python Script I Need To Run On An Agent. I Have An Agent Listening To The Default Queue And I Also Have The Script That I Need To Run On The Agent. Can You Guide Me On How To

Basically, as soon as I get the trigger that a new dataset has been published, I want to pass the dataset id to the script as an cli argument and pass the code to the agent

3 years ago
0 I'M Publishing The Model Artifact After The Tensorflow Model Is Saved. But In The Ui, The Download Button Is Grayed Out. And Get_Local_Copy() Doesn'T Seem To Work Either, Help Would Be Appreciated.

Since I want to save the model to the clearml server? What should the port be alongside the url?

And in that case, if I do, model.save('test'), it will also save the model to the clearml server?

3 years ago
0 Hey Guys, Sorry For The Rapid Fire Questions In The Past Few Days. I Have Another Issue Though. I Initially Ran A Task, Directly From A Repo. It Succesfully Installed The Requirements From The Requirements File In The Repo And Ran The Task Without Any Iss

up to date with https://fawad_nizamani@bitbucket.org/fawad_nizamani/custom_yolov5
Traceback (most recent call last):
File "train.py", line 682, in <module>
main(opt)
File "train.py", line 525, in main
assert os.path.isfile(ckpt), 'ERROR: --resume checkpoint does not exist'
AssertionError: ERROR: --resume checkpoint does not exist

3 years ago
0 Um, Is There A Way To Delete An Artifact From A Task That Is Running?

I ran a training code from a github repo. It saves checkpoints every 2000 iterations. Only problem is I'm training it for 3200 epochs and there's more than 37000 iterations in each epoch. So the checkpoints just added up. I've stopped the training for now. I need to delete all of those checkpoints before I start training again.

3 years ago
0 Um, Is There A Way To Delete An Artifact From A Task That Is Running?

AgitatedDove14 Alright I think I understand, changes made in storage will be visible in the front end directly.

Will using Model.remove, completely delete from storage as well?

3 years ago
0 I'M Publishing The Model Artifact After The Tensorflow Model Is Saved. But In The Ui, The Download Button Is Grayed Out. And Get_Local_Copy() Doesn'T Seem To Work Either, Help Would Be Appreciated.

After the previous code, I got the model uploaded by the previous code using its ID. Now when I add tags here, they were visible in the UI

3 years ago
0 I'Ve Been Using Add_Function_Step In The Pipeline Controller As Of Yet. Now I Wish To Use Add_Step For One Of The Steps And Pass In A Task Id. Only Problem Is, For Some Of The Argparse Parameters, I Want To Modify The Values In The New Task. Can You Guys

Sorry for the late reply. The situation is that when I ran the task initially, it took arguments in using ArgParse. It took in a lot of arguments. Now my understanding is that add_step() clones that task. I want that to happen but I would like to be able to modify some of the values of the args, e.g epochs or some other argument.

3 years ago
0 I Keep Facing This Issue. I'M Trying To Set Up My Own Clearml Server Using This Tutorial.

Shouldn't I get redirected to the login page if i'm not logged in instead of the dashboard? 😞

4 years ago
0 I Wanted To Ask, How To Run Pipeline Steps Conditionally? E.G If Step Returns A Specific Value, Exit The Pipeline Or Run Another Step Instead Of The Sequential Step

In another answer, I was shown that I can access it like this. How can I go about accessing the value of merged_dataset_id which was returned by merge_n_datasets and stored as an artifact.

3 years ago
0 I Wanted To Ask, I'M Versioning My Data Using Clearml Data. And I'Ll Have A Training Task With Clearml Task. My Question Is, Does Clearml Keep Track Of The Data Versions Fetched From Clearml Data? Basically I Want To See How Much Of Tracking And Informati

It does to me. However I'm proposing a situation where a user gets N number of Datasets using Dataset.get, but uses m number of datasets for training where m < n. Would it make sense to only log the m datasets that were used for training? How would that be done?

3 years ago
3 years ago
3 years ago
0 Hey Guys, Sorry For The Rapid Fire Questions In The Past Few Days. I Have Another Issue Though. I Initially Ran A Task, Directly From A Repo. It Succesfully Installed The Requirements From The Requirements File In The Repo And Ran The Task Without Any Iss

Another issue I'm having is I ran a task using clearml-task and did it using a repo. It runs fine, when I clone said task however and run it on the same queue again, it throws an error from the code. I can't seem to figure out why its happening.

3 years ago
0 Is There A Quicker Way To Abort All Running Experiments In A Project? I Have Over A Thousand Running Anonymous Data Tasks In A Specific Project And I Want To Abort Them Before Debugging Them.

Is there a code example for this.

I was going through the python api and the closest thing that resembled to my use case was task.close() however it didn't do anything.

3 years ago
4 years ago
Show more results compactanswers