Examples: query, "exact match", wildcard*, wild?ard, wild*rd
Fuzzy search: cake~ (finds cakes, bake)
Term boost: "red velvet"^4, chocolate^2
Field grouping: tags:(+work -"fun-stuff")
Escaping: Escape characters +-&|!(){}[]^"~*?:\ with \, e.g. \+
Range search: properties.timestamp:[1587729413488 TO *] (inclusive), properties.title:{A TO Z}(excluding A and Z)
Combinations: chocolate AND vanilla, chocolate OR vanilla, (chocolate OR vanilla) NOT "vanilla pudding"
Field search: properties.title:"The Title" AND text
Profile picture
VexedCat68
Moderator
60 Questions, 381 Answers
  Active since 10 January 2023
  Last activity one year ago

Reputation

0

Badges 1

371 × Eureka!
0 Votes
29 Answers
2K Views
0 Votes 29 Answers 2K Views
3 years ago
0 Votes
10 Answers
2K Views
0 Votes 10 Answers 2K Views
3 years ago
0 Votes
17 Answers
2K Views
0 Votes 17 Answers 2K Views
Should Dataset Triggers also be activated if there is no trigger condition except dataset_project and a new task starts in that project? Is this expected beh...
4 years ago
0 Votes
13 Answers
2K Views
0 Votes 13 Answers 2K Views
Another simple query guys. I've installed clearml on ubuntu. However, it says command not found when i run any command with clearml . I feel like it might be...
4 years ago
0 Votes
2 Answers
2K Views
0 Votes 2 Answers 2K Views
When saving the model, there's a label tab but it's empty.
3 years ago
0 Votes
2 Answers
2K Views
0 Votes 2 Answers 2K Views
3 years ago
0 Votes
12 Answers
2K Views
0 Votes 12 Answers 2K Views
Is there a quicker way to abort all running experiments in a project? I have over a thousand running anonymous data tasks in a specific project and I want to...
3 years ago
0 Votes
25 Answers
2K Views
0 Votes 25 Answers 2K Views
I'll just ask this question again to get some fresh attention to this. Is there any way to run a pipeline step conditionally? E.g, under certain condition, e...
3 years ago
0 Votes
2 Answers
2K Views
0 Votes 2 Answers 2K Views
How can I register a json file I'm creating as an artifact
3 years ago
0 Votes
7 Answers
2K Views
0 Votes 7 Answers 2K Views
3 years ago
0 Votes
10 Answers
2K Views
0 Votes 10 Answers 2K Views
4 years ago
0 Votes
1 Answers
2K Views
0 Votes 1 Answers 2K Views
In the configuration section of an experiment, I can see the args I passed via argparse. Is there any way to create other sections other than args and tf_define
3 years ago
0 Votes
11 Answers
2K Views
0 Votes 11 Answers 2K Views
3 years ago
0 Votes
14 Answers
2K Views
0 Votes 14 Answers 2K Views
3 years ago
0 Votes
11 Answers
2K Views
0 Votes 11 Answers 2K Views
3 years ago
0 Votes
3 Answers
2K Views
0 Votes 3 Answers 2K Views
3 years ago
0 Votes
10 Answers
2K Views
0 Votes 10 Answers 2K Views
4 years ago
0 Votes
9 Answers
2K Views
0 Votes 9 Answers 2K Views
Trying to create a data pipeline on my own. Wanted to ask, for each batch of data, do I have to create a new Dataset Object or do I just create one Dataset O...
4 years ago
0 Votes
5 Answers
2K Views
0 Votes 5 Answers 2K Views
Would appreciate some help. Getting this Error. ValueError: Node train_model, parameter '${split_dataset.split_dataset_id}', input type 'split_dataset_id' is...
3 years ago
0 Votes
2 Answers
2K Views
0 Votes 2 Answers 2K Views
Given this pipeline step, is there any way to get the return value outside of the pipeline? Like put split_dataset_id in a variable in the main pipeline file.
3 years ago
0 Votes
4 Answers
2K Views
0 Votes 4 Answers 2K Views
Is it possible to write text file and see it in results of the experiment? I want to use it to version data as in keeping a track of what images have been tr...
4 years ago
0 Votes
25 Answers
2K Views
0 Votes 25 Answers 2K Views
Um, is there a way to delete an artifact from a task that is running?
3 years ago
0 Votes
9 Answers
2K Views
0 Votes 9 Answers 2K Views
Pipeline_Controller.py doesn't exist in repo, or atleast the link to it in the docs is dead in the simple pipeline example. Any help would be appreciated.
4 years ago
0 Votes
10 Answers
2K Views
0 Votes 10 Answers 2K Views
Given I want to run a task in a pipeline using a base task id. One of my steps just finds the latest model to use. I want the task to output the id, and the ...
3 years ago
0 Votes
16 Answers
2K Views
0 Votes 16 Answers 2K Views
3 years ago
0 Votes
23 Answers
2K Views
0 Votes 23 Answers 2K Views
I'm looking at how triggers work in ClearML. Is there an example, maybe with clearml data and a dataset being uploaded or some other example?
4 years ago
0 Votes
1 Answers
2K Views
0 Votes 1 Answers 2K Views
I have a question. I'm struggling with setting up my own ClearML server. I think i've got it up and running but not sure how to send clearml-task to my own s...
4 years ago
0 Votes
13 Answers
2K Views
0 Votes 13 Answers 2K Views
3 years ago
0 Votes
8 Answers
2K Views
0 Votes 8 Answers 2K Views
Is there a direct way to get a model using its id like it works with Dataset.get?
3 years ago
0 Votes
3 Answers
2K Views
0 Votes 3 Answers 2K Views
When I upload and publish data to clearml-data, it says successful. Now when i try to get it using id, I get the following error. Error: __get_tasks() got mu...
4 years ago
Show more results questions
0 I'Ve Been Asked Requirements To Deploy Clearml Server On Our Company Server. I'M Not Exactly What To Put In The Requirements. I Know It'S A Vague Question. If We Can Exclude Project Specific Requirements For Now And Only Keep Clearml Server Deployment Req

Also I made another thread regarding clear ml agent. can you respond to that? I'm gonna be trying to set up a clear ml server properly on a server machine. Want to test how to train models and enqueue tasks and automate this whole process with GPU training included.

4 years ago
0 I'M Trying To Use Store Some Data In Clearml-Data. Then I Want To Get It Back Elsewhere Using Dataset.Get_Local_Copy. It Returns A Directory But I Don'T Know How To Actually Access And Use That Data, And Remove It When It'S Done. Help Would Be Appreciated

I hope what I said was clear. Basically in reality they both seem mutable, with just the directory downloaded to being optional in one and in the other it's downloaded to the cache folder always.

4 years ago
0 Hi Everyone, Is It Possible To Not Create A Copy Of A Dataset When Adding To Clearml? My Data Is Already In A Directory On The Clearml-Server Machine And I Do Not Want To Copy It, Just Add It To Clearml As Dataset.

How about instead of uploading the entire dataset to the clearml server, upload a text file with the location of the dataset on the machine. I would think that should do the trick.

3 years ago
0 Given I Want To Run A Task In A Pipeline Using A Base Task Id. One Of My Steps Just Finds The Latest Model To Use. I Want The Task To Output The Id, And The Next Step To Use It. How Would I Go About Doing This?

is this the correct way to upload an artifact?

checkpoint.split('.')[0] is the name that I want it assigned and the second argument is the path to the file.

3 years ago
0 I'D Been Following The Clearml Serving Example On Its Github Repo Here. It Basically Deploys A Keras Mnist Model. The Tutorial However Ends Once The Model Is Deployed However And I'Ve Tried Going Through Resources On How To Do Inference But Have Had Troub

For anyone who's struggling with this. This is how I solved it. I'd personally not worked with GRPC so I instead looked at the HTTP docs and that one was much simpler to use.

3 years ago
0 Hey Guys, Sorry For The Rapid Fire Questions In The Past Few Days. I Have Another Issue Though. I Initially Ran A Task, Directly From A Repo. It Succesfully Installed The Requirements From The Requirements File In The Repo And Ran The Task Without Any Iss

My use case is that the code using pytorch saves additional info like the state dict when saving the model. I'd like to save that information as an artifact as well so that I can load it later.

3 years ago
0 I Wanted To Ask, I'M Versioning My Data Using Clearml Data. And I'Ll Have A Training Task With Clearml Task. My Question Is, Does Clearml Keep Track Of The Data Versions Fetched From Clearml Data? Basically I Want To See How Much Of Tracking And Informati

I'm not sure about auto logging, since you might be using different datasets or you might get a dataset but might not use it based on specific conditions. However as a developer choosing to use such as ClearML who considers it more of an ecosystem instead of just a continuous training pipeline, I would want as many aspects of the MLOPS process and the information around the experiment to be able to be logged within the bounds of ClearML without having to use external databases or libraries.

3 years ago
0 Is There A Direct Way To Get A Model Using Its Id Like It Works With Dataset.Get?

Basically when I'm loading the model in InputModel, it loads it fine but I can't seem to get a local copy.

3 years ago
0 I'M Looking At How Triggers Work In Clearml. Is There An Example, Maybe With Clearml Data And A Dataset Being Uploaded Or Some Other Example?

This here shows my situation. You can see the code on the left and the tasks called 'Cassava Training' on the right. They keep getting enqueued even though I only sent a trigger once. By that I mean I only published a dataset once.

4 years ago
0 When Using Dataset.Get_Local_Copy(), Once I Get The Location, Can I Add Another Folder Inside Location Add Some Files In It, Create A New Dataset Object, And Then Do Dataset.Upload(Location)? Should This Work? Or Since Its Get_Local_Copy, I Won'T Be Able

I already have the dataset id as a hyperparameter. I get said dataset. I'm only handling one dataset right now but merging multiple ones is a simple task as well.

Also I'm not very experienced and am unsure what proposed querying is and how and if it works in ClearML here.

3 years ago
0 I'Ve Been Asked Requirements To Deploy Clearml Server On Our Company Server. I'M Not Exactly What To Put In The Requirements. I Know It'S A Vague Question. If We Can Exclude Project Specific Requirements For Now And Only Keep Clearml Server Deployment Req

SuccessfulKoala55 Sorry to ping you like this. I have to ask. What's the minimum requirements for clear ml installation. Excluding requirements for databases or file server

4 years ago
0 I'Ll Just Ask This Question Again To Get Some Fresh Attention To This. Is There Any Way To Run A Pipeline Step Conditionally? E.G, Under Certain Condition, Execute The Step Otherwise Don'T?

AnxiousSeal95 I just have a question, can you share an example of accessing an artifact of a previous step in the pre execute callback?

3 years ago
0 Given I Want To Run A Task In A Pipeline Using A Base Task Id. One Of My Steps Just Finds The Latest Model To Use. I Want The Task To Output The Id, And The Next Step To Use It. How Would I Go About Doing This?

Well yeah, you can say that. In add function step, I pass in a function which returns something. And since I've written the name of the returned parameter in add_function_step, I can use it, but I can't seem to figure out a way to do something similar using a task in add_step

3 years ago
0 I'Ll Just Ask This Question Again To Get Some Fresh Attention To This. Is There Any Way To Run A Pipeline Step Conditionally? E.G, Under Certain Condition, Execute The Step Otherwise Don'T?

AnxiousSeal95 I'm trying to access the specific value. I checked the type of task.artifacts and it's a ReadOnlyDict. Given that the return value I'm looking for is called merged_dataset_id, how would I go about doing that?

3 years ago
0 How Do I Get Args Like Epochs To Show Up In The Ui Configuration Panel Under Hyperparameters? I Want To Be Able To Change Number Of Epochs And Learning Rate From Within The Ui.

I checked and it seems when i an example from git, it works as it should. but when I try to run my own script, the draft is in read only mode.

4 years ago
0 I'Ll Just Ask This Question Again To Get Some Fresh Attention To This. Is There Any Way To Run A Pipeline Step Conditionally? E.G, Under Certain Condition, Execute The Step Otherwise Don'T?

I'm not using decorators. I have a bunch of function_steps followed by a normal task step, where I've passed a base_task_id.

I want to check the value of one of the functional steps, and if it holds true, I want to execute the task step otherwise I want the pipeline to end there, since the task step is the last one.

3 years ago
Show more results compactanswers