Examples: query, "exact match", wildcard*, wild?ard, wild*rd
Fuzzy search: cake~ (finds cakes, bake)
Term boost: "red velvet"^4, chocolate^2
Field grouping: tags:(+work -"fun-stuff")
Escaping: Escape characters +-&|!(){}[]^"~*?:\ with \, e.g. \+
Range search: properties.timestamp:[1587729413488 TO *] (inclusive), properties.title:{A TO Z}(excluding A and Z)
Combinations: chocolate AND vanilla, chocolate OR vanilla, (chocolate OR vanilla) NOT "vanilla pudding"
Field search: properties.title:"The Title" AND text
Profile picture
VexedCat68
Moderator
60 Questions, 381 Answers
  Active since 10 January 2023
  Last activity one year ago

Reputation

0

Badges 1

371 × Eureka!
0 Votes
25 Answers
2K Views
0 Votes 25 Answers 2K Views
I'll just ask this question again to get some fresh attention to this. Is there any way to run a pipeline step conditionally? E.g, under certain condition, e...
3 years ago
0 Votes
2 Answers
2K Views
0 Votes 2 Answers 2K Views
When saving the model, there's a label tab but it's empty.
3 years ago
0 Votes
2 Answers
2K Views
0 Votes 2 Answers 2K Views
How can I register a json file I'm creating as an artifact
3 years ago
0 Votes
5 Answers
2K Views
0 Votes 5 Answers 2K Views
3 years ago
0 Votes
10 Answers
2K Views
0 Votes 10 Answers 2K Views
Given I want to run a task in a pipeline using a base task id. One of my steps just finds the latest model to use. I want the task to output the id, and the ...
3 years ago
0 Votes
10 Answers
2K Views
0 Votes 10 Answers 2K Views
4 years ago
0 Votes
3 Answers
2K Views
0 Votes 3 Answers 2K Views
3 years ago
0 Votes
13 Answers
2K Views
0 Votes 13 Answers 2K Views
3 years ago
0 Votes
9 Answers
2K Views
0 Votes 9 Answers 2K Views
I have a clearml server deployed. I can see in the docs that one way of doing things is through api calls. The docs mention the end point url but does not me...
3 years ago
0 Votes
2 Answers
2K Views
0 Votes 2 Answers 2K Views
Is there a ClearML way to write a scheduler to watch a folder and publish a dataset when there are X number of files in that folder, or do I have to write a ...
3 years ago
0 Votes
7 Answers
2K Views
0 Votes 7 Answers 2K Views
When i create clearml-dataset from the cli, I get an id. The same doesn't happen when I use the Python API. is there any way to get the ID in python?
4 years ago
0 Votes
15 Answers
2K Views
0 Votes 15 Answers 2K Views
I'm publishing the model artifact after the tensorflow model is saved. But in the UI, the download button is grayed out. And get_local_copy() doesn't seem to...
3 years ago
0 Votes
2 Answers
2K Views
0 Votes 2 Answers 2K Views
Given this pipeline step, is there any way to get the return value outside of the pipeline? Like put split_dataset_id in a variable in the main pipeline file.
3 years ago
0 Votes
11 Answers
2K Views
0 Votes 11 Answers 2K Views
3 years ago
0 Votes
14 Answers
2K Views
0 Votes 14 Answers 2K Views
3 years ago
0 Votes
22 Answers
2K Views
0 Votes 22 Answers 2K Views
I've setup my own clearml server. Only problem is, I can't seem to find where the credentials are. I've attached a screenshot.
4 years ago
0 Votes
5 Answers
2K Views
0 Votes 5 Answers 2K Views
Would appreciate some help. Getting this Error. ValueError: Node train_model, parameter '${split_dataset.split_dataset_id}', input type 'split_dataset_id' is...
3 years ago
0 Votes
5 Answers
2K Views
0 Votes 5 Answers 2K Views
3 years ago
0 Votes
4 Answers
2K Views
0 Votes 4 Answers 2K Views
Is it possible to write text file and see it in results of the experiment? I want to use it to version data as in keeping a track of what images have been tr...
4 years ago
0 Votes
13 Answers
2K Views
0 Votes 13 Answers 2K Views
3 years ago
0 Votes
11 Answers
2K Views
0 Votes 11 Answers 2K Views
3 years ago
0 Votes
17 Answers
2K Views
0 Votes 17 Answers 2K Views
Should Dataset Triggers also be activated if there is no trigger condition except dataset_project and a new task starts in that project? Is this expected beh...
3 years ago
0 Votes
5 Answers
1K Views
0 Votes 5 Answers 1K Views
Is it possible to add just a string or some other object as an artifact? if yes, then how?
3 years ago
0 Votes
8 Answers
2K Views
0 Votes 8 Answers 2K Views
3 years ago
0 Votes
7 Answers
2K Views
0 Votes 7 Answers 2K Views
Hi guys, needed a bit of clarification. Every 15 minutes, the scheduler prints, "Syncing scheduler, sleeping for 15 minutes until next sync". Can you guide m...
3 years ago
0 Votes
27 Answers
2K Views
0 Votes 27 Answers 2K Views
I wanted to ask, how to run pipeline steps conditionally? E.g if step returns a specific value, exit the pipeline or run another step instead of the sequenti...
3 years ago
0 Votes
10 Answers
2K Views
0 Votes 10 Answers 2K Views
4 years ago
0 Votes
12 Answers
3K Views
0 Votes 12 Answers 3K Views
I'm training a tensorflow model and saving it in the end. I looked at the OutputModel class. How do I connect the model I'm saving to the OutputModel?
3 years ago
0 Votes
10 Answers
2K Views
0 Votes 10 Answers 2K Views
3 years ago
0 Votes
4 Answers
2K Views
0 Votes 4 Answers 2K Views
I'm trying to run a task on an agent. I've passed the requirements file but it isn't able to install it. The error is in the reply. Help would be appreciated.
3 years ago
Show more results questions
0 Trying To Create A Data Pipeline On My Own. Wanted To Ask, For Each Batch Of Data, Do I Have To Create A New Dataset Object Or Do I Just Create One Dataset Object And Add Batches To It. If Its The Latter, Then How.

I'm kind of new to developing end to end applications so I'm also learning how the predefined pipelines work as well. I'll take a look at the clear ml custom pipelines

4 years ago
0 Should Dataset Triggers Also Be Activated If There Is No Trigger Condition Except Dataset_Project And A New Task Starts In That Project? Is This Expected Behavior?

I just assumed it should only be triggered by dataset related things but after a lot of experimenting i realized its also triggered by tasks, if the only condition passed is dataset_project and no other specific trigger condition like on publish or on tags are added.

3 years ago
0 I'M Publishing The Model Artifact After The Tensorflow Model Is Saved. But In The Ui, The Download Button Is Grayed Out. And Get_Local_Copy() Doesn'T Seem To Work Either, Help Would Be Appreciated.

For anyone facing a similar issue to mine and wanting the model to uploaded just like data is uploaded,

in the Task.init, set the output_uri = True.

This basically makes it use the default file server for clearml that you define in the clearml.conf file. Ty.

3 years ago
0 Um, Is There A Way To Delete An Artifact From A Task That Is Running?

shouldn't checkpoints be uploaded immediately, that's the purpose of checkpointing isn't it?

3 years ago
0 In The Case Where I'M Passing A Schedule_Fn To Add_Task In A Taskscheduler, How Do I Pass The Function Arguments?

there are other parameters for add_task as well, I'm just curious as to how do I pass the folder and batch size in the schedule_fn=watch_folder part

3 years ago
0 Is There An Example Of Simple Pipeline Using Task Scheduler? I Want To Create A Simple Pipeline Where First A Folder Is Monitored Using A Taskscheduler, And Under Certain Conditions, Data Is Uploaded And Dataset Id Is Given To The Next Step In The Pipeli

Alright. Anyway I'm practicing with the pipeline. I have an agent listening to the queue. Only problem is, it fails because of requirement issues but I don't know how to pass requirements in this case.

3 years ago
0 Um, Is There A Way To Delete An Artifact From A Task That Is Running?

I need to both remove the artifact from the UI and the storage.

3 years ago
0 When Using Dataset.Get_Local_Copy(), Once I Get The Location, Can I Add Another Folder Inside Location Add Some Files In It, Create A New Dataset Object, And Then Do Dataset.Upload(Location)? Should This Work? Or Since Its Get_Local_Copy, I Won'T Be Able

My current approach is, watch a folder, when there are sufficient data points, just move N of them into another folder and create a raw dataset and call the pipeline with this dataset.

It gets downloaded, preprocessed, and then uploaded again.

In the final step, the preprocessed dataset is downloaded and is used to train the model.

3 years ago
0 Um, Is There A Way To Delete An Artifact From A Task That Is Running?

AgitatedDove14 CostlyOstrich36 I think that is the approach that'll work for me. I just need to be able to remove checkpoints I don't need given I know their name, from the UI and Storage.

3 years ago
0 Um, Is There A Way To Delete An Artifact From A Task That Is Running?

Is there a difference? I mean my use case is pretty simple. I have a training and it basically creates a lot of checkpoints. I just want to keep the n best checkpoints and whenever there are more than N checkpoints, I'll delete the worst performing one. Deleted both locally and from the the task artifacts.

3 years ago
0 When Using Dataset.Get_Local_Copy(), Once I Get The Location, Can I Add Another Folder Inside Location Add Some Files In It, Create A New Dataset Object, And Then Do Dataset.Upload(Location)? Should This Work? Or Since Its Get_Local_Copy, I Won'T Be Able

Sorry for the late response. Agreed, that can work, although I would prefer a way to access the data by M number of batches added instead of a certain range, since these cases aren't interchangeable. Also a simple thing that can be done is that you can create an empty Dataset in the start, and then make it the parent of every dataset you add.

3 years ago
0 Is This The Write Way To Add A Tag To An Output Model Artifact Of A Task? Torch.Save(Model, '

adding tags this way to a Dataset object works fine. This issue only occured when doing this to a model.

3 years ago
0 When Saving The Model, There'S A Label Tab But It'S Empty.

Thanks, I went through it and this seems easy

3 years ago
Show more results compactanswers