Examples: query, "exact match", wildcard*, wild?ard, wild*rd
Fuzzy search: cake~ (finds cakes, bake)
Term boost: "red velvet"^4, chocolate^2
Field grouping: tags:(+work -"fun-stuff")
Escaping: Escape characters +-&|!(){}[]^"~*?:\ with \, e.g. \+
Range search: properties.timestamp:[1587729413488 TO *] (inclusive), properties.title:{A TO Z}(excluding A and Z)
Combinations: chocolate AND vanilla, chocolate OR vanilla, (chocolate OR vanilla) NOT "vanilla pudding"
Field search: properties.title:"The Title" AND text
Profile picture
VexedCat68
Moderator
60 Questions, 381 Answers
  Active since 10 January 2023
  Last activity one year ago

Reputation

0

Badges 1

371 × Eureka!
0 Votes
10 Answers
2K Views
0 Votes 10 Answers 2K Views
Given I want to run a task in a pipeline using a base task id. One of my steps just finds the latest model to use. I want the task to output the id, and the ...
3 years ago
0 Votes
13 Answers
2K Views
0 Votes 13 Answers 2K Views
4 years ago
0 Votes
5 Answers
2K Views
0 Votes 5 Answers 2K Views
Is it possible to add just a string or some other object as an artifact? if yes, then how?
3 years ago
0 Votes
2 Answers
2K Views
0 Votes 2 Answers 2K Views
3 years ago
0 Votes
12 Answers
2K Views
0 Votes 12 Answers 2K Views
Is there a quicker way to abort all running experiments in a project? I have over a thousand running anonymous data tasks in a specific project and I want to...
3 years ago
0 Votes
4 Answers
2K Views
0 Votes 4 Answers 2K Views
I'm trying to run a task on an agent. I've passed the requirements file but it isn't able to install it. The error is in the reply. Help would be appreciated.
3 years ago
0 Votes
3 Answers
2K Views
0 Votes 3 Answers 2K Views
Is there any way to stop all clearml agent workers on a machine or stop workers from the clearml ui?
3 years ago
0 Votes
29 Answers
2K Views
0 Votes 29 Answers 2K Views
3 years ago
0 Votes
8 Answers
2K Views
0 Votes 8 Answers 2K Views
Is there a direct way to get a model using its id like it works with Dataset.get?
3 years ago
0 Votes
2 Answers
2K Views
0 Votes 2 Answers 2K Views
Is there a ClearML way to write a scheduler to watch a folder and publish a dataset when there are X number of files in that folder, or do I have to write a ...
4 years ago
0 Votes
5 Answers
2K Views
0 Votes 5 Answers 2K Views
Would appreciate some help. Getting this Error. ValueError: Node train_model, parameter '${split_dataset.split_dataset_id}', input type 'split_dataset_id' is...
3 years ago
0 Votes
7 Answers
2K Views
0 Votes 7 Answers 2K Views
Hi guys, needed a bit of clarification. Every 15 minutes, the scheduler prints, "Syncing scheduler, sleeping for 15 minutes until next sync". Can you guide m...
3 years ago
0 Votes
2 Answers
2K Views
0 Votes 2 Answers 2K Views
Given this pipeline step, is there any way to get the return value outside of the pipeline? Like put split_dataset_id in a variable in the main pipeline file.
3 years ago
0 Votes
22 Answers
2K Views
0 Votes 22 Answers 2K Views
I've setup my own clearml server. Only problem is, I can't seem to find where the credentials are. I've attached a screenshot.
4 years ago
0 Votes
12 Answers
2K Views
0 Votes 12 Answers 2K Views
I'm on the machine with ClearML Server hosted. Is there any way to see datasets uploaded to ClearML Data without downloading them using ClearML Data?
4 years ago
0 Votes
5 Answers
2K Views
0 Votes 5 Answers 2K Views
4 years ago
0 Votes
1 Answers
2K Views
0 Votes 1 Answers 2K Views
4 years ago
0 Votes
14 Answers
2K Views
0 Votes 14 Answers 2K Views
3 years ago
0 Votes
7 Answers
2K Views
0 Votes 7 Answers 2K Views
3 years ago
0 Votes
5 Answers
2K Views
0 Votes 5 Answers 2K Views
3 years ago
0 Votes
9 Answers
2K Views
0 Votes 9 Answers 2K Views
I have a clearml server deployed. I can see in the docs that one way of doing things is through api calls. The docs mention the end point url but does not me...
3 years ago
0 Votes
19 Answers
2K Views
0 Votes 19 Answers 2K Views
I keep facing this issue. I'm trying to set up my own clearml server using this tutorial. https://allegro.ai/clearml/docs/docs/deploying_clearml/clearml_serv...
4 years ago
0 Votes
7 Answers
2K Views
0 Votes 7 Answers 2K Views
When i create clearml-dataset from the cli, I get an id. The same doesn't happen when I use the Python API. is there any way to get the ID in python?
4 years ago
0 Votes
0 Answers
2K Views
0 Votes 0 Answers 2K Views
Just following instructions from the clearml-serving repo.
3 years ago
0 Votes
2 Answers
2K Views
0 Votes 2 Answers 2K Views
When saving the model, there's a label tab but it's empty.
3 years ago
0 Votes
13 Answers
2K Views
0 Votes 13 Answers 2K Views
3 years ago
0 Votes
15 Answers
2K Views
0 Votes 15 Answers 2K Views
I'm publishing the model artifact after the tensorflow model is saved. But in the UI, the download button is grayed out. And get_local_copy() doesn't seem to...
3 years ago
0 Votes
17 Answers
2K Views
0 Votes 17 Answers 2K Views
Should Dataset Triggers also be activated if there is no trigger condition except dataset_project and a new task starts in that project? Is this expected beh...
4 years ago
0 Votes
16 Answers
2K Views
0 Votes 16 Answers 2K Views
3 years ago
0 Votes
5 Answers
2K Views
0 Votes 5 Answers 2K Views
I keep getting this error when trying to upload a dataset. Anyone has any idea what might be causing it?
3 years ago
Show more results questions
0 Once I Set Up A Clear Ml Server On A Machine. I Understand I Need An Agent Listening On A Queue To Run Code. Do I Absolutely Have To Have An Agent And The Gpu On The Same Machine As The Server? Also Do The Gpu And Agent Have To Be On The Same Machine?

Just to be absolutely clear.

Agent Listening on Machine A with GPU listening to Queue X.

Task enqueued onto queue X from Machine B with no GPU.

Task runs on Machine A and experiment gets published to server?

4 years ago
0 I'Ve Been Using Add_Function_Step In The Pipeline Controller As Of Yet. Now I Wish To Use Add_Step For One Of The Steps And Pass In A Task Id. Only Problem Is, For Some Of The Argparse Parameters, I Want To Modify The Values In The New Task. Can You Guys

Sorry for the late reply. The situation is that when I ran the task initially, it took arguments in using ArgParse. It took in a lot of arguments. Now my understanding is that add_step() clones that task. I want that to happen but I would like to be able to modify some of the values of the args, e.g epochs or some other argument.

3 years ago
0 Another Question, I Have Written A Code That Includes A Task Scheduler That Calls A Function. That Function Watches A Folder And If There Are Sufficient Images, It Creates And Publishes The Dataset, After Which It Clears The Folder. Problem, For Some Rea

I'm still a bit confused around the fact that since my function runs once per hour, why are there indefinitely growing anonymous tasks, even after i've closed the main schedulers.

3 years ago
0 I Wanted To Ask, How To Run Pipeline Steps Conditionally? E.G If Step Returns A Specific Value, Exit The Pipeline Or Run Another Step Instead Of The Sequential Step

Is the only possible way to get a specific node, is to use one of the get_running_nodes or get_processed_nodes, and then checking every node in the list to see if the name matches the one we're looking for?

3 years ago
0 Is There An Example Of Simple Pipeline Using Task Scheduler? I Want To Create A Simple Pipeline Where First A Folder Is Monitored Using A Taskscheduler, And Under Certain Conditions, Data Is Uploaded And Dataset Id Is Given To The Next Step In The Pipeli

Alright. Anyway I'm practicing with the pipeline. I have an agent listening to the queue. Only problem is, it fails because of requirement issues but I don't know how to pass requirements in this case.

3 years ago
0 Is The Only Available Resource To Learn And Use Clearml-Serving, The Github Repo, Or Are Their Other Resources As Well? Also, In The Repo, Once The Model Is Served, It Says I Can Curl To The End Point And It Mentions <Serving-Engine-Ip> But I Have No Ide

Have never done something like this before, and I'm unsure about the whole process from successfully serving the model to sending requests to it for inference. Is there any tutorial or example for it?

3 years ago
0 Hi Everyone, Is It Possible To Not Create A Copy Of A Dataset When Adding To Clearml? My Data Is Already In A Directory On The Clearml-Server Machine And I Do Not Want To Copy It, Just Add It To Clearml As Dataset.

I normally just upload the data to the ClearML server and then remove it locally from my machine but I understand that isn't what you want. A quick hack was the only thing I could come up with at the moment xd. Anyway you're welcome. Hope you find a solution.

3 years ago
0 I'Ll Just Ask This Question Again To Get Some Fresh Attention To This. Is There Any Way To Run A Pipeline Step Conditionally? E.G, Under Certain Condition, Execute The Step Otherwise Don'T?

I'm not using decorators. I have a bunch of function_steps followed by a normal task step, where I've passed a base_task_id.

I want to check the value of one of the functional steps, and if it holds true, I want to execute the task step otherwise I want the pipeline to end there, since the task step is the last one.

3 years ago
0 I Have A Function That Runs Normally. Its Job Is To Monitor A Specific Folder, And When I Execute The Script Locally It Works Fine. When I Make A Taskscheduler And Pass That Function To The Scheduler, Then Run Remotely, Then I Get An Error: Clearml Resul

Set up is on a single machine, I have a nas mounted where I'm watching a folder, if there are sufficient images, it should publish the data but since I was using start_remotely, the code was running somewhere else and couldn't access folder.

3 years ago
0 Would Appreciate Some Help. Getting This Error. Valueerror: Node Train_Model, Parameter '${Split_Dataset.Split_Dataset_Id}', Input Type 'Split_Dataset_Id' Is Invalid

AgitatedDove14 Can you help me with this? Maybe something like storing the returned values or something in a variable outside the pipeline?

3 years ago
0 Hey Guys, Sorry For The Rapid Fire Questions In The Past Few Days. I Have Another Issue Though. I Initially Ran A Task, Directly From A Repo. It Succesfully Installed The Requirements From The Requirements File In The Repo And Ran The Task Without Any Iss

My use case is that the code using pytorch saves additional info like the state dict when saving the model. I'd like to save that information as an artifact as well so that I can load it later.

3 years ago
3 years ago
0 So I Decided To Re-Create My Clearml Server, I

It's probably a cookie issue I agree.

4 years ago
Show more results compactanswers