Examples: query, "exact match", wildcard*, wild?ard, wild*rd
Fuzzy search: cake~ (finds cakes, bake)
Term boost: "red velvet"^4, chocolate^2
Field grouping: tags:(+work -"fun-stuff")
Escaping: Escape characters +-&|!(){}[]^"~*?:\ with \, e.g. \+
Range search: properties.timestamp:[1587729413488 TO *] (inclusive), properties.title:{A TO Z}(excluding A and Z)
Combinations: chocolate AND vanilla, chocolate OR vanilla, (chocolate OR vanilla) NOT "vanilla pudding"
Field search: properties.title:"The Title" AND text
Profile picture
VexedCat68
Moderator
60 Questions, 381 Answers
  Active since 10 January 2023
  Last activity one year ago

Reputation

0

Badges 1

371 × Eureka!
0 I Have No Prior Devops Experience. I'Ve Been Able To Set Up A Simple Continuous Training Setup Using Clearml. I Wanted To Ask What Should I Learn Which Would Help Me Move A Project From Mlops Level 0 To Level 1, And Then Level 2, Using Clear Ml. I Would A

Honestly anything. I tried looking up on youtube but There's very little material there, especially which is up to date. It's understandable given that ClearML is still in beta. I can look at courses / docs. I just want to be pointed in the right direction as to what I should look up and study

4 years ago
0 Would Appreciate Some Help. Getting This Error. Valueerror: Node Train_Model, Parameter '${Split_Dataset.Split_Dataset_Id}', Input Type 'Split_Dataset_Id' Is Invalid

These are the pipeline steps. Basically unable to pass these .

Some more of the error.

ValueError: Node train_model, parameter '${split_dataset.split_dataset_id}', input type 'split_dataset_id' is invalid
2021-12-30 16:22:00,130 - clearml.Repository Detection - WARNING - Failed auto-generating package requirements: exception SystemExit() not a BaseException subclass

3 years ago
0 Given This Pipeline Step, Is There Any Way To Get The Return Value Outside Of The Pipeline? Like Put Split_Dataset_Id In A Variable In The Main Pipeline File.

I want to maybe have a variable in the simple-pipeline.py, which has the value returned by split_dataset

3 years ago
0 Hi Guys, Needed A Bit Of Clarification. Every 15 Minutes, The Scheduler Prints, "Syncing Scheduler, Sleeping For 15 Minutes Until Next Sync". Can You Guide Me As To It Is Syncing With What?

Alright. Can you guide me on how to edit the task configuration object? Is it done via the UI or programatically? Is there a config file and can it work with any config file I create or is it a specific config file? Sorry for the barrage of questions.

3 years ago
0 Sorry For The Barrage Of Questions. I Can'T Seem To Figure Out How Best To Get A Python Script I Need To Run On An Agent. I Have An Agent Listening To The Default Queue And I Also Have The Script That I Need To Run On The Agent. Can You Guide Me On How To

So the api is something new for me. I've already seen the sdk. Am I misremembering sending python script and requirements to run on agent directly from the cli? Was there no such way?

3 years ago
0 Is There An Example Of Simple Pipeline Using Task Scheduler? I Want To Create A Simple Pipeline Where First A Folder Is Monitored Using A Taskscheduler, And Under Certain Conditions, Data Is Uploaded And Dataset Id Is Given To The Next Step In The Pipeli

Alright. Anyway I'm practicing with the pipeline. I have an agent listening to the queue. Only problem is, it fails because of requirement issues but I don't know how to pass requirements in this case.

3 years ago
0 I Keep Facing This Issue. I'M Trying To Set Up My Own Clearml Server Using This Tutorial.

When I try to access the server with the IP I set as CLEARML_HOST_IP, it looks like this. I set that IP to the ip assigned to me by the network

4 years ago
0 I'Ll Just Ask This Question Again To Get Some Fresh Attention To This. Is There Any Way To Run A Pipeline Step Conditionally? E.G, Under Certain Condition, Execute The Step Otherwise Don'T?

AnxiousSeal95 Basically its a function step return. if I do, artifacts.keys(), there are no keys, even though the step prior to it does return the output

3 years ago
0 I'Ll Just Ask This Question Again To Get Some Fresh Attention To This. Is There Any Way To Run A Pipeline Step Conditionally? E.G, Under Certain Condition, Execute The Step Otherwise Don'T?

I checked the value is being returned, but I'm having issues accessing merged_dataset_id in the preexecute_callback like the way you showed me.

3 years ago
0 I'M Publishing The Model Artifact After The Tensorflow Model Is Saved. But In The Ui, The Download Button Is Grayed Out. And Get_Local_Copy() Doesn'T Seem To Work Either, Help Would Be Appreciated.

For anyone facing a similar issue to mine and wanting the model to uploaded just like data is uploaded,

in the Task.init, set the output_uri = True.

This basically makes it use the default file server for clearml that you define in the clearml.conf file. Ty.

3 years ago
0 I'M Publishing The Model Artifact After The Tensorflow Model Is Saved. But In The Ui, The Download Button Is Grayed Out. And Get_Local_Copy() Doesn'T Seem To Work Either, Help Would Be Appreciated.

CostlyOstrich36 I'm observing some weird behavior. Before when I added tags to the model before publishing it, it worked fine and I could see the tags in the UI.

Now when I do it this way, tags aren't set. If I then run another code which gets the model, using ID, and then set tags, it worked fine. Let me share the codes.

3 years ago
0 How Do I Get Args Like Epochs To Show Up In The Ui Configuration Panel Under Hyperparameters? I Want To Be Able To Change Number Of Epochs And Learning Rate From Within The Ui.

Quick follow up question. Once I parse args, should they be directly available for i even enque the project for the first time or will i be able to access hyperparameters after running it once?

4 years ago
4 years ago
0 In The Case Where I'M Passing A Schedule_Fn To Add_Task In A Taskscheduler, How Do I Pass The Function Arguments?

def watch_folder(folder, batch_size):
count = 0
classes = os.listdir(folder)
class_count = len(classes)
files = []
dirs = []
for cls in classes:
class_dir = os.path.join(folder, cls)
fls = os.listdir(class_dir)
count += len(fls)
files.append(fls)
dirs.append(class_dir)

if count >= batch_size:
    dataset = Dataset.create(project='data-repo')
    dataset.add_files(folder)
    dataset.upload()
    dataset.final...
3 years ago
0 In The Case Where I'M Passing A Schedule_Fn To Add_Task In A Taskscheduler, How Do I Pass The Function Arguments?

there are other parameters for add_task as well, I'm just curious as to how do I pass the folder and batch size in the schedule_fn=watch_folder part

3 years ago
0 How Do I Get Args Like Epochs To Show Up In The Ui Configuration Panel Under Hyperparameters? I Want To Be Able To Change Number Of Epochs And Learning Rate From Within The Ui.

I'm using clear-ml agent right now. I just upload the task inside a project. I've used arg parse as well however as of yet, I have not been able find writable hyperparameters in the UI. Is there any tutorial video you can recommend that deals with this or something? I was following https://www.youtube.com/watch?v=Y5tPfUm9Ghg&t=1100s this one on youtube but I can't seem to recreate his steps as he sifts through his code.

4 years ago
Show more results compactanswers