Examples: query, "exact match", wildcard*, wild?ard, wild*rd
Fuzzy search: cake~ (finds cakes, bake)
Term boost: "red velvet"^4, chocolate^2
Field grouping: tags:(+work -"fun-stuff")
Escaping: Escape characters +-&|!(){}[]^"~*?:\ with \, e.g. \+
Range search: properties.timestamp:[1587729413488 TO *] (inclusive), properties.title:{A TO Z}(excluding A and Z)
Combinations: chocolate AND vanilla, chocolate OR vanilla, (chocolate OR vanilla) NOT "vanilla pudding"
Field search: properties.title:"The Title" AND text
Profile picture
AgitatedDove14
Moderator
48 Questions, 8049 Answers
  Active since 10 January 2023
  Last activity 5 months ago

Reputation

0

Badges 1

25 × Eureka!
0 Hello! I Get The Following Error In Results->Console After A Task Is Sent For Remote Execution (Using Sdk):

I have an idea, can you try with:
task = Task.init(..., reuse_last_task_id=False)I have a suspicion it starts the Tasks in parallel, and the "reuse_last_task_id" causes them to "reuse the same task locally" which makes them overwrite the configuration of one another.

2 years ago
0 Hello! I Get The Following Error In Results->Console After A Task Is Sent For Remote Execution (Using Sdk):

What's strange is that the remote jobs, as soon as they are launched, if I compare their configs while in state pending, they have the right all different configs, but later, while running,

Wait I think I found it, since usuallyu the case with hydra you configure everything from overrides / config, when launched remotely it looks at it by default. But with the launch plugin it should be overwritten with the Task
` task = Task.init(...)
task.set_parameter(name="Hydra/_allow_omegaconf_ed...

2 years ago
0 Yesterday I Executed An Experiment In Our Hosted Clearml Cluster. After The Experiment Was Finished, We Got An Aws Guard Duty Notification About Suspicious Outbound Traffic From The Ec2 That Executed The Job. It Looks Like The Tag Being Used Is Hardcoded

It looks like the tag being used is hardcoded to 1.24-18. Was this issue identified and fixed in later versions?

BoredHedgehog47 what do you mean by "hardcoded 1.24-18" ? tag to what I think I lost context here

2 years ago
0 Or Is It Just The Ubuntu Official Image

Are you suggesting the default "ubuntu:18.04" is somehow contaminated ?

This is an official Ubuntu container (nothing to do with ClearML), this is Very Very odd...

2 years ago
0 Hi Everyone, I'M Using The

So as you say, it seems hydra kills these

Hmm let me check in the code, maybe we can somehow hook into it

2 years ago
0 Hi Everyone, I'M Using The

AttractiveCockroach17 can I assume you are working with the hydra local launcher ?

2 years ago
0 Hi, I’M Trying Out Clearml Pipelines From Decorators, And I’M Encountering A Few Problems I Don’T Know How To Solve.

I understand I can change the docker image for a component in the pipeline, but for the

it isnโ€™t possible.

you can always to Task.current_task.connect() from the pipeline function itself, to connect more configuration arguments you basically add via the function itself, all the pipeline logic function arguments become pipeline arguments, it's kind of neat ๐Ÿ™‚ regrading docker, the idea is that you use a very basic python docker (the default for services) queue for all...

one year ago
0 Hi Folks, One Question: I Have A Script That Looks Like:

I was expecting the remote experiment to behave similarly, why do I need to import pandas there?

The only problem os that the remote code did not install pandas , once the package is there we can read the artifacts
(this is in contrast to the local machine where pandas is installed and so we can create/read the object)
Does that make sense ?

2 years ago
0 Hi Folks, One Question: I Have A Script That Looks Like:

wouldn't it be possible to store this information in the clearml server so that it can be implicitly added to the requirements?

I think you are correct, and if we detect that we are using pandas to upload an artifact, we should try and make sure it is listed in the requirements
(obviously this is easier said than done)

And if instead I want to force "get()" to return me the path (e.g. I want to read the csv with a library that is not pandas) do we have an option for that?

Yes, c...

2 years ago
0 Hi All, We Have A Weird Inconsistency. We Have A Clearml Server Installed On-Prem And Started Playing With It. Using The Dataset.Create Command And The Subsequent Add_Files, And Upload Commands I Can See The Upload Action As An Experiment But The Data Is

Using the dataset.create command and the subsequent add_files, and upload commands I can see the upload action as an experiment but the data is not seen in the Datasets webpage.

ScantCrab97 it might be that you need the latest clearml package installed on the client end (as well as the new server with the UI)
What is your clearml package version ?

2 years ago
0 Suggestion

SolidSealion72 this makes sense, clearml deletes artifacts/models after they are uploaded, so I have to assume these are torch internal files

2 years ago
0 Hi Guys, I Am Having Some Trouble Running Some Training Scripts With The Agent Functionality:

Martin, if you want, feel free to add your answer in the stackoverflow so that I can mark it as a solution.

Will do ๐Ÿ™‚ give me 5

2 years ago
0 Hi Folks, Any Idea Why I Am Getting This Strange Error From Clearml-Data:

HurtWoodpecker30 could it be you hit a limit of some sort ?

2 years ago
0 Hi There, I Used

JitteryCoyote63 Are you calling clearml-task with --skip-task-init ?

2 years ago
0 Hi all :wave:! I got a problem regarding Grafana/Prometheus. When I deploy a model with clearml-serving and I add metrics like this: `clearml-serving --id *** metrics add --endpoint slm_POC --variable-scalar beds=0,1,5,10,50 bath=0,1,5,10,50 y=0,100000,50

Hmm so yes that is true, if you are changing the bucket values you will have to manually also adjust it in grafana. I wonder if there is a shortcut here, the data is stored in Prometheus, and I would rather try to avoid deleting old data, Wdyt?

one year ago
0 Hi! Is There A Way To Export The Credentials Of The Aws Account Only During The Creation Of The Docker? I Don’T Want Every User In My Team To Know The Credentials To Access S3 Buckets. I Just Want Them To Be Able To Write In The Bucket Without The Credent

it would be clearml-serverโ€™s job to distribute to each user internally?

So you mean the user will never know their own S3 access credentials?
Are those credentials unique per user or once"hidden" for all of them?

2 years ago
0 Hello Everyone.I Have No Idea Why Clearml-Serving Inference Server Tries To Get Model From That Url(Pic 1), But In Clearml Ui I Have Correct Url(Pic 2). Could You Help Me With This?

ComfortableShark77 it seems the clearml-serving is trying to Upload data to a different server (not download the model)
I'm assuming this has to do with the CLEARML_FILES_HOST, and missing credentials. It has nothing to do with downloading the model (that as you posted, will be from the s3 bucket).
Does that make sense ?

2 years ago
0 Hi! Is There Any Reason Why Integer/Float Values Are Casted To String When Connecting Arguments Dictionary To Task And Then Retrieve Them Using

GiganticTurtle0
I think that what you are looking for is:
param_dict = {'key': 1234} task.connect(param_dict, name='general')Notice that when this code runs manually (i.e. not by the agent), the dict is stored on "general" parameter section of the Task.
But when the code is executed by the Agent, the opposite happens and the parameters from the "general" section of the Task or put back into the param_dict , here the casting is done based on the type of the original values.
Generall...

3 years ago
0 I Am Running Trains=0.16.4 Python==3.7.5 , And Notice That The "Log" Page Sometimes Didn'T Capture The Console Log From My Program. Is This A Known Issue, Anyone Have Experienced Similar Behavior?

The log is missing, but the Kedro logger is print toย sys.stdout in my local terminal.

I think the issue night be it starts a new subprocess, and that subprocess is not "patched" to capture the console output.
That said if an agent is running the entire pipeline, then everything is logged from the outside, so whatever is written to stdout/stderr is captured.

3 years ago
0 Can Anyone Point Me To Web Ui Source Code Out, I’M Wondering How Can I Customizing Web App Ui A Little Bit

MagnificentPig49 that's a good question, I'll ask the guys ๐Ÿ™‚
BTW, I think the main issues is actually making sure there is enough documentation on how to compile it...
Anyhow I'll update here

4 years ago
0 "Clearml-Data Sync --Folder ." Doesn'T Work

Hi @<1631102016807768064:profile|ZanySealion18>
sorry missed that one

The cache doesn't work, it attempts to download the dataset every time.

just making sure the dataset itself contains all the files?

Once I used clearml-data add --folder * CLI everything works correctly (though all files recursively ended up in the root, I had luck all were named differently).

Not sure I follow here, is the problem the creation of the dataset of fetching it? is this a single version or multi...

3 months ago
0 Hi Folks, A Question Regarding The Clearml-Agent With K8S Glue. In The Agents We Mount An Nfs Volume So That Some Artifacts And Data Would Be Available For Training. I Have Seen That The K8S Glue Runs As Root (I Guess To Be Able To Spawn New Pods?), But

For example, for some of our models we create pdf reports, that we save in a folder in the NFS disk

Oh, why not as artifacts ? at least you will be able to access from the web UI, and avoid VFS credential hell ๐Ÿ™‚

Regrading clearml datasets:
https://www.youtube.com/watch?v=S2pz9jn26uI

2 years ago
0 Hi, Similar To Task.Set_Offline(True), Is There A Way To Simulate An Execution In An Agent? (For Testing Purposes)

Because it lives behind a VPN and github workers donโ€™t have access to it

makes sense
If this is the case, I have to admit that combining offline-mode and remote execution makes sense, no?

2 years ago
Show more results compactanswers