Examples: query, "exact match", wildcard*, wild?ard, wild*rd
Fuzzy search: cake~ (finds cakes, bake)
Term boost: "red velvet"^4, chocolate^2
Field grouping: tags:(+work -"fun-stuff")
Escaping: Escape characters +-&|!(){}[]^"~*?:\ with \, e.g. \+
Range search: properties.timestamp:[1587729413488 TO *] (inclusive), properties.title:{A TO Z}(excluding A and Z)
Combinations: chocolate AND vanilla, chocolate OR vanilla, (chocolate OR vanilla) NOT "vanilla pudding"
Field search: properties.title:"The Title" AND text
Profile picture
CostlyOstrich36
Moderator
0 Questions, 4175 Answers
  Active since 10 January 2023
  Last activity 2 years ago

Reputation

0
0 Hi! We Run Self-Hosted Clear-Ml And Use Autoscaler For Aws, But We'Re Migrating To Google Cloud And I Can Not Find Any Documentation Is It Possible To Use Autoscaler For Gcp? Thanks.

Hi @<1523701491863392256:profile|VastShells9> , the GCP autoscaler is not available in the open source I'm afraid. Only in PRO licenses and up

one year ago
0 Hi! I'M Running A Clearml Task Remotely. It Runs A Python Machine Learning Evaluation Script In Repo A. The Script Depends On A Module Which Is Defined In Another One Of My Company'S Repos: Repo B. The File Structure Of The Module In Repo B Is As Follows

Hi @<1614069770586427392:profile|FlutteringFrog26> , if I'm not mistaken ClearML doesn't support running from different repoes. You can only clone one code repository per task. Is there a specific reason these repoes are separate?

one year ago
0 Hello Everyone, I Encountered The Following Error When I Tried To

If it's deployed by you, then try running clearml-init from the same machine the server is on. Doesn't matter if it's a cloud machine really

one year ago
0 How Do You Start An Agent On The Clearml Server With A Specific

Hi @<1654294828365647872:profile|GorgeousShrimp11> , you can set it with an env var - CLEARML_CONFIG_FILE
None

one year ago
0 Hey, Do Hyperdatasets Offer The Same Features With Tabular Data? Almost All Examples On The Docs Are On Image Datasets

Basically the same capabilities that are offered for the unstructured data - ability to register files, keep track and manage them with links and ability to query into all of their metadata and then connect it to the experiment as a query on the metadata inside different versions - basically giving you a feature store.

I am of course over simplifying as the HyperDatasets feature is an extremely powerful tool for managing unstructured data.

2 years ago
0 Two Simple Lineage Related Questions:

RoughTiger69 , regarding the dataset loading, we are actually thinking of adding it as another "hyper parameter" section, and I think the idea came up a few times in the last month, so we should definitely do that. The question is how do we support multiple entries (i.e. two datasets loaded)? Should we force users to "name" the dataset when they "get it" ?

Regrading cloning, we had a lot of internal discussions on it, "Parent" is a field on a Task, so the information can be easily stored, th...

3 years ago
0 Hi, Can I Ask How I Can Make Clearml-Datasets In Comparison With Pytorch Datasets/Dataloader? In Particular, Pytorch Dataloaders Would Be Able To Batch Pull And Then Preprocess Data Using Multi-Cpus, Feed It Into The Training Loop And Achieve As High Util

Hi SubstantialElk6 ,

That's an interesting idea. I think if you want to preprocess a lot of data I think the best would be using multiple datasets (each per process) or different versions of datasets. Although I think you can also pull specific chunks of dataset and then you can use just the one - I'm not sure about the last point.

What do you think?

3 years ago
2 months ago
0 Hi All, I Wanted To Know About Saving Datasets, We Want To Specify The Path To Gs By Default, As I Understand By Default It Uses The Path To File_Server? We Tried Sdk.Development.Default_Output_Uri =

Hi @<1603198163143888896:profile|LonelyKangaroo55> , you can change the value of files_server in your clearml.conf to control it as well.

one year ago
0 Hey Everyone! I’M Currently Trying To Set Up Hyperparameter Optimization With Clearml On A Base Experiment Using Hydra. I Got Everything Working From The Examples. However, The Child Experiments Started By The Optimization Don’T Seem To Actually Get The N

In the task hyper parameters section you have a section called Hydra. In that section there should be a configuration called _allow_omegaconf_edit_ , what is it set to?

3 years ago
0 What Is The Difference Between Model And Inputmodel?

I think that Model is used to do general actions as allowed by the SDK. InputModel is for an easier interface when working with the Task object directly.

What is your use case?

2 years ago
0 Hi Team,

@<1533257278776414208:profile|SuperiorCockroach75> , excuse my ignorance, but doesn't it depend on the output model i.e. the training run that created it?

2 years ago
0 Hi, We Have Recurring Disk Space Issues On Our Clearml Server (Drop Of Many Gb In A Few Days). After Some Analysis, We Noted

What do you mean by drop of many GB? Can you please elaborate on what happens exactly?

I know that elastic can sometimes create disk corruptions and requires regular backups..

3 years ago
0 I Guess One Experiment Is Running Backwards In Time

JitteryCoyote63 , are you on a self hosted server? It seems that the issue was solved for 3.8 release and I think should be released to the next self hosted release

3 years ago
0 Hi Guys, Is There A Way, Analogous To Using

Hi GiganticMole91 ,

I see that the storage settings are also available through environment variables, but I'm worried that the environment variables have already been parsed at that time.

I'm not sure I understand. Can you elaborate? How do you run in remotely? Do you raise an instance each time or are your instances persistent?

3 years ago
0 Hello! I'M Trying To Figure Out How To Deploy A Scheduled Pipeline. I Have A Sample Pipeline Here

It would work from your machine as well, but the machine needs to be turned on... like when an ec2 instance that is running.

2 years ago
0 Hello, I Have Clearml Autoscaler Setup. Previously, When A New Task Comes Up, An Already Running Worker (If There Is), Will Take It, Apply The New Commit And Run The Task. Now, I Get An Error. So, I Can'T Run A Task On An Already Running Worker. It Has To

Hi @<1523704674534821888:profile|SourLion48> , making sure I understand - You push a job into a queue that an autoscaler is listening to. A machine is spun up by the autoscaler and takes the job and it runs. Afterwards during the idle time, you push another job to the same queue, it is picked up by the machine that was spun up by the autoscaler and that one will fail?

one year ago
0 Hello, Is It Commo That

It can take some time if the file is very large or if the folder is very large. This can also depend on the connectivity. If the folder is very large please consider the fact that zipping it also can be resource demanding.

How long has it been hanging and how large is it the folder?

3 years ago
0 How Can I Check Which Experiments Are Using The Most Metrics? And Does Archiving Those Experiments Save Some Space?

@<1581454875005292544:profile|SuccessfulOtter28> , I don't think there is such a capability currently. I'd suggest opening a GitHub feature request for this.

one year ago
0 Hey All, There Is A Way To Compare All The Experiments In A Project

Hi @<1523706700006166528:profile|DizzyHippopotamus13> , you can simply do it in the experiments dashboard in table view. You can rearrange columns, add custom columns according to metrics and hyper parameters. And of course you can sort the columns

2 years ago
Show more results compactanswers