Examples: query, "exact match", wildcard*, wild?ard, wild*rd
Fuzzy search: cake~ (finds cakes, bake)
Term boost: "red velvet"^4, chocolate^2
Field grouping: tags:(+work -"fun-stuff")
Escaping: Escape characters +-&|!(){}[]^"~*?:\ with \, e.g. \+
Range search: properties.timestamp:[1587729413488 TO *] (inclusive), properties.title:{A TO Z}(excluding A and Z)
Combinations: chocolate AND vanilla, chocolate OR vanilla, (chocolate OR vanilla) NOT "vanilla pudding"
Field search: properties.title:"The Title" AND text
Profile picture
AgitatedDove14
Moderator
49 Questions, 8126 Answers
  Active since 10 January 2023
  Last activity one year ago

Reputation

0

Badges 1

25 × Eureka!
0 Bug?

Yep the automagic only kick in with Task.init... The main difference and the advantage of using a Dataset object is the underlying Task resides in a specific structure that is used when searching based on project/name/version, but other than that, it should just work

3 years ago
0 My Agent (Running On Gcp In Docker Mode) Is Having Trouble With Git Fetch --All. I'M Using Ssh For Authentication, However, Known_Hosts Doesn'T Seem To Be Passed To The Docker So It Prompts For Authentification/Fingerprint. Any Ideas?

Hi @<1631102016807768064:profile|ZanySealion18>

I'm using SSH for authentication, however, known_hosts doesn't seem to be passed to the docker so it prompts for authentification/fingerprint. Any ideas?

Hmm it is supposed to automatically mount your ~/.ssh folder into the docker to solve for that.
First try to set force_git_ssh_protocol: true
None
If that does not he...

one year ago
0 Hello People

Hmm not sure, try the latest anyhow 🙂

3 years ago
0 Can Someone Confirm That

instead of the one that I want or the one of the env which it is started from.

The default is the python that is used to run the agent.
agent.ignore_requested_python_version = true agent.python_binary = /my/selected/python3.8

4 years ago
0 Are There Any Particular System Dependencies Needed To Enable

Oh that is odd. Is this reproducible? @<1533620191232004096:profile|NuttyLobster9> what was the flow that required another task.init?

one year ago
0 Hi, How Can I Remove A Tag From A Task Via Code In A Non-Barbaric Way?

Ooops 😞
task.get_tags()
task.set_tags()

5 years ago
0 <image>

No sure what O'm seeing here

4 years ago
3 years ago
0 Hi! I Need Help Debugging The Following Issue Please. I'M Training A Cnn And Plotting The Confusion Matrices For Train And Val In Each Epoch. When I Get To Epoch 101, The Ui Kind Of Breaks..It Starts Showing Me The Images For Epoch 1. When I Right Click O

I was using clearml == 0.17.5 and I also had this issue

I think it was introduced when we moved to subprocess reporting, with 0.17.5
You can disable it with the following in clearml.conf:
sdk.development.report_use_subprocess = false

4 years ago
0 Anyone Deployed Trains On Azure, I Am Interested To Know About Your Experience.

For setting trains-server I would recommend the docker-compose, it is very easy to setup, and you just need a single fixed compute instance, details https://github.com/allegroai/trains-server/blob/master/docs/install_linux_mac.md With regards to the "low prio clusters", are you asking how they could be connected with the trains-agent or if running code that uses trains will work on them?

5 years ago
0 Hi, I Am Trying To Use The Config Values From A Experiment, But

Hi SkinnyPanda43
Are you trying to access the same Task or an external one ?

4 years ago
5 years ago
0 I Have A Little Bit Of Code That Goes Like:

great 🙂
two things:
I'm not sure argparse supports dict as a type (I mean it will take anything but I'm not sure it will parse your arguments as dict) I know there was an issue with argparsing, but I think it was solvedbtw: Basically the way clearml-agent works, it does not actually pass the arguments in commandline but directly to the argparser at runtime
What happens if you clone the Task (the one with Args showing and without the explicit task.connect(_args) and send it to the age...

4 years ago
0 Hello, I'M Running A Ml Training Using

Hi FancyWhale93 you can disable the auto model uploading with
@PipelineDecorator.component(..., auto_connect_frameworks={'pytorch': False}) def step(): pass

3 years ago
0 Quest About

clearml-task

 seems does not allow me passing the 

run

 argument without value

EnviousStarfish54 did you try --args run=True
I'm assuming run is a boolean of a sort ?

a. The submitted job would automatically download data from internal data repository, but it will be time consuming if data is re-downloaded every time. Does ClearML caching the data somewhere?

What do you mean by the agent will download the data ? are you referring to Dataset ?

4 years ago
0 Hi, Can We Upload Our Project Repository To Trains Server? If We Can, How Should We Do? I Know When We Write "Task.Init()", It Uploads Our Experiment Into Server, But It Also Run The Experiment. However, I Want To Upload All My Experiments In Draft Status

MysteriousBee56 I would do Task.create()
you can get the full Task internal representation with task.data
Then call task._edit(script={'repo': ...}) to edit/update all the Task entries.
You can check the dull details of the task object here: https://github.com/allegroai/trains/blob/master/trains/backend_api/services/v2_8/tasks.py#L954
BTW: when you have a sample script working, consider PR-ing it, I'm sure it will be useful for others 🙂 (also a great way to get us involved with debuggin...

5 years ago
0 Hi, Guys! Thank You A Lot For Your Great Software, But I'Ve Got A Problem. I Have Got Two Remotes: Gitlab And Gitea. The Branch From Which I Run The Code Is Upstreamed With Gitea. However, In The Clearml Experiment, Gitlab Repository Is Automatically Sele

How does ClearML select reference branch? Could it be that ClearML only checks "origin" branch?

Yes 😞 I think we can quickly fix that, I'm just trying to realize if there are down sides to running "git ls-remote --get-url" without origin

3 years ago
0 Hi All! Question Around Resource Management Using

Oh that makes sense, This depends on how you setup the clearml k8s glue, (becuase the resource allocation is done by k8s) a good hack to limit the number of containers per GPU is to set a RAM limitation per pod, then k8s will know to limit the number of pods on the same GPU machine,
wdty?

3 years ago
0 Does The New 2.0 Helm Charts (App Ver 1.1.0) Not Support Nfs?

I think this is the only mount you need:

Data persisted in every Kubernetes volume by ClearML will be accessible in /tmp/clearml-kind folder on the host.

SuccessfulKoala55 is this correct ?

4 years ago
0 Hi I'M Using Latest Versions And Experiencing A Bug. I Created A Controller In A New Project. Just After That I Clone A Task And Give It The New Project Name To Clone Into. It Fails To Create The Task: Code:

Notice the error code:
Action failed <400/401: tasks.create/v1.0 (Invalid project id: id=first_attempt)>If that is the case, The project ID is incorrect (project id is not the project name)

5 years ago
0 Is It Possible To Add A Callback For A Pipeline From A Step?

Ephemeral Dataset, I like that! Is this like splitting a dataset for example, then training/testing, when done deleting. Making sure the entire pipeline is reproducible, but without storing the data long term?

4 years ago
0 Is There A Way To Configure The File Server To Use Minio Storage, Or Does Every Individual User Have To Configure Their Own Minio Credentials?

will my datasets be stored on the same machine that hosts the clearml server?

By default yes, they will be stored to the files-server (but you can change it, this is an argument for both the CLI and the python interface)

3 years ago
0 Hello People

so I guess this could be one reason to start about thinking upgrading ....

Wait you mean the clearml-server ? (there is no reason not to upgrade the python package)

3 years ago
0 Hi All! I Have A Question About Pipelines. My Pipeline Consists Of Several Steps:

GrotesqueDog77 this should just work, decorate the functions with @PipelineDecorator.component and call the functions one after the other
paths = step_one() step_two(paths)ClearML will make sure it serializes the strings and pass them to step two (of course step two should actually run on a machine with access to the same folder, but this is another issue 🙂 )

3 years ago
0 Currently, To Provide Ssh Access To The Docker Images For A Task,

By default SSH server is not running in a lot of scenarios (k8s for example, Windows, MacOS)...

4 years ago
0 Hi, I Am Getting An Error While Running

Hi JitteryCoyote63
could you check if the problem exists in the latest RC?
pip install clearml==1.0.4rc1

4 years ago
0 Hi All, I Am Starting To Use Clearml-Agent. Run It With

I am creating this user

Please explain, I think this is the culprit ...

4 years ago
0 Hi There - I Am Attempting To Use The Hp Optimization Feature, But Keep Getting The Following Error:

CharmingBeetle38 try adding "General/" before the arguments. This means batch_size becomes General/batch_size. This is only because we are accessing the parameters externally, when the task is executed it is resolved automatically

5 years ago
Show more results compactanswers