Examples: query, "exact match", wildcard*, wild?ard, wild*rd
Fuzzy search: cake~ (finds cakes, bake)
Term boost: "red velvet"^4, chocolate^2
Field grouping: tags:(+work -"fun-stuff")
Escaping: Escape characters +-&|!(){}[]^"~*?:\ with \, e.g. \+
Range search: properties.timestamp:[1587729413488 TO *] (inclusive), properties.title:{A TO Z}(excluding A and Z)
Combinations: chocolate AND vanilla, chocolate OR vanilla, (chocolate OR vanilla) NOT "vanilla pudding"
Field search: properties.title:"The Title" AND text
Profile picture
SoreSparrow36
Moderator
4 Questions, 47 Answers
  Active since 21 July 2023
  Last activity one month ago

Reputation

0

Badges 1

22 × Eureka!
0 Votes
19 Answers
1K Views
0 Votes 19 Answers 1K Views
one year ago
0 Votes
10 Answers
176 Views
0 Votes 10 Answers 176 Views
one month ago
0 Votes
3 Answers
395 Views
0 Votes 3 Answers 395 Views
is there somewhere I can track upcoming releases by any chance? Trying to plan an upgrade of our services. namely I'm wondering if I need to continue using m...
3 months ago
0 Votes
5 Answers
1K Views
0 Votes 5 Answers 1K Views
How can I control the ~/clearml.conf file being used by agent-services in the docker-compose stack for clearml-server ? namely, if I enqueue a task, I notice...
one year ago
0 Can Anyone Recommend A Good Workflow For

Oh yes. I see. Yeah, no ML here actually (doing the testing infra of endpoints), but certainly when there is its an issue.

How does clearml session avoid it? I guess only if autoscaling is used (one worker one machine)?

one year ago
0 Can Anyone Recommend A Good Workflow For

I'm guessing this is done through code-server?

I'm currently rolling a JupyterHub instance (multiuser, with codeserver inside) on the same machine as clearml-server. That’s where tasks are executed etc. so, all browser dev env.

It sounds like there’s an option to basically bypass this latter step and just use clearml’s credentialing to accomplish much the same thing? Am I understanding clearml-session correctly?

one year ago
one year ago
0 I Just Encountered A Really Frightening Bug. Best I Can Explain What Happened Was This: Data Scientist Created New Venv, Installed Clearml==1.11.0 Instead Of Clearml[S3]==1.11.1, And Upon Re-Running A Pipeline From Cli, The Entire Project "Disappeared" (W

the dataset, task, and pipeline were under the same project name. i'm seeing what happens if the dataset project name was different ( f"{project_name}_data" ). which project would get deleted... the dataset one or the project of the task that kicked it off?

and the answer is...
the project is preserved, the dataset's project hidden.

so ... empty dataset names due to a small typo in parameter override + the choice for the dataset to have the same project name as the task that created it (...

one year ago
0 I Just Encountered A Really Frightening Bug. Best I Can Explain What Happened Was This: Data Scientist Created New Venv, Installed Clearml==1.11.0 Instead Of Clearml[S3]==1.11.1, And Upon Re-Running A Pipeline From Cli, The Entire Project "Disappeared" (W

i think we may have found the frankenbug?

the argument to the dataset name was not being overridden correctly (mistyped), so the default value of an empty string (instead of a placeholder like "CHANGE_ME") in the parent task caused the dataset to basically get created with an empty name, and somehow that hid the whole project, despite hundreds of existing tasks in it.

and no way to un-hide it as far as I can tell?

one year ago
0 Can Anyone Recommend A Good Workflow For

Oh neat! I want to take a look at this. Only a few more weeks at the client but it’d be nice to reduce the complexity of the software stack if I can before handoff.

Can you please elaborate on the latter point? My jupyterhub’s fully containerized and allows users to select their own containers (from a list i built) at launch, and launch multiple containers at the same time, not sure I follow how toes are stepped on.

one year ago
0 I Just Encountered A Really Frightening Bug. Best I Can Explain What Happened Was This: Data Scientist Created New Venv, Installed Clearml==1.11.0 Instead Of Clearml[S3]==1.11.1, And Upon Re-Running A Pipeline From Cli, The Entire Project "Disappeared" (W

then back to CLI, updated the pipeline to point the tasks to the new queue. run it, shows up in the UI (same container as default worker, just replicated w a new docker-compose and CMD to point to the new queue).

one year ago
0 Hello! I Created A

the clearml github, search for a file named cleanup service dot py (or something to that effect)

one year ago
0 Hey, the <https://clear.ml/docs/latest/docs/references/api/#request-format|api reference> says that the url should be ```https://&lt;base_url&gt;/auth.login``` but to make it actually work I have to do ```https://&lt;base_url&gt;/api/v1.0/auth.login``` Th

Weird . I recently implemented a function that talked to this exact endpoint and found it had to exclude the version and api paths . Is there some sort of redirect that happens?

10 months ago
0 Hello

👀 following.
I have much the same issue, and it's mission-critical that I resolve it soon.

one year ago
0 Hello! I Created A

Might be under examples

one year ago
0 I Am Still Going Through All The Docs And Intro Videos … But: Is The Only Way To Create A New Experiment To Run The Script That Contains The Experiment At Least Once? I Wonder About This B.C. Most Of What I Want To Run Are Quite Long Jobs, So Even Running

Yup if you scroll through the logs in the console, near the top (post config dump), you’ll see a git clone and checkout to the specific hash.

PS You can actually change this parameter in an experiment’s configuration if it is in draft mode.

one year ago
0 Hello! I Created A

credentials for the server to do things with s3 will be in /opt/clearml/apiserver.conf.

one year ago
0 Hi All

oh i see. you're talking about the agent-services, not a separate agent in a container.
yup, I've got the same thing going there.
fwiw...
for me, HOST_IP is 0.0.0.0 and the other "HOSTS" env vars don't contain "http" in them.
and my server is publicly reachable, not sure if that matter either.
image

one year ago
0 Hey All, Very New To Clearml! I Am Trying To Design An Hpo Setup Using The Optuna Configuration, And I'M Working On Getting My Template Trainer Set Up. The Issue I'M Having Is It'S Unclear To Me How To Define One Of My Hyperparameters Whose Size Is Dynami

youre basically asking to sample from a distribution where not all parameters are mutually independent .

the short answer is no- this is not directly supported . optuna needs each hyperparam to be independent, so its up to you to handle the dependencies between parameters yourself unfortunately .

your solution of defining them independently and then using num_layers to potentially ignore other parameters is a valid one .

one month ago
0 Hi All

hm, you should be able to hit None if docker networking is working properly. it shouldn't need to go through the internet to get back to your machine.

one year ago
0 I Just Encountered A Really Frightening Bug. Best I Can Explain What Happened Was This: Data Scientist Created New Venv, Installed Clearml==1.11.0 Instead Of Clearml[S3]==1.11.1, And Upon Re-Running A Pipeline From Cli, The Entire Project "Disappeared" (W

yeah let's step through this, i'm having her execute these steps as we speak.

create a task with the new project name. its created as a draft. can see it in the UI under the new project.

pipeline script is updated with new project name for. execute script to create pipeline. now see in UI under this new project name. nothing hidden.

the pipeline is running. when the queue is default (only serviced by one container with agent in it ( clearml-agent==1.5.2 ). abort it. everything is still ...

one year ago
0 Hi Everybody! I'M Running An Example Pipeline From A Web Ui. I Notice Very Strange Behavior. After The First Local Run, I Can Create A New Run And Pass Initialization Parameters There, But After A Successful Run, I Lose The Ability To Create New Runs With

tasks that create pipelines feels like a hack and i found they dont show up in the UI (have to use the link in the console).

I've found that sometimes i need to right click "Run" a couple of times before the parameters are filled in properly.

one year ago
0 I Just Encountered A Really Frightening Bug. Best I Can Explain What Happened Was This: Data Scientist Created New Venv, Installed Clearml==1.11.0 Instead Of Clearml[S3]==1.11.1, And Upon Re-Running A Pipeline From Cli, The Entire Project "Disappeared" (W

one note is that it happened after I tried deploying a set of workers to a new queue, which she tried to use to run the tasks in parallel instead of our default queue which is only serviced by one worker (a container i built)

one year ago
0 Hello! I Created A

I think you’d have to run the cleanup service. That’s what seems to be what is controlling deletion based on archived status and some other temporal filters

one year ago
0 I Just Encountered A Really Frightening Bug. Best I Can Explain What Happened Was This: Data Scientist Created New Venv, Installed Clearml==1.11.0 Instead Of Clearml[S3]==1.11.1, And Upon Re-Running A Pipeline From Cli, The Entire Project "Disappeared" (W

the project wasn't hidden before. I'm aware of the pipeline tasks being hidden, that makes sense for organization. but the actual project itself as an entirety has a ghost icon.

she created a new project and started working in there, it was visible in the UI... and just now it disappeared again. it's kind of like running the pipeline makes it disappear.

one year ago
0 Hey All, Very New To Clearml! I Am Trying To Design An Hpo Setup Using The Optuna Configuration, And I'M Working On Getting My Template Trainer Set Up. The Issue I'M Having Is It'S Unclear To Me How To Define One Of My Hyperparameters Whose Size Is Dynami

you could also take the route of NOT specifying num_layers, and instead write your own code to create a set of viable layer designs to choose from and pass that as a parameter, so optuna selects from a countable set instead of suggesting integer values .

the downside of this is the lack of gradient information in the optimization process

one month ago
0 Hi, I Noticed That When I Commit Changes And Not Push Them And Try To Run A Job I Am Getting

if you commit but do not push, the metadata tells clearml that it needs to pull a non-existant commit. any changes you made on top may be saved as a diff, but they'd fail to apply.

for clearml to work on un-pushed commits, it'd have to wait for a push to register a new diff target, which can become a problem (what if you have multiple remotes? which one will it wait for?) so rather, it assumes it can access the most recent commit from your remote repo, and records this as the "base" upon whi...

one year ago
Show more results compactanswers