Examples: query, "exact match", wildcard*, wild?ard, wild*rd
Fuzzy search: cake~ (finds cakes, bake)
Term boost: "red velvet"^4, chocolate^2
Field grouping: tags:(+work -"fun-stuff")
Escaping: Escape characters +-&|!(){}[]^"~*?:\ with \, e.g. \+
Range search: properties.timestamp:[1587729413488 TO *] (inclusive), properties.title:{A TO Z}(excluding A and Z)
Combinations: chocolate AND vanilla, chocolate OR vanilla, (chocolate OR vanilla) NOT "vanilla pudding"
Field search: properties.title:"The Title" AND text
Profile picture
SoreSparrow36
Moderator
2 Questions, 39 Answers
  Active since 21 July 2023
  Last activity 11 months ago

Reputation

0

Badges 1

16 × Eureka!
0 Votes
19 Answers
649 Views
0 Votes 19 Answers 649 Views
11 months ago
0 Votes
5 Answers
562 Views
0 Votes 5 Answers 562 Views
How can I control the ~/clearml.conf file being used by agent-services in the docker-compose stack for clearml-server ? namely, if I enqueue a task, I notice...
12 months ago
0 Hi Guys, I'M Trying To Deploy An Image Segmentation Model, So I Expect That The Front-End Of The Endpoint Will Allow Users To Upload Images, Get Their Segmented Images & Option To Annotate The Images If The Results Are Not Good Enough. My Question Is: How

If you can hit the endpoint with curl, you for sure can hook it up to many frontend frameworks.

Personal recs: gradio, streamlit

Abstract the interaction into a function call, and wrap it all in some UI elements using python.

11 months ago
0 Hi All

oh i see. you're talking about the agent-services, not a separate agent in a container.
yup, I've got the same thing going there.
fwiw...
for me, HOST_IP is 0.0.0.0 and the other "HOSTS" env vars don't contain "http" in them.
and my server is publicly reachable, not sure if that matter either.
image

11 months ago
0 Is There Any Documentation From Clearml On Best Practices For Mounting/Using External Ebs Volumes For The Clearml Server? We Would Like To Mount An External Ebs Volume To The

my approach was to spin up an EC2 and run the deployment there from within the EBS volume mount.

I symlinked /opt/clearml to /mnt/xvda/clearml to minimize docker-compose changes. been working out fine so far.

with aws-cdk, the deployment steps can be automated (format the volume, clone a repo with the config, etc). I can link you to a resource that may help with that if you're interested.

11 months ago
0 Hi All

I ran into something similar during deployment. Hopefully this helps with your debugging: if the agent was launched separately from the rest of the stack, it may not have proper docker-DNS resolution to None . (e.g. if in the same docker-compose, perhaps you didnt add the backend network field, or if it was launched separately through docker run without an explicit external network defined)

if the agent's on the same machine, try docker network connect to add...

11 months ago
0 I Am Still Going Through All The Docs And Intro Videos … But: Is The Only Way To Create A New Experiment To Run The Script That Contains The Experiment At Least Once? I Wonder About This B.C. Most Of What I Want To Run Are Quite Long Jobs, So Even Running

Yup if you scroll through the logs in the console, near the top (post config dump), you’ll see a git clone and checkout to the specific hash.

PS You can actually change this parameter in an experiment’s configuration if it is in draft mode.

10 months ago
0 Hi Everybody! I'M Running An Example Pipeline From A Web Ui. I Notice Very Strange Behavior. After The First Local Run, I Can Create A New Run And Pass Initialization Parameters There, But After A Successful Run, I Lose The Ability To Create New Runs With

tasks that create pipelines feels like a hack and i found they dont show up in the UI (have to use the link in the console).

I've found that sometimes i need to right click "Run" a couple of times before the parameters are filled in properly.

11 months ago
11 months ago
0 Hello! I Created A

the clearml github, search for a file named cleanup service dot py (or something to that effect)

10 months ago
0 I Am Still Going Through All The Docs And Intro Videos … But: Is The Only Way To Create A New Experiment To Run The Script That Contains The Experiment At Least Once? I Wonder About This B.C. Most Of What I Want To Run Are Quite Long Jobs, So Even Running

For reproducibility, it kind of makes sense though. The existence of the file is contingent on the worker cloning the source code. I'm sure things can be done to maintain state differently but I personally adapted to the git-based workflow for managing files pretty quickly.

though yes I will admit I had the same thought first: why must I run it each time?

Beware: squash merges will ruin the ability to reproduce the experiment at that time since the git commit will be lost (presuming th...

10 months ago
0 I Am Still Going Through All The Docs And Intro Videos … But: Is The Only Way To Create A New Experiment To Run The Script That Contains The Experiment At Least Once? I Wonder About This B.C. Most Of What I Want To Run Are Quite Long Jobs, So Even Running

you can put task.execute_remotely() to create it in draft mode. I've taken to configuring defaults to run things very quickly just in case i forget though (e.g. placeholder string for dataset, bail out early if not changed… or just do one epoch on a small subset of samples, etc).

10 months ago
0 Hello! I Created A

I think you’d have to run the cleanup service. That’s what seems to be what is controlling deletion based on archived status and some other temporal filters

10 months ago
0 Hello! I Created A

credentials for the server to do things with s3 will be in /opt/clearml/apiserver.conf.

10 months ago
0 Hello! I Created A

Might be under examples

10 months ago
0 I Just Encountered A Really Frightening Bug. Best I Can Explain What Happened Was This: Data Scientist Created New Venv, Installed Clearml==1.11.0 Instead Of Clearml[S3]==1.11.1, And Upon Re-Running A Pipeline From Cli, The Entire Project "Disappeared" (W

the dataset, task, and pipeline were under the same project name. i'm seeing what happens if the dataset project name was different ( f"{project_name}_data" ). which project would get deleted... the dataset one or the project of the task that kicked it off?

and the answer is...
the project is preserved, the dataset's project hidden.

so ... empty dataset names due to a small typo in parameter override + the choice for the dataset to have the same project name as the task that created it (...

11 months ago
0 Can Anyone Recommend A Good Workflow For

Oh neat! I want to take a look at this. Only a few more weeks at the client but it’d be nice to reduce the complexity of the software stack if I can before handoff.

Can you please elaborate on the latter point? My jupyterhub’s fully containerized and allows users to select their own containers (from a list i built) at launch, and launch multiple containers at the same time, not sure I follow how toes are stepped on.

11 months ago
0 Can Anyone Recommend A Good Workflow For

I'm guessing this is done through code-server?

I'm currently rolling a JupyterHub instance (multiuser, with codeserver inside) on the same machine as clearml-server. That’s where tasks are executed etc. so, all browser dev env.

It sounds like there’s an option to basically bypass this latter step and just use clearml’s credentialing to accomplish much the same thing? Am I understanding clearml-session correctly?

11 months ago
0 Can Anyone Recommend A Good Workflow For

Oh yes. I see. Yeah, no ML here actually (doing the testing infra of endpoints), but certainly when there is its an issue.

How does clearml session avoid it? I guess only if autoscaling is used (one worker one machine)?

11 months ago
0 How Can I Control The

dug deeper. if i'm to make a guess...
/root/clearml.conf -> used on startup of agent-services as a template of sorts to create .clearml_agent.<id>.cfg on demand -> this task-specific file is used to mount to /tmp/clearml_default.conf in a new container (docker in docker bc of the socket mounted to the agent-services) -> used to execute the task

12 months ago
0 How Can I Control The

I tried mounting a config file (in the structure of the one on github but with just the relevant s3 section) into the agent-services container at /root/clearml.conf and after restarting the container, it seems to have had an impact. thank you!

When I inspect the console of the task I'm trying to run, I see there's a call to cp /tmp/clearml.conf ~/default_clearml.conf in the docker command and that the volume /tmp/clearml.conf is picked up from the host at some custom-named file ...

12 months ago
0 How Can I Control The

thank you!
I'll add a volume mount to the services-agent container, and from what I understand that will become the template it uses?

is this the structure of the file?
None

or is it the "dot" syntax (like what shows up in the console when the task executes / your snippet)?

12 months ago
0 Hi, I Noticed That When I Commit Changes And Not Push Them And Try To Run A Job I Am Getting

if you commit but do not push, the metadata tells clearml that it needs to pull a non-existant commit. any changes you made on top may be saved as a diff, but they'd fail to apply.

for clearml to work on un-pushed commits, it'd have to wait for a push to register a new diff target, which can become a problem (what if you have multiple remotes? which one will it wait for?) so rather, it assumes it can access the most recent commit from your remote repo, and records this as the "base" upon whi...

12 months ago
0 Hey, the <https://clear.ml/docs/latest/docs/references/api/#request-format|api reference> says that the url should be ```https://&lt;base_url&gt;/auth.login``` but to make it actually work I have to do ```https://&lt;base_url&gt;/api/v1.0/auth.login``` Th

Weird . I recently implemented a function that talked to this exact endpoint and found it had to exclude the version and api paths . Is there some sort of redirect that happens?

2 months ago
0 I Just Encountered A Really Frightening Bug. Best I Can Explain What Happened Was This: Data Scientist Created New Venv, Installed Clearml==1.11.0 Instead Of Clearml[S3]==1.11.1, And Upon Re-Running A Pipeline From Cli, The Entire Project "Disappeared" (W

yeah let's step through this, i'm having her execute these steps as we speak.

create a task with the new project name. its created as a draft. can see it in the UI under the new project.

pipeline script is updated with new project name for. execute script to create pipeline. now see in UI under this new project name. nothing hidden.

the pipeline is running. when the queue is default (only serviced by one container with agent in it ( clearml-agent==1.5.2 ). abort it. everything is still ...

11 months ago
Show more results compactanswers