Examples: query, "exact match", wildcard*, wild?ard, wild*rd
Fuzzy search: cake~ (finds cakes, bake)
Term boost: "red velvet"^4, chocolate^2
Field grouping: tags:(+work -"fun-stuff")
Escaping: Escape characters +-&|!(){}[]^"~*?:\ with \, e.g. \+
Range search: properties.timestamp:[1587729413488 TO *] (inclusive), properties.title:{A TO Z}(excluding A and Z)
Combinations: chocolate AND vanilla, chocolate OR vanilla, (chocolate OR vanilla) NOT "vanilla pudding"
Field search: properties.title:"The Title" AND text
Profile picture
AgitatedDove14
Moderator
49 Questions, 8122 Answers
  Active since 10 January 2023
  Last activity one year ago

Reputation

0

Badges 1

25 × Eureka!
0 Hello, Is There A Way To Update A Task Diff Programatically? Eg, I'M Creating A Task Using

ShakyJellyfish91 can you check if version 1.0.6rc2 can find the changes ?

3 years ago
4 years ago
0 Hi All, I Am Having Trouble Using The

Notice both needs to be str
btw, if you need the entire folder just use StorageManager.upload_folder

3 years ago
0 Hi, I Encountered A Few Problems:
  1. Artifacts and models will be uploaded to the output URI, debug images are uploaded to the default file server. It can be changed via the Logger.
  2. Hmm is this like a configuration file?
    You can do.
    local_text_file = task.connect_configuration('filenotingit.txt')
    Then open the 'local_text_file' it will create a local copy of the data in runtime, and the content will be stored on the Task itself.
  3. This is how the agent installs the python packages, but if the docker already contactains th...
5 years ago
0 Hi Folks, We Are Trying To Find A Tool To Help With Workflow Orchestration. This Is Our Stack So Far (Label Studio/Clearml/Seldon). Does Anyone Have Any Experience With Using Any Workflow Which Is Most Compatible Esp Wrt To Clearml.

TenseOstrich47 / PleasantGiraffe85
The next version (I think releasing today) will already contain scheduling, and the next one (probably RC right after) will include triggering. That said currently the UI wizard for both (i.e. creating the triggers), is only available in the community hosted service. That said I think that creating it from code (triggers/schedule) actually makes a lot of sense,

pipeline presented in a clear UI,

This is actually actively worked on, I think Anxious...

4 years ago
0 Hi, Is There Any Document About Migration Clearml-Server. Currently, I Have Clearml-Server Running On Servera But I Want To Move All Data (Including Artifacts, Task, Dataset) From Servera To Serverb.

VictoriousPenguin97 basically spin down sereverA (this should flush all DBs) then copy /opt/clearml to the new server and spin it with docker-compose. As long as the new server is on the same address as the previous one, everything should work out of the box

3 years ago
0 Hey All

. And I saw that it upload the notebook it self as notebook. Does it is normal? There is a way to disable it?

Hi FriendlyElk26
Yes this is normal, it backups your notebook as well as converts it into python code (see "Execution - uncommitted changes" so that later the clearml-agent will be able to run it for you on remote machines.
You can also use task.connect({"param": "value") to expose arguments to use in the notebook so that later you will be able to change them from the U...

2 years ago
0 Hello Friends! I Am Trying To Play Around With The Configs For

We are working on 1.3.0 so this is right in time

2 years ago
0 Hey, I'M Probably Being Thick Here But I Would Like To Pull Some Data From A Database And Write It To A Particular Bucket In S3 Within A Task I'M Doing. I'M Using Task.Upload_Artifact But Can'T Understand Where I Write The Bucket Path.

It should have been:
output_uri="s3://company-clearml/artifacts/bethan/sales_journeys/artifacts/examples/load_artifacts.f0f4d1cd5eb54795b11508dd1e739145/artifacts/filename.csv.gz/filename.csv.gz

4 years ago
0 Sometimes I Notice That At The End Of An Experiment Clearml Keeps Hanging (Something With Repository Detection?) And The Script Does Not End. Do More People See This? Especially In Our Continuous Integration Pipeline This Give Problems Because Tests Are G

Quick update, I might have been able to reproduce the issue ( GreasyPenguin14 working "offline" is a great hack to accelerate debugging this issue, thank you!)
It seems it is related to the known and very annoying Python forking issue (and this is why changing to "spawn" method solves the issue):
https://bugs.python.org/issue6721
Long story short, in some cases when forking (i.e. ProcessPoolExecutor), python can copy locks in a "bad" state, this means that you can end up with a lock acquir...

3 years ago
0 Another Question: Is It Possible To Read The Dependencies Manually From A Conda Environment.Yml? It Seems Like Clearml Is Not Able To Fetch The Dependencies Correctly When

ReassuredTiger98 regrading the agent error, can you see the package some_packge in the "Installed Packages" in the UI? Was it installed ? are you using pip or conda as package manager in the agent (check the clearml.conf) is the agent running in docker mode ?

4 years ago
0 Hello All! Quick Question, Do Any Of You Know Of A Clean Way To Access The Clearml Logger Inside Of A

That being said it returns none for me when I reload a task but it's probably something on my side.

MistakenDragonfly51 just making sure, you did call Task.init, correct ?
What dues
from clearml import Task task = Task.current_task()returns ?

Notice that you need to create the Task before actually calling Logger.current_logger() or Task.current_task()

2 years ago
4 years ago
0 I Am Working Up With The Autoscaler, After Setting Up The Autoscaler Instance I Am Getting The Following Error When I Launch The Autoscaler Googleapiclient.Errors.Httperror: <Httperror 404 When Requesting

Thanks!
Hmm from here : None
Could it be you do not have privileges to the resource, or that you did not provide credentials ?
Did that autoscaler work before ?

one year ago
0 Hi, I Have Been Getting The Following For A While. Is There A More Detailed Log I Can Look Into? This Happens On Both Https And Http.

In our case this is not possible due to client security (e.g. training data from clients can potentially be 'reverse engineered' from trained models in future).

Hmm I see, wouldn't it make more sense to separate clients like a multi-tenant SAAS solution ?

4 years ago
0 Hi! I Was Wondering Regarding This Issue:

Thanks WittyOwl57 ! let me check

3 years ago
4 years ago
0 I Assume I Can Ask A Question Here. The Clearml Orchestrator Looks Interesting. But The Website Suggests That K8S Is Required. We Have A Linux Training Box (Lambdabox) Where We Want To Run Training. Can We Place The Clearml Orchestrator Agent On The M

Hi RobustFlamingo1

The ClearML Orchestrator looks interesting. But the website suggests that K8S is required

No k8s is not a must, only an option 🙂

We have a Linux training box (LambdaBox) where we want to run training. Can we place the ClearML orchestrator agent on the machine without needing K8S?

Yes should be quite easy.
If you intent to use containers, make sure you have docker installed.
Then just pip install clearml-agent and configure it:
https://clear.ml/doc...

3 years ago
0 Hi There, I’Ve Been Trying To Play Around With The Model Inference Pipeline Following

I did change the

instead of 8080?

So this is the issue

2 years ago
0 Hmm Is There Any Clear (Pun Intended) Documentation On The Roles Of Storagemanager, Dataset And Artefacts? It Seems To Me There Are Various Overlapping Roles And I'M Not Sure I Fully Grasp The Best Way Of Using Them. Especially When Looking At The Way Da

Btw I sometimes get a gzip error when I am accessing artefacts via the '.get()' part.

Hmm this is odd, is this a download issue? if this is reproducible maybe we should investigate further...

4 years ago
0 Good Day! I Ran Into A Problem When When Running Two Or More Identical Nodes In A Pipeline (Multi_Instance_Support=True ), Only One Of Them Uses An Already Created Venv From Cache For This Task. And The Other Node Starts To Re-Create The Same Virtual Envi

Hi @<1598487094601191424:profile|MysteriousCow84>

only one of them uses an already created venv from cache for this task. And the other node starts to re-create the same virtual environment.

Just be clear, the second one is running, but it does not use the same venv as the other one (that is running in parallel), is that correct?

2 years ago
0 Hi Guys, I Am Having Some Problems About Parsing Arguments To A Task When Running On Agent. I Have A Task A That Creates Another Task B Via Subprocess. Some Arguments Of A And B Share The Same Names. The Problem Is That They'Re Visible To B Even Though I'

I was wondering about what i can do with the agent's argparse magic

You mean how to pass arguments to components a pipeline? btw did you check the pipeline example here?
None

one year ago
Show more results compactanswers