Examples: query, "exact match", wildcard*, wild?ard, wild*rd
Fuzzy search: cake~ (finds cakes, bake)
Term boost: "red velvet"^4, chocolate^2
Field grouping: tags:(+work -"fun-stuff")
Escaping: Escape characters +-&|!(){}[]^"~*?:\ with \, e.g. \+
Range search: properties.timestamp:[1587729413488 TO *] (inclusive), properties.title:{A TO Z}(excluding A and Z)
Combinations: chocolate AND vanilla, chocolate OR vanilla, (chocolate OR vanilla) NOT "vanilla pudding"
Field search: properties.title:"The Title" AND text
Profile picture
AgitatedDove14
Moderator
48 Questions, 8051 Answers
  Active since 10 January 2023
  Last activity 8 months ago

Reputation

0

Badges 1

25 × Eureka!
0 Hello! Since Today I Get

But I do not have anything linked correctly since I rely in conda installing cuda/cudnn for me

From the log it installed:
cudatoolkit==11.1.1
based on the CUDA it found on the host machine: agent.cuda_version = 110
But for some reason it installed the pytorch from the conda "pytorch" repo without the cuda support.

3 years ago
0 Hello! Since Today I Get

And the one with the CPU version? is it with "~=" or "="?

3 years ago
0 Hi Team, I'M Currently Trying To Install Clearml-Server On A Powerpc Server With Redhat7. The Issue Is That The Clearml-Server Pre-Built Images Doesn'T Run On The Powerpc, So The Docker Containers Need To Be Rebuild On The Powerpc Host. Is There Dockerfil

Hi Team, I'm currently trying to install ClearML-Server on a Powerpc server with RedHat7.

You are a brave man LividCrab90 !

s there dockerfiles for the ClearML-Server stack somewhere ?

The main issue is replacing the DB containers, do you have elastic/mongo/redis for powerpc ?

3 years ago
0 Hi Everyone, Is There A Way To Avoid The Environment Setup When Running A Task Using A Worker? I Am Currently Using A Custom Docker Image That Already Has All The Require Packages Installed. I Tried Setting The Env Var

SteepDeer88
Try the following:
` Task.add_requirements("pycocotools-windows", "; platform_system == "Windows"")
Task.add_requirements("pycocotools", "; platform_system != "Windows"")

Task.init(...) You should see in your "installed packages" something like: pycocotools-windows ; platform_system == "Windows"
pycocotools ; platform_system != "Windows" `

2 years ago
0 Wondering Why This Is The Case When Deploying The Clearml Server Locally

ScantMoth28 where are you seeing this warning ?

2 years ago
0 I Have A Question About The Clearml Self Hosted Instance, I Notice There Is Elastic Search, Mondodb, And Redis In The Helm Chart Are These Required Or Can We Bring Our Own? I'M Wondering What Happens If I Were To Host The Instance And One Of These Were

I'm wondering what happens if i were to host the instance and one of these were to go down from time to time in production, as the deployments provided by the helm chart are not redundant.

Long story short, it will break the clearml-server, please do not take them down, if you do need to do that, also take down the clearml-server. The python clients will wait until it is up again, so no session would be destroyed

2 years ago
0 Hmm Is There Any Clear (Pun Intended) Documentation On The Roles Of Storagemanager, Dataset And Artefacts? It Seems To Me There Are Various Overlapping Roles And I'M Not Sure I Fully Grasp The Best Way Of Using Them. Especially When Looking At The Way Da

Hi JealousParrot68
This is the same as:
https://clearml.slack.com/archives/CTK20V944/p1627819701055200
and,
https://github.com/allegroai/clearml/issues/411

There is something odd happening in the files-server as it replaces the header (i.e. guessing the content o fthe stream) and this breaks the download (what happens is the clients automatically ungzip the csv).
We are working on a hit fix to he issue (BTW: if you are using object-storage / shared folders, this will not happen)

3 years ago
0 I Have Set

from task pick-up to "git clone" is now ~30s, much better.

This is "spent" calling apt update && update install && pip install clearml-agent
if you have those preinstalled it should be quick

though as far as I understand, the recommendation is still to not run workers-in-docker like this:

if you do not want it to install anything and just use existing venv (leaving the venv as is) and if something is missing then so be it, then yes sure that the way to go

7 months ago
0 I Have Set

BTW: you can also just add -e " CLEARML_AGENT_SKIP_PYTHON_ENV_INSTALL=1" to the docker args (under the Execution tab) to override the setting of the docker.
you can also add " export; " to the docker startup bash script section (do not add "#/bin/bash" , just the actual script) to get a list of all the environment variables inside the docker, just in case

7 months ago
0 Hello Guys, Not Sure If This Is The Right Place To Ask About Clearml Serving. May I Know If An Updated Readme Will Be Released Soon? I Did Not Manage To Get Clearml Serving Work With My Own Clearml Server And Triton Setup.

Hi OddShrimp85

right place to ask about clearml serving.

It is πŸ™‚

I did not manage to get clearml serving work with my own clearml server and triton setup.

Yes it should have been updated already, apologies.
Until we manage to sync the docs, what seems to be your issue, maybe we can help here?

3 years ago
0 Any Plans To Add Unpublished State For Clearml-Serving?

Any plans to add unpublished stateΒ for clearml-serving?

Hmm OddShrimp85 do you mean like flag, not being served ?
Should we use archive ?
The publish state, basically locks the Task/Model so they are not to be changed, should we enable unlocking (i.e. un-publish), wdyt?

2 years ago
0 Hi, Is There A Way To Pull Clearml Datasets To A Mounted Pv Instead Of The Pod'S Local Directory.

By the way, will downloading still happen if the datasets is available in the cache folder?

If it is cached, then there is no need to re-download πŸ™‚

one year ago
0 Hi, Is There A Way To Pull Clearml Datasets To A Mounted Pv Instead Of The Pod'S Local Directory.

Hi @<1523701304709353472:profile|OddShrimp85>
Do you mean Dataset.get_local_copy() ?

one year ago
0 Hi, Is There A Way To Pull Clearml Datasets To A Mounted Pv Instead Of The Pod'S Local Directory.

When you set the pod make sure you mount the clearml local cache folder to the PV
basically /root/.clearml/cache/

one year ago
0 Am Reading Through The New Docs…. How Do I Go About Creating The Cron Jobs In The Ui

Scheduled training is what I’m looking forward to

I'll try to remember to update here once we pushed into the GitHub repo, feedback is always appropriated πŸ™‚
If in the next two weeks you hear nothing, please ping here to make sure I did not forget πŸ˜‰

3 years ago
0 Hi, Can I Use Clearml As A Tool For Deploying Models In A Private Network? Did Not Manage To Understnd From The Docs.

does the clearml server is a worker i can serve on models?

The serving is done by one of the clearml-agents.
Basically you spin an agent, then this agent is spinning the model serving engine container (fully managed).
(1) install run run clearml-agent (2) run clearml-session CLI to configure and spin the serving engine

3 years ago
0 Hi, Does Anyone Use Mlflow / Weight & Biases /

Could you disable the windows anti-virus firewall and test?

4 years ago
4 years ago
0 Am Reading Through The New Docs…. How Do I Go About Creating The Cron Jobs In The Ui

Hi ComfortableHorse5
Yes this is more of a suggestion that you should write them using the platform capabilities, the UI implementation is being worked on, as well as a few helpers classes, I thin you'll be able to see a few in the next release πŸ™‚

3 years ago
0 Hello, This Is The Following Python Code I Had Saved As Main.Py.

Seems like a okay clearml.conf file
Notice this is the error:
404can you curl to this address ? are you sure you have httpS and not http ? was the dns configured ?

3 years ago
0 Hello, This Is The Following Python Code I Had Saved As Main.Py.

No worries πŸ™‚ glad to hear it worked out πŸ™‚

3 years ago
0 Hey - Is There A Known Issue With Tf.Summary.Image Not Being Logged When There'S A 'Description' Argument?

Hi BattyLizard6
Not that I'm aware of, which TF version are you using, and which clearml version?

3 years ago
4 years ago
0 Hi, Does Anyone Use Mlflow / Weight & Biases /

Nice debugging experience

Kudos on the work !
BTW, I feel weird to add an issue on their github, but someone should, this generic setup will break all sorts of things ...

4 years ago
0 Does Artifact Track Per File Base? What If Only Some File Is Updated, Does It Knows Only Uploading The New Files? Also, Wonder What Is The Best Way To Setup Storage For Teams To Share? (Not Prefer Using Cloud As Network Cost Can Be Significant Since We Do

EnviousStarfish54 regrading file server, you have one built into the trains-server, and this will be the default location to store all artifacts. You can also use external solutions like S3 GS Azure etc.
Regarding the models, any model store / load is automatically logged as long as you are using one of the supported frameworks (TF Keras PyTorch scikit learn)
If you want your model to be automatically uploaded, just add outpu_uri:
task=Task.init('examples', 'model', output_uri=' http://trai...

4 years ago
0 Hello, I'M Really Enjoying The Clearml Experience And We'Re Using It Very Successfully At Work. I Have A Small Question Though: I'M Trying To Set The "File_History_Size" Sdk Parameter From Python Code Instead Of The Conf File. Is That Possible? Thanks!

Hi DisturbedWalrus17
This is a bit of a hack, but will work:
from clearml.backend_interface.metrics.events import UploadEvent UploadEvent._file_history_size = 10Maybe we should expose it somewhere, what do you think?

3 years ago
0 Hi All! Does Anyone Know A Solution To My Issue With Deploying Models Saved On Azure On The Clearml-Serving Docker Container?

how specifically I map my clearml.conf to the containers?

You add:

    volumes:
      - $HOME/clearml.conf:/root/clearml.conf

to here:
None
and here:
None

one year ago
Show more results compactanswers