Examples: query, "exact match", wildcard*, wild?ard, wild*rd
Fuzzy search: cake~ (finds cakes, bake)
Term boost: "red velvet"^4, chocolate^2
Field grouping: tags:(+work -"fun-stuff")
Escaping: Escape characters +-&|!(){}[]^"~*?:\ with \, e.g. \+
Range search: properties.timestamp:[1587729413488 TO *] (inclusive), properties.title:{A TO Z}(excluding A and Z)
Combinations: chocolate AND vanilla, chocolate OR vanilla, (chocolate OR vanilla) NOT "vanilla pudding"
Field search: properties.title:"The Title" AND text
Profile picture
SweetBadger76
Moderator
1 Question, 239 Answers
  Active since 10 January 2023
  Last activity 2 years ago

Reputation

0

Badges 1

4 × Eureka!
0 Votes
8 Answers
2K Views
0 Votes 8 Answers 2K Views
hello TartSeagull57 This is a bug introduced with version 1.4.1, for which we are working on a patch. The fix is actually in test, and should be released ver...
3 years ago
0 Hi, I Would Like To Log Locally Each Link To My Experiments, How Can I Get The Link To The Experiment (The One Created At The Beginning Of The Run And Printed To The Console), From The Task Object? Is It Always Going To Be:

hi DizzyHippopotamus13
Yes you can generate a link to the experiments using this format.
However I would suggest you to use the SDK for more safety :
task = Task.get_task(project_name=xxx, task_name=xxx)
url = task.get_output_log_web_page()

Or in one line
url = Task.get_task(project_name=xxx, task_name=xxx).get_output_log_web_page()

3 years ago
0 Hi, I Am Trying The Triggerscheduler To Catch When A User Add Specific Tag To A Task. I Used The Below Code But The Schedule_Function Is Not Called When Adding Tags To Task (It Seems The Task.Last_Update Is Not Modified After Adding Tag)

hi ApprehensiveSeahorse83
i am working too on your issue. It seems that there is a wrong behavior here, so we need to get a bit deeper to understand what's happening. We will keep you updated asap, thanks for your contribution ! 🙏

3 years ago
0 Hey Guys, Is There An E2E Working Example Of Writing A Pipeline With 2-3 Tasks? Just An Hello World. I Am The First One Who Tries To Make Clearml Pipeline To Work I Wasn'T Able To Make It:

Agent is a process that pulls task from a queue and assigns ressources (worker) to them. In the pipeline, when not runned locally, steps are enqueued tasks

3 years ago
0 Hi All. I Was Using Clearml Server Hosted On A Box That I Reach Behind Traefik Using Alias For Web, File And Api. After Migration It Works Perfect For New Experiments. I Changed The Name Of The Alias From

Hi MotionlessCoral18
You need to run some scripts when migrating, to update your old experiments. I am going to try to find you soem examples

3 years ago
0 Hi I'M Looking Into How Clearml Supports Datasets And Dataset Versioning And I'M A Bit Confused. Is Dataset Versioning Not Supported At All In The Non-Enterprise Or Is Versioning Available By A Different Mechanism? I See That

Hi PanickyMoth78
There is indeed a versioning mechanism available for the open source version 🎉

The datasets keep track of their "genealogy" so you can easily access the version that you need through its ID

In order to create a child dataset, you simply have to use the parameter "parent_datasets" when you create your dataset : have a look at
https://clear.ml/docs/latest/docs/clearml_data/clearml_data_sdk#datasetcreate

You also alternatively squash datasets together to create a c...

3 years ago
0 I’M Trying To Get The Meta-Information About The Code (Section Execution) To Be Auto-Filled, However When I Run The Script With The Pycharm Testrunner, It Is Missing. If I Use

Hi EnormousWorm79

The Pycharm testrunner wraps the script into a local cript, and thats what you are getting.
(jb pytest runner). Because it is local, you lose the source info

Let me check if I have a workaround or solution for you. I keep you updated

3 years ago
0 Hey.

Hi Max
you can configure a clearml agent to pull your docker image from ECR and run the experiment into it. Is that answering your question ?

3 years ago
0 Hey,

btw here is the content of the imported file:

import torch
from torchvision import datasets, transforms
import os
MY_GLOBAL_VAR = 32

def my_dataloder ():
return torch.utils.data.DataLoader(
datasets.MNIST(os.path.join('./', 'data'), train=True, download=True,
transform=transforms.Compose([
transforms.ToTensor()
` ...

3 years ago
0 Hi Everyone

Hey LuckyKangaroo60
So far there isnt a CLI command to check the conf file format : if there is an error, it is detected from the beginning of the execution and the program fails. Here is what i use as a conf for accessing my local docker based minio :

`
s3 {
# S3 credentials, used for read/write access by various SDK elements

        # Default, used for any bucket not specified below
        region: ""
        # Specify explicit keys
        key: "david"
...
3 years ago
0 Need

When the pipeline or any step is executed, a task is created, and it name will be taken from the decorator parameters. Additionally, for a step, the name parameter is optional : if not provided, the function name will be used instead.

It seems to me that your script fails creating the pipeline controller task because it fails pulling the name parameter. which is weird ... Weird because in the last error line, we can see that name !

3 years ago
0 Hi, I Am Trying To Use The Parameterset For Hyper-Parameter Tuning With Dependencies, An Example Of How I Use It: Parameterset([{“Prm1”:1, “Prm2": 1},{“Prm1”:2, “Prm2":2}]) But I Get A Warning :

hi MoodySheep3
I think that you use ParameterSet the way it is supposed to be 🙂
When I run my examples, I also get this warning - which is weird ! because
This is just a warning, the script continues anyway (and reaches end without issue) Those HP exists - and all the sub tasks corresponding to a given parameters set find them !

3 years ago
0 Need

can you also check that you can access the servers ?
try to do curl http://<my server>:port for your different servers ? and share the results 🙂

3 years ago
0 Hello Clearml, I Am Curious To Know How Does The Clearml

it is basically auto-generated when you do clearml-init
there are a bench of optional configurations that are not in the auto generated file though.
Have a look here it is pretty detailed https://clear.ml/docs/latest/docs/configs/clearml_conf

3 years ago
0 Hello Community! How I Can Add S3 Credentials To S3 Bucket In Example.Env For Clearml-Serving-Triton? I Need To Add Bucket Name, Keys And Endpoint

hi AbruptHedgehog21
clearml-serving will use your clearml.conf file
Configure it to access your s3 bucket - that is the place for bucket, host etc

3 years ago
0 Hey,

Hey
unfortunatly there is no way to manage the CPU cores allocation for an agent

3 years ago
0 Hi, Very New To Clearml And This Concept In Overall. Does Clearml-Data Track Feature In Excel File? I.E. Can It Give Me An Evoution Of The Excel File? Thanks

Hi JitteryParrot8
clearml-data is a tool for improving datasets management, including versioning. However it is not tracking modifications at the file level , so it would not track the evolution of a single file.
The idea is to create a dataset with your excel file, and each time you will change that file, you will create a new version of the dataset with the updated file. So that you could register the different versions of your file through the dataset version.

3 years ago
0 Hi, Love What You Guys Did With The New Datasets! I Need Some Help Though. I Assume There Will Be A No-Code Way To Do This, Maybe Not Now But In The Future. But Anyway, I Have Three Different Datasets, And I Want To Create A Merged Version Of All Three Of

hi GrittyStarfish67
"Hi, love what you guys did with the new datasets!" Thanks 🙂 !

you can squash the datasets together : it will result in the creation of a child dataset, that will contain its parents data merged together. Note that there will be no duplicate upload of the parents data : when a dataset inherits from parents datasets, it receives the references to the data uploaded by the parents.
SDK: https://clear.ml/docs/latest/docs/clearml_data/clearml_data_sdk#datasetsquash
...

3 years ago
0 I Have An Inference Task In Clearml Where I Apply A Model (Defined In Input Params) To A Dataset. Clearml Registers The Model As An Input Model, Which Is Nice. But When I Clone The Task And Modify Input Param To Apply Another Model To The Same Dataset, Th

hi FiercePenguin76
Can you also send your clearml packages versions ?
I would like to sum your issue up , so that you could check i got it right

you have a task that has a model, that you use to make some inference on a dataset you clone the task, and would like to make inferences on the dataset, but with another modelthe problem is that you have a cloned task with the first model....

How have you registered the second model ? Also can you share your logs ?

3 years ago
0 Does Clearml Have Any Suggestions On Gpu And Non-Gpu Amis For The Autoscaler? The Two Default Ones In The Documentation Are Either:

Hi UnevenDolphin73
you can find AMI that could satisfy your needs in the AWS website. Simply enter a console, choose your region (upper-right) , enter EC2 -> launch instance
There you can choose your os, or type "deep learning" if you need GPU machines

If you need some help to find more specific machines, do not hesitate to ask me :)

3 years ago
0 Hello

Great !
Concerning the running status :
in the first case, the program failed so the server has no way to be informed of a status change in the second case, this is not a task status but a dataset, so the status would change when you will publish the datasetThe fix is in last phases of testing, I hope that it will be released very soon

3 years ago
0 Hi, I Am Trying The Triggerscheduler To Catch When A User Add Specific Tag To A Task. I Used The Below Code But The Schedule_Function Is Not Called When Adding Tags To Task (It Seems The Task.Last_Update Is Not Modified After Adding Tag)

yes i have the behavior. I think that we have a bug. We will release a fix and will keep you updated 🙂
can you please open a github issue ? Of course if it is a problem, i can do it for you, just let me know 🙂
http://github.com/allegroai/clearml/issues

3 years ago
0 Need

can you show the logs ?

3 years ago
Show more results compactanswers