Examples: query, "exact match", wildcard*, wild?ard, wild*rd
Fuzzy search: cake~ (finds cakes, bake)
Term boost: "red velvet"^4, chocolate^2
Field grouping: tags:(+work -"fun-stuff")
Escaping: Escape characters +-&|!(){}[]^"~*?:\ with \, e.g. \+
Range search: properties.timestamp:[1587729413488 TO *] (inclusive), properties.title:{A TO Z}(excluding A and Z)
Combinations: chocolate AND vanilla, chocolate OR vanilla, (chocolate OR vanilla) NOT "vanilla pudding"
Field search: properties.title:"The Title" AND text
Profile picture
ExcitedSeaurchin87
Moderator
2 Questions, 11 Answers
  Active since 10 January 2023
  Last activity one year ago

Reputation

0

Badges 1

5 × Eureka!
0 Votes
9 Answers
1K Views
0 Votes 9 Answers 1K Views
Hi everyone! I've noticed that if I run an experiment and it fails, the ClearML agent will delete all datasets that have been downloaded during the run. Is i...
2 years ago
0 Votes
5 Answers
995 Views
0 Votes 5 Answers 995 Views
Hello everyone! Do you know how to run several agents simultaneously on the same PC? I launched the first one and if I try to run the second one I get: clear...
2 years ago
0 Hello Everyone! Do You Know How To Run Several Agents Simultaneously On The Same Pc? I Launched The First One And If I Try To Run The Second One I Get:

Thank you for your response. It works. I want to run several workers simultaneously on the same GPU, because I have to train several, relatively simple and small, neural networks. It would be faster to train several of them at the same time on the same GPU, rather than do it consequently.

2 years ago
0 Hi! I'M Using Func

Hi AnxiousSeal95 ! Thank you for the parallel download feature you have added. We have tested it with ClearML 1.5.0 and it seems that it is not really helpful. For a big dataset the time it takes to download does not really changes with the new feature. We indeed can download several chunks in parallel, but it turns out that while N workers are downloading N chunks, downloading speed for each worker is N times less than for a consequent download. Since that most of the chunks have the same s...

2 years ago
0 Hi! I'M Using Func

Could you please point me to the piece of ClearML code related to the downloading process?

2 years ago
0 Hi! I'M Using Func

Yes AgitatedDove14 , I mean multithreading. I did not quite understand your question about single Dataset version. Could you clarify the question for me, please?

At least from the logs I see in my terminal I assume that right now downloading works as on the scheme 2. This one:

2 years ago
0 Hi! I'M Using Func

Here is the schemes of discussed variants (1. download process before ClearML 1.5.0; 2. current version; 3. proposed method). I hope they will be helpful.

2 years ago
0 Hi Everyone! I'Ve Noticed That If I Run An Experiment And It Fails, The Clearml Agent Will Delete All Datasets That Have Been Downloaded During The Run. Is It Correct Behavior? How Can I Force The Agent To Preserve Such Datasets?

Sorry. I probably misunderstood you. I just downloaded the clearml-agent package to my machine and ran the agent with the following command: python -m clearml_agent daemon --queue default dinara --docker --detached

2 years ago
0 Hello Everyone! Do You Know How To Run Several Agents Simultaneously On The Same Pc? I Launched The First One And If I Try To Run The Second One I Get:

Yes, I would like to run several agents on the same GPU. I use command python -m clearml_agent daemon --queue default queue_name --docker --gpus 0 --detached

2 years ago
0 Hi Everyone! I'Ve Noticed That If I Run An Experiment And It Fails, The Clearml Agent Will Delete All Datasets That Have Been Downloaded During The Run. Is It Correct Behavior? How Can I Force The Agent To Preserve Such Datasets?

SuccessfulKoala55
I initialized the task with Python:

task = Task.init(project_name=args.project_name, task_name=args.task_name)

and downloaded set of datasets later in the code:

for dataset_name in datasets_list:
clearml_dataset = clearml.Dataset.get(dataset_project=dataset_project, dataset_name=dataset_name)
clearml_dataset_path = clearml_dataset.get_local_copy()

Then I go through the resulting directories in search of the files I need, and send their pa...

2 years ago
0 Hi! I'M Using Func

Or, we can download chunks in parallel, like we do it right now, but we have to prioritize the download of the earlier chunks to make extraction and downloading run in parallel.

2 years ago
0 Hi! I'M Using Func

Hi DepressedFish57 AgitatedDove14 AnxiousSeal95 ! It's me again. I created a https://github.com/allegroai/clearml/pull/713 where I slightly changed the dataset loading code, ran some tests to estimate dataset loading times using the current and proposed approaches and tried to explain in detail the problem I see.

2 years ago