Examples: query, "exact match", wildcard*, wild?ard, wild*rd
Fuzzy search: cake~ (finds cakes, bake)
Term boost: "red velvet"^4, chocolate^2
Field grouping: tags:(+work -"fun-stuff")
Escaping: Escape characters +-&|!(){}[]^"~*?:\ with \, e.g. \+
Range search: properties.timestamp:[1587729413488 TO *] (inclusive), properties.title:{A TO Z}(excluding A and Z)
Combinations: chocolate AND vanilla, chocolate OR vanilla, (chocolate OR vanilla) NOT "vanilla pudding"
Field search: properties.title:"The Title" AND text
Profile picture
MelancholyElk85
Moderator
28 Questions, 174 Answers
  Active since 10 January 2023
  Last activity one year ago

Reputation

0

Badges 1

149 × Eureka!
0 Votes
8 Answers
2K Views
0 Votes 8 Answers 2K Views
Hi, a question about dataset storage Suppose I create a dataset like this dataset = Dataset.create(dataset_name='my_name', dataset_project='my_project') What...
2 years ago
0 Votes
21 Answers
2K Views
0 Votes 21 Answers 2K Views
Hi, I'm trying to make use of new capabilities of DAG creation in ClearML. Seems that api has changed pretty much since a few versions back. There seems to b...
3 years ago
0 Votes
4 Answers
2K Views
0 Votes 4 Answers 2K Views
Dear clearml practitioners, I'd like to share my experiment to another company, in order for them to clone and fine-tune it on their own clearml server. Is t...
2 years ago
0 Votes
1 Answers
2K Views
0 Votes 1 Answers 2K Views
Hi! Is there a way to connect multiple S3 storages to ClearML? As far as I understand, I can only specify one set of S3 credentials in clearml.conf
2 years ago
0 Votes
2 Answers
1K Views
0 Votes 2 Answers 1K Views
Hi! Our ClearML server keeps crashing because of some weird task with the = symbol in the ID. After restoring the database to the latest working state, it wo...
one year ago
0 Votes
19 Answers
2K Views
0 Votes 19 Answers 2K Views
What is set_base_docker() equivalent for PipelineController ?
3 years ago
0 Votes
15 Answers
2K Views
0 Votes 15 Answers 2K Views
Is there any documentation for StorageHelper class?
3 years ago
0 Votes
13 Answers
1K Views
0 Votes 13 Answers 1K Views
Hi! Is there a way to get_local_copy but to download only the latest diff in dataset genealogy (that is, the files listed by list_added_files ), omitting eve...
2 years ago
0 Votes
9 Answers
2K Views
0 Votes 9 Answers 2K Views
A suggestion to add task_name parameter to InputModel.__ init__ . As for now I can search by project name and model name, but what if I had 2 models within t...
3 years ago
0 Votes
12 Answers
2K Views
0 Votes 12 Answers 2K Views
fatal: Could not read from remote repository. Please make sure you have the correct access rights and the repository exists.I have strong deja vu feeling. Cr...
3 years ago
0 Votes
9 Answers
2K Views
0 Votes 9 Answers 2K Views
Hi! Please, could anyone explain this sort of behaviour? First, we save a model checkpoint during training (both in .pth and .onnx format) output_model_pth =...
3 years ago
0 Votes
7 Answers
2K Views
0 Votes 7 Answers 2K Views
Is there a place in web UI to find all Models I created?
3 years ago
0 Votes
3 Answers
2K Views
0 Votes 3 Answers 2K Views
Hi! I see the following in the docs https://clear.ml/docs/latest/docs/references/sdk/model_model/#modelquery_models But I get AttributeError: type object 'Mo...
3 years ago
0 Votes
1 Answers
2K Views
0 Votes 1 Answers 2K Views
What's the difference between http://allegro.ai/clearml/docs and http://clear.ml/docs ? Which one is the real documentation? I need to look up something very...
3 years ago
0 Votes
6 Answers
2K Views
0 Votes 6 Answers 2K Views
Has anybody used albumentations with pytorch in remote execution on clearml? The task is freezing during DataLoader initialization every time I put any augme...
3 years ago
0 Votes
13 Answers
2K Views
0 Votes 13 Answers 2K Views
When I pass invalid key to task_overrides parameter in PipelineController.add_step method, it fails but with COMPLETED status
3 years ago
0 Votes
18 Answers
2K Views
0 Votes 18 Answers 2K Views
Hi! How to add files locally to dataset and then upload to a custom s3 location? The location should be specified within python code, NOT clearml.conf . defa...
3 years ago
0 Votes
1 Answers
1K Views
0 Votes 1 Answers 1K Views
Can I manually delete .zip files with datasets in .clearml/cache/storage_manager/datasets directory? seems that the datasets are stored both in packed and un...
3 years ago
0 Votes
19 Answers
2K Views
0 Votes 19 Answers 2K Views
Hi! Which method is used to delete OutputModel from the storage? https://clear.ml/docs/latest/docs/references/sdk/model_outputmodel/
3 years ago
0 Votes
2 Answers
2K Views
0 Votes 2 Answers 2K Views
3 years ago
0 Votes
1 Answers
1K Views
0 Votes 1 Answers 1K Views
Hi everyone! How do I disable automatic reporting of matplotlib plots and images?
2 years ago
0 Votes
10 Answers
2K Views
0 Votes 10 Answers 2K Views
Is there any canonical way to migrate all data between different clearml-server s? Some time ago we were using trains server with quite a few valuable experi...
3 years ago
0 Votes
27 Answers
1K Views
0 Votes 27 Answers 1K Views
I .
I add_step with clone_base_task=False . However, it still clones the task, wtf
3 years ago
0 Votes
2 Answers
2K Views
0 Votes 2 Answers 2K Views
3 years ago
0 Votes
15 Answers
2K Views
0 Votes 15 Answers 2K Views
How to use task_overrides in PipelineController.add_step method to set latest commit in some branch?
3 years ago
0 Votes
5 Answers
2K Views
0 Votes 5 Answers 2K Views
3 years ago
0 Votes
3 Answers
1K Views
0 Votes 3 Answers 1K Views
Hi everyone. I'm facing a problem when connecting to Azure storage. When the task tries to download a dataset, it fails with the message 2024-08-29 11:02:19,...
one year ago
0 Votes
23 Answers
2K Views
0 Votes 23 Answers 2K Views
Hi! For clearml.Dataset we can provide target storage (for example, in S3) with output_url parameter in the upload method. Is there any similar way to provid...
3 years ago
0 I .

I think it's still an issue, not critical though, because we have another way to do it and it works

3 years ago
0 I .

@<1523701205467926528:profile|AgitatedDove14> this last advice works, thank you!

3 years ago
0 I .

Suppose I have the following scenario (real-world project, real ML pipeline scenario)

  • I have separate projects for different steps (ETL, train, test, tensorrt conversion...). Every step has it's own git repository, docker image, branch etc
  • For quite a long time all the steps were not functioning as parts of an automated pipeline. For example, collaborative experimentation (training and validation steps). We were just focusing on reproducibility/versioning etc
  • After some time, we decided...
3 years ago
0 I .

so I had a base_task for every step. Then I wanted to modify a shitton of things, and I found no better way than to export, modify, and import to another task. Then this second task serves as step without cloning (that was my intention)

3 years ago
0 I .

@<1523701070390366208:profile|CostlyOstrich36> On the screenshot, the upper task has the lower task as parent

3 years ago
0 I .

Before the code I shared, there were some lines like this

step_base_task = Task.get_task(project_name=cfg[name]['base_project'],
                               task_name=cfg[name]['base_name'])
export_data = step_base_task.export_task()

... modify export_data in-place ...

task = Task.import_task(export_data)
pipe.add_step(base_task_id=task.id, clone_base_task=False, ...)
3 years ago
0 I .

@<1523701205467926528:profile|AgitatedDove14> yeah, I'll try calling task.reset() before add_step
No, IMO it's better to leave task_overrides arguments with "." - the same structure as in the dictionary we get from export_data - this is more intuitive

3 years ago
0 I .

In principle, I can modify almost everything with task_overrides , omitting export part, and it's fine. But seems that by exporting I can change more things, for example project_name

3 years ago
0 I .

base task is in draft status, so when I call import_data it imports draft status as well, am I right?

3 years ago
0 I .

but maybe add a lot more examples of using it to documentation

3 years ago
0 I .

so yeah, in short, target_project cannot do that

3 years ago
0 I .

We digressed a bit from the original thread topic though 😆 About clone_base_task=False .

I ended up using task_overrides for every change, and this way I only need 2 tasks (a base task and a step task, thus I use clone_base_task=True and it works as expected - yay!)

So, the problem I described in the beginning can be reproduced only this way:

  • to have a base task
  • export_data - modify - import_data - have a second task
  • pass the second task to add_step with `cl...
3 years ago
0 I .
task = Task.import_task(export_data)
pipe.add_step(
    name=name,
    base_task_id=task.id,
    parents=parents,
    task_overrides={'script.branch': 'main', 'script.version_num': '', },
    execution_queue=pipe_cfg['step_queue'],
    cache_executed_step=True,
    clone_base_task=False
)
3 years ago
0 I .

@<1523701205467926528:profile|AgitatedDove14> clearml 1.1.1
Yeah, of course it is in draft mode ( Task.import_task creates a task in draft mode, it is the lower task on the screenshot)

3 years ago
0 I .

maybe being able to change 100% of things with task_overrides would be the most convenient way

3 years ago
0 I .

doesn't target_project force the same project on all pipeline steps?

3 years ago
0 Is There Any Canonical Way To Migrate All Data Between Different

Or maybe there is a log with something more informative which I could check up

3 years ago
0 How To Use

I specifically set is as empty with export_data['script']['requirements'] = {} in order not to reduce overhead during launch. I have everything installed inside the container

3 years ago
0 How To Use

It doesn't install anything with pip during launch, I'm assuming it should take everything from the container itself (otherwise there would be a huge overhead). It simply fails trying to import things in the script

File "preprocess.py", line 4, in <module> from easydict import EasyDict as edict ModuleNotFoundError: No module named 'easydict'

3 years ago
0 How To Use

When I launch tasks with a pipeline, they keep complaining about missing pip packages. I run it inside a docker container, and I'm sure these packages are present inside it (when I launch the container locally, run python3 and import them, it works like charm). Any ideas how to fix this?

3 years ago
0 How To Use

You're right, I re-launched the agents, now it pulls images

3 years ago
0 How To Use

It worked, thank you!

3 years ago
0 How To Use

I have a base task for each pipeline step. When I initialize a pipeline, for each step I clone the corresponding task, modify it and add it as a step. Tasks are launched from a pipeline, not cli. I'm absolutely sure docker argument is not empty (I specify it with export_data['container']['image'] = ' http://registry.gitlab.com/cherrylabs/ml/clearml-demo:clearml ' , and it shows on Web UI)

3 years ago
Show more results compactanswers