Examples: query, "exact match", wildcard*, wild?ard, wild*rd
Fuzzy search: cake~ (finds cakes, bake)
Term boost: "red velvet"^4, chocolate^2
Field grouping: tags:(+work -"fun-stuff")
Escaping: Escape characters +-&|!(){}[]^"~*?:\ with \, e.g. \+
Range search: properties.timestamp:[1587729413488 TO *] (inclusive), properties.title:{A TO Z}(excluding A and Z)
Combinations: chocolate AND vanilla, chocolate OR vanilla, (chocolate OR vanilla) NOT "vanilla pudding"
Field search: properties.title:"The Title" AND text
Profile picture
RoughTiger69
Moderator
28 Questions, 101 Answers
  Active since 10 January 2023
  Last activity one month ago

Reputation

0

Badges 1

90 × Eureka!
0 Votes
6 Answers
690 Views
0 Votes 6 Answers 690 Views
autoscaler from saas (pro version). I attempted to use the autoscaler “application” from clearml UI. here is what I get in the logs of the autoscaler screen ...
one year ago
0 Votes
5 Answers
767 Views
0 Votes 5 Answers 767 Views
hi, I am running a pipeline from decorators. the pipeline runs fine. Then I try to clone it by clicking the (successful) run and launching. The pipeline fail...
one year ago
0 Votes
18 Answers
621 Views
0 Votes 18 Answers 621 Views
Is there a case-study or ref. architecture for interacting with CI/CD i.e. exposing mature pipelines to be triggered upon code pushes (taking latest git hash...
2 years ago
0 Votes
4 Answers
743 Views
0 Votes 4 Answers 743 Views
hi, I created a dataset with 20K files, total of 20GB, with storage pointing to S3. When I upload (or close) the dataset, during the compression phase, the c...
one year ago
0 Votes
0 Answers
645 Views
0 Votes 0 Answers 645 Views
Did more digging, seems that you need to start the agent with CLEARML_AGENT_SKIP_PIP_VENV_INSTALL=1
one year ago
0 Votes
8 Answers
716 Views
0 Votes 8 Answers 716 Views
one year ago
0 Votes
5 Answers
635 Views
0 Votes 5 Answers 635 Views
I have a logical task that I want to split to multiple workers. The task involves processing media files (not training). The optimal design for me would be: ...
one year ago
0 Votes
5 Answers
600 Views
0 Votes 5 Answers 600 Views
Avoiding http://Clear.ml glue code spaghetti - community best practices? Say I have training pipeline : Task 1 - data preprocessing -> create a dataset artif...
2 years ago
0 Votes
5 Answers
792 Views
0 Votes 5 Answers 792 Views
Hi, I am trying to use the aws autoscaler to assign instance profiles to new machines. This is a better way than managing credentials. I added the configurat...
one year ago
0 Votes
3 Answers
693 Views
0 Votes 3 Answers 693 Views
FYI I am getting a lot of read timeouts from the community server: Retrying (Retry(total=235, connect=240, read=235, redirect=240, status=240)) after connect...
2 years ago
0 Votes
7 Answers
626 Views
0 Votes 7 Answers 626 Views
one year ago
0 Votes
0 Answers
574 Views
0 Votes 0 Answers 574 Views
Hi, I am catching up with http://clear.ml for stuff beyond exp. tracking, and have a few questions. Will ask them separately to allow threading:
2 years ago
0 Votes
0 Answers
555 Views
0 Votes 0 Answers 555 Views
autoscaler 101 questions: What is the best practice for managing credentials so that they don’t get saved in clearml webapp? When the https://clear.ml/docs/l...
one year ago
0 Votes
9 Answers
696 Views
0 Votes 9 Answers 696 Views
one year ago
0 Votes
12 Answers
59 Views
0 Votes 12 Answers 59 Views
Is there a reference implmentation for a task in a pipeline that awaits user input?
2 years ago
0 Votes
10 Answers
694 Views
0 Votes 10 Answers 694 Views
one year ago
0 Votes
3 Answers
629 Views
0 Votes 3 Answers 629 Views
2 years ago
0 Votes
4 Answers
596 Views
0 Votes 4 Answers 596 Views
hi, When running a training script from pycharm, it seems that clearml logs only those packages that are explicitly imported by my .py files; it seems to not...
2 years ago
0 Votes
14 Answers
678 Views
0 Votes 14 Answers 678 Views
question about pipeline and long-waiting tasks: Say I want to generate a dataset. The workflow I have requires query to a DB Creating a labeling assigment in...
2 years ago
0 Votes
3 Answers
675 Views
0 Votes 3 Answers 675 Views
what’s a good ami to use for the clearml autoscaler on AWS? the defaults offered confidently by the various auto scaler installers don’t seem to exist…| e.g....
one year ago
0 Votes
10 Answers
689 Views
0 Votes 10 Answers 689 Views
2 years ago
0 Votes
3 Answers
666 Views
0 Votes 3 Answers 666 Views
one year ago
0 Votes
11 Answers
693 Views
0 Votes 11 Answers 693 Views
I have a local folder a, and a dataset B. a: a a/.DS_Store a/1.txt a/b a/b/.DS_Store a/b/1.txt a/b/c a/b/c/1.txtDataset B: b b/2.txt b/c b/c/2.txtI want to “...
one year ago
0 Votes
10 Answers
631 Views
0 Votes 10 Answers 631 Views
hi folks, is there a way to force clear-ml agent with --docker to not create a virtualenv at all? And perhaps not even attempt to install requirements even? ...
one year ago
0 Votes
14 Answers
581 Views
0 Votes 14 Answers 581 Views
Two simple lineage related questions: Task B is a clone of Taks A. Does B store the information that it was cloned from A somewhere? Training task X loads Da...
2 years ago
0 Votes
9 Answers
722 Views
0 Votes 9 Answers 722 Views
one year ago
0 Votes
6 Answers
627 Views
0 Votes 6 Answers 627 Views
Regarding the “classic” datasets (not hyper datasets): Is there an option to do something equivalent to dvc’s “ https://dvc.org/doc/user-guide/managing-exter...
2 years ago
0 Votes
8 Answers
670 Views
0 Votes 8 Answers 670 Views
Hi, I am using PipelineDecorator to create tasks. is there a way to force it to use the entire git repo it is created from on the pythonpath? vs. just the de...
one year ago
0 Question About Pipeline And Long-Waiting Tasks: Say I Want To Generate A Dataset. The Workflow I Have Requires

AgitatedDove14

What was important for me was that the user can define the entire workflow and that I can see its status as one ‘pipeline’ in the UI (vs. disparate tasks).

perform query process records into a labeling assignment Call labeling system API wait for and external hook when labels are ready clean the labels upload them to a dataset
Do you know what specific API do I need to signal “resume” after “abort”?
not “reset” I presume?

2 years ago
0 I Have Code That Does Torch.Load(Path) And Deserializes A Model. I Am Performing This In Package A.B.C, And The Model’S Module Is Available In In A.B.C.Model Unfortunately, The Model Was Serialized With A Different Module Structure - It Was Originally Pla

I think it has something to do with clearml since I can run this code as pure python without clearml, and when I activate clearml, I see that torch.load() hits the

import_bind

.

__patched_import3

when trying to deserialize the saved model

one year ago
0 <no title>

could work! is there a way to visualize the pipeline such that this step is “stuck” in executing?

2 years ago
0 <no title>

sounds workable, will start with that

2 years ago
0 <no title>

yes and no.
This is a pseudo flow:
Data download -> pre-processing -> model training (e.g. HPT) - > model evaluation (per variant) -> model comparison dashboard -> human selects the best model using a heuristic and the status of the weather -> model packaging -> inference tests etc.

I could divide it into two pipelines:
Data download --> dashboard
Packaging --> …

Where packaging takes a parameter which is the human selected ID of the model.
However, this way, I lose the context of the ent...

2 years ago
0 Autoscaler From Saas (Pro Version). I Attempted To Use The Autoscaler “Application” From Clearml Ui. Here Is What I Get In The Logs Of The Autoscaler Screen Itself (Consistent):

CostlyOstrich36 from what I gather the UI creates a task in the background, in status “hidden”, and it has like 10 fields of json configurations…

one year ago
0 Cannot Upload A Dataset With A Parent - Seems Very Odd! Clearml Versions I Tried: 1.6.1, 1.6.2 Scenario: * Create Parent Dataset (With Storage On S3) * Upload Data * Close Dataset * Create Child Dataset (Tried With Storage On Both S3 Or On Clearml Serv

I tested it again with much smaller data and it seems to work.
I am not sure what is the difference between the use-cases. it seems like something specifically about the particular (big) parent doesn’t agree with clearml…

one year ago
0 Cannot Upload A Dataset With A Parent - Seems Very Odd! Clearml Versions I Tried: 1.6.1, 1.6.2 Scenario: * Create Parent Dataset (With Storage On S3) * Upload Data * Close Dataset * Create Child Dataset (Tried With Storage On Both S3 Or On Clearml Serv

Tried with 1.6.0, doesn’t work

#this is the parent clearml-data create --project xxx --name yyy --output-uri `
clearml-data add folder1
clearml-data close

#this is the child, where XYZ is the parent's id
clearml-data create --project xxx --name yyy1 --parents XYZ --output-uri
clearml-data add folder2
clearml-data close
#now I get the error above `

one year ago
0 I Have A Logical Task That I Want To Split To Multiple Workers. The Task Involves Processing Media Files (Not Training). The Optimal Design For Me Would Be:

AgitatedDove14 it’s pretty much similar to your proposal but with pipelines instead of tasks, right?

one year ago
0 <no title>

yeah, its a tradeoff that is dependent on parameters that lie outside the realm of human comprehension.
Let’s call if voodoo.
Yes, the manual selection can be done via tagging a model.
The main thing is that I want the selection to be part of the overall flow.
I want the task of human tagging a model to be “just another step in the pipeline”

2 years ago
0 Regarding The “Classic” Datasets (Not Hyper Datasets): Is There An Option To Do Something Equivalent To Dvc’S “

AgitatedDove14 I haven’t done a full design for this 😉
Just referring to how DVC claims it can detect and invalidate changes in large remote files.
So I take it there is no such feature in http://clear.ml 🙂

2 years ago
0 Avoiding

AgitatedDove14 Not sure the pipeline decorator is what I need.

Here’s a very simplified example to my question.

Say I want to train my model on some data.
Before adding http://clear.ml , the code looks something like:
def train(data_dir, ...): ...
Now I want to leverage the data versioning capability in http://clear.ml
So now, the code needs to fetch dataset by ID, save it locally, and let the model train on it as before:
` from clearml import Dataset
def train_clearml(dataset_id...

2 years ago
0 Avoiding

I mean that there will be no task created, and no invocation of any http://clear.ml API whatsoever including no imports in the “core ML task” This is the direction - add very small wrappers of http://clear.ml code around the core ML task. The http://clear.ml wrapper is “aware’ of the core ML code, and never the other way. For cases where the wrapper is only “before” and “after” the core ML task, its somewhat easier to achieve. For reporting artifacts etc. which is “mid flow” - it’s m...

2 years ago
0 <no title>

that’s the thing. I want to it to appear like one long pipeline, vs. trigger a new set of steps after the approval. So “wait” is a better metaphore for me

2 years ago
0 Regarding The “Classic” Datasets (Not Hyper Datasets): Is There An Option To Do Something Equivalent To Dvc’S “

AgitatedDove14 nope… you can run md5 on the file as stored in the remote storage (nfs or s3)

2 years ago
0 Ui Suggestion:

which configuration are you passing? are you using any framework for configuration?

2 years ago
2 years ago
Show more results compactanswers