Examples: query, "exact match", wildcard*, wild?ard, wild*rd
Fuzzy search: cake~ (finds cakes, bake)
Term boost: "red velvet"^4, chocolate^2
Field grouping: tags:(+work -"fun-stuff")
Escaping: Escape characters +-&|!(){}[]^"~*?:\ with \, e.g. \+
Range search: properties.timestamp:[1587729413488 TO *] (inclusive), properties.title:{A TO Z}(excluding A and Z)
Combinations: chocolate AND vanilla, chocolate OR vanilla, (chocolate OR vanilla) NOT "vanilla pudding"
Field search: properties.title:"The Title" AND text
Profile picture
RoughTiger69
Moderator
28 Questions, 101 Answers
  Active since 10 January 2023
  Last activity 8 months ago

Reputation

0

Badges 1

90 × Eureka!
0 Votes
6 Answers
1K Views
0 Votes 6 Answers 1K Views
autoscaler from saas (pro version). I attempted to use the autoscaler “application” from clearml UI. here is what I get in the logs of the autoscaler screen ...
2 years ago
0 Votes
8 Answers
1K Views
0 Votes 8 Answers 1K Views
2 years ago
0 Votes
0 Answers
979 Views
0 Votes 0 Answers 979 Views
autoscaler 101 questions: What is the best practice for managing credentials so that they don’t get saved in clearml webapp? When the https://clear.ml/docs/l...
2 years ago
0 Votes
14 Answers
1K Views
0 Votes 14 Answers 1K Views
question about pipeline and long-waiting tasks: Say I want to generate a dataset. The workflow I have requires query to a DB Creating a labeling assigment in...
2 years ago
0 Votes
10 Answers
1K Views
0 Votes 10 Answers 1K Views
3 years ago
0 Votes
3 Answers
1K Views
0 Votes 3 Answers 1K Views
FYI I am getting a lot of read timeouts from the community server: Retrying (Retry(total=235, connect=240, read=235, redirect=240, status=240)) after connect...
2 years ago
0 Votes
5 Answers
1K Views
0 Votes 5 Answers 1K Views
Hi, I am trying to use the aws autoscaler to assign instance profiles to new machines. This is a better way than managing credentials. I added the configurat...
2 years ago
0 Votes
10 Answers
1K Views
0 Votes 10 Answers 1K Views
hi folks, is there a way to force clear-ml agent with --docker to not create a virtualenv at all? And perhaps not even attempt to install requirements even? ...
2 years ago
0 Votes
9 Answers
1K Views
0 Votes 9 Answers 1K Views
2 years ago
0 Votes
10 Answers
1K Views
0 Votes 10 Answers 1K Views
2 years ago
0 Votes
18 Answers
1K Views
0 Votes 18 Answers 1K Views
Is there a case-study or ref. architecture for interacting with CI/CD i.e. exposing mature pipelines to be triggered upon code pushes (taking latest git hash...
3 years ago
0 Votes
0 Answers
1K Views
0 Votes 0 Answers 1K Views
Hi, I am catching up with http://clear.ml for stuff beyond exp. tracking, and have a few questions. Will ask them separately to allow threading:
3 years ago
0 Votes
5 Answers
1K Views
0 Votes 5 Answers 1K Views
Avoiding http://Clear.ml glue code spaghetti - community best practices? Say I have training pipeline : Task 1 - data preprocessing -> create a dataset artif...
3 years ago
0 Votes
4 Answers
1K Views
0 Votes 4 Answers 1K Views
hi, When running a training script from pycharm, it seems that clearml logs only those packages that are explicitly imported by my .py files; it seems to not...
3 years ago
0 Votes
7 Answers
1K Views
0 Votes 7 Answers 1K Views
2 years ago
0 Votes
0 Answers
1K Views
0 Votes 0 Answers 1K Views
Did more digging, seems that you need to start the agent with CLEARML_AGENT_SKIP_PIP_VENV_INSTALL=1
2 years ago
0 Votes
5 Answers
1K Views
0 Votes 5 Answers 1K Views
hi, I am running a pipeline from decorators. the pipeline runs fine. Then I try to clone it by clicking the (successful) run and launching. The pipeline fail...
2 years ago
0 Votes
8 Answers
1K Views
0 Votes 8 Answers 1K Views
Hi, I am using PipelineDecorator to create tasks. is there a way to force it to use the entire git repo it is created from on the pythonpath? vs. just the de...
2 years ago
0 Votes
9 Answers
1K Views
0 Votes 9 Answers 1K Views
2 years ago
0 Votes
3 Answers
1K Views
0 Votes 3 Answers 1K Views
2 years ago
0 Votes
4 Answers
1K Views
0 Votes 4 Answers 1K Views
hi, I created a dataset with 20K files, total of 20GB, with storage pointing to S3. When I upload (or close) the dataset, during the compression phase, the c...
2 years ago
0 Votes
5 Answers
1K Views
0 Votes 5 Answers 1K Views
I have a logical task that I want to split to multiple workers. The task involves processing media files (not training). The optimal design for me would be: ...
2 years ago
0 Votes
6 Answers
1K Views
0 Votes 6 Answers 1K Views
Regarding the “classic” datasets (not hyper datasets): Is there an option to do something equivalent to dvc’s “ https://dvc.org/doc/user-guide/managing-exter...
2 years ago
0 Votes
3 Answers
1K Views
0 Votes 3 Answers 1K Views
what’s a good ami to use for the clearml autoscaler on AWS? the defaults offered confidently by the various auto scaler installers don’t seem to exist…| e.g....
2 years ago
0 Votes
11 Answers
1K Views
0 Votes 11 Answers 1K Views
I have a local folder a, and a dataset B. a: a a/.DS_Store a/1.txt a/b a/b/.DS_Store a/b/1.txt a/b/c a/b/c/1.txtDataset B: b b/2.txt b/c b/c/2.txtI want to “...
2 years ago
0 Votes
14 Answers
1K Views
0 Votes 14 Answers 1K Views
Two simple lineage related questions: Task B is a clone of Taks A. Does B store the information that it was cloned from A somewhere? Training task X loads Da...
3 years ago
0 Votes
3 Answers
1K Views
0 Votes 3 Answers 1K Views
3 years ago
0 Votes
12 Answers
497 Views
0 Votes 12 Answers 497 Views
Is there a reference implmentation for a task in a pipeline that awaits user input?
3 years ago
0 <no title>

could work! is there a way to visualize the pipeline such that this step is “stuck” in executing?

3 years ago
0 <no title>

thanks so much for taking the time by the way!

3 years ago
0 Hi, I Am Using Pipelinedecorator To Create Tasks. Is There A Way To Force It To Use The Entire Git Repo It Is Created From On The Pythonpath? Vs. Just The Decorated Function And Perhaps The Helper_Function=[Some_Function]?

sure CostlyOstrich36
I have something like the following:

@PipelineDecorator.component(....) def my_task(...) from my_module1 import my_func1 from my_modeul2 import ....my_module1 and 2 are modules that are a part of the same project source. they don’t come as a separate package.

Now when I run this in clearml, these imports don’t work.

These functions may require transitive imports of course, so the following doesn’t work:
` PipelineDecorator.component(helper_function=[my_fu...

2 years ago
3 years ago
0 Hi, I Am Using Pipelinedecorator To Create Tasks. Is There A Way To Force It To Use The Entire Git Repo It Is Created From On The Pythonpath? Vs. Just The Decorated Function And Perhaps The Helper_Function=[Some_Function]?

AgitatedDove14 the emphasis is that the imports I am doing are not from external/pipe packages, they are just neighbouring modules to the function I am importing. Imports that rely on pip installed packages work well

2 years ago
0 Hello Folks! I Have A Pipeline With Three Tasks: A, B, And C I Want To Set It Up So That: A Gets Assigned A Machine (E.G. Based On The Queue) B Always Gets Assigned To The Same Machine As A (But May Run In A Different Docker Etc.) C Will Be Submitted To

I don’t think so.
In most cases I woudl have multiple agents pulling from the same queue. I can’t have a queue per pipeline execution.
So if I submit A and B to the same queue, it still doesn’t gurantee that they will be pulled by the same agent….

2 years ago
0 Is It Possible To Use In Clearml

As far I know storage can be https://clear.ml/docs/latest/docs/integrations/storage/#direct-access .
typical EBS is limited to being mounted to 1 machine at a time.
so in this sense, it won’t be too easy to create a solution where multiple machines consume datasets from this storage type

PS https://docs.aws.amazon.com/AWSEC2/latest/UserGuide/ebs-volumes-multi.html is possible under some limitations

2 years ago
0 Autoscaler From Saas (Pro Version). I Attempted To Use The Autoscaler “Application” From Clearml Ui. Here Is What I Get In The Logs Of The Autoscaler Screen Itself (Consistent):

CostlyOstrich36 from what I gather the UI creates a task in the background, in status “hidden”, and it has like 10 fields of json configurations…

2 years ago
0 Two Simple Lineage Related Questions:

I think that in principal, if you “intercept” the calls to Model.get() or Dataset.get() from within a task, you can collect the ID’s and do various stuff with them. You can store and visualize it for lineage, or expose it as another hyper parameter I suppose.

You’ll just need the user to name them as part of loading them in the code (in case they are loading multiple datasets/models).

3 years ago
0 Two Simple Lineage Related Questions:

Re. “which task did I clone from” - to my understanding “parent’ field is used for “runtime parent” - i.e. what task started me.
This is not the same as “which task was I cloned from”

3 years ago
0 Two Simple Lineage Related Questions:

so I think it will just be confusing

3 years ago
0 Two Simple Lineage Related Questions:

CostlyOstrich36 Lineage information for datasets - oversimplifying but bare with me:
Task should have a section called “input datasets”)
each time I do a Dataset.get() inside a current_task, add the dataset ID to this section

Same can work with InputModel()

This way you can have a full lineage graph (also queryable/visualizable)

3 years ago
0 Two Simple Lineage Related Questions:

I mean, if it’s not tracked, I think it would be a good feature!

3 years ago
0 2. Is There A Case-Study Or Ref. Architecture For Interacting With Ci/Cd I.E. Exposing Mature Pipelines To Be Triggered Upon Code Pushes (Taking Latest Git Hash) Or With Manual Ci Triggers?

I want to have a CI/CD pipeline that, upon Engineer A commit, ensures that the pipeline is re-deployed such that with Engineer B uses it as template, it’s definitely the latest version of the code and process

3 years ago
Show more results compactanswers