Examples: query, "exact match", wildcard*, wild?ard, wild*rd
Fuzzy search: cake~ (finds cakes, bake)
Term boost: "red velvet"^4, chocolate^2
Field grouping: tags:(+work -"fun-stuff")
Escaping: Escape characters +-&|!(){}[]^"~*?:\ with \, e.g. \+
Range search: properties.timestamp:[1587729413488 TO *] (inclusive), properties.title:{A TO Z}(excluding A and Z)
Combinations: chocolate AND vanilla, chocolate OR vanilla, (chocolate OR vanilla) NOT "vanilla pudding"
Field search: properties.title:"The Title" AND text
Profile picture
FierceHamster54
Moderator
29 Questions, 178 Answers
  Active since 10 January 2023
  Last activity one year ago

Reputation

0

Badges 1

131 × Eureka!
0 Votes
8 Answers
2K Views
0 Votes 8 Answers 2K Views
Hey, is there a shortcut on the Dataset SDK to directly get the latest version of a dataset ?
2 years ago
0 Votes
20 Answers
2K Views
0 Votes 20 Answers 2K Views
Hey currently trying to run a pipeline locally to test a pipeline component with PipelineDecorator.run_locally() , first try returned a random pandas error, ...
2 years ago
0 Votes
1 Answers
2K Views
0 Votes 1 Answers 2K Views
Heya, good day to everyone, I'm finding myself facing this random error with a very opaque backtrace when attempting to squash two distinct versions of the s...
2 years ago
0 Votes
6 Answers
2K Views
0 Votes 6 Answers 2K Views
Hey guys, i hope you all have a nice days, I had to use the Task method .setup_aws_upload(bucket=...,region=...) to overcome a incorrect region specified for...
2 years ago
0 Votes
4 Answers
2K Views
0 Votes 4 Answers 2K Views
2 years ago
0 Votes
9 Answers
2K Views
0 Votes 9 Answers 2K Views
Hello everyone! I setup a GCP autoscaler on my Pro SaaS deployment but I keep getting this error: clearml_agent: ERROR: Server does not support --use-owner-t...
3 years ago
0 Votes
15 Answers
2K Views
0 Votes 15 Answers 2K Views
Heya, trying to setup a GCP autoscaler for general-purpose CPU instances ( e2-standard-4 ) but I get into that error: googleapiclient.errors.HttpError: And I...
2 years ago
0 Votes
15 Answers
2K Views
0 Votes 15 Answers 2K Views
Hey having an issue passing parameters to a component in a pipeline, the parameters appear to be None inside the component function: @PipelineDecorator.compo...
2 years ago
0 Votes
7 Answers
2K Views
0 Votes 7 Answers 2K Views
Heya, is there any plan for ClearML to leverage the new https://developer.nvidia.com/blog/getting-the-most-out-of-the-a100-gpu-with-multi-instance-gpu/ tech ...
2 years ago
0 Votes
8 Answers
1K Views
0 Votes 8 Answers 1K Views
Hey everyone, I am having some difficulties passing environment variables on my piepeline components running on agents ( 1.6.1 ) without docker mode: - I set...
one year ago
0 Votes
1 Answers
2K Views
0 Votes 1 Answers 2K Views
Hey is ClearML MLFlow based ? Is it exposed to CVE-2023-1176 and CVE-2023-1177 ?
2 years ago
0 Votes
13 Answers
2K Views
0 Votes 13 Answers 2K Views
2 years ago
0 Votes
5 Answers
2K Views
0 Votes 5 Answers 2K Views
Hey, would it be possible to add a way to edit autoscaler configs without having to clone them ? This is really frustrating especially when you reached the q...
2 years ago
0 Votes
8 Answers
2K Views
0 Votes 8 Answers 2K Views
2 years ago
0 Votes
6 Answers
2K Views
0 Votes 6 Answers 2K Views
one year ago
0 Votes
1 Answers
2K Views
0 Votes 1 Answers 2K Views
2 years ago
0 Votes
4 Answers
2K Views
0 Votes 4 Answers 2K Views
2 years ago
0 Votes
9 Answers
2K Views
0 Votes 9 Answers 2K Views
Hey just wanting to know: what is the recommended best practice to write ClearML Pipelines between controller and decorators ?
3 years ago
0 Votes
7 Answers
1K Views
0 Votes 7 Answers 1K Views
Hey everyone, As a Pro-tier SaaS user, I'm experiencing a very high latency when finalizing a dataset, it is attached in a big dataset version hierarchy and ...
one year ago
0 Votes
10 Answers
2K Views
0 Votes 10 Answers 2K Views
2 years ago
0 Votes
25 Answers
2K Views
0 Votes 25 Answers 2K Views
Hey, trying to figure out how to create an https://clear.ml/docs/latest/docs/clearml_sdk/model_sdk#output-models , the doc says it needs a TaskId but my trai...
2 years ago
0 Votes
7 Answers
2K Views
0 Votes 7 Answers 2K Views
Hey I hope everyone is having a good day, two quick questions about datasets: Does squashing two datasets deletes the two original datasets ? Is it possible ...
2 years ago
0 Votes
2 Answers
2K Views
0 Votes 2 Answers 2K Views
Hey, it is said in the pipeline decorator example that requirements for executing a pipeline component is inferred from the imports inside the component func...
2 years ago
0 Votes
7 Answers
2K Views
0 Votes 7 Answers 2K Views
Hey, is there a way to pass docker args for the execution of a pipeline controller defined through decorator @PipelineDecorator.pipeline the same way we can ...
2 years ago
0 Votes
12 Answers
2K Views
0 Votes 12 Answers 2K Views
Hey has anyone managed to capture Darts logging with ClearML when using the temporal fusion transformers ? Even when overriding their trainer with a custom P...
2 years ago
0 Votes
13 Answers
2K Views
0 Votes 13 Answers 2K Views
Hey all, hope you're having a great day, having an unexpected behavior with a training task of a YOLOv5 model on my pipeline, I specified a task in my traini...
2 years ago
0 Votes
4 Answers
1K Views
0 Votes 4 Answers 1K Views
2 years ago
0 Votes
1 Answers
1K Views
0 Votes 1 Answers 1K Views
Hey is .get_local_copy() thread-safe ? I mean can I concurrently download several datasets without breaking the cache and StorageManager ?
2 years ago
0 Votes
15 Answers
2K Views
0 Votes 15 Answers 2K Views
Heya, I hope you're all well in this beautiful day, my GCP Autoscaler just died with that strange but short backtrace, wondered if it rang a bell to any of y...
2 years ago
2 years ago
0 Hey, Is There A Way To Pass Docker Args For The Execution Of A Pipeline Controller Defined Through Decorator

Yes but not in the controller itself, which is also remotely executed in a docker container

2 years ago
0 Heya, Trying To Setup A Gcp Autoscaler For General-Purpose Cpu Instances (

AnxiousSeal95 Okay it seems to work with a compute optimized c2-standard-4 instance

2 years ago
0 Hey, Is There A Shortcut On The Dataset Sdk To Directly Get The Latest Version Of A Dataset ?

AgitatedDove14 I have annotation logs from the end-user that I fetch periodically, I process it and I want to add it as a new version of my dataset where all versions correspond to the data collected during a precise time window, currently I'm doing it by fetching the latest dataset, incrementing the versionmm and creating a new dataset version

2 years ago
0 Hey, Is There A Shortcut On The Dataset Sdk To Directly Get The Latest Version Of A Dataset ?

And by extension is there a way to upsert a dataset by automatically creating an entry wich a incremented version or create it if it does not exists ? Or am I forced to do a get, check if the latest version is fainallyzed, then increment de version of that version and create my new version ?

2 years ago
0 Hey Has Anyone Managed To Capture Darts Logging With Clearml When Using The Temporal Fusion Transformers ? Even When Overriding Their Trainer With A Custom Pytorch Lightning Trainer It Seems That Clearml Cannot Retrieve The Iteration Log...

The expected behavior is that the task would capture the iteration scalar of the PL trainer but nothing is recorded

import clearml
from darts.models import TFTModel

model = TFTModel(
    input_chunk_length=28,
    output_chunk_length=14,
    n_epochs=300,
    batch_size=4096,
    add_relative_index=True,
    num_attention_heads=4,
    dropout=0.3,
    full_attention=True,
    save_checkpoints=True,
)

task = Task.init(
    project_name='sales-prediction',
    task_name='TFT Training 2'...
2 years ago
0 Hey Has Anyone Managed To Capture Darts Logging With Clearml When Using The Temporal Fusion Transformers ? Even When Overriding Their Trainer With A Custom Pytorch Lightning Trainer It Seems That Clearml Cannot Retrieve The Iteration Log...

No I was was pointing out the lack of one, but turns out on some model the iteration is so slow even on GPUs when training on a lots of time serie that you have to set the pytorch lightning trainer argument log_every_n_steps to 1 (default 50 ) to prevent the ClearML iteration logger from timing-out

2 years ago
0 Hey Currently Trying To Run A Pipeline Locally To Test A Pipeline Component With

I have a pipeline with a single component:
` @PipelineDecorator.component(
return_values=['dataset_id'],
cache=True,
task_type=TaskTypes.data_processing,
execution_queue='Quad_VCPU_16GB'
)
def generate_dataset(start_date: str, end_date: str, input_aws_credentials_profile: str = 'default'):
"""
Convert autocut logs from a specified time window into usable dataset in generic format.
"""
print('[STEP 1/4] Generating dataset from autocut logs...')
import os
...

2 years ago
0 Hey All, Hope You'Re Having A Great Day, Having An Unexpected Behavior With A Training Task Of A Yolov5 Model On My Pipeline, I Specified A Task In My Training Component Like This:

SmugDolphin23 But the training.py has already a CLearML task created under the hood since its integration with ClearML, beside initing the task before the execution of the file like in my snippet is not sufficient ?

2 years ago
0 Hey Having An Issue Passing Parameters To A Component In A Pipeline, The Parameters Appear To Be

Nice it works 😍
I'll try to update the version in the image I provide to the workers of th autoscaler app (but sadly I don't control the version of those in itself since it's CLearML managed)

2 years ago
0 Hey, Is There A Way To Pass Docker Args For The Execution Of A Pipeline Controller Defined Through Decorator

Takling about that decorator which shouyld also have a docker_arg param since it is executed as an "orchestration component" but the param is missing https://clear.ml/docs/latest/docs/references/sdk/automation_controller_pipelinecontroller/#pipelinedecoratorpipeline

2 years ago
0 Hey Having An Issue Passing Parameters To A Component In A Pipeline, The Parameters Appear To Be

Hey SuccessfulKoala55 currently using the clearml package version 1.7.1 and my server is a PRO SaaS deployment

2 years ago
0 Heya, Trying To Setup A Gcp Autoscaler For General-Purpose Cpu Instances (

As specified in the initial message, the instance type used is e2-standard-4

2 years ago
0 Hi There, Currently I Have A Clearml Pipeline That Takes In A Bunch Of Parameters For Various Tasks And Passes These Parameters Via Parameter_Override For Every Pipe.Add_Step(). However, I Have A Lot Of Parameters, And So My Pipeline Code Is A Little Unwi

Well if you have:

ret_obj = None
for in in range(5):
    ret_obj = step_x(ret_obj)

SInce the orchestration automatically determine the order of execution using the logic of return objects the controller will execute them sequentially.

However, if your steps don't have dependencies like this:

for i in range(5):
    step_x(...)

It will try to execute them concurrently

one year ago
0 Hey Guys, I Hope You All Have A Nice Days, I Had To Use The

ClearML package version used: 1.9.1
ClearML Server: SaaS - Pro Tier

2 years ago
0 Heya, Is There Any Plan For Clearml To Leverage The New

Oh wow, would definitely try it out if there were an Autoscaler App integrating it with ClearML

2 years ago
0 Hey Everyone, I'M Using Clearml Pipeline, I Have A Pre-Processing Step Which I Would Like To Use There Small Instance, And Processing Step Which Executed On A Gpu Queue. The Initialisation Time Of The Gpu Instance Is Very Long. Is There A Way To Configure

You can set a dummy step which is executed in parallel of your pre-processing step and which is set to be executed in your GPU queue, provided that your autoscaler doesn't scale back your compute before your pre-processing is complete that should do the trick

one year ago
0 Hey Currently Trying To Run A Pipeline Locally To Test A Pipeline Component With

Would gladly try to run it on a remote instance to verify the thesis on some local cache acting up but unfortunately also ran into an issue with the GCP autoscaler https://clearml.slack.com/archives/CTK20V944/p1665664690293529

2 years ago
0 Hey Currently Trying To Run A Pipeline Locally To Test A Pipeline Component With

So basically CostlyOstrich36 I feel like debug_pipeline() use the latest version of my code as it is defined on my filesystem but the run_locally() used a previous version it cached somehow

2 years ago
0 Hey Currently Trying To Run A Pipeline Locally To Test A Pipeline Component With

The value of start_date and end_date seems to be None

2 years ago
0 Heya, Is There Any Plan For Clearml To Leverage The New

Hey, I'm a SaaS user in PRO tier and I was wondering if it was a feature available on the auto-scaler apps so I could improve the cost-efficiency of my provisionned GCP A100 instances

2 years ago
0 Hey Currently Trying To Run A Pipeline Locally To Test A Pipeline Component With

print(f"start_date: {start_date} end_date: {end_date}") time_range = pd.date_range(start=start_date, end=end_date, freq='D').to_pydatetime().tolist()

2 years ago
0 Hey Currently Trying To Run A Pipeline Locally To Test A Pipeline Component With

Component's prototype seems fine:
@PipelineDecorator.component( return_values=['dataset_id'], cache=False, task_type=TaskTypes.data_processing, execution_queue='Quad_VCPU_16GB', ) def generate_dataset(start_date: str, end_date: str, input_aws_credentials_profile: str = 'default'):

2 years ago
0 Hey Just Wanting To Know: What Is The Recommended Best Practice To Write Clearml Pipelines Between Controller And Decorators ?

Nice, that's a great feature! I'm also trying to have a component executing Giskard QA test suites on model and data, is there a planned feature when I can suspend execution of the pipeline, and display on the UI that this pipeline "steps" require a human confirmation to go on or stop while displaying arbitrary text/plot information ?

3 years ago
Show more results compactanswers