Examples: query, "exact match", wildcard*, wild?ard, wild*rd
Fuzzy search: cake~ (finds cakes, bake)
Term boost: "red velvet"^4, chocolate^2
Field grouping: tags:(+work -"fun-stuff")
Escaping: Escape characters +-&|!(){}[]^"~*?:\ with \, e.g. \+
Range search: properties.timestamp:[1587729413488 TO *] (inclusive), properties.title:{A TO Z}(excluding A and Z)
Combinations: chocolate AND vanilla, chocolate OR vanilla, (chocolate OR vanilla) NOT "vanilla pudding"
Field search: properties.title:"The Title" AND text
Profile picture
PanickyMoth78
Moderator
34 Questions, 167 Answers
  Active since 10 January 2023
  Last activity 21 days ago

Reputation

0

Badges 1

166 × Eureka!
0 Votes
7 Answers
715 Views
0 Votes 7 Answers 715 Views
Hi. I am experimenting with clearml.Dataset and encountering an error. LockException: [Errno 11] Resource temporarily unavailable In my experiment, I make a ...
2 years ago
0 Votes
14 Answers
863 Views
0 Votes 14 Answers 863 Views
Hi. I have a job that processes images and creates ~5 GB of processed image files (lots of small ones). At the end - it creates a clearml.Dataset and perform...
one year ago
0 Votes
2 Answers
729 Views
0 Votes 2 Answers 729 Views
Hi. I'm using @PipelineDecorator.component to define a task from a function (to run in a pipeline) I'd like to get the task object within this function so th...
2 years ago
0 Votes
11 Answers
746 Views
0 Votes 11 Answers 746 Views
Hi. I have a few questions about the snippet attached re-running this code produces the same printouts... I chose 47 out of 100 in the pipeline ... I chose 8...
2 years ago
0 Votes
16 Answers
903 Views
0 Votes 16 Answers 903 Views
Hi. Question about Dataset upload errors: When uploading a clearml.Dataset created with output_uri=" gs://lavi_test/datasets after adding 20 files of size 50...
gcp
one year ago
0 Votes
3 Answers
699 Views
0 Votes 3 Answers 699 Views
2 years ago
0 Votes
2 Answers
816 Views
0 Votes 2 Answers 816 Views
I am using the AWS autoscaler and I wish to set my files server to be gs. I tried to do so by having this in the ADDITIONAL CLEARML CONFIGURATION window: api...
2 years ago
0 Votes
25 Answers
863 Views
0 Votes 25 Answers 863 Views
Autoscaler parallelization issue: I have an AWS Autoscaler set up with a resource that has a max of 3 instances assigned to the default queue I've given it a...
2 years ago
0 Votes
8 Answers
815 Views
0 Votes 8 Answers 815 Views
2 years ago
0 Votes
7 Answers
700 Views
0 Votes 7 Answers 700 Views
I have 5 unarchived pipeline runs that were defined with this decorator: @PipelineDecorator.pipeline( name="fastai_image_classification_pipeline", project="l...
2 years ago
0 Votes
3 Answers
775 Views
0 Votes 3 Answers 775 Views
2 years ago
0 Votes
27 Answers
805 Views
0 Votes 27 Answers 805 Views
Hi. I'm running this little pipeline: from clearml.automation.controller import PipelineDecorator from clearml import TaskTypes @PipelineDecorator.component(...
2 years ago
0 Votes
14 Answers
706 Views
0 Votes 14 Answers 706 Views
Bug? dataset name is ignored if use_current_task=True
one year ago
0 Votes
9 Answers
799 Views
0 Votes 9 Answers 799 Views
Hi. I have a question about pipelines and their generated dependency graphs. I took the code of the clearml pipeline from decorator example: https://github.c...
2 years ago
0 Votes
3 Answers
739 Views
0 Votes 3 Answers 739 Views
Hi. First time user here 👋 I have experienced a problem following the getting started documentation. I opened an account on https://app.clear.ml/ I then fol...
2 years ago
0 Votes
2 Answers
867 Views
0 Votes 2 Answers 867 Views
I have a training task that auto-magically saves a model for me to GCS task = Task.init( project_name=project_name, task_name=f"Image classification training...
one year ago
0 Votes
8 Answers
851 Views
0 Votes 8 Answers 851 Views
2 years ago
0 Votes
22 Answers
915 Views
0 Votes 22 Answers 915 Views
I started two pipelines (using AWS autoscaler in app.clear.ml ). The pipelines ran concurrently, using the same pipeline code. Both failed in the same compon...
2 years ago
0 Votes
1 Answers
724 Views
0 Votes 1 Answers 724 Views
suppose I use a pipeline decorator to define a pipeline: @PipelineDecorator.pipeline(name='my-pipeline', project='my-project', version='0.2') def my_pipeline...
2 years ago
0 Votes
22 Answers
772 Views
0 Votes 22 Answers 772 Views
Hi. I'm encountering a problem with model.name At least, for models that where auto-magically uploaded. I see it in my own code but you can see it if you run...
one year ago
0 Votes
8 Answers
777 Views
0 Votes 8 Answers 777 Views
Hi (again... sorry for asking so many questions) Question about using google cloud storage in a clearml agent running in AWS ec2 instance. my clearml.conf ha...
2 years ago
0 Votes
7 Answers
732 Views
0 Votes 7 Answers 732 Views
Hi. I have a problem accessing repo code in pipeline components running in an AWS autoscaler (first attempts at doing this) My local clearml.conf file has ag...
2 years ago
0 Votes
6 Answers
742 Views
0 Votes 6 Answers 742 Views
Is there some built-in way in clearml to trigger further action on task fail (or pipeline fail)?
one year ago
0 Votes
7 Answers
760 Views
0 Votes 7 Answers 760 Views
Hi I'm looking into how clearml supports datasets and dataset versioning and I'm a bit confused. Is dataset versioning not supported at all in the non-enterp...
2 years ago
0 Votes
13 Answers
725 Views
0 Votes 13 Answers 725 Views
Another question on the topic of how a remote execution of a pipeline kills the calling process (previously discussed https://clearml.slack.com/archives/CTK2...
2 years ago
0 Votes
3 Answers
757 Views
0 Votes 3 Answers 757 Views
Hi. Shoulf this command succeed in the presence of project lavi-testing and absence of dataset tmp_datset within it? from clearml import Dataset tmp_dataset ...
2 years ago
0 Votes
2 Answers
735 Views
0 Votes 2 Answers 735 Views
Hi. I've noticed that my clearml.conf has both: agent.git_user="" agent.git_pass=""and agent { ... git_user: "" git_pass: "" ... }What's the difference? Shou...
2 years ago
0 Votes
9 Answers
720 Views
0 Votes 9 Answers 720 Views
Hi. Help 🥺 I have a clearml.Datase which I can't get
2 years ago
0 Votes
14 Answers
745 Views
0 Votes 14 Answers 745 Views
Hi there. I'm trying to switch pipeline code from a local run using PipelineDecorator.run_locally()to a slightly-less-local run using PipelineDecorator.set_d...
2 years ago
0 Votes
2 Answers
808 Views
0 Votes 2 Answers 808 Views
Hi. Suppose I want to report on what my task has done by having it generate a markdown (.md) file with links to some "local" figure files. looking at the rep...
one year ago
Show more results questions
0 Hi. I Have A Job That Processes Images And Creates ~5 Gb Of Processed Image Files (Lots Of Small Ones). At The End - It Creates A

I ran another version of the above code where
output_uri="./random_dataset_local_target"
(i.e. db target on local disk instead of gcp).
I still see large memory usage.
I also find it worrisome that while generating the random dataset and writing it to disk took under 3 minutes, generating the hash took 9 minutes and saving the files to a dataset target in an adjacent folder took 30 minutes (10 times longer than writing the original files)! Simply copying the files to an adjacent folde...

one year ago
0 Hi. I'M Running This Little Pipeline:

Thanks TimelyPenguin76 .
From your reply I understand that I have control over what the destination is but that all files generated in a task get transferred regardless of the return_values decorator argument. Is that correct? Can I disable auto-save of artifacts?
Ideally, I'd like to have better control over what gets auto-saved. E.g. I'm happy for tensorboard events to be captured and shown in clearml and for matplotlib figures to be uploaded (perhaps to gcs) but I'd like to avoid ...

2 years ago
0 Hi There. I'M Trying To Switch Pipeline Code From A Local Run Using

first, thanks for having these discussions. I appreciate this kind of support is an effort 🙏
Yes. i perfectly understand that once a pipeline job (or a task) is sent off in this manner, it executes separately (and, most likely in a different machine) from the process that instantiated it.
I still feel strongly that such a command should not be thought of as a fire and exit operation. I can think of several scenarios where continued execution of the instantiating process is desired:
I ...

2 years ago
0 Hi. I'D Like To Try The Gcp Autoscaler.

Hi TimelyPenguin76
Thanks for working on this. The clearml gcp autoscaler is a major feature for us to have. I can't really evaluate clearml without some means of instantiating multiple agents on GCP machines and I'd really prefer not to have to set up a k8 cluster with agents and manage scaling it myself.

I tried the settings above with two resources, one for default queue and one for the services queue (making sure I use that image you suggested above for both).
The autoscaler started up...

2 years ago
0 Hi. Question About Dataset Upload Errors: When Uploading A

If

Dataset.upload()

does not crash or return a success value that I can check and

Are you saying that with this error showing upload data does not crash? (edited)

Unfortunately that is correct. It continues as if nothing happened!

To replicate this in linux (even with max_workers=1 ):
https://averagelinuxuser.com/limit-bandwidth-linux/ to throttle your connection: sudo apt-get install wondershaper
Throttle your connection to 1mb/s with somethin...

one year ago
0 I Started Two Pipelines (Using Aws Autoscaler In App.Clear.Ml ). The Pipelines Ran Concurrently, Using The Same Pipeline Code. Both Failed In The Same Component Half-Way Though The Pipeline Run With:

start a training task. From what I can tell from the console log, the agent hasn't actually started running the component.
This is the component code. It is a wrapper around a non-component training function
` @PipelineDecorator.component(
return_values=["run_model_path", "run_info"],
cache=True,
task_type=TaskTypes.training,
repo="git@github.com:shpigi/clearml_evaluation.git",
repo_branch="main",
packages="./requirements.txt",
)
def train_image_classifier_component(
...

2 years ago
0 Hi. I'D Like To Try The Gcp Autoscaler.

I noticed that the base docker image does not appear in the autoscaler task' configuration_object
which is:
` [{"resource_name": "cpu_default", "machine_type": "n1-standard-1", "cpu_only": true, "gpu_type": "", "gpu_count": 1, "preemptible": false, "num_instances": 5, "queue_name": "default", "source_image": "projects/ubuntu-os-cloud/global/images/ubuntu-1804-bionic-v20220131", "disk_size_gb": 100}, {"resource_name": "cpu_services", "machine_type": "n1-standard-1", "cpu_only": true, "gp...

2 years ago
2 years ago
0 Another Question On The Topic Of How A Remote Execution Of A Pipeline Kills The Calling Process (Previously Discussed

I've also not figured out how to modify the examples above to wait for one pipline to end before the next begins

2 years ago
0 Hi. I'D Like To Try The Gcp Autoscaler.

I'll try a more carefully checked run a bit later but I know it's getting a bit late in your time zone

2 years ago
0 Hi. I'M Encountering A Problem With

BTW:

If I try to find the right model in the

task.models["output"]

(this time there is just one but in my code there may be several) it appears with the

(see other attached screenshot).

What would make sense here ? (I have to be honest I'm not sure).

If the model was saved with a file name (is that the trigger for auto-upload?), I think it makes sense for the model name to match the file name (not the task name), especially when there may be ...

one year ago
0 Hi. I'M Encountering A Problem With

To be specific there is "model name" which is not unique , and there is model-key which is unique to the Task

not sure why the two fields don't simply match. I guess that there may be situations where file name (without the full path) may be used several times.

one year ago
0 Hi. I'M Encountering A Problem With

anyhow - looks like the keys are simple enough to use (so I can just ignore the model names)

one year ago
2 years ago
0 Hi. I'M Encountering A Problem With

Ooh nice.
I wasn't aware task.models["output"] also acts like a dict.
I can get the one I care about in my code with something like task.models["output"]["best_model"]
however can you see the inconsistency between the key and the name there:

one year ago
0 Hi. I'M Encountering A Problem With

yes. several checkpoints + the one that did best on validation data.

one year ago
0 Hi There. I'M Trying To Switch Pipeline Code From A Local Run Using

actually, re-running pipeline_from_decorator.py a second time (and a third time) from the command line seem to have executed without the that ValueError so maybe that issue was some fluke.
Nevertheless, those runs exit prior to line
print('process completed')
and I would definitely prefer the command executing_pipeline to not kill the process that called it.
For example, maybe, having started the pipeline I'd like my code to also report having started the pipeline to som...

2 years ago
0 Hi. I'D Like To Try The Gcp Autoscaler.

I'll give it a try.
And if I wanted to support GPU in the default queue, are you saying that I'd need a different machine from the n1-standard-1 ?

2 years ago
0 Hi There I'M Trying Out Clearml. I Saw Mention That Clearml Can Capture Tensorboard Output So I Tried It With This Little Script (Image Below). The Events File Is Filled, The Clearml Task Is Created, And Marked Complete However There Is Nothing In The Sc

here is the code in text if you feel like giving it a try:
import tensorboard_logger as tb_logger from clearml import Task task = Task.init(project_name="great project", task_name="test_tb_logging") task_tb_logger = tb_logger.Logger(logdir='./tb/run1', flush_secs=2) for i in range(10): task_tb_logger.log_value("some_metric", 42, i) task.close()

2 years ago
0 Hi There. I'M Trying To Switch Pipeline Code From A Local Run Using

I'm on clearml 1.6.2
The jupyter notebook service and two clear-ml agents ( version1.3.0, one in queue "default" and one in queue "services" and with --cpu-only flag) ) are all running inside a docker container

2 years ago
0 Hi. I'M Encountering A Problem With

another weird thing:
Before my training task is done:
print(task.models['output'].keys())outputs
odict_keys(['Output Model #0', 'Output Model #1', 'Output Model #2'])
after task.close()
I can do:
task = Task.get_task(task_id) for i in range(100): print(task.models["output"].keys())which prints
odict_keys(['Output Model #0', 'Output Model #1', 'Output Model #2'])in the first iteration
and prints the file names in the latter iterations:
` od...

one year ago
0 Hi. I'D Like To Try The Gcp Autoscaler.

Trying to switch to a resources using gpu-enabled VMs failed with that same error above.
Looking at spawned VMs, they were spawned by the autoscaler without gpu even though I checked that my settings ( n1-standard-1 and nvidia-tesla-t4 and https://console.cloud.google.com/compute/imagesDetail/projects/ml-images/global/images/c0-deeplearning-common-cu113-v20220701-debian-10?project=ml-tooling-test-external image for the VM) can be used to make vm instances and my gcp autoscaler...

2 years ago
2 years ago
Show more results compactanswers