Examples: query, "exact match", wildcard*, wild?ard, wild*rd
Fuzzy search: cake~ (finds cakes, bake)
Term boost: "red velvet"^4, chocolate^2
Field grouping: tags:(+work -"fun-stuff")
Escaping: Escape characters +-&|!(){}[]^"~*?:\ with \, e.g. \+
Range search: properties.timestamp:[1587729413488 TO *] (inclusive), properties.title:{A TO Z}(excluding A and Z)
Combinations: chocolate AND vanilla, chocolate OR vanilla, (chocolate OR vanilla) NOT "vanilla pudding"
Field search: properties.title:"The Title" AND text
Profile picture
EnthusiasticShrimp49
Moderator
0 Questions, 96 Answers
  Active since 18 February 2023
  Last activity one year ago

Reputation

0
0 Hi, I Would Like To Bring Awareness

I think you can set the cuda version in the clearml.conf , alternatively you can have the agent use a docker image with your required version of cuda instead of setting the environment directly on the machine

11 months ago
0 Hi Guys, I Have A Question Regarding Model Tracking. I Have Pipelines That Use Xgboost Through The Scikit-Learn Api To Perform:

Hey @<1569858449813016576:profile|JumpyRaven4> , about your first point, what exactly is the question?

About your second point - you can try to manually save the final model and give it a proper file name, that way we will show it in the UI with the name you provided. Make sure to use xgboost.save_model and not raw pickle.

For your final question , given that your models have customised code, I can suggest trying to use clearml.OutputModel which will register the file you provide ...

11 months ago
one year ago
0 Hi, I Configured An On-Prem File Server For Clearml Which Is Mounted On My Pc.

What happens if you comment or remove the pipe.set_default_execution_queue('default') and use run_locally instead of start_locally ?

Because in the current setup, you are basically asking to run the pipeline controller task locally, while the rest of the steps need to run on an agent machine. If you do the changes I suggested above, you will be able to run everything on your local machine.

one year ago
0 My Project Pipeline Is Runnung Well And Good But Instead Of Completed It Is Coming As Aborted After Complete Execution

That is not specific enough. Can you show the code? And ideally also the console log of the pipeline

one year ago
0 My Project Pipeline Is Runnung Well And Good But Instead Of Completed It Is Coming As Aborted After Complete Execution

Is this a jupyter notebook or something ? Can you download it properly as either a .ipynb or .py file?

one year ago
0 Hey Guys

Hey, yes, the reason for this issue seems to be our currently limited support for lightning 2.0. We will improve the support in the following releases. Right now one way to circumvent this issue, that I can recommend, is to use torch.save if possible, because we fully support automatic model capture on torch.save calls.

12 months ago
0 Hi There,

Hey @<1523701066867150848:profile|JitteryCoyote63> , could you please open a GH issue on our repo too, so that we can more effectively track this issue. We are working on it now btw

12 months ago
0 Quick Question - Does Clearml'S Task Support Subprocesses Launched Within A Script? I Have This Scenario

Hey @<1535069219354316800:profile|PerplexedRaccoon19> , yes it does. Take a look at this example, and let me know if there are any more questions: None

one year ago
0 Quick Question - Does Clearml'S Task Support Subprocesses Launched Within A Script? I Have This Scenario

Yes, you can do that. But it may make it harder to identify the task later on

one year ago
0 Hi People! I Think The Clearml

Hello @<1523710243865890816:profile|QuaintPelican38> , could you try Dataset.get ing an existent dataset and tell whether there are any errors or not?

one year ago
0 Hi! I Have A Dataset Like This: V1.0.0

You can create a new dataset and specify the parent datasets as all the previous ones. Is that something that would work for you ?

11 months ago
0 Hello Guys I Have A Question About Local Cache Right Now Im Trying To Store In Cache A Pretty Large Dataset (Over 20Mil Files And 3Tb Of Data) I Use A

Hey @<1577468626967990272:profile|PerplexedDolphin99> , yes, this method call will help you limit the number of files you have in your cache, but not the total size of your cache. To be able to control the size, I’d recommend checking the ~/clearml.conf file in the sdk.storage.cache section

11 months ago
0 Hello Guys I Have A Question About Local Cache Right Now Im Trying To Store In Cache A Pretty Large Dataset (Over 20Mil Files And 3Tb Of Data) I Use A

Thanks for pointing this out, we will need to update our documentation. Still, if you manually inspect the ~/clearml.conf file you will see the available configurations

11 months ago
11 months ago
0 Hello, I Saw, That Clearml Data Was Integrated Into Yolov5

Hey Pawel, thanks for opening the PR on Ultralytics’ side. The full support should come from them, so if it’s missing for YOLOv8 it means they didn’t enable it. Still , you can try clearml-task for auto-logging support in case of remote execution .

Also, I’d say you could easily have the possibility to use a ClearML dataset id as input to YOLOv8 with a few lines of code by basically downloading/ get ing the dataset by id yourself and passing the path to it as input to the ultralytics...

10 months ago
0 Hello. I Want To Update An Artifact In A Task (A Pandas Data Frame). I Do This With

Hey @<1547390444877385728:profile|ThickSnake12> , how exactly do you access the artifact next time? Can you provide a code sample?

one year ago
0 Hi, When Running A Task, Is There A Way To Run A Script Already In The Container? For Example There Is A Script /Home/Root/Entrypoint.Sh I Would Like To Avoid Specifying A Repo Or A Local Script On My Machine.

Hey @<1681836314334334976:profile|GrotesqueSeaturtle83> , yes, it is possible to do so, but you must configure the docker --entrypoint argument (as part of the docker_arguments ) and the docker image of for said task. In general this isn't a recommended approach. Rather than that, prefer a setup where your task code invokes the functionalities defined in other scripts that are pre-baked in the image.

See docker args here:
[None](https://clear.ml/docs/latest/docs/references/sdk/task/...

one month ago
0 Hey Guys

It may indeed be, thanks for letting us know, we’ll try to replicate it

one year ago
0 Is There An External Way To Access Pipelinecontroller._Relaunch_Node(Node) ?

Hey @<1639799308809146368:profile|TritePigeon86> , given that you want to retry on connection error, wouldn't it be easier to use retry_on_failure from PipelineController / PipelineDecorator.pipeline None ?

one month ago
0 Hello Guys, I Have 4 Workers (2 In Default And 2 In Service Queue On Same Machine) And Running A Cron Job Of Data Preparation.It Works Well For About 3 Days But After That Tasks Are Getting Failed By Their Own With Given Below Error.Can Anyone Help Me O

Hey @<1526734437587357696:profile|ShaggySquirrel23> , what version of the clearml-agent are you using? Also, if I were you I’d check how much free disk there’s on the machine running the agents

11 months ago
0 Aws { S3 { # S3 Credentials, Used For Read/Write Access By Various Sdk Elements # The Following Settings Will Be Used For Any Bucket Not Specified Below In The "Credentials" Section # ----------------------

@<1637624992084529152:profile|GlamorousChimpanzee22> using localhost I'm assuming it's minio, is the s3 path you're trying to access something like this: None <some file or dir> ?

5 months ago
0 Hi Everyone, I’M Trying To Create A Pipeline From Tasks Without Uploading The Data Into Clearml Server Because

Hey Yasir, to use tensorflow prefetch your data needs to be (1) chunked and (2) stored on some server/bucket/network-attached FS. If both conditions are not satisfied, TF prefetch won't help you.

How large is the dataset we're talking about?

2 months ago
0 Hi Everyone, I’M Trying To Create A Pipeline From Tasks Without Uploading The Data Into Clearml Server Because

That's not that much. You can use the AWS autoscaler and provision a spot g4dn GPU instance with a bit more disk. This should cost you less than 50 cents an hour

2 months ago
0 Hi All ! I'M Trying To Run My Tasks On Open-Source Clearml-Server. But Everytime I Try To Run A Task On It With Clearml-Agent, I Get The Same Error : Clearml_Agent: Error: 'Utf-8' Codec Can'T Decode Byte 0Xe9 In Position 26: Invalid Continuation Byte Even

Hey @<1546303293918023680:profile|MiniatureRobin9> , to help narrow down the problem, could you try to manually download None and open it with pickle ?

Also, is your agent running on the same machine as your server and the example pipeline code? And what Python version are you using for all three components? Because I see there's a warning `could not locate requested Python version 3.11, reverting t...

one year ago
0 Hi There, Does Anyone Have Suggestions For Best Practice For Deploying A Pipeline So That It Can Run Remotely On A Clearml Server Using A Docker Image? I Am Finding The Clearml Docs And Videos Insufficient To Get The Pipeline To Actually Run To Completion

Which gives me an idea. Could you please remove the entrypoint from the docker image altogether and try again ?

Overriding the entrypoint in the image can lead to docker run/docker exec failing to work properly , because instead of a shell it will use your entrypoint to run everything

one month ago
0 Hi Guys, I Have A Question Regarding Model Tracking. I Have Pipelines That Use Xgboost Through The Scikit-Learn Api To Perform:

This is the method you're looking for None . But make sure you have a model saved on disk before using it. And if you don't want the model to be deleted from disk after it, make sure to set auto_delete_file=False

11 months ago
Show more results compactanswers