Examples: query, "exact match", wildcard*, wild?ard, wild*rd
Fuzzy search: cake~ (finds cakes, bake)
Term boost: "red velvet"^4, chocolate^2
Field grouping: tags:(+work -"fun-stuff")
Escaping: Escape characters +-&|!(){}[]^"~*?:\ with \, e.g. \+
Range search: properties.timestamp:[1587729413488 TO *] (inclusive), properties.title:{A TO Z}(excluding A and Z)
Combinations: chocolate AND vanilla, chocolate OR vanilla, (chocolate OR vanilla) NOT "vanilla pudding"
Field search: properties.title:"The Title" AND text
Profile picture
ExasperatedCrab78
Moderator
2 Questions, 221 Answers
  Active since 10 January 2023
  Last activity one year ago

Reputation

0

Badges 1

2 × Eureka!
0 Votes
0 Answers
1K Views
0 Votes 0 Answers 1K Views
A little something else: Using ClearML, an OAK-1 AI camera and a raspberry pi to create a pushup counter that locks my PC every hour and only unlocks again w...
2 years ago
0 Votes
5 Answers
1K Views
0 Votes 5 Answers 1K Views
We're working on ClearML serving right now and are very interested in what you all are searching for in a serving engine, so we can make the best serving eng...
2 years ago
0 Hello Everyone ! I Tried To Reproduce Your Tutorial :

am I getting it right that alias = dataset id which can be found in the clearml dashboard?

Not really. It's Alias + dataset_id that will be found in the clearml dashboard 🙂 I'm attaching a screenshot of what that looks like in both the code and the dashboard

2 years ago
0 Another Question On The Topic Of How A Remote Execution Of A Pipeline Kills The Calling Process (Previously Discussed

Also, the answer to blocking on the pipeline might be in the .wait() function: https://clear.ml/docs/latest/docs/references/sdk/automation_controller_pipelinecontroller#wait-1

TimelyPenguin76 I can't seem to make it work though, on which object should I run the .wait() method?

2 years ago
0 Another Question On The Topic Of How A Remote Execution Of A Pipeline Kills The Calling Process (Previously Discussed

Hey PanickyMoth78
Here is an easy to reproduce, working example. Mind the multi_instance_support=True parameter in the pipeline itself. This code launches 3 pipelines for me just as it should 🙂
` from clearml.automation.controller import PipelineDecorator
import time

PipelineDecorator.set_default_execution_queue("default")

@PipelineDecorator.component()
def step_one():
time.sleep(2)

@PipelineDecorator.component()
def step_two():
time.sleep(2)

@PipelineDecorator.pipel...

2 years ago
0 Another Question On The Topic Of How A Remote Execution Of A Pipeline Kills The Calling Process (Previously Discussed

Hi PanickyMoth78 , I have made a minimal example and indeed adding multi_instance_support=True prevents ClearML from killing the process, allowing you to launch pipelines in a loop 🙂

2 years ago
2 years ago
0 We'Re Working On Clearml Serving Right Now And Are Very Interested In What You All Are Searching For In A Serving Engine, So We Can Make The Best Serving Engine We Can

Hi Jax! Thanks for the feedback, we really appreciate it 😄

MLFlow is able to support a multitude of models from dev to deployment. Is ClearML and ClearML-Serving going to support as much as well?

Do you mean by this that you want to be able to seamlessly deploy models that were tracked using ClearML experiment manager with ClearML serving?

I believe in such scenarios, a custom engine would be required. I would like to know, how difficult is it to create a custom engine wit...

2 years ago
0 Hi All, I'M Having Some Issues With Syncing Modified Files Using

Well I'll be had, you're 100% right, I can recreate the issue. I'm logging it as a bug now and we'll fix it asap! Thanks for sharing!!

one year ago
0 Hi All, I'M Having Some Issues With Syncing Modified Files Using

Hi LackadaisicalDove24 !

Does this happen with every csv file? If so, I can reproduce it to check if it is a bug 🙂

one year ago
0 When Dumping Model Via Clearml Serving, What Are The Things That The Clearml Will Look At To Populate The Input_Size And Output_Size? I Tried To Dump An Sklearn Model, And The Input_Size And Output_Size Is Null. I Prefer Not To Update It Separately Using

Just to be sure I understand you correctly: you're saving/dumping an sklearn model in the clearml experiment manager, then want to serve it using clearml serving, but you do not wish to specify the model input and ouput shapes in the CLI?

one year ago
0 Hi. I Am Experimenting With

AgitatedDove14 I was able to recreate the error. Simply by running Lavi's example on clearml==1.6.3rc1 in a fresh env. I don't know what is unique to the flow itself, but it does seem reproducible

2 years ago
0 Hi Team,When Clearml-Agent Is Used To Run The Code,I T Will Setup The Environment ,How It Take The Python Package Version?

Hi @<1533257278776414208:profile|SuperiorCockroach75> , the clearml experiment manager will try to detect your package requirements from its original environment. Meaning that if you run the code and it imports e.g. SQLAlchemy, then it will log the exact version of SQLAlchemy you have installed locally.

When you run only get_data,py locally and have the experiment manager track it, can you then look at the task that is made in the clearml webUI and check the installed packages section? ...

one year ago
0 Hi Everyone! I Faced The Problem With Clearml-Serving. I'Ve Deployed Onnx Model From Higgingface In Clearml-Serving, But

Hey! Sorry, didn't fully read your question and missed that you already did it. It should not be done inside the clearm-serving-triton service but instead inside the clearml-serving-inference service. This is where the preprocessing script is ran and it seems to be where the error is coming from.

one year ago
0 Hello

Can you run

ls -la /opt/clearml

and paste the results here?

one year ago
0 Hello

What kind of machine are you running this on?

one year ago
0 Hey Guys! I Was Wondering If There Are Any Good Examples In Deployment Of Cv (Detection) Models Through Clearml-Serving? The Examples I Have Seen So Far Are Classification, And So The Post-Processing From The Triton Output Is Minimal. As An Example Mayb

That's a good idea! I think the YOLO models would be a great fit for a tutorial/example like this. We can add it to our internal list of TODOs, or if you want, you could take a stab at it and we'll try to support you through it 🙂 It might take some engineering though! Serving is never drag and drop 🙂

That said, I think it should be quite easy to do since YOLOv8 supports exporting to tensorrt format, which is native to Triton serving which underlies ClearML serving. So the process shoul...

one year ago
0 Hey, We Are Using Clearml 1.9.0 With Transformers 4.25.1… And We Started Getting Errors That Do Not Reproduce In Earlier Versions (Only Works In 1.7.2 All 1.8.X Don’T Work):

Hey @<1523701949617147904:profile|PricklyRaven28> I'm checking! Have you updated anything else and on which exact commit of transformers are you now?

one year ago
0 Hi There, Another Triton-Related Question: Are We Able To Deploy

Hey! Thanks for all the work you're putting in and the awesome feedback 😄

So, it's weird you get the shm error, this is most likely our fault for not configuring the containers correctly 😞 The containers are brought up using the docker-compose file, so you'll have to add it in there. The service you want is called clearml-serving-triton , you can find it [here](https://github.com/allegroai/clearml-serving/blob/2d3ac1fe63637db1978df2b3f5ea4903ef59788a/docker/docker-...

one year ago
0 Hi There, Another Triton-Related Question: Are We Able To Deploy

Hi @<1547028116780617728:profile|TimelyRabbit96> Awesome that you managed to get it working!

one year ago
0 Hi All! I Recently Started Working With Clearml Serving. I Got This Example Working

Hey @<1526371965655322624:profile|NuttyCamel41> Thanks for coming back on this and sorry for the late reply. This looks like a bug indeed, especially because it seems to be working when coming from the clearml servers.

Would you mind just copy pasting this info into a github issue on clearml-serving repo? Then we can track the progress we make at fixing it 🙂

one year ago
0 Side Note For Documentation,

I've passed it through, we'll update the docs, thanks!!

one year ago
0 Hi There, Another Triton-Related Question: Are We Able To Deploy

Yes, you will indeed need to add all ensemble endpoints separately 🙂

one year ago
0 Hello! Is There Any Way To Access The The

I'm not quite sure what you mean here? From the docs it seems like you should be able to simply send an HTTP request to the localhost url to get the metrics. Is this not working for you? Otherwise, all the metrics end up in Prometheus, so you can also query that instead or use something like Grafana to visualize it

one year ago
0 Hello, I Am Using Datasets In Community Server. I Have Been Trying To Create A Child Dataset The Following Way: Dataset_Name = "Training2501" Dataset_Project = "Datasets" Dataset_Path = Dataset.Create( Dataset_Name=Dataset_Name, Dataset_Project=Da

That would explain why it reports the task id to be 'a' in the error. It tried to index the first element in a list, but took the first character of a string instead.

2 years ago
0 Hi All! I Recently Started Working With Clearml Serving. I Got This Example Working

Wow awesome! Really nice find! Would you mind compiling your findings to a github issue, then we can help you search better :) this info is enough to get us going at least!

one year ago
2 years ago
Show more results compactanswers