Examples: query, "exact match", wildcard*, wild?ard, wild*rd
Fuzzy search: cake~ (finds cakes, bake)
Term boost: "red velvet"^4, chocolate^2
Field grouping: tags:(+work -"fun-stuff")
Escaping: Escape characters +-&|!(){}[]^"~*?:\ with \, e.g. \+
Range search: properties.timestamp:[1587729413488 TO *] (inclusive), properties.title:{A TO Z}(excluding A and Z)
Combinations: chocolate AND vanilla, chocolate OR vanilla, (chocolate OR vanilla) NOT "vanilla pudding"
Field search: properties.title:"The Title" AND text
Profile picture
RattySparrow90
Moderator
4 Questions, 8 Answers
  Active since 23 December 2023
  Last activity one month ago

Reputation

0

Badges 1

7 × Eureka!
0 Votes
4 Answers
144 Views
0 Votes 4 Answers 144 Views
2 months ago
0 Votes
2 Answers
211 Views
0 Votes 2 Answers 211 Views
Hi everyone, is there a possibility to point the ClearML Server to an already existing elasticsearch instance instead of deploying and connecting to its 'own...
4 months ago
0 Votes
4 Answers
110 Views
0 Votes 4 Answers 110 Views
Hi, we are currently seeing the following error in our logs of the ClearML apiserver pod: [2024-03-20 15:33:32,089] [8] [WARNING] [elasticsearch] POST None [...
one month ago
0 Votes
2 Answers
106 Views
0 Votes 2 Answers 106 Views
Hi, we recently deployed ClearML server and ClearML serving to our test infrastructure where it is running lika a breeze 🙂 . Moving forward in the productio...
one month ago
0 Hi Guys, I Have Been Running The Clearml-Serving For A While Now And I Realize That From Time To Time After A Couple Of Hours The Serving Task (Control Plane) That Is Configured Through The Cli Goes Into Status Abort. This Happens Even Though All The Pods

@<1523701205467926528:profile|AgitatedDove14> I experience the exact same behaviour for the clearml-serving (version 1.3.0). Status of the serving-task goes to Aborted, status message is also "Forced stop (non-responsive)" and also after a while of no incoming traffic

2 months ago
0 Hi Everyone, I Have A Few Questions To Understand The Clearml-Serving A Little Better (And How Much Resources To Allocate For The Serving Pods): Are The Models I Defined To Be Served E.G. Via The Cli Downloaded To The Serving Pod? So That They Are Physica

Thanks again, I was able to locate the files. And it was indeed (as most of the time with k8s) a general routing issue. After fixing this everything works fine 🙂

2 months ago
0 Hi Everyone, I Have A Few Questions To Understand The Clearml-Serving A Little Better (And How Much Resources To Allocate For The Serving Pods): Are The Models I Defined To Be Served E.G. Via The Cli Downloaded To The Serving Pod? So That They Are Physica

@<1523701205467926528:profile|AgitatedDove14> Thanks for the explanations. Where exactly are the model files stored on the pod? I was not able to find them.
The reason I ask is that the clearml serving pod is up and running and from the logs and the logs of the fileserver ist seems that the model and the preprocessing code was loaded.
Currently I encounter the problem that I always get a 404 HTTP error when I try to access the model via the defaultBaseServeUrl + model endpoint and I would li...

2 months ago
one month ago
0 Hi, We Are Currently Seeing The Following Error In Our Logs Of The Clearml Apiserver Pod:

@<1523701087100473344:profile|SuccessfulKoala55> Only one

one month ago
0 Hi, We Are Currently Seeing The Following Error In Our Logs Of The Clearml Apiserver Pod:

By the way, is there the possibility to decrease the log level of the api server and the file server? In the ClearML serving deployment a uvicorn log level environment variable can be set. Is there something similar available for the ClearML api and file server? I searched a little bit in the code and did not really find a place where the log level is defined

one month ago