Examples: query, "exact match", wildcard*, wild?ard, wild*rd
Fuzzy search: cake~ (finds cakes, bake)
Term boost: "red velvet"^4, chocolate^2
Field grouping: tags:(+work -"fun-stuff")
Escaping: Escape characters +-&|!(){}[]^"~*?:\ with \, e.g. \+
Range search: properties.timestamp:[1587729413488 TO *] (inclusive), properties.title:{A TO Z}(excluding A and Z)
Combinations: chocolate AND vanilla, chocolate OR vanilla, (chocolate OR vanilla) NOT "vanilla pudding"
Field search: properties.title:"The Title" AND text
Profile picture
NuttyCamel41
Moderator
12 Questions, 42 Answers
  Active since 18 January 2023
  Last activity 5 months ago

Reputation

0

Badges 1

42 × Eureka!
0 Votes
3 Answers
1K Views
0 Votes 3 Answers 1K Views
one year ago
0 Votes
6 Answers
1K Views
0 Votes 6 Answers 1K Views
Hi there! :) I have an issue regarding the get_local_copy(..) function of the Model class. Whenever this function is called in the course of a hyperparameter...
one year ago
0 Votes
4 Answers
1K Views
0 Votes 4 Answers 1K Views
Hi all! Does anyone know a solution to my issue with deploying models saved on azure on the clearml-serving docker container?
one year ago
0 Votes
3 Answers
591 Views
0 Votes 3 Answers 591 Views
Hi all! I am in the process of setting up clearml-serving on my kubernetes cluster using the provided helm charts. Currently I am stuck with running the cont...
8 months ago
0 Votes
5 Answers
1K Views
0 Votes 5 Answers 1K Views
Hi all, I have the same problem as stated in this Thread. The file encoding of all files in my project is utf-8 and I already set the environment variable PY...
one year ago
0 Votes
6 Answers
1K Views
0 Votes 6 Answers 1K Views
one year ago
0 Votes
0 Answers
786 Views
0 Votes 0 Answers 786 Views
Hi all, I am trying to add a model to my recently setup k8s self-hosted clearml-serving. The command looks like this and worked whith my previous docker setu...
8 months ago
0 Votes
2 Answers
1K Views
0 Votes 2 Answers 1K Views
Hi all, I again have a problem which was already reported here. It seems like the agent ignores the reqirements even when I am explicitely adding them by Tas...
one year ago
0 Votes
22 Answers
1K Views
0 Votes 22 Answers 1K Views
Hi all! I recently started working with clearML serving. I got this example working https://github.com/allegroai/clearml-serving/tree/main/examples/pytorch a...
one year ago
0 Votes
8 Answers
1K Views
0 Votes 8 Answers 1K Views
one year ago
0 Votes
13 Answers
562 Views
0 Votes 13 Answers 562 Views
Hi there! Can anybody help me with specifying the 'platform' for a model in clearml-serving. I am using the k8s clearml-serving setup (version 1.3.1). I alre...
6 months ago
0 Votes
6 Answers
635 Views
0 Votes 6 Answers 635 Views
8 months ago
0 Hi There! Can Anybody Help Me With Specifying The 'Platform' For A Model In Clearml-Serving. I Am Using The K8S Clearml-Serving Setup (Version 1.3.1). I Already Tried A Bunch Of Variants Like

What do you mean by "How are you creating the model?"? I executed a pytorch model training saved a traced version of the model so that saved with the executed task. This was also no problem with the docker container setup.

6 months ago
0 Hi There! Can Anybody Help Me With Specifying The 'Platform' For A Model In Clearml-Serving. I Am Using The K8S Clearml-Serving Setup (Version 1.3.1). I Already Tried A Bunch Of Variants Like

Hi @<1523701205467926528:profile|AgitatedDove14> you are right for the docker setup. But with the k8s setup I get the error Poll failed for model directory 'advanced_basic_classifier.pytorch': unexpected 'platform' and 'backend' pair, got:, pytorch when I do not specify the platform, which sounds like I should specify the platform.

Btw if I do not name the model after the 'model.<backend_name>' convention then I get this error
`Poll failed for model directory 'advanced_basic_classifi...

6 months ago
0 Hi There! Can Anybody Help Me With Specifying The 'Platform' For A Model In Clearml-Serving. I Am Using The K8S Clearml-Serving Setup (Version 1.3.1). I Already Tried A Bunch Of Variants Like

Hi @<1523701205467926528:profile|AgitatedDove14> , exactly!
I just tried the pytorch example from the clearml-serving repo and got the error about the wrong model name Poll failed for model directory 'test_model_pytorch': Invalid model name: Could not determine backend for model 'test_model_pytorch' with no backend in model configuration. Expected model name of the form 'model.<backend_name>'.

6 months ago
0 Hi There! Can Anybody Help Me With Specifying The 'Platform' For A Model In Clearml-Serving. I Am Using The K8S Clearml-Serving Setup (Version 1.3.1). I Already Tried A Bunch Of Variants Like

Hi @<1523701205467926528:profile|AgitatedDove14> the config.pbtxt for 1. looks like this: (because I do not specify input and output type and size within the command)

  backend: "pytorch"
  platform: "pytorch_libtorch"
  input [
    {
      name: "INPUT__0"
      data_type: TYPE_FP32
      dims: [1, 64]
    }
  ]
  output [
    {
      name: "OUTPUT__0"
      data_type: TYPE_FP32
      dims: [1, 11]
    }
  ]

while the config.ptxt for 2. looks like this: (because everything else ...

6 months ago
0 Hi There :) Can Anybody Tell Me What The Best Practice Is For Performing A Normalization In The Preprocess.Py Script Used By Clearml-Serving? Currently I Use A Sklearn Minmaxscaler Which Is Loaded And Applied Before And After The Data Is Send To The Model

Hi @<1523701205467926528:profile|AgitatedDove14> , I serialized a sklearn MinMaxScaler object which I created on the training data using pickle. So when serving the model I would like to load that pickle file in the preprocess script such that I can perform the same normalization as done during training. Unless there is a better practice applying the same normalization during training and serving time.

one year ago
0 Hi There :) Can Anybody Tell Me What The Best Practice Is For Performing A Normalization In The Preprocess.Py Script Used By Clearml-Serving? Currently I Use A Sklearn Minmaxscaler Which Is Loaded And Applied Before And After The Data Is Send To The Model

Hi @<1523701205467926528:profile|AgitatedDove14> thanks for your answer! 🙂 I think my case is a bit different. I do not want to load a custom model but I want to load a custom object used for preprocessing. So I think the load method would not fit, as the local_file_name parameter I get in the load function would lead to the model file. And as far as I can see there is no mechanism installed to load other objects than the model file inside the Preprocess class, right?

one year ago
0 Hi All! I Was Just Wondering What Is The Best Way To Log Additional Information? Right Now I'M Only Printing It To The Console, But That'S Not The Most Pleasant Way To Retrieve The Information Later On. As Far As I Can See, The 'Logger.Report_Text(...)' M

Hi @<1523701087100473344:profile|SuccessfulKoala55> , thanks for your message! 🙂 I am aware that the console is also logged on the server, but I somehow find it not optimal to look for relevant information in the console log and would like to place the information in a more structured way.

one year ago
0 Hi All! I Was Just Wondering What Is The Best Way To Log Additional Information? Right Now I'M Only Printing It To The Console, But That'S Not The Most Pleasant Way To Retrieve The Information Later On. As Far As I Can See, The 'Logger.Report_Text(...)' M

Hi @<1523701323046850560:profile|OutrageousSheep60> , thanks for your message as well. So far I have actually been using these exact functions until I noticed the following: when I run a task with these calls, everything works as expected. However, if I do a hyperparameter tuning and change some of the hyperparameters so that the additional information that is not a hyperparameter also changes, they are not adjusted. For better understanding again my concrete example: I have 3 parameters/inf...

one year ago
0 Hi All! I Was Just Wondering What Is The Best Way To Log Additional Information? Right Now I'M Only Printing It To The Console, But That'S Not The Most Pleasant Way To Retrieve The Information Later On. As Far As I Can See, The 'Logger.Report_Text(...)' M

Hello CostlyOstrich36 , thanks for your question. At the moment I am training a MLP for a regression problem and in one case I want to store the number of neurons per layer. Note that in my case it is not a hyperparameter because I calculate the number of neurons based on the number of layers and the number of model parameters. Another case is that I want to store some local paths where the models are stored, since I currently don't have any remote storage set up for my models.

one year ago
0 Hi All! I Recently Started Working With Clearml Serving. I Got This Example Working

I think you are correct with your guess that the services were not shut down properly. I noticed that some services were still shown as running on the clear ml dashboard. I aborted all and at least got rid of the error ValueError: triton-server process ended with error code 1 . But the two errors you named are still there and I also got these two warnings:
` clearml-serving-triton | Warning: more than one valid Controller Tasks found, using Task ID=4709b0b383a04bb1a033e99fd325dc...

one year ago
0 Hi There! Can Anybody Help Me With Specifying The 'Platform' For A Model In Clearml-Serving. I Am Using The K8S Clearml-Serving Setup (Version 1.3.1). I Already Tried A Bunch Of Variants Like

Hi @<1523701205467926528:profile|AgitatedDove14> thanks for your hint! I already convert it to torch script using tracing. Everything around the model should be fine, since it already worked with the docker clearml-serving setup.
I think the real issue is that I am not able to specify a platform for the model, as the error above tells me that no platform is given no matter how I try to pass it.

6 months ago
0 Hi There :) Can Anybody Tell Me What The Best Practice Is For Performing A Normalization In The Preprocess.Py Script Used By Clearml-Serving? Currently I Use A Sklearn Minmaxscaler Which Is Loaded And Applied Before And After The Data Is Send To The Model

Hi @<1523701205467926528:profile|AgitatedDove14> , that is an interesting idea! But wouldn't it be better to load the model in the load() function, so that the model doesn't have to be loaded again with every request? Or is there kind of internal link that when the load() method is implemented it is expected that there was a custom model loaded and applied in the process() function?

one year ago
0 Hi All! I Recently Started Working With Clearml Serving. I Got This Example Working

Ok, I have some weird update... I shut down and restarted the docker container just to get fresh logs and now I am getting the following error message by clearml-serving-triton
` clearml-serving-triton | clearml-serving - Nvidia Triton Engine Controller
clearml-serving-triton | Warning: more than one valid Controller Tasks found, using Task ID=433aa14db3f545ad852ddf846e25dcf0
clearml-serving-triton | ClearML Task: overwriting (reusing) task id=350a5a919ff648148a3de4483878...

one year ago
0 Hi All! I Recently Started Working With Clearml Serving. I Got This Example Working

Ok, I have found the issue. 🙌 When I try to serve a model which is saved on azure (generated by Task.init(..., output_uri='azure://...') ) I get the poll failed for model directory 'test_model_pytorch': failed to open text file for read /models/test_model_pytorch/config.pbtxt: No such file or directory error. A model which was saved on the clearML server (generated by Task.init(..., output_uri=True) ) can be served without any problems.
For now I am not sure why th...

one year ago
0 Hi All! I Recently Started Working With Clearml Serving. I Got This Example Working

I got the last bit of my issue solved. I thought for a start it would be easier to provide the AZURE_STORAGE_ACCOUNT and AZURE_STORAGE_KEY in my 'example.env' in plain text and not access my environment variables because I was not sure about the syntax. Turns out the syntax is not AZURE_STORAGE_KEY="mystoragekey123" but AZURE_STORAGE_KEY=mystoragekey123 . Same for AZURE_STORAGE_ACCOUNT . Also the syntax for accessing my environment variables is just the same as in the clear...

one year ago
0 Hi All! I Recently Started Working With Clearml Serving. I Got This Example Working

Hi ExasperatedCrab78 , thanks for your answer! In fact I used your recommended format for passing input and output size before and changed it in my debugging process. I have just tried again but got the same error message.
Also thanks for the hint to check the log for warnings I wil do this in a moment.

one year ago
0 Hi All! I Recently Started Working With Clearml Serving. I Got This Example Working

Hi @<1523701118159294464:profile|ExasperatedCrab78> , thanks for your answer. 🙂 Yes sure! I will create the issue right away.

one year ago
0 Hi All! I Recently Started Working With Clearml Serving. I Got This Example Working

Ok, so I killed all docker containers (the proposal by chatgpt did not work for me, but your commands did). The result is that we have one less warning. The warning clearml-serving-triton | Warning: more than one valid Controller Tasks found, using Task ID=4709b0b383a04bb1a033e99fd325dcbf seems to be solved. All remaining errors come up in the clearml-serving-triton service and this is the log I get

CLEARML_SERVING_TASK_ID=9309c20af9244d919b0f063642198c57
CLEARML_TRITON_POLL...
one year ago
0 Hi All! I Recently Started Working With Clearml Serving. I Got This Example Working

By the way, the example which worked for me in the beginning also produces the same error now poll failed for model directory 'test_model_pytorch': failed to open text file for read /models/test_model_pytorch/config.pbtxt: No such file or directory . So there really seems to be something wrong with the docker containers.

one year ago
0 Hi All! I Recently Started Working With Clearml Serving. I Got This Example Working

I got it working!! For now I am not sure what did the trick because I tried a bunch of different things. But I will try to reproduce it and come back to this thread for other users facing this problem. So big thanks for your help, @<1523701118159294464:profile|ExasperatedCrab78> !

one year ago
0 Hi All! I Recently Started Working With Clearml Serving. I Got This Example Working

Hi @<1523701118159294464:profile|ExasperatedCrab78> , I have a sad update on this issue. It does not seem to be completely solved yet. 😕 But I think I can at least describe it a bit better now:

  • Models which are located on the clearML servers (created by Task.init(..., output_uri=True) ) still run perfectly.
  • Models which are located on azure blob storage make different problems in different scenarios (which made me think we resolved this issue):- When I start the docker con...
one year ago
0 Hi There! Can Anybody Help Me With Specifying The 'Platform' For A Model In Clearml-Serving. I Am Using The K8S Clearml-Serving Setup (Version 1.3.1). I Already Tried A Bunch Of Variants Like

Hi @<1523701205467926528:profile|AgitatedDove14> , thanks for coming back to my issue. Unfortunately I have a lot of other stuff on my desk right now so I have to postpone finishing this issue. I will reach out to you again as soon as possible (especially if I was able to find a solution).

5 months ago
0 Hi There! :) I Have An Issue Regarding The

Hi @<1523701070390366208:profile|CostlyOstrich36> , thanks for your answer! I just updated the 'azure_storage_blob' package to the newest version and got some strange behaviour. When running the BOHB hyperparameter optimization, there is only one job executed and not stopped. I aborted the job after 3500 epochs because I set the the max_iteration_per_job parameter to 1000 and the job seems to run infinitely long. I just downgraded the package back to version 12.14.1 and everything works as b...

one year ago
0 Hi There! :) I Have An Issue Regarding The

Yes, I also find that very weird... I start the hyperparameter optimization via python code using the HyperParameterOptimizer class of clearml. Which configurations are you explicitely interested in?
When comparing the logs of the two hpo tasks it seems like no logs of the subtasks are getting to the hpo task. So maybe this is the reason for the infinitely long running subtask? But what does the azure package have to do with that?

one year ago
0 Hi There! :) I Have An Issue Regarding The

This is the log of the hpo task with the newest azure azure-storage-blob version

one year ago
0 Hi There! :) I Have An Issue Regarding The

This is the hpo task log with the azure-storage-blob in version 12.14.1

one year ago
0 Hey All, I Want To Purchase The Pro Version Of Clearml But Would Like To Have A Better Understanding Of The Metric Events And Api Calls That Are Performed When Using Clearml-Serving. For Example: I Have No Understanding Which Docker Container Calls The Ap

Hi @<1523701205467926528:profile|AgitatedDove14> , thanks for your answer!
I reached over 1M API calls in about one week using clearml-serving on one machine with only calling the deployed model a few hundred times for testing purpose. So I wanted to dig a little bit deeper on that. Thanks for the channel suggestion, I will repost my question there. :)

one year ago
0 Hi All, I Have The Same Problem As Stated In This Thread. The File Encoding Of All Files In My Project Is Utf-8 And I Already Set The Environment Variable

Yes I am running the agent by calling clearml-agent daemon --queue default in my virtual environment on my local computer.

one year ago
0 Hi All! Does Anyone Know A Solution To My Issue With Deploying Models Saved On Azure On The Clearml-Serving Docker Container?

Hi @<1523701205467926528:profile|AgitatedDove14> , thanks for your answer! Can you tell me, how specifically I map my clearml.conf to the containers? By the way, the credentials are already set (and working) in the clearml.conf.

one year ago
Show more results compactanswers