Reputation
Badges 1
8 × Eureka!Hi @<1523701205467926528:profile|AgitatedDove14> .,
Of course! The output of curl -X POST
command is at least reassuring, it shows that the automatic endpoint works. As you say, the RPC error when sending request seems to be returned from the GPU backend.
Nothing gets printed in docker compose log when sending the curl -X POST
, but beforehand following log is displayed for clearml-serving-triton
container with among others `WARNING: [Torch-TensorRT] - Unable to read CUDA capab...
Hi @<1523701205467926528:profile|AgitatedDove14> !
Thank you again for coming back to me with your support!
- ๐ Thank you, I have noticed that (when using
model auto-update
) different model versions (with their ownmodel_id
) appear under "model_monitoring_eps" section of the serving service. - ๐ It's now clear to me that I was always accessing the static endpoint (that I created with
model add
) with mycurl
command.
I retested automatic model deployment with you...
Well, after testing, I observed two things:
- When using automatic model deployment and training several models to which tag "released" was added, the
model_id
in the "endpoints" section of the Serving Service persistently presents the ID of the initial model that was used to create the endpoint (and NOT the one of the latest trained model) (see first picture below โคต ). This is maybe the way it is implemented in ClearML, but a bit non-intuitive since, when using automa...
Hi @<1523701205467926528:profile|AgitatedDove14> .,
Thanks a lot for your quick reply! ๐ In fact, I am more interested in using the same endpoint with latest model version than effectively creating an endpoint on tagging.
Your statement makes sense, it seems that we have anyway to create an endpoint with model add
prior to set up automatic model deployment with model auto-update
. This seems to work since section "LINEAGE" under my latest trained model gets updated with infor...
Hi @<1523701205467926528:profile|AgitatedDove14> !
Thank you for having a look at this log file ๐ .
Effectively, the Triton backend was not able to load my model. I will investigate this issue that is surely related to my own GPU backend (GeForce RTX 4070 Ti), I suppose ClearML PyTorch example works for other users. I am not sure this is related to the fact the model is not correctly converted to Tor...
Okay, so that's surely the reason why the model is not found, I will investigate that, thank you again for your insight! ๐
Hi @<1523701205467926528:profile|AgitatedDove14> ,
Just for verifying which model is actually called by the endpoint when using model auto-update
for automatic model deployment I performed following steps with ClearML Serving PyTorch example :
- I modified the code of
train_pytorch_mnist.py
in thetrain
function withtarget = torch.zeros(data.shape[0]).long()
in order for the model to bel...
Thank you so much for your reply Martin!
It's clear to me now.
Let's see if this works! I will try waiting those 5 minutes at the beginning of next week and let you know if I can obtain an updated endpoint with the new model id of the latest trained model!
Have a nice weekend!
Best regards.