Hi @<1523701205467926528:profile|AgitatedDove14> !
Thank you for having a look at this log file 🙏 .
Effectively, the Triton backend was not able to load my model. I will investigate this issue that is surely related to my own GPU backend (GeForce RTX 4070 Ti), I suppose ClearML PyTorch example works for other users. I am not sure this is related to the fact the model is not correctly converted to TorchScript since I am directly testing ClearML PyTorch example (with no modifications, hence also including this line torch.jit.script(model).save('
serving_model.pt ')
) 🤔 . At least, the automatic endpoint exists! 😅 👍
That's weird that I have no problem to make inferences with static endpoint and however face this issue with automatic endpoint... 🙄
I tried to making use of docker-compose-triton-gpu.yml
instead of docker-compose-triton.yml
, but I still get the issue clearml-serving-triton | I1127 09:55:19.366794 57 libtorch.cc:349] Inference Mode is disabled for model instance 'test_model_pytorch_auto_1'
.
Again, thank you very much for your help and insight! 🙇