Well from the error it seems there is no layer called "dense" , hence triton failing to find the layer returning the reult. Does that make sense?
Okay, here is the full docker-compose log AgitatedDove14 Thank You.
clearml-serving --id cd4c615583394719b9019667068954bd model add --engine triton --endpoint "test_model_lstm2" --preprocess "preprocess.py" --name "train lstmae model - serving_model" --project "serving examples" --input-size 1 60 1 --input-name "lstm_input" --input-type float32 --output-size -1 60 1 --output-name "dense" --output-type float32
This is how i add my model and set the endpoint, is it right? AgitatedDove14
GG AgitatedDove14 IT WORKS!!! I changed from "dense" to "time_distributed"
THANK YOU SO MUCH!!
NICE! MoodyCentipede68 this is awesome 🙂
MoodyCentipede68 can you post the full docker-compose log (from spinning it until you get the error?)
You can just pipe the output to a file with :docker-compose ... up > log.txt
MoodyCentipede68 from your log
clearml-serving-triton | E0620 03:08:27.822945 41 model_repository_manager.cc:1234] failed to load 'test_model_lstm2' version 1: Invalid argument: unexpected inference output 'dense', allowed outputs are: time_distributed
This seems the main issue of triton failing to.load
Does that make sense to you? how did you configure the endpoint model?