Unanswered
Hi Guys, I Am Trying To Upload And Serve A Pre-Existing 3-Rdparty Pytorch Model Inside My Clearml Cluster. However, After Proceeding With The Suggested Sequence Of Operations By Official Docs And Later Even Gpt O3, I Am Having Errors Which I Cannot Solve.
Hi, SuccessfulKoala55 Yeah, sure, please, wait a sec - I will rerun the command. :)
Here's the command and output:
clearml-serving model add --endpoint deepl_query --engine triton --model-id 8df30222595543d3a3ac55c9e5e2fb15 --input-size 7 1 --input-type float32 --output-size 6 --output-type float32 --input-name layer_0 --output-name layer_99
clearml-serving - CLI for launching ClearML serving engine
Notice! serving service ID not provided, selecting the first active service
Warning: more than one valid Controller Tasks found, using Task ID=ccb7bafba16e416ba5590ca717f05de0
Serving service Task ccb7bafba16e416ba5590ca717f05de0, Adding Model endpoint '/deepl_query/'
Info: syncing model endpoint configuration, state hash=ce7bbe44e5dead79f03e9ca8e28d45a6
Warning: Model endpoint 'deepl_query' overwritten
Updating serving service
Note: I would gladly avoid triton as it requires parameters I don't even understand, but seems there is no other option to run the pytorch or other neural network models otherwise.
Also, GPT suggested, that there must be some preprocessing for model file itself to convert it from PTH to something called ONNX, but I have no idea what it is and whether it is actually needed.
32 Views
0
Answers
one month ago
one month ago