But you have to do config.pbtxt stuff right?
Ah, just saw from the example that even that is doing the config pbtxt stuff - https://github.com/allegroai/clearml-serving/blob/main/examples/keras/keras_mnist.py#L51
Found the custom backend aspect of Triton - https://github.com/triton-inference-server/python_backend
Is that the right way?
Sure, got it. Will play around with it 🙂
Here’s an example error I get trying it out on one of the example models:Error: Requested Model project=ClearML Examples name=autokeras imdb example with scalars tags=None not found. 'config.pbtxt' could not be inferred. please provide specific config.pbtxt definition.
Hey SuccessfulKoala55 Like I mentioned, I have a spacy ner model that I need to serve for inference.
AlertBlackbird30 - i don’t understand why it can’t be a focus though. Probably missing some context.
For now that's a quick thing, but for actual use I will need a proper model (pkl) and the .py
The agent ip? Generally what’s the expected pattern to deploy and scale this for multiple models?
AgitatedDove14 - i had not used the autoscaler since it asks for access key. Mainly looking for GPU use cases - with sagemaker one can choose any instance they want and use it, autoscaler would need set instance configured right? need to revisit. Also I want to use the k8s glue if not for this. Suggestions?
Generally like the kedro project and pipeline setup that I have seen so far, but haven’t started using it in anger yet. Been looking at clearml as well, so wanted to check how well these two work together
Task.add_requirements
would fit the bill yeah, thanks
Maybe two thing here:
If Task.init() is called in an already running task, don’t reset auto_connect_frameworks? (if i am understanding the behaviour right) Option to disable these in the clearml.conf
Was asking about using iam roles without keys
Thanks AgitatedDove14 - i get overall what you are saying. Have to get glue setup, which I couldn’t understand fully, so that’s a different topic 🙂
Yeah was planning to use nested projects for that
Do people generally update the same model “entry”? That feels so wrong to me…how do you reproduce a older model version or do a rollback etc?
Ok, got it thanks. Would be cool to let it get untracked as well, especially if we want to as an option
That's cool AgitatedDove14 , will try it out and pester you a bit more. 🙂
AgitatedDove14 - apologies for late reply. So to give context this in a Sagemaker notebook which has conda envs.
I use a lifecycle like this to pip install a package (a .tar.gz downloaded from s3) in a conda env- https://github.com/aws-samples/amazon-sagemaker-notebook-instance-lifecycle-config-samples/blob/master/scripts/install-pip-package-single-environment/on-start.sh
In the notebook I can do things like create experiments and so on. Now the problem is in running the cloned experimen...
Thanks for the fast responses as usual AgitatedDove14 🙂