Reputation
Badges 1
383 × Eureka!I just want to change git remote like https://gitserver.com/path/to.git -> mailto:git@gitserver.com :path/to.git
As of now solving by updating the git config locally before creating the task
Yeah please if you can share some general active ones to discuss both algos and engineering side
A channel here would be good too 🙂
AgitatedDove14 - it does have boto but the clearml-serving installation and code refers to older commit hash and hence the task was not using them - https://github.com/allegroai/clearml-serving/blob/main/clearml_serving/serving_service.py#L217
Initially it was complaining about it, but then when I did the connect_configuration it started working
The agent ip? Generally what’s the expected pattern to deploy and scale this for multiple models?
For now that's a quick thing, but for actual use I will need a proper model (pkl) and the .py
forking and using the latest code fixes the boto issue at least
don’t know what’s happening there
AgitatedDove14 - looks like the serving is doing the savemodel stuff?
https://github.com/allegroai/clearml-serving/blob/main/clearml_serving/serving_service.py#L554
Also btw, is this supposed to be screenshot from community verison? https://github.com/manojlds/clearml-serving/blob/main/docs/webapp_screenshots.gif
Model says PACKAGE, that means it’s fine right?
It did pick it from the task?
That makes sense - one part I am confused on is - The Triton engine container hosts all the models right? Do we launch multiple gorups of these in different projects?
I used .update_weights(path)
with path being the model
dir containing the model.py annd the config.pbtxt. Should I use update_weights_package
?
Got the engine running.
curl <serving-engine-ip>:8000/v2/models/keras_mnist/versions/1
What’s the serving-engine-ip supposed to be?
Yes, I have no experience with triton does it do lazy loading? Was wondering how it can handle 10s, 100s of models. If we load balance across a set of these engine containers with say 100 models and all of these models get traffic but distribution is not even, each of those engine container will load all those 100 models?
I am also not understanding how clearml-serving is doing the version for models in triton.
Think I will have to fork and play around with it 🙂
Without some sort of automation on top feels a bit fragile
Progress with boto3 added, but fails:
I am essentially creating a EphemeralDataset abstraction and creating controlled lifecycle for it such that the data is removed after a day in experiments. Additionally and optionally, data created during a step in a pipeline can be cleared once the pipeline completes
So General would have created a General instead of Args?
What happens if I do blah/dataset_url
?
Yes, for datasets where we need GDPR compliance