Hi @<1838387863251587072:profile|JealousCrocodile85>
I'm assuming this is with clearml-serving, notice that it cannot install the correct scikit learn package per endpoint, you have to specify it in the docker compose or k8s helm, see example here
Example
https://github.com/clearml/clearml-serving/blob/5c7077537ad46439f864f24e99e2ea5d4d5b35b3/docker/docker-compose.yml#L103
'''
services:
clearml-serving-inference:
image: allegroai/clearml-serving:latest
environment:
- CLEARML_API_ACCESS_KEY=${CLEARML_API_ACCESS_KEY}
- CLEARML_API_SECRET_KEY=${CLEARML_API_SECRET_KEY}
- CLEARML_SERVING_TASK_ID=${CLEARML_SERVING_TASK_ID}
# Add your extra packages here (space‑separated)
- CLEARML_EXTRA_PYTHON_PACKAGES=scikit-learn==1.7.1
ports:
- "8080:8080"
'''
Thank you for your answer, @<1523701205467926528:profile|AgitatedDove14> !
I've managed to compose a docker with the needed version.
Should I be deploying the entire docker file for every model, with the updated requirements?
Or, can I deploy everything (Prometheus, Grafana, etc.) once and make a serving docker yml for each model with a different port?
Eventually, I want many models (with different package versions) served within a single machine