Hi @<1838387863251587072:profile|JealousCrocodile85> , sure!
Is there a way to easily deploy a model using the UI, or should it be done via the CLI with providing the model id etc.?
For none LLM models there is only the CLI (Which you can automate). For LLMs there is an app engine that allows you to launch them directly - None
I want the models to be isolated because of packages versions. So, for every model deployed, will it serve it on a different docker and I need a different port?
I think it will serve it on a different endpoint, the port will most likely be the same.
I'm currently using the same server for ClearML Server and clearml-serving (in the docker-compose.yml, I set the clearml-serving-inference ports to 8082:8080). Should it work well? or should I expect issues and move to 2 different servers?
So do you also have a GPU on that machine or are you serving models that do not require a GPU? The baseline assumption is that serving ("worker") is done on a machine separate from the ClearML server (control plane)
In the clearml_serving_setup documentation, it says "Edit the environment variables file (
docker/example.env
) with your clearml-server credentials". Do they mean to create a ".env" file? Or is it being created with "example.env" file?
These are the API credentials you get from the webUI (Settings -> Workspace)
Thank you very much, @<1523701070390366208:profile|CostlyOstrich36> !