
Reputation
Badges 1
12 × Eureka!@<1523701087100473344:profile|SuccessfulKoala55> Also, there's one more thing that is bugging me: I have my model files on a remote host in the same LAN (.68 machine), so I try to push them to the model storage of clearml server (.69 machine).
But as far as I understand, I must provide either the URL or local path to the model file in order for ClearML SDK to send it to server machine. So I provide the absolute local path on my .68 device.
However, when I open the model storage on .69 and...
Hi, @<1523701087100473344:profile|SuccessfulKoala55> Yeah, sure, please, wait a sec - I will rerun the command. :)
Here's the command and output:
clearml-serving model add --endpoint deepl_query --engine triton --model-id 8df30222595543d3a3ac55c9e5e2fb15 --input-size 7 1 --input-type float32 --output-size 6 --output-type float32 --input-name layer_0 --output-name layer_99
clearml-serving - CLI for launching ClearML serving engine
Notice! serving service...
Ok, @<1523701087100473344:profile|SuccessfulKoala55> , I was partially able to find one of the incorrect parts of my serving setup:
- Pytorch models inference require me to have .ENV file and clearml-serving-triton-gpu docker configured and running.
- Configuration of .ENV requires me to provide the clearml-serving Service ID, which was created by clearml-serving create.
- I have multiple services created via that command, as there is no command to remove the others, only to create add...
Hi, @<1523701205467926528:profile|AgitatedDove14> , host OS is Ubuntu, I connect there via ssh.
The docker compose is of version 2 (the one that uses "docker compose" instead of older "docker-compose").
I did not pass anything to or from docker manually, only used the commands according to the official guide for clearml-serving:
pip install clearml-serving
clearml-serving create --name deeplog-inference-test --project LogSentinel
git clone
nano .env # here I added my...
Hi @<1523701087100473344:profile|SuccessfulKoala55> , thank you for the reply!
Yes, I am talking about clearml-serving.
I will be near my pc in nearest couple of hours and will send the list of commands as well as a visual scheme of an architecture. :)
@<1523701087100473344:profile|SuccessfulKoala55> Thank you once again, I extracted the scripts and commands, that seemingly were responsible for model registration and its inference on GPU worker server:
register_model.py
from clearml import Task, OutputModel
task = Task.init(project_name="LogSentinel", task_name="Model Registration")
model_path = "~/<full_local_path_to_model>/deeplog_bestloss.pth"
# Register the model
output_model = OutputModel(task=task)
output_model....
Here's the simplified diagram of the architecture:
@<1523701205467926528:profile|AgitatedDove14> ClearML server itself and all of its components (API server etc.) are on x.x.x.69 machine.
Agents and serving are on x.x.x.68 worker machine. My model files are also there, just placed in some usual non-shared linux directory.
And I didn't do any specific configurations of the clearml fileserver docker - everything is on its defaults without a single line changed except the IP address of the ClearML server.
I tried a couple of approaches to u...
Also, @<1523701205467926528:profile|AgitatedDove14> , thank you very much for your advice regarding archive - I did that, removed all current clearml-serving services, created a new one, attached its ID to the ENV file, disabled all running serving dockers and then restarted the clearml-serving-triton-gpu
docker, adding a model file afterwards.
I don't see any docker run errors now in clearml webui tasks console, but now serving is not able to locate the model file itself, and that file...
Hi @<1523701205467926528:profile|AgitatedDove14> , I don't remember it well as I initially installed ClearML about half a year ago, but as far as I remember, I didn't preconfigure any specific queue.
HOWEVER, the first thing I have done in webui - I accidentally deleted the "default" queue, and later, when my clearml agents began to fail due to its abscence, had to use api to create another queue named "my_default_queue" with "default" system tag - then it was fixed.
Here are the logs from...
Also, I tested out the reachability of an endpoint with a CURL query made from modifying the example in Clearml-serving tutorial: https://clear.ml/docs/latest/docs/clearml_serving/clearml_serving_tutorial ,
and it returns error 405: method not allowed
:
curl -X POST -H "accept: application/json" -H "Content-Type: application/json" -d '{"log_sequence": [1, 2, 3, 4, 5, 6, 7, 8, 9, 10]}'
<htm...
@<1523701205467926528:profile|AgitatedDove14> Please, correct me, if I am wrong: are you currently proposing the following sequence:
- On a device, that hosts clearml server, I should have my model file in any directory.
- Then, I should upload it to the clearml model repository as
OutputModel
directly?
Because today I did try to upload the model using the following script:
from clearml import Task, OutputModel
# Step 1: Initialize a Task
task = Task.init(project_name="LogSentine...