Reputation
Badges 1
51 × Eureka!I don't thing WEB_HOST is important, but what about FILE_HOST?
do I need to change it accordingly?
I have to step away for a couple of hours
please let me know if you find something wrong
I tried that, it didn't work. I was confused by the separate port parameter:
CLEARML_SERVING_PORT: ${CLEARML_SERVING_PORT:-8080}
which is only one port related in docker-compose-triton.yml
Can I test /auth.login
somehow independently? Using curl or any other way. Which address does it suppose to have and which creds should I use?
curl -X POST "
" -H "accept: application/json" -H "Content-Type: application/json" -d '{"url": "
"}' curl: (56) Recv failure: Connection reset by peer
seems true
root@9f6a74ab9a27:~/clearml# curl
<!DOCTYPE HTML PUBLIC "-//W3C//DTD HTML 3.2 Final//EN">
<title>405 Method Not Allowed</title>
<h1>Method Not Allowed</h1>
<p>The method is not allowed for the requested URL.</p>
root@9f6a74ab9a27:~/clearml# curl
curl: (7) Failed to connect to localhost port 8081: Connection refused
root@9f6a74ab9a27:~/clearml#
it suppose to have access_key and secret_key which should correspond to this file
I got only smth like this:
clearml-serving-triton | I0701 08:32:58.580705 46 server.cc:250] Waiting for in-flight requests to complete.
clearml-serving-triton | I0701 08:32:58.580710 46 server.cc:266] Timeout 30: Found 0 model versions that have in-flight inferences
clearml-serving-triton | I0701 08:32:58.580713 46 server.cc:281] All models are stopped, unloading models
clearml-serving-triton | I0701 08:32:58.580717 46 server.cc:288] Timeout 30: Found 0 live ...
how d you start a docker-compose?
docker-compose --env-file example.env -f docker-compose.yml up -d
and my ~/clearml.conf
api {
web_server:
api_server:
files_server:
# test 3
credentials {
"access_key" = "91SFEX4BYUQ9YCZ9V6WP"
"secret_key" = "4WTXT7tAW3R6tnSi8hzSKNjgkmgUoyv22lYT2FIzIfLoeGERRO"
}
}
hi WickedElephant66
I have the same issue, but port is not the only problem
https://clearml.slack.com/archives/CTK20V944/p1656446563854059
In my understanding requests still go through clearml-server
which configuration I left intact. Maybe due to the port change in clearml-serving
I need to adjust smth.
I made it working with full port reassignment to 9090 in clearml-serving-inference
which still send me an error that the format of my request is somehow wrong
but then I started from scratch by creating completely new project and new endpoint
curl -X POST "
" -H "accept: application/json" -H "Content-Type: application/json" -d '{"url": "
"}' {"detail":"Error processing request: Error: Failed loading preprocess code for 'py_code_test_model_pytorch2': 'NoneType' object has no attribute 'loader'"}