
Reputation
Badges 1
51 × Eureka!I don't know why it requests localhost
basicaly I don't want to train a new model and I try to create an endpoint following the example but I finally get$ curl -X POST "
" -H "accept: application/json" -H "Content-Type: application/json" -d '{"url": "
` "}'
<html> <head><title>405 Not Allowed</title></head> <body> <center><h1>405 Not Allowed</h1></center> <hr><center>nginx/1.20.1</center> </body> </html> `
and my ~/clearml.conf
api {
web_server:
api_server:
files_server:
# test 3
credentials {
"access_key" = "91SFEX4BYUQ9YCZ9V6WP"
"secret_key" = "4WTXT7tAW3R6tnSi8hzSKNjgkmgUoyv22lYT2FIzIfLoeGERRO"
}
}
I have to step away for a couple of hours
please let me know if you find something wrong
CLEARML_FILES_HOST="
"
curl -X POST "
" -H "accept: application/json" -H "Content-Type: application/json" -d '{"url": "
"}' {"detail":"Error processing request: Error: Failed loading preprocess code for 'py_code_test_model_pytorch2': 'NoneType' object has no attribute 'loader'"}
I changed port here:
clearml-serving-inference:
image: allegroai/clearml-serving-inference:latest
container_name: clearml-serving-inference
restart: unless-stopped
ports:
- "9090:8080"
clearml-serving-inference | 2022-07-03 22:06:26,893 - clearml.storage - ERROR - Could not download
, err: HTTPConnectionPool(host='localhost', port=8081): Max retries exceeded with url: /DevOps/serving%20example%2010.0a76d264e30940c2b600375fa839f1a2/artifacts/py_code_test_model_pytorch2/preprocess.py (Caused by NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fc3f41b1790>: Failed to establish a new connection: [Errno 111] Connection refused'))
seems true
root@9f6a74ab9a27:~/clearml# curl
<!DOCTYPE HTML PUBLIC "-//W3C//DTD HTML 3.2 Final//EN">
<title>405 Method Not Allowed</title>
<h1>Method Not Allowed</h1>
<p>The method is not allowed for the requested URL.</p>
root@9f6a74ab9a27:~/clearml# curl
curl: (7) Failed to connect to localhost port 8081: Connection refused
root@9f6a74ab9a27:~/clearml#
I got only smth like this:
clearml-serving-triton | I0701 08:32:58.580705 46 server.cc:250] Waiting for in-flight requests to complete.
clearml-serving-triton | I0701 08:32:58.580710 46 server.cc:266] Timeout 30: Found 0 model versions that have in-flight inferences
clearml-serving-triton | I0701 08:32:58.580713 46 server.cc:281] All models are stopped, unloading models
clearml-serving-triton | I0701 08:32:58.580717 46 server.cc:288] Timeout 30: Found 0 live ...
seems like an issue about 2 compose apps using different networks which are not accessible from each other
I wonder if I just need to join 2 docker-compose files to run everything in one session
I don't thing WEB_HOST is important, but what about FILE_HOST?
do I need to change it accordingly?
you are right, for some reason it doesn't resolve inside a container
root@dd0252a8f93e:~/clearml# curl
curl: (7) Failed to connect to localhost port 8008: Connection refused
root@dd0252a8f93e:~/clearml# curl
curl: (7) Failed to connect to 127.0.0.1 port 8008: Connection refused
root@dd0252a8f93e:~/clearml#
same thing
clearml-serving-inference | Retrying (Retry(total=236, connect=236, read=240, redirect=240, status=240)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7f899dc4e8b0>: Failed to establish a new connection: [Errno 111] Connection refused')': /auth.login