
Reputation
Badges 1
76 × Eureka!I had an old cookie from an old deployment that’s why it didn’t recognize my user…
The cookie has a ttl of 3000+ years…
why locally it doesn’t try to run docker?
Can you point me to the direction maybe?
I already made changes in the clearml-server repo and made my own image for the helm-charts… maybe I could fix this as well
I see some keys from apiserver that are a couple of hundreds of mb
simple helm install, didn’t change anything in values
Ok thanks, I’m searching a way to not be dependent on the aws auto scaler script but rather use a k8s job which will run and die after it’s finished thus saving resources
Some way more k8s native like a job with a base image and the exported task
oooo my mistake it’s ubuntu…
I thought it’s a custom image that you guys did
I see the hosts.conf file in clearml-server, I’m trying to add args and build the image again
So it’s enough to add hosts.conf inside additionalConfigs?
One of the keys:
b”<class ‘apiserver.database.model.base.GetMixin.GetManyScrollState’>/f4010ba7df0f45dbbea10a71fe568a94"
Any idea why?
I’m running “docker build -t clearml-server:1.0 -f docker/build/Dockerfile .”
Do I need to build it on linux maybe?
it has 3 “-” in the DAEMON_OPTIONS
PLus I don’t understand where is the repo for that image because the repo allegroai/clearml-agent has a different entrypoint file
Also added the “CLEARML_SUPPRESS_UPDATE_MESSAGE” env which didn’t work…
Thanks for the reply 👍
I only enabled the pod and created services queue
When running with --local True
everything works btw… but locally 🙃