Reputation
Badges 1
70 × Eureka!I want to cache as much as possible and /clearml-cache/venvs-cach
(on the host) does contain caches venvs. But /clearml-cache/venvs-builds
is empty. My question was how to also cache venvs_builds
AgitatedDove14 one more thing regarding the initial question,apt-cache
, pip-cache
, pip-download-cache
, vcs-cache
and venvs-cache
contain data on the shared clearml-cache
but venvs-build
does not? What sort of data would be stored in the venvs-build
folder? I do have venvs_dir = /clearml-cache/venvs-builds
specified in the clearml.conf
it appears at multiple places. Seems like the mapping of pip and apt cache does work but the access rights are now an issue
Ok it is more a docker issue, I guess it is not feasible reading the thread.
Executing: ['docker', 'run',......] chown: changing ownership of '/root/.cache/pip': Operation not permitted Get:1
focal-security InRelease [114 kB] Get:2
focal InRelease [265 kB] Get:3
focal-updates InRelease [114 kB
It is at the top of the logs
Hi AgitatedDove14 one more question about efficient caching, is it possible to cache/share docker images between agents?
so now there is the user conflict between the host and the agent inside the container
the cache on the host is mounted as nfs and the nfs server was configured to not allow the clients to do root operations
or only not for apt and pip?
` # pip cache folder mapped into docker, used for python package caching
docker_pip_cache = /clearml-cache/pip-cache
# apt cache folder mapped into docker, used for ubuntu package caching
docker_apt_cache = /clearml-cache/apt-cache
docker_internal_mounts {
apt_cache: "/clearml-cache/apt-cache"
pip_cache: "/clearml-cache/pip-cache"
vcs_cache: "/clearml-cache/vcs-cache"
venv_build: "/clearml-cache/venvs-builds"
pip_download: "/cl...
probably found the issue
W: chown to _apt:root of directory /var/cache/apt/archives/partial failed - SetupAPTPartialDirectory (1: Operation not permitted) W: chmod 0700 of directory /var/cache/apt/archives/partial failed - SetupAPTPartialDirectory (1: Operation not permitted) Collecting pip==20.1.1
So I don't need docker_internal_mounts
at all?
I do have this setting in my clearml.conf filevenvs_cache: { free_space_threshold_gb: 50.0 path: /clearml-cache/venvs-cache }
So it should cache the venvs right? I also see content in the /clearml-cache/venvs-cache
folder. Because I have venvs_cache configured there is nothing in venvs-build, since it uses the cache?
The agents also share the clearml.conf
file which causes some issue with the worker_id/worker_name. They all want to be ubuntu:gpu0. Any idea how I can randomize it? Setting the CLEARML_WORKER_ID env var somehow does not work
are they in conflict?
I think Anna means that if artifacts and models are stored on the clearml fileserver their path will contain the IP or domain of the fileserver. If you then move the fileserver to a different host, all the urls are broken since the host changed.
SuccessfulKoala55 do you have any example? I guess a lot of people face this issue
SuccessfulKoala55 Hey, for us artifact download urls, model download urls, images in plots and debug image urls are broken. In the linked example I can see a solution for the debug images and potentially plot images but cant find the artifacts and model urls inside ES. Are those urls maybe stored inside the mongodb? Any idea where to find them?
That's it? no apparent error?
After the logs on the top there was only logs on "info" level with PluginsService
Very good news!
` elasticsearch:
networks:
- backend
container_name: clearml-elastic
environment:
ES_JAVA_OPTS: -Xms8g -Xmx8g
bootstrap.memory_lock: "true"
cluster.name: clearml
cluster.routing.allocation.node_initial_primaries_recoveries: "500"
cluster.routing.allocation.disk.watermark.low: 500mb
cluster.routing.allocation.disk.watermark.high: 500mb
cluster.routing.allocation.disk.watermark.flood_stage: 500mb
discovery.zen.minimum_master_no...
curl -XPUT -H 'Content-Type: application/json' 'localhost:9200/_settings' -d '{"index" : {"number_of_replicas" : 0}}
This command made all my indices beside the broken one which is still red, come green again. It comes from https://stackoverflow.com/questions/63403972/elasticsearch-index-in-red-health/63405623#63405623 .
so you say deleting other old indices that I don't need could help?
This did not help, I still have the same issue