
Reputation
Badges 1
35 × Eureka!you can either add it manually to the installed packages, or remove the installed packages and use a setup.py file to manage the installation process
there is no /usr/share/elasticsearch/logs/clearml.log
file (neither inside the container nor in my server)
oh but docker-ps
shows me 8081 ports for webserver, apiserver and fileserver containers
` CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES
0b3f563d04af allegroai/clearml:latest "/opt/clearml/wrappe…" 7 minutes ago Up 7 minutes 8008/tcp, 8080-8081/tcp, 0.0.0.0:8080->80/tcp, :::8080->80/tcp clear...
can you share your clearml.conf
file (remove the critical information first)?
ok, I entered the container, replaced all 8081 to 8085 in every file, commited the container and changed the docker-compose.yml
to use that image instead of the allegroai/clearml:latest
and now it works 🙂
you would, but I’d advise against it, since that is not the intended way
would the same experiment be called in either clearml server?
Is not direcly cached in the ~/.clearml
folder. There are some directories inside (one for storage, one for pip, another for venvs, etc.
So in your case it would be stored in ~/.clearml/cache/storage_manager/datasets/ds_{ds_id}/my_file.json
I also changed the permissions of /usr/share/elasticsearch
according to this post: https://techoverflow.net/2020/04/18/how-to-fix-elasticsearch-docker-accessdeniedexception-usr-share-elasticsearch-data-nodes/ , but I’m getting the same error
each person has its laptop..and they do a pip install clearml and set their credentials in the clearml.conf
file
mmm, can you try the following:
create a new folder with no git repo, and copy those two notebooks launch the notebook with the base task and copy the task id launch the notebook with the hyperopt task modifying the TEMPLATE_TASK_ID
variable accordingly
so I removed the entrypoint, and now I can see that it tries to install the packages, but it fails because it can’t download the repo
great! and I saw that there were some system packages needed for opencv that were installed automatically that could be turned off. Now I’m just wondering if I could remove the PIP install at the very beginning, so it starts straightaway
That’s why I’m suggesting him to do that 🙂
ClearML downloads/caches datasets to ~/.clearml/
folder so yes, you need to modify your code.dataset_folder = Dataset.get(project_name=, dataset_name=, version=).get_local_copy() file_json_path = os.path.join(dataset_folder, 'file.json')
if I were to run an agent that would require to install pandas at some point I’d run it:OPENBLAS="$(brew --prefix openblas)" clearml-agent daemon --queue default
Hi ExasperatedCrocodile76 , I guess that you were able to install Scikit-learn and you were able to run it locally, and now you want to try it with an agent on the same machine.
The error is that it can’t find OpenBLAS:
` Run-time dependency openblas found: NO (tried pkgconfig and cmake)
Run-time dependency openblas found: NO (tried pkgconfig)
../../scipy/meson.build:130:0: ERROR: Dependency "OpenBLAS" not found, tried pkgconfig `My question is: did you export some env variabl...
it would be easier for a sysadmin to center the credentials of the bucket in the clearml-server, without the need to distribute them…every user in the server has the same credentials, and they don’t need to know them..makes sense?
please remove rmdatasets == 0.0.1
no because every user that is trying to write in the bucket has the same credentials
Hey! When you say it wasn’t enough, what do you mean? Can you launch the web UI?
right, I’m saying I had to do that in my MAC. In your case you would have to point it to somewhere else. Please check where openblas is installed on your ubuntu