Reputation
Badges 1
45 × Eureka!isnt brew for mac ? i am on ubuntu
OPENBLAS="$(brew --prefix openblas)" pip install pandas
Hello , I passed following variables to the agent followingly: CLEARML_AGENT_SKIP_PIP_VENV_INSTALL=1 CLEARML_AGENT_SKIP_PYTHON_ENV_INSTALL=1 clearml-agent daemon --queue "default" --foreground
and it works perfectly ! Thanks for advise.
Hello ExasperatedCrab78 ! I tried to add the PR. I hope it will be sufficient and clear. Have a nice day and thanks for help ! :)
Oh thanks ! 🙂 I understand now. Please let me know about the other problem.
Hi ExasperatedCrab78 ,
am I getting it right that alias = dataset id which can be found in the clearml dashboard?
I also tried the changed url (github) you sent. I can successfully run all the scripts, but I do not see any results from log_dataset_statistics
. Maybe I am wrong but in the https://github.com/allegroai/clearml-blogs/blob/master/urbansounds8k/preprocessing.py this function is not even called. Could you help me with that please? I just want to replicate -> clearml dashboa...
Anyways no histogram / table is reported in dashboard however in tutorial video @ https://www.youtube.com/watch?v=quSGXvuK1IM it is.
Server: 3.13.0-832. I just waited for some time and suddenly it was possible to delete it again. Idk what was that issue. No matter what , if it will appear again I can announce you.
I went into app.clear.ml dashboard -> project -> archived experiments -> and i wanted to delete experiment from there -> that is the output i got
Thanks for help ! I will try and let you know.
So, I just just define all the requirements for the docker in clearml.conf
in default_docker
part?
Do I have to do some port forwarding / add extra parameters ? Copy clearml.conf inside of the docker ? And all the stuff ? Because it does not seems to be done automatically.
So probably you are right - > nc -vz localhost 8080
Output when run locally not in docker: Connection to localhost (127.0.0.1) 8080 port [tcp/http-alt] succeeded!
Output when inside docker bash: localhost [127.0.0.1] 8080 (http-alt) : Connection refused
Yes api server is on the same machine -> running in container
web_server: http://localhost:8080
api_server: http://localhost:8008
files_server: http://localhost:8081
I'm at the point where it looks like the clearml-agent is stuck (How i execute the agent: clearml-agent daemon --queue "default" --gpus 0 --foreground --docker. After the last message: " Successfully installed:<dependencies>" nothing really happens. I do attach logs from experiment. And I also do provide the config:
I just set up my server from following url : https://clear.ml/docs/latest/docs/deploying_clearml/clearml_server_linux_mac/
you can edit the requirements section directly <- where ? if i create requirements.txt it seems to be ignored
I understand what you mean. I am just describing different case. Lets assume i have my docker image already (all dependencies , data solved). Right now I run my task and it automatically looks for requirements.txt file in the repository. My question is -> is there any way to avoid this (simplest solution for now will be to -> rename requirements.txt to different filename)? I tried the things you sent already. The thing is that in requirements.txt in this repos can not be installed that easily...
The agent simply try to install requirements from requirements.txt , however i dont want to do that because i have my docker image ready.
For future you can put in requirements.txt :detectron @
To describe the use-case. Lets say we have someapp which can export specific training script. I would like to create this as a specific "draft" task and later execute it.
Probably i figured out everything. For me it is better to work with https and token (for git).
Insight 2 <- repository is cloned correctly but the uncomitted changes are not.
reinstalling realy solved the problem , thank you. Have a very nice day.
Hi @<1523701070390366208:profile|CostlyOstrich36> , the worker cloned the repo correctly, however in the nested scipt if you use task.init
it wont work / wont overwrite anything.