Reputation
Badges 1
45 × Eureka!OPENBLAS="$(brew --prefix openblas)" pip install pandas
Thanks ! That is exactly what I meant 🙂
Have a nice day!
For future you can put in requirements.txt :detectron @
I understand what you mean. I am just describing different case. Lets assume i have my docker image already (all dependencies , data solved). Right now I run my task and it automatically looks for requirements.txt file in the repository. My question is -> is there any way to avoid this (simplest solution for now will be to -> rename requirements.txt to different filename)? I tried the things you sent already. The thing is that in requirements.txt in this repos can not be installed that easily...
but for the local execution virtual env (conda) named detectron2 was used
Hello ExasperatedCrab78 ! I tried to add the PR. I hope it will be sufficient and clear. Have a nice day and thanks for help ! :)
reinstalling realy solved the problem , thank you. Have a very nice day.
Hello , I passed following variables to the agent followingly: CLEARML_AGENT_SKIP_PIP_VENV_INSTALL=1 CLEARML_AGENT_SKIP_PYTHON_ENV_INSTALL=1 clearml-agent daemon --queue "default" --foreground and it works perfectly ! Thanks for advise.
command i run:learml-agent daemon --queue default --foreground
response i get:clearml_agent: ERROR: create.<locals>.Validator.__init__() got an unexpected keyword argument 'types'
you can edit the requirements section directly <- where ? if i create requirements.txt it seems to be ignored
I went into app.clear.ml dashboard -> project -> archived experiments -> and i wanted to delete experiment from there -> that is the output i got
Still not solved, idk if these dependencies are cached somewhere but when i change requirements.txt or i add it manually into code it still have problems with the torch and is looking for 'torch==1.10.0+cu113'
Is the solution simply to change the binary_path in config and restart the worker or just use docker mode ? Cant i just forward this path ?
i can try , any guide online ? or is it totaly easy
Hi @<1523701070390366208:profile|CostlyOstrich36> , the worker cloned the repo correctly, however in the nested scipt if you use task.init it wont work / wont overwrite anything.
when i run it locally it was python script.py and for the remote you are right
Insight 2 <- repository is cloned correctly but the uncomitted changes are not.
I just set up my server from following url : https://clear.ml/docs/latest/docs/deploying_clearml/clearml_server_linux_mac/
Anyways no histogram / table is reported in dashboard however in tutorial video @ https://www.youtube.com/watch?v=quSGXvuK1IM it is.
RE: When people do not install the server on the same machine, how is it possible for them then ? I cant reach apiserver / clearml-apiserver.
After new installation of clearml-agent and clearml I still do have the same problem.
Example: I have a simple python script and defined default_docker in clearml.config. When i clone this experiment and run it from clearml dashboard my clearml-agent running in docker mode should execute this task in docker. However, it is stucked after dependencies in...
Do I have to do some port forwarding / add extra parameters ? Copy clearml.conf inside of the docker ? And all the stuff ? Because it does not seems to be done automatically.
So probably you are right - > nc -vz localhost 8080
Output when run locally not in docker: Connection to localhost (127.0.0.1) 8080 port [tcp/http-alt] succeeded!
Output when inside docker bash: localhost [127.0.0.1] 8080 (http-alt) : Connection refused
I can see the docker in docker ps but it seems like it never gets to code execution. I do not have an idea where it got from. Seems like somewhere it gets "pip" + "pip".
Yes api server is on the same machine -> running in container
web_server: http://localhost:8080
api_server: http://localhost:8008
files_server: http://localhost:8081
Original experiment has 1.10.0 pytorch and 113 cuda ['1.10.0+cu113']. Everything was run on the my local computer. In the virutal env i have these versions (however the system itself has little bit newer).
