Reputation
Badges 1
45 × Eureka!Thanks ! That is exactly what I meant 🙂
Have a nice day!
command i run:learml-agent daemon --queue default --foreground
response i get:clearml_agent: ERROR: create.<locals>.Validator.__init__() got an unexpected keyword argument 'types'
Do I have to do some port forwarding / add extra parameters ? Copy clearml.conf inside of the docker ? And all the stuff ? Because it does not seems to be done automatically.
I'm at the point where it looks like the clearml-agent is stuck (How i execute the agent: clearml-agent daemon --queue "default" --gpus 0 --foreground --docker. After the last message: " Successfully installed:<dependencies>" nothing really happens. I do attach logs from experiment. And I also do provide the config:
Hello , I passed following variables to the agent followingly: CLEARML_AGENT_SKIP_PIP_VENV_INSTALL=1 CLEARML_AGENT_SKIP_PYTHON_ENV_INSTALL=1 clearml-agent daemon --queue "default" --foreground
and it works perfectly ! Thanks for advise.
Is the solution simply to change the binary_path in config and restart the worker or just use docker mode ? Cant i just forward this path ?
I can see the docker in docker ps
but it seems like it never gets to code execution. I do not have an idea where it got from. Seems like somewhere it gets "pip" + "pip".
Hi ExasperatedCrab78 ,
am I getting it right that alias = dataset id which can be found in the clearml dashboard?
I also tried the changed url (github) you sent. I can successfully run all the scripts, but I do not see any results from log_dataset_statistics
. Maybe I am wrong but in the https://github.com/allegroai/clearml-blogs/blob/master/urbansounds8k/preprocessing.py this function is not even called. Could you help me with that please? I just want to replicate -> clearml dashboa...
For future you can put in requirements.txt :detectron @
isnt brew for mac ? i am on ubuntu
I went into app.clear.ml dashboard -> project -> archived experiments -> and i wanted to delete experiment from there -> that is the output i got
Server: 3.13.0-832. I just waited for some time and suddenly it was possible to delete it again. Idk what was that issue. No matter what , if it will appear again I can announce you.
i can try , any guide online ? or is it totaly easy
Original experiment has 1.10.0 pytorch and 113 cuda ['1.10.0+cu113']. Everything was run on the my local computer. In the virutal env i have these versions (however the system itself has little bit newer).
I understand what you mean. I am just describing different case. Lets assume i have my docker image already (all dependencies , data solved). Right now I run my task and it automatically looks for requirements.txt file in the repository. My question is -> is there any way to avoid this (simplest solution for now will be to -> rename requirements.txt to different filename)? I tried the things you sent already. The thing is that in requirements.txt in this repos can not be installed that easily...
Oh thanks ! 🙂 I understand now. Please let me know about the other problem.
Hello ! I would like to just skip whole clearML process (e.g. someone do not wants to log everything for this time). The problem I have is that if you do not provide task name and project name <- it should skip clearml process. However, when the task is cloned also project name and task name are whole Task is None before task init. I am asking if there is any way to know we are running a clone before task.init
after we call task = Task.current_task()
Probably i figured out everything. For me it is better to work with https and token (for git).
RE: When people do not install the server on the same machine, how is it possible for them then ? I cant reach apiserver / clearml-apiserver.
After new installation of clearml-agent and clearml I still do have the same problem.
Example: I have a simple python script and defined default_docker in clearml.config. When i clone this experiment and run it from clearml dashboard my clearml-agent running in docker mode should execute this task in docker. However, it is stucked after dependencies in...
when i run it locally it was python script.py and for the remote you are right
for the requirements how do you mean it please? To add requirements.txt into root directory ith the description of packages is enough ? or do you have to put somewhere you want to use this file? Thanks
Thanks for help ! I will try and let you know.
Anyways no histogram / table is reported in dashboard however in tutorial video @ https://www.youtube.com/watch?v=quSGXvuK1IM it is.
OPENBLAS="$(brew --prefix openblas)" pip install pandas