Reputation
Badges 1
45 × Eureka!Hi ExasperatedCrab78 ,
am I getting it right that alias = dataset id which can be found in the clearml dashboard?
I also tried the changed url (github) you sent. I can successfully run all the scripts, but I do not see any results from log_dataset_statistics
. Maybe I am wrong but in the https://github.com/allegroai/clearml-blogs/blob/master/urbansounds8k/preprocessing.py this function is not even called. Could you help me with that please? I just want to replicate -> clearml dashboa...
Thanks for help ! I will try and let you know.
Do I have to do some port forwarding / add extra parameters ? Copy clearml.conf inside of the docker ? And all the stuff ? Because it does not seems to be done automatically.
command i run:learml-agent daemon --queue default --foreground
response i get:clearml_agent: ERROR: create.<locals>.Validator.__init__() got an unexpected keyword argument 'types'
OPENBLAS="$(brew --prefix openblas)" pip install pandas
isnt brew for mac ? i am on ubuntu
The agent simply try to install requirements from requirements.txt , however i dont want to do that because i have my docker image ready.
Server: 3.13.0-832. I just waited for some time and suddenly it was possible to delete it again. Idk what was that issue. No matter what , if it will appear again I can announce you.
Hello ! I would like to just skip whole clearML process (e.g. someone do not wants to log everything for this time). The problem I have is that if you do not provide task name and project name <- it should skip clearml process. However, when the task is cloned also project name and task name are whole Task is None before task init. I am asking if there is any way to know we are running a clone before task.init
after we call task = Task.current_task()
To describe the use-case. Lets say we have someapp which can export specific training script. I would like to create this as a specific "draft" task and later execute it.
Probably i figured out everything. For me it is better to work with https and token (for git).
Insight 2 <- repository is cloned correctly but the uncomitted changes are not.
reinstalling realy solved the problem , thank you. Have a very nice day.
i can try , any guide online ? or is it totaly easy
I understand what you mean. I am just describing different case. Lets assume i have my docker image already (all dependencies , data solved). Right now I run my task and it automatically looks for requirements.txt file in the repository. My question is -> is there any way to avoid this (simplest solution for now will be to -> rename requirements.txt to different filename)? I tried the things you sent already. The thing is that in requirements.txt in this repos can not be installed that easily...
Original experiment has 1.10.0 pytorch and 113 cuda ['1.10.0+cu113']. Everything was run on the my local computer. In the virutal env i have these versions (however the system itself has little bit newer).
Still not solved, idk if these dependencies are cached somewhere but when i change requirements.txt or i add it manually into code it still have problems with the torch and is looking for 'torch==1.10.0+cu113'
you can edit the requirements section directly <- where ? if i create requirements.txt it seems to be ignored
when i run it locally it was python script.py and for the remote you are right
but for the local execution virtual env (conda) named detectron2 was used
for the requirements how do you mean it please? To add requirements.txt into root directory ith the description of packages is enough ? or do you have to put somewhere you want to use this file? Thanks
For future you can put in requirements.txt :detectron @
Thanks ! That is exactly what I meant 🙂
Have a nice day!
So probably you are right - > nc -vz localhost 8080
Output when run locally not in docker: Connection to localhost (127.0.0.1) 8080 port [tcp/http-alt] succeeded!
Output when inside docker bash: localhost [127.0.0.1] 8080 (http-alt) : Connection refused
Yes api server is on the same machine -> running in container
web_server: http://localhost:8080
api_server: http://localhost:8008
files_server: http://localhost:8081
I can see the docker in docker ps
but it seems like it never gets to code execution. I do not have an idea where it got from. Seems like somewhere it gets "pip" + "pip".