Reputation
Badges 1
981 × Eureka!Alright SuccessfulKoala55 I was able to make it work by downgrading clearml-agent to 0.17.2
AgitatedDove14 yes but I don't see in the docs how to attach it to the logger of the earlystopping handler
AgitatedDove14 Is it possible to shut down the server while an experiment is running? I would like to resize the volume and then restart it (should take ~10 mins)
I guess I’ll get used to it 😄
Nice, thanks!
The file /tmp/.clearml_agent_out.j7wo7ltp.txt  does not exist
might be worth documenting 😄
Will it freeze/crash/break/stop the ongoing experiments?
AgitatedDove14 I am actually considering rolling back to 1.1.0, so 1.3.0 is not really an option for now
yes what happens in the case of the installation with pip wheels files?
When installed with http://get.docker.com , it works
as it's also based on pytorch-ignite!
I am not sure to understand, what is the link with pytorch-ignite?
We're in the brainstorming phase of what are the best approaches to integrate, we might pick your brain later on
Awesome, I'd be happy to help!
AnxiousSeal95 Any update on this topic? I am very excited to see where this can go 🤩
Yes, not sure it is connected either actually - To make it work, I had to disable both venv caching and set use_system_packages to off, so that it reinstalls the full env. I remember that we discussed this problem already but I don't remember what was the outcome, I never was able to make it update the private dependencies based on the version. But this is most likely a problem from pip that is not clever enough to parse the tag as a semantic version and check whether the installed package ma...
torch==1.7.1 git+ .
Is it because I did not specify --gpu 0 that the agent, by default pulls one experiment per available GPU?
Ho nice, thanks for pointing this out!
Thanks AgitatedDove14 ! I created a project with a default output destination to a s3 bucket but I don't have local access to this bucket (only agents have access to it for security reasons). Because of that, I cannot create a task in this project programmatically locally because it tries to access the bucket and fails. And there is no easy way to change the default output location (not in the web UI, not in the sdk)
/data/shared/miniconda3/bin/python /data/shared/miniconda3/bin/clearml-agent daemon --services-mode --detached --queue services --create-queue --docker ubuntu:18.04 --cpu-only
I would probably leave it to the ClearML team to answer you, I am not using the UI app and for me it worked just well with different regions. Maybe check permissions of the key/secrets?
but if you do that and the package is already installed it will not install using the git repo, this is an issue with pip
Exactly, that’s my problem: I want to remove it to make sure it is reinstalled (because the version can change)
I think that since the agent installs everything from scratch it should work for you. Wdyt?
With env caching enabled, it won’t reinstall this private dependency, right?
I have two controller tasks running in parallel in the trains-agent services queue
If I don’t start clearml-session , I can easily connect to the agent, so clearml-session is doing something that messes up the ssh config and prevent me from ssh into the agent afterwards
Should I try to disable dynamic mapping before doing the reindex operation?
Adding back clearml logging with matplotlib.use('agg') , uses more ram but not that suspicious
So I created a symlink in /opt/train/data -> /data
Or even better: would it be possible to have a support for HTML files as artifacts?