
Reputation
Badges 1
979 × Eureka!This is what I get, when I am connected and when I am logged out (by clearing cache/cookies)
yes -> but I still don't understand why the post_packages didn't work, could be worth investigating
Task.get_project_object().default_output_destination = None
then print(Task.get_project_object().default_output_destination)
is still the old value
Yes, not sure it is connected either actually - To make it work, I had to disable both venv caching and set use_system_packages to off, so that it reinstalls the full env. I remember that we discussed this problem already but I don't remember what was the outcome, I never was able to make it update the private dependencies based on the version. But this is most likely a problem from pip that is not clever enough to parse the tag as a semantic version and check whether the installed package ma...
Awesome! (Broken link in migration guide, step 3: https://allegro.ai/docs/deploying_trains/trains_server_es7_migration/ )
So I changed ebs_device_name = "/dev/sda1"
, and now I correctly get the 100gb EBS volume mounted on /
. All good 👍
I think this is because this API is not available in elastic 5.6
SuccessfulKoala55 Thanks to that I was able to identify the most expensive experiments. How can I count the number of documents for a specific series? Ie. I suspect that the loss, that is logged every iteration, is responsible for most of the documents logged, and I want to make sure of that
Stopping the server Editing the docker-compose.yml file, adding the logging section to all services Restarting the serverDocker-compose freed 10Go of logs
Ok, I could reproduce with Firefox and Chromium. Steps:
Add creds (either via the popup or in the settings) Go the /settings/webapp-configuration -> Creds should be there Hit F5 Creds are gone
I hitted enter too fast ^^
Installing them globally via$ pip install numpy opencv torch
will install locally with warning:Defaulting to user installation because normal site-packages is not writeable
, therefore the installation will take place in ~/.local/lib/python3.6/site-packages
, instead of the default one. Will this still be considered as global site-packages
and still be included in experiments envs? From what I tested it does
Hi CostlyOstrich36 , I mean insert temporary access keys
I just read, I do have the trains version 0.16 and the experiment is created with that version
ok, thanks SuccessfulKoala55 !
Or even better: would it be possible to have a support for HTML files as artifacts?
Some context: I am trying to log an HTML file and I would like it to be easily accessible for preview
AgitatedDove14 WOW, thanks a lot! I will dig into that 🚀
Sorry, what I meant is that it is not documented anywhere that the agent should run in docker mode, hence my confusion
If I remove security_group_ids
and just let subnet_id
in the configuration, it is not taken into account (the instances are created in a the default subnet)
Hi SuccessfulKoala55 , there it is > https://github.com/allegroai/clearml-server/issues/100
Also maybe we are not on the same page - by clean up, I mean kill a detached subprocess on the machine executing the agent
Yes, it works now! Yay!
I think we should switch back, and have a configuration to control which mechanism the agent uses , wdyt? (edited)
That sounds great!
The jump in the loss when resuming at iteration 31 is probably another issue -> for now I can conclude that:
I need to set sdk.development.report_use_subprocess = false
I need to call task.set_initial_iteration(0)
So it looks like the agent, from time to time thinks it is not running an experiment