Reputation
Badges 1
533 × Eureka!cool, didn't know about the PAT
you can use pgrep -af "trains-agent"
Okay SuccessfulKoala55 , problem solved! Indeed the problem was that there is not .git folder. I updated necessary things to make the checkout action get the actual repo and now it works
It wasn't really clear to me what "standalone" means, maybe it will be better to add to the error
Error: Standalone
(no .git folder found)
script detected 'tasks/hp_optimization.py', but no requirements provided
checking and will let you know
It's kind of random, it works sometimes and sometimes it doesn't
If you want we can do live zoom or something so you can see what happens
I know I can configure the file server on trains-init - but that only touches the client side, what about the container on the trains server?
Yes, I'll prepare something and send
Okay so regarding the version - we are using 1.1.1
The thing with this error it that it happens sometimes, and when it happens it never goes away...
I don't know what causes it, but we have one host where it works okay, then someone else checks out the repo and tried and it fails for this error, while another guy can do the same and it will work for him
I'll check the version tomorrow, about the current_task call, I tried before and after - same result
The weirdest thing, is that the execution is "completed" but it actually failed
I'm asking that because the DSes we have are working on multiple projects, and they have only one trains.conf file, I wouldn't want them to edit it each time they switch project
UptightCoyote42 - How are these images avaialble to all agents? Do you host them on Docker hub?
Okay, so the agent automatically inherits the launching environment's variables?
I set it to true, I have more packages installed now, but it still fails.. here is the log TimelyPenguin76
` Successfully installed clearml-1.0.5 cloudpickle-1.6.0 cycler-0.10.0 hyperopt-0.2.5 kiwisolver-1.3.2 matplotlib-3.4.3 networkx-2.6.2 pandas-1.3.2 patsy-0.5.1 plotly-5.3.0 python-dateutil-2.8.2 statsmodels-0.12.2 tenacity-8.0.1 tqdm-4.62.2
Adding venv into cache: /home/elior/.clearml/venvs-builds/3.8
Running task id [24a54a473c234b00a126ec805d74046f]:
[.]$ /home/elior/.clearml/venvs...
That's awesome, but my problem right now is that I have my own cronjob deleting the contents of /tmp each interval, and it deletes the cfg files... So I understand I must skip deleting them from now on
So how do I solve the problem? Should I just relaunch the agents? Because they can't execute jobs now
after you create the pipeline object itself , can you get Task.current_task() ?
AgitatedDove14 no I can't... Just checked this. This is a huge problem for us, it used to work before and it just stopped working and I can't figure out why.
It's a problem for us because we made it a methodology of running some tasks under a pipeline task and saving summary iunfo to the pipeline task - but now since Task.current_task() doesn't work on the pipeline object we have a serious problem
Cool - what kind of objects are returned by .artifacts. getitem ? I want to check their docs
Continuing on this discussion... What is the relationship between configuring files_server and all the rest we just talked about and the the default_output_uri ?
that will require restarting the agent again?