Reputation
Badges 1
533 × Eureka!This is the pip freeze of the environment I don't know why it differs from what the agent has... the agent only has a subset of these google libs
Martin: In your trains.conf, change the valuefiles_server: ' s3://ip :port/bucket'
Isn't this a client configuration ( trains-init )? Shouldn't be any change to the server configuration ( /opt/trains/config... )?
alabaster==0.7.12 appdirs==1.4.4 apturl==0.5.2 attrs==21.2.0 Babel==2.9.1 bcrypt==3.1.7 blinker==1.4 Brlapi==0.7.0 cachetools==4.0.0 certifi==2019.11.28 chardet==3.0.4 chrome-gnome-shell==0.0.0 clearml==1.0.5 click==8.0.1 cloud-sptheme==1.10.1.post20200504175005 cloudpickle==1.6.0 colorama==0.4.3 command-not-found==0.3
I don't have ifconfig
and also in the extra_vm_bash_script variables, I ahve them under export TRAINS_API_ACCESS_KEY and export TRAINS_API_SECRET_KEY
I assume trains passes it as is, so I think the quoting I mentioned might work
what if i want it to use ssh creds?
Even assuming it suspects me, why doesn't the captcha prove my innocence? Isn't it what it is for O_O
cool, didn't know about the PAT
you can use pgrep -af "trains-agent"
Okay SuccessfulKoala55 , problem solved! Indeed the problem was that there is not .git folder. I updated necessary things to make the checkout action get the actual repo and now it works
It wasn't really clear to me what "standalone" means, maybe it will be better to add to the error
Error: Standalone
(no .git folder found)
script detected 'tasks/hp_optimization.py', but no requirements provided
checking and will let you know
It's kind of random, it works sometimes and sometimes it doesn't
If you want we can do live zoom or something so you can see what happens
I know I can configure the file server on trains-init - but that only touches the client side, what about the container on the trains server?
Yes, I'll prepare something and send
Okay so regarding the version - we are using 1.1.1
The thing with this error it that it happens sometimes, and when it happens it never goes away...
I don't know what causes it, but we have one host where it works okay, then someone else checks out the repo and tried and it fails for this error, while another guy can do the same and it will work for him
I'll check the version tomorrow, about the current_task call, I tried before and after - same result
The weirdest thing, is that the execution is "completed" but it actually failed