Reputation
Badges 1
533 × Eureka!I want to get the instances of the tasks executed by this controller task
Continuing on this line of thought... Is it possible to call task.execute_remotely
on a CPU only machine (data scientists' laptop for example) and make the agent that fetches this task to run it using GPU? I'm asking that because it is mentioned that it replicates the running environment on the task creator... which is exactly what I'm not trying to do 😄
and in the UI configuration I didn't understand where does permission management came into play
Gotcha, didn't think of an external server as Service Containers are part of Github's offering, I'll consider that
moreover, in each pipeline I have 10 different settings of task A -> Task b (and then task C), each run 1-2 fails randomly
does the services mode have a separate configuration for base image?
ClearML results page:
`
Launching step: 2019-09-03_2021-01-25_choose_best
Parameters:
{***}
Configurations:
None
Overrides:
None
Launching step: 2019-10-23_2021-01-15_choose_best
Parameters:
{********}
Configurations:
None
Overrides:
None
Launching step: 2019-05-26_2020-12-26_choose_best
Parameters:
{******}
Configurations:
None
Overrides:
None
Launching step: 2019-07-15_2021-01-05_choose_best
Parameters:
{************}
Configurations:
None
Overrides:
None
Launching step...
In standard docker TimelyPenguin76 this quoting you mentioned is wrong, since the whole argument is being passed - hence the double tricky quotation I posted above
Not sure I understand, if i run pipe.start_locally(run_pipeline_steps_locally=True|False)
what is the difference betwee ntrue and false? assuming I want to execute locally
this is the df -h
output
TimelyPenguin76 , this can safely be set to s3://
right?
I mean, I barely have 20 experiments
doesn't contain the number 4
so in my code, I'll use this environment variable to read from disk
AgitatedDove14 clearml version on the Cleanup Service is 0.17.0
It's working! 😄
That's awesome, but my problem right now is that I have my own cronjob deleting the contents of /tmp
each interval, and it deletes the cfg files... So I understand I must skip deleting them from now on
So how do I solve the problem? Should I just relaunch the agents? Because they can't execute jobs now