Hi BoredPigeon26
what do you mean by "reuse the task" ? is this manual execution (i.e. from code)?
How about archiving the old version?
You can also force Task.init to always create a new Task (which preserves the previous run alongside the execution tab)
Basically what's the specific use case ?
I'm saying that because in the task under "INSTALLED PACKAGES" this is what appears
This is exactly what I was looking for. Thanks!
Yes that makes sense, I think this bug was fixed a long time ago, and this is why I could not reproduce it.
I also think you can use a later version of clearml ๐
Hi SmallDeer34
On the SaaS you can right click on an experimenter and publish it ๐
This will make the link available for everyone, would that help?
Hi AstonishingSwan80 , what do you mean by "ec2 API"?
EnviousStarfish54 whats your matplotlib version ?
Basically just change the helm yamlqueue: my_second_queue_name_here
I think, this all ties into the none-standard git repo definition. I cannot find any other reason for it. Is it actually stuck for 5 min at the end of the process, waiting for the repo detection ?
Is there any way to debug these sessions through clearml? Thanks!
Yes this is a real problem, AWS does not allow to get the data very easily...
Can you check the AWS console, see what you have there ?
In theory this should have worked.
Maybe we you are missing some escaping for the "extra_vm_bash_script" ?
I'm hoping the console output will tell us
Hiย SmoggyGoat53
There is a storage limit on the file server (basically 2GB per file limit), thisย is the cause of the error.
You can upload the 10GB to any S3 alike solution (or a shared folder). Just set the "output_uri" on the Task (either at Task.init or with Task.output_uri = " s3://bucket ")
it would be clearml-serverโs job to distribute to each user internally?
So you mean the user will never know their own S3 access credentials?
Are those credentials unique per user or once"hidden" for all of them?
JitteryCoyote63 with pleasure ๐
BTW: the Ignite TrainsLogger will be fixed soon (I think it's on a branch already by SuccessfulKoala55 ) to fix the bug ElegantKangaroo44 found. should be RC next week
The file itslef is csv.gz compressed, it's actually sending from the file-server back that messes things
(you can test with output_uri=/tmp/folder
)
Hi VexedCat68
can you supply more details on the issue ? (probably the best is to open a github issue, and have all the details there, so we have better visibility)
wdyt?
Okay verified, it won't work with the demo server. give me a minute ๐
Actually scikit implies joblib ๐ (so you should use scikit, anyhow I'll make sure we add joblib as it is more explicit)
Hi SuperiorCockroach75
You mean like turning on caching ? What do you mean by taking too long?
should I update nodejs in centos image ?
I think so, it might have been forgotten
Hi @<1547028074090991616:profile|ShaggySwan64>
I'm guessing just copying the data folder with rsync is not the most robust way to do that since there can be writes into mongodb etc.
Yep
Does anyone have experience with something like that?
basically you should just backup the 3 DBs (mongo, redis, elastic) each one based on their own backup workflows. Then just rsync the files server & configuration.
command line ๐
cmd.exe / bash
MelancholyElk85 I'm assuming you have the agent setup and everything in the example code works, is that correct ?
Where is it failing on your pipelines ?
but never executes/enqueues them (they are all inย
Draft
ย mode).
All pipeline steps are not enqueued ?
Is the pipeline controller itself running?
how did you install trains?pip install git+
I can share some code
Please do ๐
MelancholyElk85 assuming we are running with clearml 1.1.1 , let's debug the pipeline and instead of pipeline start/wait/stop :
Let's do:pipeline.start_locally(run_pipeline_steps_locally=False)
Would it also be possible to query based on
multiple
user properties
multiple key/value I think are currently not that easy to query,
but multiple tags are quite easy to do
tags=["__$all", "tag1", "tag2],
Hi @<1544853721739956224:profile|QuizzicalFox36>
Sure just change the ports on the docker compose
What's the difference between the example pipeeline and this code ?
Could it be the "parents" argument ? what is it?
๐
Okay But we should definitely output an error on that