Reputation
Badges 1
383 × Eureka!Will try it out. A weird one this.
AgitatedDove14 - this was an interesting one. I think I have found the issue, but verifying the fix as of now.
One of the devs was using shutil.copy2
to copy parts of dataset to a temporary directory in a with
block - something like:
with TemporaryDirectory(dir=temp_dir) as certificates_directory: for file in test_paths: shutil.copy2(f"{dataset_local}/{file}", f"{certificates_directory}/file")
My suspicion is since copy2 copies with full data and symlin...
AgitatedDove14 - worked with mutable copy! So was definitely related to the symlinks in some form
Ah ok. Kind of getting it, will have to try the glue mode
I would like to create a notebook instance and start using it without having to do anything on a dev box
AlertBlackbird30 :
--remote-gateway [REMOTE_GATEWAY] Advanced: Specify gateway ip/address to be passed to interactive session (for use with k8s ingestion / ELB)
I see this in clearml-session - what’s the intent here?
But ok the summary is I guess it doesn’t work in a k8s env
IF there’s a post-task script, I can add a way to zip and upload pip cache etc to s3 - as in do any caching that I want without having first class support in clearml
Initially thought use_current_task in Dataset.create
does it, but that seems to set the DataprocessingTask itself to the current task
I would prefer controlled behavior than some available version being used. Here triggered a bunch of jobs that all went fine and even evaluations were fine and then when we triggered a inference deploy it failed
Ah thanks for the pointer AgitatedDove14
Sounds great!
Is nested projects in the release? I see it in the community server but no mention in the blog or the release notes?
Yes, for datasets where we need GDPR compliance
I only see published getting preference, not a way to filter only to published
Sorry if it was confusing. Was asking if people have setup pipelines automatically triggered on update to datasets
As in if there are jobs, first level is new pods, second level is new nodes in the cluster.
AgitatedDove14 - the actual replication failed. When we run a task by cloning and enqueueing, there is a current task even if I am yet to do a Task.init right?
Can you point me at the relevant code?
Looks like Task.current_task() is indeed None in this case. Bit of log below where I print(Task.current_task()) as first step in the script
Environment setup completed successfully Starting Task Execution: None