Reputation
Badges 1
42 × Eureka!Ahh. This is a shame. I really want to use ClearML to efficiently compute features but it’s proving a challenge!
Thanks
The Pipeline is defined using PipelineDecorators, so currently to “build and run” it would just involve running the script it is defined in (which enqueues it and runs it etc).
This is not ideal, as I need to access the Task ID and the only methods I can see are for use within the Task/Pipeline ( Task.current_task
and PipelineDecorator.get_current_pipeline
)
The reason I want to check completion etc outside the Pipeline Task is that I want to run data validation etc once when the pipe...
The return objects were stored to S3 but PipelineDecorator.upload_artifact
still uploaded to the file server. Not sure what was up with that but as explained in my next comment it did work when I tried again.
It also seems that PipelineDecorator.upload_artifact
is not compatible with caching, sadly, but that is another issue for another thread that I will be starting on Monday.
Have a good weekend
Hi John, we are using a self-hosted server with:
WebApp 1.9.2-317
Server: 1.9.2-317
API: 2.23
edit: clearml==1.11.0
I used task.flush(wait_for_uploads=True)
in the final cell of the notebook
Yes, sorry, the final cell has the flush
followed by the close
I have added a lot of detail to this, sorry.
The inline comments in the code talk about that specific script/implementation.
I have added a lot of context in the doc string at the top.
Ahh that’s great, thank you.
And then I could use storage manager or whatever to get the files. Perfect
To illustrate, here’s an example repo:
repo/
package1/
package2/ # installed separately to package1
task_script.py # requires package1 and package2 to have been pip installed
my colleague, @<1534706830800850944:profile|ZealousCoyote89> has been looking at this – I think he has used the relevant kwarg in the component decorator to specify the packages, and I think it worked but I’m not 100%. Connah?
The issue here is I don’t have the pipeline ID as it is a new version of the pipeline - i.e. the code has been updated, I want to run the updated pipeline (for the first time), get its ID, and then analyse the run/perform updates to tags (for example)