Reputation
Badges 1
662 × Eureka!It does, but I don't want to guess the json structure (what if ClearML changes it or the folder structure it uses for offline execution?). If I do this, I'm planning a test that's reliant on ClearML implementation of offline mode, which is tangent to the unit test
Or well, because it's not geared for tests, I'm just encountering weird shit. Just calling task.close()
takes a long time
This is with:Task.set_offline_mode(True) task = Task.init(..., auto_connect_streams=False)
I'm working on the config object references ๐
Either one would be nice to have. I kinda like the instant search option, but could live with an ENTER to search.
I opened this meanwhile - https://github.com/allegroai/clearml-server/issues/138
Generally, it would also be good if the pop-up presented some hints about what went wrong with fetching the experiments. Here, I know the pattern is incomplete and invalid. A less advanced user might not understand what's up.
I'm trying to decide if ClearML is a good use case for my team ๐
Right now we're not looking for a complete overhaul into new tools, just some enhancements (specifically, model repository, data versioning).
We've been burnt by DVC and the likes before, so I'm trying to minimize the pain for my team before we set out to explore ClearML.
Okay trying again without detached
That's fine for the current use-case I believe.
Once the team is happy with the logging functionality, we'll move on to remote execution and things will update.
If I add the bucket to that (so CLEARML_FILES_HOST=
s3://minio_ip:9000/minio/bucket ), I then get the following error instead --
2021-12-21 22:14:55,518 - clearml.storage - ERROR - Failed uploading: SSL validation failed for
... [SSL: WRONG_VERSION_NUMBER] wrong version number (_ssl.c:1076)
I guess the thing that's missing from offline execution is being able to load an offline task without uploading it to the backend.
Or is that functionality provided by setting offline mode and then importing an offline task?
Nope, no other config files
Ah, uhhhh whatever is in the helm/glue charts. I think itโs the allegroai/clearml-agent-k8s-base
, but since I hadnโt gotten a chance to try it out, itโs hard to say with certainty which would be the best for us ๐
I'll give it a shot. Honestly, the SDK documentation for both InputModel and OutputModel is (sorry) horrible ...
Can't wait for the documentation revamping.
Is it CLEARML_CONFIG_FILE
? (I had to dig this from the GH code ๐
)
Or if it wasn't clear, that chunk of code is from clearml's dataset.py
That could be a solution for the regex search; my comment on the pop-up (in the previous reply) was a bit more generic - just that it should potentially include some information on what failed while fetching experiments ๐
Is there a preferred way to stop the agent?
That will come at a later stage
I've also followed https://clearml.slack.com/archives/CTK20V944/p1628333126247800 but it did not help
Anyway sounds good! ๐
I will TIAS, but maybe worthwhile to also mention if it has to be the absolute path or if relative path is fine too!
Honestly I wouldn't mind building the image myself, but the glue-k8s setup is missing some documentation so I'm not sure how to proceed
I think so, it was just missing from the official documentation ๐ Thanks!
Any updates @<1523701087100473344:profile|SuccessfulKoala55> ? ๐ซฃ
AFAICS it's quite trivial implementation at the moment, and would otherwise require parsing the text file to find some references, right?
https://github.com/allegroai/clearml/blob/18c7dc70cefdd4ad739be3799bb3d284883f28b2/clearml/task.py#L1592
Why not give ClearML read-only access credentials to the repository?
Okay so the only missing thing of the puzzle I think is that it would be nice if this propagates to the autoscaler as well; that then also allows hiding some of the credentials etc ๐ฎ
Right so this is checksum based? Are there plans to only store delta changes for files (i.e. store the changed byte instead of the entire file)?
Just because it's handy to compare differences and see how the data changed between iterations, but I guess we'll work with that ๐
We'll probably do something like:
When creating a new dataset with a parent (or parents), look at immediate parents for identically-named files If those exist, load those with matching framework (pyarrow, pandas, etc), and log differences to the new dataset ๐
I also tried switching to dockerized mode now, getting the same issue ๐ค