Reputation
Badges 1
46 × Eureka!@<1523701070390366208:profile|CostlyOstrich36> , as written above, I've done that. It still tries to send to 8081
Also @<1523701070390366208:profile|CostlyOstrich36> - are these actions available for on prem OSS clearml-server deployments too?
nice! I was wondering whether we can trigger it by the UI, like "on publishing" an experiment
is the agent execution dependent on some CMD in my docker file?
With respect to unstructured data, do hyperdatasets work well with audio data (and associated metadata) ?
@<1537605940121964544:profile|EnthusiasticShrimp49> , now that I have run the task on remote, can I copy the artefacts/files it creates back to my local fs?
Lets say the artefacts are something likeartefacts = [checkpoint.pth, dvc.lock, some_other_dynamically_generated_file]
I do change the task and the project name, the task name change works fine but the project name change silently fails
So I am deploying clearml-server on an on-prem server, and the checkpoints etc. are quite large for the experiments I will do.
Instead I want to periodically upload / back up this data to s3, and free up local disk space. Is that something that is supported?
I see that in my docker-compose installation, most of the big files are in /opt/clearml/data
That makes sense, but that would mean that each client/user has to manage the upload themselves, right?
(I'm trying to use clearml to create an abstraction over the compute / cloud)
I'm thinking of using s3fs on the entire /opt/clearml/data folder. What do you think?
Thanks, I can have docker
+ poetry
execution modes then?
I set it up like this: clearml-agent daemon --detached --gpus 0,1,2 --queue single-gpu-24 --docker
but when I create the session : clearml-session --docker xyz --git-credentials
and I run nvidia-smi
I only see one gpu
We have some scenario where a group of clearml experiments might represent a logical experiment. We then want to use all the trained models in a pipeline to generate some output.
With that output, we probably want to some third party like mechanical turk, do some custom evaluations - and some times more than once. We then want to connect (and present) these evaluations alongwith ClearML experiments.
we have various services internally to do this --> however, we have to manually link it up w...
because, otherwise it becomes a bit of a chicken and egg problem
- update code
- git push
- docker build and push on CI
- use new docker sha for task execution
- update code
- git push
- repeat?
I've also overriden CLEARML_FILES_HOST= None , and configured it in clearml.conf file. Don't know where its picking 8081 😕
where is it persisted? if I have multiple sessions I want to persist, is that possible?
How does it work with k8s? how can I request the two pods to sit on the same gpu?
is it in the OSS version too?