Reputation
Badges 1
82 × Eureka!in my case using self-hosted and agent inside a docker container:
47:45 : taks foo pulled
[ git clone, pip install, check that all requirements satisfied, and nothing is downloaded]
48:16 : start training
Are you talking about this: None
It seems to not doing anything aboout the database data ...
In summary:
Spin down the local server
Backup the data folder
In the cloud, extract the data backup
Spin up the cloud server
wow , did not know that vscode have a http "interface" !!! Make kind of sense as vscode is just a Chrome rendering webpage behind the scene ?
@<1523701868901961728:profile|ReassuredTiger98> I found that you an set the file_server
in your local clearml.conf
to your own cloud storage. In our case, we use something like this in our clearml.conf:
api {
file_server: "azure://<account>..../container"
}
All non artifact model are then store in our azure storage. In our self-hosted clearml setup, we don't even have a file server running alltogether
all good. Just wanted to know in case I missed it
About the caching: how does it work ? ClearML maintain it own cache and monitor if any of you code changes? Even code that get change inside an import ?
Should i open a feature request?
In the web UI, in the queue/worker tab, you should see a service queue and a worker available in that queue. Otherwise the service agent is not running. Refer to John c above
even it's just a local image ? You need a docker repository even if it will only be local PC ?
(I never played with pipeline feature so I am not really sure that it works as I imagined ...)
once you install manually your package inside the docker container, check that your file module_b/templates/my_template.yml
is where it should be
What about migrating existing expriment in the on prem server?
what is the difference between vscode via clearml-session and vscode via remote ssh extension ?
you should be able to use as many agent as you want.
On the same or different queue
Oh, I was assuming you are passing the entire DB backups to the cloud.
Yes, that is what I want to do.
So I need to migrate both the MongoDB database and elastic search database from my local docker instance to the equivalent in the cloud ?
1.12.2 because some bug that make fastai lag 2x
1.8.1rc2 because it fix an annoying git clone bug
Clear. Thanks @<1523701070390366208:profile|CostlyOstrich36> !
with ssh public key, if from a terminal, I can do git clone, then so do the clearml agent, as it run on behalf of an local user. That apply to both local and VM
nevermind, all the database files are in data folder
can you make train1.py
use clearml.conf.server1
and train2.py
use clearml.conf2
?? In which case I would be intersted @<1523701087100473344:profile|SuccessfulKoala55>
nice !! That is exactly what I am looking for !!
I understand for cleaml-agent
What I mean is that I have 2 self deployed server. I want to switch between the 2 config when running the code locally, not inside the agent