It also could be something on your machine that is blocking this. Either way, I would start with them.
Hi @<1749965229388730368:profile|UnevenDeer21> , can you add the log of the job that failed?
Also, note that you can set these arguments from the webUI on the task level itself as well, Execution tab and then container section
AbruptWorm50 , please provide a log of the task 🙂
Hi @<1546303293918023680:profile|MiniatureRobin9> , can you please add logs of the tasks + controller?
Hi @<1523701295830011904:profile|CluelessFlamingo93> , I think this is what you're looking for:
None
This looks more appropriate if the username itself is "ubuntu"
SubstantialElk6 , do you mean the dataset task version?
SubstantialElk6 , can you view the dataset in the UI? Can you please provide a screenshot so I can mark it down for you
Are you seeing any errors in the webserver container?
SmugTurtle78 , I think so. Can you verify on your end?
Hi @<1523701842515595264:profile|PleasantOwl46> , I think you can add a PR here - None
Hi @<1546665634195050496:profile|SolidGoose91> , you can actually create multiple resources inside the same autoscaler application. I would suggest attaching each resource to different queues as well.
Hi @<1523701842515595264:profile|PleasantOwl46> , In the info section you can see the user name but not ID (However it is returned in the API request the webUI sends)
What is your use case?
Why not give an option to provide their user name and then convert it in the code?
Hi @<1523701842515595264:profile|PleasantOwl46> , I think that is what happening. If server is down, code continues running as if nothing happened and ClearML will simply cache all results and flush them once server is back up
CluelessElephant89 , I've added screenshots. Tell me if those help 🙂
I think you would have to re-register it
Hi @<1523701842515595264:profile|PleasantOwl46> , I think its a docker hub limitation for non paying accounts pulling too many images and unrelated to ClearML. You can always host your own artifactory as well.
Meaning that you should configure your host as follows host: "somehost.com:9000"
As I wrote, you need to remove the s3 from the start of the host section..
Hi @<1761199244808556544:profile|SarcasticHare65> , and if you run locally for the same amount of iterations this does not happen?
What do you mean by repo cleaml.conf ?
Just to make sure, clearml.conf
by default should sit in your home folder
Please try like Kirill mentioned. Also please note that there is no file target in the snippet you provided 🙂
Hi @<1731483438642368512:profile|LoosePigeon2> , you need to set the following:
sdk: {
development: {
store_code_diff_from_remote: false
store_uncommitted_code_diff: false
On the machine you're running your pipeline from
Hi @<1731483438642368512:profile|LoosePigeon2> , I think what you're looking for is this env variable - CLEARML_AGENT_FORCE_CODE_DIR
- None
You can simply point to the local repo and skip the cloning process.
@<1754676270102220800:profile|AlertReindeer55> , I think what @<1523701087100473344:profile|SuccessfulKoala55> means is that you can set the docker image on the experiment level itself as well. If you go into the "EXECUTION" tab of the experiment, in the container section you might see an image there
Hi @<1742355077231808512:profile|DisturbedLizard6> , you can achieve this using the following env var:
CLEARML_AGENT_FORCE_EXEC_SCRIPT
None
Hi @<1772795696529805312:profile|LethalCoral80> , I think what you're looking for is the offline mode - None
How do you normally mount the ssh keys?