Reputation
Badges 1
64 × Eureka!can you test what happens if you pass the credentials in the global scope as well, i.e. here:
That didn’t help
AgitatedDove14 its running inside a docker based worker.
Are you interested in the full pip freeze of that docker?
can you provide some mode details please ? Do you intend to store your artefacts locally or remotely ?
Does the manual reporting also fails ?
If you could also give your clearml packages versions it could help
I store the artifacts on a minio server (in my LAN).
If I run the python script locally (i.e. no execute remotely() it works fine).
I use the latest clearml 1.6.2
Did you by any chance save the checkpoint without any file extention? Or with a weird name containing sl...
Because we want all our data to be stored on premises.
Thanks ExasperatedCrab78
AgitatedDove14 - attached
Thanks @<1523701087100473344:profile|SuccessfulKoala55> @<1523701070390366208:profile|CostlyOstrich36> .
Fixed with the RC.
This is actually in runtime (i.e. when running the code),
My script looks like that
` import clearml
clearml.Task.force_requirements_env_freeze(force=False, requirements_file="requirements.txt")
task = clearml.Task.init(...)
task.set_script(working_dir=dir, entry_point="my_script.py")
task.execute_remotely(queue_name='default')
rest of script goes here.... `When you refer to runtime, do you mean when the script is executed remotely, or when I run my_script.py locally (in order to ...
Thanks AgitatedDove14 !
I’ll use clearml 1.4.1 until the fix is out.
Oh!
That was so silly on my side...
And I'm using the latest clearml / clearml-agent
@<1523701070390366208:profile|CostlyOstrich36>
AgitatedDove14 , did you test it using a worker, or with local execution?
I just tested https://github.com/allegroai/clearml/blob/master/examples/frameworks/pytorch/pytorch_mnist.py with a (docker based) worker and it yields the same error
` 2022-07-17 07:59:40,330 - clearml.Task - INFO - Waiting to finish uploads
2022-07-17 07:59:40,330 - clearml.storage - INFO - Starting upload: /tmp/.clearml.upload_model_0_4d_ikk.tmp => tapsff.local:9000/clearml/examples/PyTorch MNIST train.02ed1df11bf54...
After signing with google, the login page is stuck at this
Im afraid there's something with my account, since it won't login on other browser as well, but logging with a different email works
clearml
numpy
pandas
pycocotools
tqdm
boto3
torch
torchvision
bbox_utils
xmltodict
scikit-video
onnxruntime
flask
AgitatedDove14 .
Note that the actual error is /workspace/miniconda/bin/python3: No module named clearml_agent since all the packages (including clearml_agent) were already installed by the agent on the default (non conda) python binary.
Anyway, when I add the binary's path to PATH, it still won't work.
I calltask.set_base_docker(docker_image='my/docker/image', docker_setup_bash_script= ["export PATH=""/workspace/miniconda/bin:$PATH"])just after Task.init.
On execution, the agent installs all pip requirements with python binary at /usr/bin/python3.6
but eventually the task fails with, /workspace/miniconda/bin/python3: No module named clearml_agent
Oh wow AgitatedDove14 . Appreciate it!
Are you sure it’s just a matter of the python version?
The same experiment script, was working on the exact docker image in the past (with older clearml versions though…).
For example this experiment log:
That would be a very useful feature.
What is the status of that issue? I havn't found it on github.
AgitatedDove14
Yes, I'd like to point to a specific binary, which is in a conda environment.
(b.t.w how can I specify the python version on the Task?)
SweetBadger76
It’s not a credential issue, because I do upload artifacts manually with tsk.upload_artifact(...)
I’ll try changing the extension, but I have to admit that in the past (I havn’t used clearml for a while and updated it recently to the latest verion) it did got this file extension right
The script is intended to be executed remotely.
Can I declare an absolute path in this case?
SuccessfulKoala55
Were using the community server
Not sure I understand the purpose of this.
it meant pip will look for wheels at thus url?
AgitatedDove14
I'm not sure.
In my case I'm not trying to reproduce a local environment in the agent, but to run a script inside a docker which already has the environment built in.
The environment is conda based.
AgitatedDove14 ,
From the experiment’s console log:
` - boto3==1.16.2
- botocore==1.19.2 `
It didn't work either. Still same error.
conda command was already in PATH and the conda activate is executed properly, but it prompts to run conda init (i.e conda wasn't configures at that shell):
` You should consider upgrading via the '/usr/bin/python3 -m pip install --upgrade pip' command.
CommandNotFoundError: Your shell has not been properly configured to use 'conda activate'.
To initialize your shell, run
$ conda init <SHELL_NAME>
Currently supported shells are:
- ba...