
Reputation
Badges 1
40 × Eureka!TimelyPenguin76 I've been using this for a bit now, I would like to set it from code, just like I set docker image, for example. Can you point me in the right direction? I couldn't find anything in the docs
TimelyPenguin76 what version of clearml are you using? my task.set_base_docker only has a single positional command. am I using an old version?
This is a great feature for debugging setup. kinda feels like a superpower 🙂 I already used it to work around another issue with my docker setup. now I'll only need to update the docker file after I iron everything out and I will already have the startup shell script as documentation for what should be fixed. awesome.
I understand. Thing is I already have a bunch of tasks where I logged the tables and did not upload an artifact. If I can get them using the SDK, as something that I can possibly extract the values from as JSON (as in the web GUI) that would be great. Currently I'm just manually downloading the json one by one as I need them.
AgitatedDove14 I don't understand how that's related. I am working on localhost, why should I have a problem communicating with the jupyter server? It's local. Also, the kernel has no problem communicating with the allegro server, obviously (since results are logged).
but this gives me an idea, I will try to check if the notebook is considered as trusted, perhaps it isn't and that causes issues?
(and as you said, running as a module didn't change anything)
create a notebook, add the following lines to it's first cell:from clearml import Task task=Task.init(project_name='proj', task_name='notebook', task_type=Task.TaskTypes.custom, continue_last_task=True,)
run the cell
in the mean time, I installed a new clean virtualenv (no conda) and got the exact same behavior. I'll try running as a module and we'll see
I'm having some issues with my github access, I'll guess I'll update later/tomorrow when I solve them
TimelyPenguin76 ok, I'll try it out, thanks.
AgitatedDove14 same thing happens to me when I run via git bash
AgitatedDove14 Looks goo, thanks! I see that I can also find the plot index based on the 'metric' key, so I can write something that would choose the plot by name rather than ordinal position.
ok, that's a difference, I did not start with python -m, as a module. I'll try that
args_string = ['--{} {}'.format(k, v) for k, v in task.get_parameters_as_dict().get('Args', {}).items()] args_strings = [a.replace("'", '').replace("[", '').replace("]", '').replace(",", '') for a in reduce(lambda l, s: l + s.split(' '), args_string, []) if a]
and then just parser.parse_args(ars_strings)
It's not even very clean, I could have replaced the multiple replace calls with a regex etc. just a quick hack to work around it.
Actually, I found out that if I use exe...
same same. I ran inside clearml git repo and got the same warnings
I checked and it now seems to work. Thanks!
Yeah, I know how to do it manually from the web GUI using the button, that's just not scalable. What I need is the Python SDK code. It doesn't have to be a single liner.
AgitatedDove14 is there some way I can update the script file manually and retrigger the git discovery process?
It's working as expected. Thanks!
hi TimelyPenguin76 I tried doing this, but it didn't work. When enqueueing the task the contents of the textbox were emptied and the script was not run. I did make sure that it was saved before clicking on enqueue (by changing to another task and back and making sure the script appeared).
AgitatedDove14 - nailed it! it was due to my notebook being pwd protected rather than token protected. I shared that with Moshik. I wonder whether that could be worked around.
I can also send you a link to the task this created on our hosted allegro web server, to look at the logs, if that helps
I'm on windows, this is a python 3.6 conda venv, I think you can see the name of the env in the logs...
TimelyPenguin76 setting what appears in the GUI as "SETUP SHELL SCRIPT"
btw, I see the same thing when I start the notebook directly, i.e. "jupyter notebook path/to/notebook.ipynb" and when I start the notebook server "jupyter notebook" and then open the notebook from the jupyter web interface.