YummyWhale40 from the code snippet, it seems like the argument is passed.
"reuse_last_task_id=True" is the default, and it means that if the previous run of the task did not create any artifacts/models and was executed 72 hours ago (configurable), The Task will be reset (i.e. all logs cleared) and will be reused in the current run.
if you have any idea to reuse id even if models are outputted, please tell me thx
I would like to confirm just in case.
In the desired behavior, reuse_last_task_id=True
forces it for any intervals?
oh I got it. my codes output models and the task catch it automatically.
In my case, I write codes and run single batch train-val, which contains model saving, in developing phase. I want TRAINS to overwrite the dev runs for keeping dashboard clean.
YummyWhale40 you mean like continue training?
https://github.com/allegroai/trains/issues/160
YummyWhale40 no idea what the pytorch-lighting guys did there. let me check a the actual code.
maybe the arguments is simply passed to Task.init()
self._trains = Task.init( project_name=project_name, task_name=task_name, task_type=task_type, reuse_last_task_id=reuse_last_task_id, output_uri=output_uri, auto_connect_arg_parser=auto_connect_arg_parser, auto_connect_frameworks=auto_connect_frameworks, auto_resource_monitoring=auto_resource_monitoring )
I don't mean continuous training but I want to know about your plans for it 😋