Reputation
Badges 1
195 × Eureka!I have a small question about the response structure, each of the metrics has this structure:metric_id: { ... "value": 0.0006447011, "min_value": 8.6326945e-06, "max_value": 0.001049518, ... } what does value refer to? the last reported?
TimelyPenguin76 , I generate it manually , clone some task -> adjust config -> enqueue
then when the agent pulls it I get the following behaviorremote_dict = task.connect(local_dict) # matching keys are overridden from remote configwhich I can't reproduce as I described above
there seem to be an additional logic to:str_value = str(value)str('.') -> '\.'
however in the configuration I see: '.' for non nested strings
worked like a charm
Hi AgitatedDove14 , the initialization of task happens once before the multiple trainings..
` Task.init
trainer.fit(model)
something
trainer.fit(model)
... `
it is missing the status that I'm looking for, namely is this worker is running a task or not
but I don't get the same when I try to reproduce it for debug as described above (with task.connect)
AgitatedDove14 no it has an offset of the value that it started with, so for example you stopped at n, then when you are running the n+1 epoch you get the 2*n+1 reported
let me try to explain myself again
OS:Fedora
browser: brave but also on chrome
reproduce like in the example I gave above.. drag the right corner across more then a single column
and which one if it does
nice, saved some electrons from running 🙂
something like in the example I shared<Machine 1> #Init Optimizer <Machine 2> **heavy one time Common Initialization** while True: #sample Optimizer # init task # Execute Something # report results <Machine i> **heavy one time Common Initialization** while True: #sample **same** Optimizer # init task # Execute Something # report results
not really.. did my previous comment made sense to you?
well when returning None it works as expected, no model uploads
Hi AgitatedDove14 , if you don't mind having a look too, I think its probably just a small misunderstanding
according to the above I was expecting the config to be auto-magically updated with the new yaml config I edited in the UI, however it seems like an additional step is required.. probably connect_dict? or am I missing something
the report media workaround I suggested was referring to having the preview directly in the debug_samples of the given task
AgitatedDove14 I am actually curious now, why is the default like this? maybe more people are facing similar bottlenecks?
something like in the snippet I shared above
very strange that you don't see the same in the community server
I can reproduce the same behavior in the community server.. but for some of the columns I get a slightly different behavior, it does something but I'm not sure exactly what
CumbersomeCormorant74 displaying a 3D Field varying in time, or anything varying in time really
Thanks it can be super cool, I hope it will happen
Hi AgitatedDove14 ,
clearml Server 1.1.1
|column1|column2|column3|
trying to extend column1 by dragging the right edge of column1 over column2 (column3 for example) will result with no extension
in 4. I am using the task.connect
Hi AgitatedDove14 , I am not uploading anything explicitly, and when I look at the UI Models tab I can only see the regular "{Project Name} - epoch={#}" and in addition "{Project Name} - {project_id}" so I am not sure what is really uploaded.. from the name of it it sounds like model weights and buffers (non-trainable)