yes. As you can see this one has the hydra
section reported in the config
GloriousPanda26 Are you getting multiple Tasks or is it a single Task ?
ClearML does
Thanks for doing that ! :i_love_you_hand_sign:
the import order does is not related to the problem
GloriousPanda26 wouldn't it make more sense that multi run would create multiple experiments ?
on one experiment it overlays the same metrics (not taking into account the run number)
but I have no idea what's behing 1
, 2
and 3
compare to the first execution
Right, I think the naming is a by-product of Hydra / TB
it's a single taks which contains metrics for all 4 executions
but despite the naming it's working quite well actually
So the naming is a by product of the many TB created (one per experiment), if you add different naming ot the TB files, then this is what you'll be seeing in the UI. Make sense ?
I think it would make sense to have one task per run to make the comparison on hyper-parameters easier
I agree. Could you maybe open a GitHub issue on it, I want to make sure we solve this issue 🙂
I am not really familiar with TB internal mechanics. For this project we are using Pytorch Lightning
Below is an example with one metric reported using multirun. This is taken from a single experiment result page as all runs feed the same experiment. Unfortunately I have no idea what 1
refers to for example. Is it possible to name each run or to break them into several experiments ?
but I have no idea what's behingÂ
1
,Â
2
 andÂ
3
 compare to the first execution
This is why I would think multiple experiments, since it will store all the arguments (and I think these arguments are somehow being lost.
wdyt?
between Hydra, PL, TB and clearml I am not quite sure who is adding the prefix for each run
It's a running number because PL is creating the same TB file for every run
but when I compare experiments the run numbers are taken into account comparing "1:loss" with "1:loss" and putting "2:loss"s in a different graph
but to go back to your question, I think it would make sense to have one task per run to make the comparison on hyper-parameters easier
the previous image was from the dashboard of one experiment