For example, I want to compare accuracy (the metrics I'm interested) among different experiments. This metrics isn't automatically recorded by ClearML so I want to manually record it.
I've found a workaround to achieve it (as mentioned in the original message), but I'm still wondering if there is any suggestion except using
Hi ScaryBluewhale66 , I believe the new server that's about the be released soon (this \ next week), we'll allow you to report a "single value metric". so if you want to report just a number per experiment you can, then you can also compare between runs.
report_scalar pernits to manually report a scalar series. This is the dedicated function. There could be other ways to report a scalar, for example through tensorboard - in this case you would have to report to tensorboard, and clearML will automatically report the values