hmmm… probably not if I don’t have a reference that clearml can update right?….
What about:
hpo_params = OmegaConf.to_object(cfg)
...
task.connect(hpo_params)
And then I use hpo_params
in the code. This way I give clearml a chance to update the object.
Would this work? Thanks
Hi @<1523701205467926528:profile|AgitatedDove14> , I see _allow_omegaconf_edit_
under HYPERPARAMETERS > Hydra
Will this work?
task.connect(OmegaConf.to_object(cfg))
assuming cfg
is my Hydra dict
Hey @<1523701205467926528:profile|AgitatedDove14> in the WebUI the hydra configuration object is under CONFIGURATION OBJECTS > OmegaConf
So should this be OmegaConf/trainer.batch_size
?
so if I want to refer to batch_size
in my_hydra_config.yaml
:
# dummy config file
trainer:
params:
batch_size: 32
do I pass this to the HyperParameterOptimizer
as:
Hydra/trainer/params/batch_size
??
@<1523701205467926528:profile|AgitatedDove14> 👆 ? Thanks
hmm….. probably simpler/cleaner if I do
hpo_params = {
'param1':cfg.param_1, ...
}
task.connect(hpo_params)
Thoughts?
Thanks @<1523701205467926528:profile|AgitatedDove14> happy to PR on the docs 😉
so it’s not intuitive to me to try Hydra/params.batch_size
I will try it nonetheless as you suggested.
OmegaConf
is the configuration, the overrides are in the Hyperparameters "Hydra" section
None
Hi @<1603198134261911552:profile|ColossalReindeer77>
Hello! does anyone know how to do
HPO
when your parameters are in a
Hydra
Basically hydra parameters are overridden with "Hydra/param"
(this is equivalent to the "override" option of hydra in CLI)
@<1523701205467926528:profile|AgitatedDove14> Got the overrides working with Hydra/params.batch_size
thank you 🙏
Glad to hear!
(yeah @<1603198134261911552:profile|ColossalReindeer77> I'm with you the override is not intuitive, I'll pass the info to the technical writers, hopefully they can find a way to make it easier to understand)
from this video tutorial None :
“…the name of the hyperparameter consist of the section is reported to followed by a slash then its name…”
So following that confuses me because I can’t see my Hydra parameters under Hyperparameters > Hydra
and this is why I thought, ok well, perhaps use OmegaConf/params.batch_size
Is this another opportunity to improve the documentation? Happy to help if so.
try Hydra/trainer.params.batch_size
hydra separates nesting with "."