Examples: query, "exact match", wildcard*, wild?ard, wild*rd
Fuzzy search: cake~ (finds cakes, bake)
Term boost: "red velvet"^4, chocolate^2
Field grouping: tags:(+work -"fun-stuff")
Escaping: Escape characters +-&|!(){}[]^"~*?:\ with \, e.g. \+
Range search: properties.timestamp:[1587729413488 TO *] (inclusive), properties.title:{A TO Z}(excluding A and Z)
Combinations: chocolate AND vanilla, chocolate OR vanilla, (chocolate OR vanilla) NOT "vanilla pudding"
Field search: properties.title:"The Title" AND text
Answered
Hello! There Is Great Alternative For Argparse Developed By Facebook For Ml Named

Hello!
There is great alternative for argparse developed by facebook for ml named https://hydra.cc/ . But unfortunately it doesn't work with trains in-browser arguments changing feature. Suppose it's because hydra gets parameters like param=value, instead of --param value.
Is it possible to add support of hydra to trains? 🙃

  
  
Posted 4 years ago
Votes Newest

Answers 11


TrickyRaccoon92 actually Click is on the to do list as well ...

  
  
Posted 4 years ago

I want the support for click as well, or is there any adhoc solution?

  
  
Posted 4 years ago

Best thing ever, thanks AgitatedDove14 !

  
  
Posted 4 years ago

AgitatedDove14 from this thread I understand hydra is not supported and therefore overriding the parameters from the UI wont work, but is there still a way to track and add the parameters to the experiment? Will task.connect_configuration work with the yaml files?

  
  
Posted 4 years ago

GrievingTurkey78 Actually it is in progress, see the GitHub issue for details:
https://github.com/allegroai/trains/issues/219

  
  
Posted 4 years ago

Not sure if you've tried this, but would click work? Otherwise the simplest thing I know is possible in Click, is to send the Dict of arguments further down the pipeline for the Allegro callback listener.

  
  
Posted 4 years ago

AgitatedDove14 do you have any ETA on that feature? Can I do anything to speed up the devopment?

  
  
Posted 4 years ago

GrievingTurkey78 yes, you are correct on both.

Will the sweep functionality work?

Yes it should, that said, it will not use the trains-agent so you are limited to the machine running the sweep.
If you want to do HPO on multi-node, checkout this example 🙂
https://github.com/allegroai/trains/blob/master/examples/optimization/hyper-parameter-optimization/hyper_parameter_optimizer.py

  
  
Posted 4 years ago

Hi ProudMosquito87
My apologies there is still no concrete ETA ...
That said I think a good toy example would really help accelerate this process.
How about opening a PR with a nice hydra example, then we can start discussing implementation details based on the toy example ?

  
  
Posted 4 years ago

Hi PungentLouse55 ,
Yes we have integration with hydra on the todo list since it was first released, we actually know the guy behind Hydra, he is awesome!
What are your thoughts on integration, we would love to get feedback and pointers (Hydra itself is quite capable, and we waiting until we have multiple configuration support, and with v0.16 it was added, so now it is actually possible)

  
  
Posted 4 years ago

Just to make sure get everything right AgitatedDove14 :
We have to define the Task inside the function decorated with the @hydra.main We can modify the parameters that are overridden on UI on : configuration tab -> Args -> overrides -> modify the listAdditional question:
Will the sweep functionality work?

  
  
Posted 4 years ago
1K Views
11 Answers
4 years ago
one year ago
Tags