Skip to content

Latest commit

 

History

History
19 lines (15 loc) · 1.58 KB

hyper-parameter-optimization.md

File metadata and controls

19 lines (15 loc) · 1.58 KB

Hyper Parameter Optimization

  1. Using HyperOpt

    Random Search
    Tree of Parzen Estimators (TPE)

    Hyperopt has been designed to accommodate Bayesian optimization algorithms based on Gaussian processes and regression trees, but these are not currently implemented.

    All algorithms can be run either serially, or in parallel by communicating via MongoDB.

    Mlflow, Hyperparameterhunter,hyperopt, concept drift, unit tests.
    Hyperopt for hyperparameter search\

  2. HyperparameterHunter
    provides a wrapper for machine learning algorithms that saves all the important data. Simplify the experimentation and hyperparameter tuning process by letting HyperparameterHunter do the hard work of recording, organizing, and learning from your tests — all while using the same libraries you already do. Don't let any of your experiments go to waste, and start doing hyperparameter optimization the way it was meant to be.\

  3. Implementation and comparison - HH slower than HO due to usage of skopt.

  4. HumpDay - a package that compares optimization algorithms and ranks them