nabenabe0928 / reading-list

1 stars 0 forks source link

Frugal Optimization for Cost-related Hyperparameters #65

Closed nabenabe0928 closed 1 year ago

nabenabe0928 commented 1 year ago

Frugal Optimization for Cost-related Hyperparameters

In principle, Line search with some modifications. The main contribution is the theoretical analysis based on convex optimization.

Main points

Condition 3 states that if the costs for $x_1, x_2$ has the relationship of $g(x^\star) < g(x_1) < g(x_2)$, $f(x^\star) < f(x_1) < f(x_2)$ also holds. Note that $x^\star$ is a local optima. Intuitively speaking, longer training time does not improve the model if the model achieved good performance with the cost of $g(x^\star)$, but we would rather end up overfitting with more cost. Although this is not always the case, they cannot provide the theoretical analysis without this condition.

Experiments

They used the following methods:

  1. Random search
  2. BOHB
  3. GPEI
  4. GPEIPS (Expected improvement per second)
  5. SMC
  6. Their methods

Their benchmarks are self-made.

The visualizations are:

  1. Performance over time
  2. dataset-wise final performance distribution of each method
  3. dataset-wise time-to-reach-best distribution of each method ==> I am not convinced of the distribution stuffs.