Advances in Optimization and Generalization

The field of machine learning is moving towards more efficient and effective optimization techniques, with a focus on generalization performance. Researchers are exploring new kernel functions, such as the Beta kernel, that can better model functions on bounded domains. There is also a growing interest in understanding the implicit bias of optimization algorithms and their impact on generalization performance. Furthermore, the issue of overtuning in hyperparameter optimization is being addressed, with studies showing that it can lead to overfitting and poor generalization. New methods, such as cross-regularization, are being developed to adaptively control model complexity and improve generalization. Notable papers in this area include:

  • Bayesian Optimization over Bounded Domains with the Beta Product Kernel, which introduces a new kernel function that outperforms existing methods in certain problems.
  • Optimal Implicit Bias in Linear Regression, which provides a precise asymptotic analysis of the generalization performance of interpolators and finds the optimal implicit bias that achieves the best generalization error.
  • Overtuning in Hyperparameter Optimization, which investigates the phenomenon of overtuning and its prevalence in hyperparameter optimization benchmarks.

Sources

Bayesian Optimization over Bounded Domains with the Beta Product Kernel

Bandwidth Selectors on Semiparametric Bayesian Networks

Optimal Implicit Bias in Linear Regression

Overtuning in Hyperparameter Optimization

Cross-regularization: Adaptive Model Complexity through Validation Gradients

Thumb on the Scale: Optimal Loss Weighting in Last Layer Retraining

Built with on top of