Sampling-based optimization with mixtures - IN2P3 - Institut national de physique nucléaire et de physique des particules Accéder directement au contenu
Poster De Conférence Année : 2009

Sampling-based optimization with mixtures

Résumé

Sampling-based Evolutionary Algorithms (EA) are of great use when dealing with a highly non-convex and/or noisy optimization task, which is the kind of task we often have to solve in Machine Learning. Two derivative-free examples of such methods are Estimation of Distribution Algorithms (EDA) and techniques based on the Cross-Entropy Method (CEM). One of the main problems these algorithms have to solve is finding a good surrogate model for the normalized target function, that is, a model which has sufficient complexity to fit this target function, but which keeps the computations simple enough. Gaussian mixture models have been applied in practice with great success, but most of these approaches lacked a solid theoretical founding. In this paper we describe a sound mathematical justification for Gaussian mixture surrogate models, more precisely we propose a proper derivation of an EDA/CEM algorithm with mixture updates using Expectation Maximization techniques. It will appear that this algorithm resembles the recent Population MCMC schemes, thus reinforcing the link between Monte- Carlo integration methods and sampling-based optimization. We will concentrate throughout this paper on continuous optimization.
Fichier principal
Vignette du fichier
OPT2009-Bardenet.pdf (315.08 Ko) Télécharger le fichier
Origine : Fichiers produits par l'(les) auteur(s)
Loading...

Dates et versions

in2p3-00580580 , version 1 (28-03-2011)

Identifiants

  • HAL Id : in2p3-00580580 , version 1

Citer

R. Bardenet, Balázs Kégl. Sampling-based optimization with mixtures. OPT 2009: 2nd NIPS Workshop on Optimization for Machine Learning, Dec 2009, Whistler, Canada. ⟨in2p3-00580580⟩
73 Consultations
98 Téléchargements

Partager

Gmail Facebook X LinkedIn More