A. Appendix, Notations where a more sophisticated quality measure is dened points in O become triplets (D i , x i , y i ), since each point is attached to a single dataset D i Notations for Chapter

P. Abreu, Update on the correlation of the highest energy cosmic rays with nearby extragalactic matter, Astroparticle Physics, vol.34, issue.5, pp.314326-67, 2012.
DOI : 10.1016/j.astropartphys.2010.08.010

URL : https://hal.archives-ouvertes.fr/in2p3-00523359

B. Adenso-diaz and M. Laguna, Fine-Tuning of Algorithms Using Fractional Experimental Designs and Local Search, Operations Research, vol.54, issue.1, p.99114, 2006.
DOI : 10.1287/opre.1050.0243

M. Aglietta, A direct measurement of the photoelectron number per vertical muon in the Capisa SD detector, 2005.

J. Alvarez-muñiz, G. Rodríguez-fernández, I. Valiño, and E. Zas, An alternative method for tank signal response and S(1000) calculation, 2005.

J. Alvarez-muñiz, G. Rodríguez-fernández, I. Valiño, and E. Zas, Update on the method for tank signal response (SdSignalUSC code), 2009.

C. Andrieu, E. Moulines, and P. Priouret, Stability of stochastic approximation under veriable conditions, SIAM Journal on Control and Optimization, vol.44, pp.283312-118, 2005.

C. Andrieu and C. P. Robert, Controlled Markov chain Monte Carlo methods for optimal sampling, Cahiers du Ceremade, 2001.

C. Andrieu and J. Thoms, A tutorial on adaptive MCMC, Statistics and Computing, vol.61, issue.3, p.343373, 2008.
DOI : 10.1007/s11222-008-9110-y

Y. Atchadé, G. Fort, E. Moulines, and P. Priouret, Bayesian Time Series Models, chapter Adaptive Markov chain Monte Carlo: Theory and Methods, pages 3353, pp.102-149, 2011.

C. Audet and D. Orban, Finding Optimal Algorithmic Parameters Using Derivative???Free Optimization, SIAM Journal on Optimization, vol.17, issue.3, pp.642-664, 2006.
DOI : 10.1137/040620886

P. Auer, N. Cesa-bianchi, and P. Fischer, Finite-time analysis of the multiarmed bandit problem, Machine Learning, vol.47, issue.151, p.235256, 2002.

P. Auger, Extensive Cosmic-Ray Showers, Reviews of Modern Physics, vol.11, issue.3-4, pp.288-291, 1939.
DOI : 10.1103/RevModPhys.11.288

R. Bardenet, O. Cappé, G. Fort, and B. Kégl, Adaptive MCMC with online relabeling. Submitted, preprint available as arXiv:1210.2601
URL : https://hal.archives-ouvertes.fr/in2p3-01115785

R. Bardenet, O. Cappé, G. Fort, and B. Kégl, An adaptive Metropolis algorithm with online relabeling, Proceedings of the International Conference on Articial Intelligence and Statistics (AISTATS), pp.9199-117, 2012.

R. Bardenet and B. Kégl, Surrogating the surrogate: accelerating Gaussianprocess-based global optimization with a mixture cross-entropy algorithm, Proceedings of the 27th International Conference on Machine Learning (ICML), 2010.

R. Bardenet and B. Kégl, An adaptive Monte Carlo Markov chain algorithm for inference from mixture signals, Proceedings of ACAT'11 Journal of Physics: Conference series, 2012. (Cited on pages vii, pp.76-80
URL : https://hal.archives-ouvertes.fr/in2p3-00714497

R. Bardenet, B. Kégl, and D. Veberi£, Single muon response: The signal model, p.73, 2010.

R. Benassi, J. Bect, and E. Vazquez, Bayesian optimization using sequential Monte Carlo, Proceedings of the International Conference on Learning and Intelligent Optimization (LION), p.2012
URL : https://hal.archives-ouvertes.fr/hal-00717195

Y. Bengio, Learning deep architectures for AI. Foundations and Trends, Machine Learning, p.1127, 2009.

J. Bergstra, R. Bardenet, B. Kégl, and Y. Bengio, Algorithms for hyperparameter optimization, Advances in Neural Information Processing Systems, pp.31-38, 2011.
URL : https://hal.archives-ouvertes.fr/hal-00642998

J. Bergstra, R. Bardenet, B. Kégl, and Y. Bengio, Implementations of algorithms for hyper-parameter optimization, NIPS Workshop on Bayesian optimization, 2011.

J. Bergstra and Y. Bengio, Random search for hyper-parameter optimization, Journal of Machine Learning Research, pp.35-36, 2012.

M. Birattari, T. Stützle, L. Paquete, and K. Varrentrapp, A racing algorithm for conguring metaheuristics, Proceedings of the Genetic and Evolutionary Computation Conference (GECCO), 2002.

C. M. Bishop, Neural Networks for Pattern Recognition, 1995.

C. B. Bonifazi, P. Bauleo, A. Ferrero, A. Filevich, and A. Reguera, Response of a water Cherenkov detector to oblique and quasi-horizontal muons, 2001.

B. Boser, I. Guyon, and V. Vapnik, A training algorithm for optimal margin classiers, Fifth Annual Workshop on Computational Learning Theory, p.144152, 1992.

S. Boyd and L. Vandenberghe, Convex Optimization, 1991.

L. Breiman, Random forests, Machine Learning, 2001.

M. Brendel and M. Schoenauer, Instance-based parameter tuning for evolutionary AI planning, Proceedings of the 20th Genetic and Evolutionary Computation Conference, p.28, 2011.

J. W. Brewer, Kronecker products and matrix calculus in system theory, IEEE Transactions on Circuits and Systems, vol.25, issue.129, pp.772781-127, 1978.

S. Bubeck, R. Munos, and G. Stoltz, Pure exploration in multi-armed bandits problems, Proceedings of the international conference on Algorithmic Learning Theory (ALT), p.27, 2009.

B. P. Carlin and T. A. Louis, Bayes and Empirical Bayes Methods for Data Analysis, 2000.

M. A. Carreira-perpiñan and G. E. Hinton, On contrastive divergence learning, Proceedings of the Tenth International Workshop on Articial Intelligence and Statistics (AISTATS), 2005.

A. Castellina and G. Navarra, Separating the electromagnetic and muonic components in the FADC traces of the Auger Surface Detectors, 2006.

G. Celeux, Bayesian inference for mixtures: The label-switching problem, COMPSTAT 98. Physica-Verlag, pp.94-106, 1998.

G. Celeux, M. A. Hurn, and C. P. Robert, Computational and inferential diculties with mixture posterior distributions, J. American Statist. Assoc, vol.95, pp.957970-93, 1995.

K. Chalupka, C. K. Williams, and I. Murray, A framework for evaluating approximation methods for Gaussian process regression. pre-print, pp.54-152, 2012.

A. S. Chou, Vertical equivalent muon study with the Fermilab tank, 2002.

W. Chu and Z. Ghahramani, Preference learning with Gaussian processes, Proceedings of the 22nd International Conference on Machine Learning, pp.137144-137191, 2005.

A. Coates, H. Lee, and A. Y. Ng, An analysis of single-layer networks in unsupervised feature learning, NIPS Deep Learning and Unsupervised Feature Learning Workshop, 2010.

A. Coates, H. Lee, and A. Y. Ng, An analysis of single-layer networks in unsupervised feature learning, NIPS Deep Learning and Unsupervised Feature Learning Workshop, pp.26-32, 2010.

A. Coates and A. Y. Ng, The importance of encoding versus training with sparse coding and vector quantization, Proceedings of the International Conference on Machine Learning (ICML), 2011.

A. Coates and A. Y. Ng, The importance of encoding versus training with sparse coding and vector quantization, Proceedings of the International Conference on Machine Learning (ICML), pp.26-32, 2011.

S. P. Coy, B. L. Golden, G. C. Runger, and E. A. , Using experimental design to nd eective parameter settings for heuristics, Journal of Heuristics, vol.7, issue.1, pp.77-97, 2001.

N. Cristianini and J. Shawe-taylor, Kernel methods for pattern recognition, 2004.

A. J. Cron and M. West, Ecient classication-based relabeling in mixture models. The American Statistician, p.1620, 2011.

D. Ravignani, Calculation of the number of photoelectrons with the water Cherenkov detector model, pp.74-77, 1997.

A. Dempster, N. Laird, and D. Rubin, Maximum likelihood from incomplete data via the EM algorithm, Journal of the Royal Statistical Society, Series B, vol.39, issue.1, p.138, 1977.

P. N. Diep, Comments on muon counting in the FADC traces of the Auger surface detector, 2008.

D. Dornic, Développement et caractérisation de photomultiplicateurs hémisphériques pour les expériences d'astroparticules d'étalonnage des détecteurs de surface et analyse des gerbes horizontales de l'Observatoire Pierre Auger, pp.74-77, 2006.

D. Dornic, F. Arneodo, I. Lhenry-yvon, X. Bertou, C. Bonifazi et al., Calibration analysis: Capisa data, 2005.

A. Etchegoyen, Track geometry and smearing of the bump calibration, 2002.

T. Fceyn and . Groups, Simulations with GEANT, 1996.

G. R. Fernandez, A. Tripathi, and K. Arisaka, Simulation of the Pierre Auger surface detector response using GEANT4, 2003.

G. R. Fernandez, E. Zas, T. Ohnuki, A. Tripathi, D. Barnhill et al., Surface detector response using lookup table based on GEANT4 simulation, 2004.

D. Garcia-pinto-for-the-pierre-auger-collaboration, Measurements of the longitudinal development or air showers with the Pierre Auger observatory, Proceedings of the International Cosmic Ray Conference, p.70, 2012.

F. Salamida-for-the-pierre-auger-collaboration, Update on the measurement of the CR energy spectrum above 10 18 eV made using the Pierre Auger observatory, Proceedings of the International Cosmic Ray Conference, p.67, 2012.

G. Fort, E. Moulines, and P. Priouret, Convergence of adaptive and interacting Markov chain Monte Carlo algorithms, Annals of Statistics, vol.39, issue.143, pp.32623289-118, 2012.
URL : https://hal.archives-ouvertes.fr/hal-00695649

S. Frühwirth-schnatter, Markov chain Monte Carlo estimation of classical and dynamic switching in mixture models, Journal of the American Statistical Association, vol.96, issue.453, 2001.

X. Garrido, A. Cordier, S. Dagoret-campagne, B. Kégl, D. Monnier-ragaigne et al., Measurement of the number of muons in Auger tanks by the FADC jump counting method, 2007.

X. Garrido, B. Kégl, A. Cordier, S. Dagoret-campagne, D. Monnier-ragaigne et al., Update and new results from the FADC jump counting method, 2009.

B. Genolini, T. Nguyen-trunc, J. Pouthas, P. Lavoute, C. Meunier et al., Photonis XP1805 and PAO SD bases: eects of the temperature and of the Earth's magnetic eld, p.77, 2003.

P. W. Goldberg, C. K. Williams, and C. M. Bishop, Regression with inputdependent noise: A Gaussian process treatment, Advances in Neural Information (NIPS), 1998.

S. Graf and H. Luschgy, Foundations of Quantization for Probability Distributions, 2000.

R. Gramacy, Bayesian treed Gaussian process models, 2005.

J. Gratch and G. Dejong, Composer: A probabilistic solution to the utility problem in speed-up learning, Proceedings of the Tenth National Conference on Articial Intelligence (AAAI), p.240, 1992.

P. J. Green, Reversible jump Markov chain Monte Carlo computation and Bayesian model determination, Biometrika, vol.82, issue.4, pp.711732-711755, 1995.

P. Hall and C. C. Heyde, Martingale limit theory and its application, 1980.

N. Hansen, The CMA evolution strategy: a comparing review, Towards a new evolutionary computation. Advances on estimation of distribution algorithms, p.75102

P. Hennig and C. J. Schuler, Entropy search for information-ecient global optimization, Journal of Machine Learning Research, 2012.

V. Hess, Observations of the penetrating radiation on seven balloon ights

G. E. Hinton, S. Osindero, and Y. Teh, A fast learning algorithm for deep belief nets, Neural Computation, vol.18, p.15271554, 2006.

S. Hornus and J. Boissonnat, An ecient implementation of delaunay triangulations in medium dimensions, 2008.

M. A. Hurn, A. Justel, and C. P. Robert, Estimating mixtures of regressions, Journal of Computational and Graphical Statistics, vol.12, p.5579, 2003.

F. Hutter, Automated Conguration of Algorithms for Solving Hard Computational Problems, pp.28-29, 2009.

A. Hyvärinen and E. Oja, Independent component analysis: Algorithms and applications, Neural Networks, vol.13, issue.45, p.411430, 2000.

A. Jasra, Bayesian inference for mixture models via Monte Carlo, pp.94-102, 2005.

A. Jasra, C. C. Holmes, and D. A. Stephens, Markov chain Monte Carlo methods and the label switching problem in Bayesian mixture modelling, Statistical Science, vol.20, issue.1, pp.5067-93, 2005.

T. Joachims, Optimizing search engines using clickthrough data, Proceedings of the ACM Conference on Knowledge Discovery and Data Mining (KDD), p.47, 2002.

D. R. Jones, A taxonomy of global optimization methods based on response surfaces, Journal of Global Optimization, vol.21, issue.2, pp.345383-345393, 2001.

E. Kaufmann, O. Cappé, and A. Garivier, On Bayesian upper condence bounds for bandit problems, Proceedings of the International Conference on Articial Intelligence and Statistics (AISTATS), pp.2012-2042

B. Kégl and R. Busa-fekete, Boosting products of base classiers, International Conference on Machine Learning, p.497504, 2009.

B. Kégl, R. Busa-fekete, K. Louedec, R. Bardenet, X. Garrido et al., Reconstructing N µ19 (1000), pp.73-83, 2011.

B. Kégl, M. Unger, and R. Busa-fekete, A nonparametric approach to estimate X max using Gaussian process regression, 2010.

B. Kégl and D. Veberi£, Single muon response: Tracklength, p.75, 2009.

H. Larochelle, D. Erhan, A. Courville, J. Bergstra, and Y. Bengio, An empirical evaluation of deep architectures on problems with many factors of variation, Proceedings of the International Conference on Machine Learning, pp.473480-473512, 2007.

P. Larrañaga and J. Lozano, Estimation of Distribution Algorithms: A New Tool for Evolutionary Computation, pp.16-19, 2001.

Y. Lecun, L. Bottou, Y. Bengio, and P. Haner, Gradient-based learning applied to document recognition, Proceedings of the IEEE, p.22782324, 1998.

D. Lizotte, Practical Bayesian Optimization, 2008.

K. Louedec, Atmospheric aerosols at the Pierre Auger Observatory: characterization and eect on the energy estimation for ultra-high energy cosmic rays, 2011.

N. Mahendran, Z. Wang, F. Hamze, and N. De-freitas, Adaptive MCMC with Bayesian optimization, Proceedings of the International Conference on Articial Intelligence and Statistics (AISTATS), p.2012

J. M. Marin, K. Mengersen, and C. P. Robert, Bayesian modelling and inference on mixtures of distributions, Handbook of Statistics, vol.25, issue.96, pp.94-102, 2004.

A. Mchutchon and C. E. Rasmussen, Gaussian process training with input noise, Advances in Neural Information (NIPS), 2011.

S. P. Meyn and R. L. Tweedie, Markov chains and stochastic stability, 1993.

J. Mockus, V. Tiesis, and A. Zilinskas, The application of Bayesian methods for seeking the extremum, Towards Global Optimization, p.117129, 1978.

M. Muja and D. G. Lowe, Fast approximate nearest neighbors with automatic algorithm conguration, International Conference on Computer Vision Theory and Applications, 2009.

V. Nannen, S. Smit, and A. Eiben, Costs and benets of tuning parameters of evolutionary algorithms, Parallel Problem Solving from Nature (PPSN), 2008.

G. Pagès, A space quantization method for numerical integration, Journal of Computational and Applied Mathematics, vol.89, p.138, 1997.

P. Papastamoulis and G. Iliopoulos, An articial allocations based solution to the label switching problem in Bayesian analysis of mixtures of distribution

P. Papastamoulis and G. Iliopoulos, On the convergence rate of random permutation sampler and ECR algorithm in missing data models, Methodology and Computing in Applied Probability, 2011.

N. Pinto, D. Doukhan, J. J. Dicarlo, and D. D. Cox, A high-throughput screening approach to discovering good forms of biologically inspired visual representation, PLoS Comput Biol, vol.5, issue.11 1, pp.26-32, 2009.

F. P. Preparata and M. I. Shamos, Computational Geometry, an Introduction. Texts and Monographs in Computer Science, 1988.

C. Pryke, Geometrical design studies for water Cherenkov detectors via simulation, 1996.

C. Pryke, Performance simulations of a 10 m 2 water Cherenkov detector and comparison with experiment, 1997.

C. Pryke, Self calibration of the water Cherenkov tanks: Simulation, 1997.

C. E. Rasmussen and C. K. Williams, Gaussian Processes for Machine Learning, pp.37-47, 2006.

S. Richardson and P. J. Green, On Bayesian analysis of mixtures with an unknown number of components, Journal of the Royal Statistical Society, Series B, vol.59, issue.4, p.731792, 1997.

C. P. Robert and G. Casella, Monte Carlo Statistical Methods, pp.15-17, 2004.

G. Roberts, A. Gelman, and W. Gilks, Weak convergence of optimal scaling of random walk Metropolis algorithms, The Annals of Applied Probability, vol.7, issue.3, pp.110120-102, 1997.

G. O. Roberts and J. S. Rosenthal, Optimal scaling for various Metropolis- Hastings algorithms, Statistical Science, vol.16, issue.3, pp.351367-102, 2001.

G. O. Roberts and J. S. Rosenthal, Coupling and ergodicity of adaptive MCMC, Journal of Applied Probability, vol.44, issue.153, pp.486475-118, 2007.

G. O. Roberts and J. S. Rosenthal, Examples of adaptive MCMC, Journal of Computational and Graphical Statistics, vol.18, p.349367, 2009.

A. Roodaki, Signal decompositions using trans-dimensional Bayesian methods
URL : https://hal.archives-ouvertes.fr/tel-00765464

A. Roodaki, J. Bect, and G. Fleury, Summarizing posterior distributions in signal decomposition problems when the number of components is unknown, IEEE Int. Conf. Acoust., Speech, Signal Processing (ICASSP), pp.2012-94
URL : https://hal.archives-ouvertes.fr/hal-00725296

R. Y. Rubinstein and D. P. Kroese, The Cross-Entropy Method: A Unied Approach to Combinatorial Optimization, Monte Carlo Simulation and Machine Learning, pp.15-16, 2004.

R. E. Schapire and Y. Singer, Improved boosting algorithms using condencerated predictions, Machine Learning, p.297336, 1999.

W. E. Slater, A. Tripathi, and K. Arisaka, A GEANT3 simulation of Pierre Auger Surface Detector response to muons, 2002.

J. Snoek, H. Larochelle, and R. P. Adams, Practical Bayesian optimization of machine learning algorithms, Advances in Neural Information Processing Systems (NIPS), 2012. (Cited on pages iv, pp.10-13

M. Sperrin, T. Jaki, and E. Wit, Probabilistic relabelling strategies for the label switching problem in Bayesian mixture models, Statistics and Computing, vol.20, issue.99, pp.357366-94, 2010.

N. Srinivas, A. Krause, S. Kakade, and M. Seeger, Gaussian process optimization in the bandit setting: No regret and experimental design, Proceedings of the 27th International Conference on Machine Learning, pp.12-30, 2010.

M. Stephens, Dealing with label switching in mixture models, Journal of the Royal Statistical Society, Series B, vol.62, issue.97, pp.795809-98, 2000.

D. Supanitsky and X. Bertou, Semi-analytical model of the three-fold charge spectrum in a water Cherenkov tank, 2003.

H. Terashima-marín, P. Ross, and M. Valenzuela-réndon, Evolution of constraint satisfaction strategies in examination timetabling, Proceedings of the Genetic and Evolutionary Computation Conference (GECCO), pp.635-642, 1999.

C. Thornton, F. Hutter, H. H. Hoos, and K. Leyton-brown, Auto-weka: Automated selection and hyper-parameter optimization of classication algorithms, pp.2012-2041

B. A. Tolson and C. A. Shoemaker, Dynamically dimensioned search algorithm for computationally ecient watershed model calibration, Water Resources Research, vol.43, 2007.

E. Vazquez and J. Bect, Convergence properties of the expected improvement algorithm with xed mean and covariance functions, Journal of Statistical Planning and Inference, vol.140, issue.11, p.30883095, 2010.

J. Villemonteix, E. Vazquez, and E. Walter, An informational approach to the global optimization of expensive-to-evaluate functions, Journal of Global Optimization, pp.10-12, 2006.
URL : https://hal.archives-ouvertes.fr/hal-00354262

P. Vincent, H. Larochelle, I. Lajoie, Y. Bengio, and P. A. , Stacked denoising autoencoders: Learning useful representations in a deep network with a local denoising criterion, Machine Learning Research, vol.11, p.33713408, 2010.