On the marchenko-pastur and circular laws for some classes of random matrices with dependent entries, Electronic Journal of Probability, vol.16, pp.1065-1095, 2011. ,
, Special functions, vol.71, 2000.
, Naum Ilich Akhiezer and Izrail Markovich Glazman. Theory of linear operators in Hilbert space, Courier Corporation, 2013.
Evolutionary formalism for products of positive random matrices, The Annals of Applied Probability, vol.4, issue.3, pp.859-901, 1994. ,
Indications of nonlinear deterministic and finite-dimensional structures in time series of brain electrical activity: Dependence on recording region and brain state, Physical Review E, vol.64, issue.6, p.61907, 2001. ,
Generalized optimization framework for graph-based semisupervised learning, Proceedings of the 2012 SIAM International Conference on Data Mining, pp.966-974, 2012. ,
URL : https://hal.archives-ouvertes.fr/inria-00633818
Handbook of mathematical functions: with formulas, graphs, and mathematical tables, Courier Corporation, vol.55, 1965. ,
High-dimensional dynamics of generalization error in neural networks, 2017. ,
Dynamical behavior of artificial neural networks with random weights, vol.6, pp.17-22, 1996. ,
A convergence theory for deep learning via over-parameterization, 2018. ,
Generalized discriminant analysis using a kernel approach, Neural computation, vol.12, issue.10, pp.2385-2404, 2000. ,
Phase transition of the largest eigenvalue for nonnull complex sample covariance matrices. The Annals of Probability, vol.33, pp.1643-1697, 2005. ,
Convex analysis and monotone operator theory in Hilbert spaces, vol.408, 2011. ,
Representation learning: A review and new perspectives, IEEE transactions on pattern analysis and machine intelligence, vol.35, pp.1798-1828, 2013. ,
On the difficulty of approximately maximizing agreements, Journal of Computer and System Sciences, vol.66, issue.3, pp.496-514, 2003. ,
Spectral analysis of the gram matrix of mixture models, ESAIM: Probability and Statistics, vol.20, pp.217-237, 2016. ,
URL : https://hal.archives-ouvertes.fr/hal-01215342
The eigenvalues and eigenvectors of finite, low rank perturbations of large random matrices, Advances in Mathematics, vol.227, issue.1, pp.494-521, 2011. ,
URL : https://hal.archives-ouvertes.fr/hal-00423593
The singular values and vectors of low rank perturbations of large rectangular random matrices, Journal of Multivariate Analysis, vol.111, pp.120-135, 2012. ,
URL : https://hal.archives-ouvertes.fr/hal-00575203
Neural networks and principal component analysis: Learning from examples without local minima, Neural networks, vol.2, issue.1, pp.53-58, 1989. ,
Probability and measure, 2012. ,
, Pattern Recognition and Machine Learning, 2007.
Eigenvalue distribution of nonlinear models of random matrices, 2019. ,
Training a 3-node neural network is NP-complete, Advances in neural information processing systems, pp.494-501, 1989. ,
Random forests. Machine learning, vol.45, pp.5-32, 2001. ,
Eigenvalues of large sample covariance matrices of spiked population models, Journal of multivariate analysis, vol.97, issue.6, pp.1382-1408, 2006. ,
Clt for linear spectral statistics of large-dimensional sample covariance matrices, Advances In Statistics, pp.281-333, 2008. ,
Spectral analysis of large dimensional random matrices, vol.20, 2010. ,
Convex optimization, 2004. ,
Kernel spectral clustering of large dimensional data, Electronic Journal of Statistics, vol.10, issue.1, pp.1393-1454, 2016. ,
URL : https://hal.archives-ouvertes.fr/hal-01215343
Solving random quadratic systems of equations is nearly as easy as solving linear systems, Advances in Neural Information Processing Systems, pp.739-747, 2015. ,
Optimal rates for the regularized least-squares algorithm, Foundations of Computational Mathematics, vol.7, issue.3, pp.331-368, 2007. ,
Hyperspectral imaging: techniques for spectral detection and classification, vol.1, 2003. ,
Random matrix improved subspace clustering, Signals, Systems and Computers, 2016 50th Asilomar Conference on, pp.90-94, 2016. ,
URL : https://hal.archives-ouvertes.fr/hal-01633444
Optimization and Nonsmooth Analysis, Society for Industrial and Applied Mathematics, 1990. ,
Nonconvex optimization meets low-rank matrix factorization: An overview, 2018. ,
Classification asymptotics in the random matrix regime, 26th European Signal Processing Conference (EUSIPCO), pp.1875-1879, 2018. ,
URL : https://hal.archives-ouvertes.fr/hal-01957686
Tight oracle inequalities for low-rank matrix recovery from a minimal number of noisy random measurements, IEEE Transactions on Information Theory, vol.57, issue.4, pp.2342-2359, 2011. ,
Beyond simple features: A large-scale feature search approach to unconstrained face recognition, Face and Gesture, pp.8-15, 2011. ,
The spectrum of random inner-product kernel matrices, Random Matrices: Theory and Applications, vol.2, p.1350010, 2013. ,
The phase transition for the existence of the maximum likelihood estimate in high-dimensional logistic regression, 2018. ,
A unified architecture for natural language processing: Deep neural networks with multitask learning, Proceedings of the 25th international conference on Machine learning, pp.160-167, 2008. ,
The asymptotic performance of linear echo state neural networks, The Journal of Machine Learning Research, vol.17, issue.1, pp.6171-6205, 2016. ,
URL : https://hal.archives-ouvertes.fr/hal-01322809
Gradient descent can take exponential time to escape saddle points, Advances in neural information processing systems, pp.1067-1077, 2017. ,
Explicit bounds for the ?ojasiewicz exponent in the gradient inequality for polynomials, Annales Polonici Mathematici, vol.1, pp.51-61, 2005. ,
Aarti Singh, and Barnabas Poczos. Gradient descent learns one-hidden-layer cnn: Don't be afraid of spurious local minima, International Conference on Machine Learning, pp.1338-1347, 2018. ,
High dimensional robust Mestimation: asymptotic variance via approximate message passing, Probability Theory and Related Fields, vol.166, pp.935-969, 2016. ,
The spectrum of random kernel matrices: universality results for rough and varying kernels, Random Matrices: Theory and Applications, vol.2, p.1350005, 2013. ,
Graph analysis of functional brain networks: practical issues in translational neuroscience, Philosophical Transactions of the Royal Society B: Biological Sciences, vol.369, p.20130521, 1653. ,
Manfred Salmhofer, and Fred Hamprecht. Essentially no barriers in neural network energy landscape, International Conference on Machine Learning, pp.1308-1317, 2018. ,
Concentration of measure and spectra of random matrices: Applications to correlation matrices, elliptical distributions and beyond, The Annals of Applied Probability, vol.19, issue.6, pp.2362-2405, 2009. ,
On Information Plus Noise Kernel Random Matrices, The Annals of Statistics, vol.38, issue.5, pp.3191-3216, 2010. ,
The spectrum of kernel random matrices. The Annals of Statistics, vol.38, pp.1-50, 2010. ,
On robust regression with high-dimensional predictors, Proceedings of the National Academy of Sciences, vol.110, pp.14557-14562, 2013. ,
Image representations for object detection using kernel classifiers, Asian Conference on Computer Vision, pp.687-692, 2000. ,
Topology and geometry of halfrectified network optimization, 2016. ,
The elements of statistical learning, Springer series in statistics, vol.1, 2001. ,
The spectral norm of random innerproduct kernel matrices. Probability Theory and Related Fields, vol.173, pp.27-85, 2019. ,
A short introduction to boosting, Journal-Japanese Society For Artificial Intelligence, vol.14, p.1612, 1999. ,
Understanding the difficulty of training deep feedforward neural networks, Proceedings of the Thirteenth International Conference on Artificial Intelligence and Statistics, pp.249-256, 2010. ,
Deep Learning, 2016. ,
Learning in the recurrent random neural network, Neural computation, vol.5, issue.1, pp.154-164, 1993. ,
Extremely randomized trees, Machine learning, vol.63, issue.1, pp.3-42, 2006. ,
URL : https://hal.archives-ouvertes.fr/hal-00341932
Escaping from saddle points-online stochastic gradient for tensor decomposition, Conference on Learning Theory, pp.797-842, 2015. ,
Loss surfaces, mode connectivity, and fast ensembling of dnns, Advances in Neural Information Processing Systems, pp.8789-8798, 2018. ,
No spurious local minima in nonconvex low rank problems: A unified geometric analysis, Proceedings of the 34th International Conference on Machine Learning, vol.70, pp.1233-1242, 2017. ,
, Izrail Solomonovich Gradshteyn and Iosif Moiseevich Ryzhik. Table of integrals, series, and products, 2014.
, Deep neural networks with random gaussian weights: A universal classification strategy? IEEE Transactions Signal Processing, vol.64, pp.3444-3457, 2016.
Bayesian framework for least-squares support vector machine classifiers, gaussian processes, and kernel fisher discriminant analysis, Neural computation, vol.14, issue.5, pp.1115-1147, 2002. ,
Matrix analysis. Cambridge university press, 2012. ,
Deterministic equivalents for certain functionals of large random matrices, The Annals of Applied Probability, vol.17, issue.3, pp.875-930, 2007. ,
URL : https://hal.archives-ouvertes.fr/hal-00621793
Invariant manifolds, vol.583, 2006. ,
A powerful generative model using random weights for the deep image representation, Advances in Neural Information Processing Systems, pp.631-639, 2016. ,
Extreme learning machine for regression and multiclass classification, IEEE Transactions on Systems, Man, and Cybernetics, vol.42, issue.2, pp.513-529, 2012. ,
Identity mappings in deep residual networks, European conference on computer vision, pp.630-645, 2016. ,
The "echo state" approach to analysing and training recurrent neural networks-with an erratum note, German National Research Center for Information Technology GMD Technical Report, vol.148, issue.34, p.13, 2001. ,
What is the best multi-stage architecture for object recognition?, 2009 IEEE 12th international conference on computer vision, pp.2146-2153, 2009. ,
Concentration of measure and spectra of random matrices: Applications to correlation matrices, elliptical distributions and beyond, The Annals of Applied Probability, pp.2362-2405, 2009. ,
Deep learning without poor local minima, Advances In Neural Information Processing Systems, pp.586-594, 2016. ,
Subspace kernel spectral clustering of large dimensional data, 2017. ,
Random feature maps for dot product kernels, In Artificial Intelligence and Statistics, pp.583-591, 2012. ,
Spectral redemption in clustering sparse networks, Proceedings of the National Academy of Sciences, vol.110, issue.52, pp.20935-20940, 2013. ,
Imagenet classification with deep convolutional neural networks, Advances in neural information processing systems, pp.1097-1105, 2012. ,
Gradientbased learning applied to document recognition, Proceedings of the IEEE, vol.86, issue.11, pp.2278-2324, 1998. ,
On the spectrum of random features maps of high dimensional data, Proceedings of the 35th International Conference on Machine Learning, vol.80, pp.3063-3071, 2018. ,
URL : https://hal.archives-ouvertes.fr/hal-01954933
A random matrix and concentration inequalities framework for neural networks analysis, IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp.4214-4218, 2018. ,
URL : https://hal.archives-ouvertes.fr/hal-01957729
Inner-product kernels are asymptotically equivalent to binary discrete kernels, 2019. ,
A large dimensional analysis of least squares support vector machines, IEEE Transactions on Signal Processing, vol.67, issue.4, pp.1065-1074, 2019. ,
URL : https://hal.archives-ouvertes.fr/hal-02048984
Concentration of measure and large random matrices with an application to sample covariance matrices, 2019. ,
URL : https://hal.archives-ouvertes.fr/hal-02020287
Random matrix theory and financial correlations, International Journal of Theoretical and Applied Finance, vol.3, issue.03, pp.391-397, 2000. ,
The concentration of measure phenomenon. Number 89, 2005. ,
A random matrix approach to neural networks, The Annals of Applied Probability, vol.28, issue.2, pp.1190-1248, 2018. ,
URL : https://hal.archives-ouvertes.fr/hal-01957656
, How to scale up kernel methods to be as good as deep neural nets, 2014.
URL : https://hal.archives-ouvertes.fr/hal-01666934
Sur les trajectoires du gradient d'une fonction analytique. Seminari di geometria, pp.115-117, 1982. ,
Central limit theorem for linear eigenvalue statistics of random matrices with independent entries. The Annals of Probability, vol.37, pp.1778-1840, 2009. ,
Gradient descent only converges to minimizers, Conference on learning theory, pp.1246-1257, 2016. ,
Acoustic modeling using deep belief networks, IEEE transactions on audio, speech, and language processing, vol.20, pp.14-22, 2011. ,
Some NP-complete problems in quadratic and nonlinear programming. Mathematical programming, vol.39, pp.117-129, 1987. ,
Distribution of eigenvalues for some sets of random matrices, Mathematics of the USSR-Sbornik, vol.1, issue.4, p.457, 1967. ,
Fisher discriminant analysis with kernels, Proceedings of the 1999 IEEE signal processing society workshop, pp.41-48, 1999. ,
On the design of loss functions for classification: theory, robustness to outliers, and savageboost, Advances in neural information processing systems, pp.1049-1056, 2009. ,
On spectral clustering: Analysis and an algorithm, Advances in neural information processing systems, pp.849-856, 2002. ,
Cubic regularization of newton method and its global performance, Mathematical Programming, vol.108, issue.1, pp.177-205, 2006. ,
A note on the marchenko-pastur law for a class of random matrices with dependent entries, Electronic Communications in Probability, vol.17, 2012. ,
A simple approach to the global regime of gaussian ensembles of random matrices, Ukrainian Mathematical Journal, vol.57, issue.6, pp.936-966, 2005. ,
Asymptotics of sample eigenstructure for a large dimensional spiked covariance model, Statistica Sinica, pp.1617-1642, 2007. ,
A highthroughput screening approach to discovering good forms of biologically inspired visual representation, PLoS computational biology, vol.5, issue.11, p.1000579, 2009. ,
No eigenvalues outside the support of the limiting empirical spectral distribution of a separable covariance matrix, Journal of Multivariate Analysis, vol.100, issue.1, pp.37-57, 2009. ,
Eigenvalue distribution of large random matrices. Number 171, 2011. ,
Resurrecting the sigmoid in deep learning through dynamical isometry: theory and practice, Advances in neural information processing systems, pp.4785-4795, 2017. ,
Nonlinear random matrix theory for deep learning, Advances in Neural Information Processing Systems, pp.2634-2643, 2017. ,
Imagenet large scale visual recognition challenge, International journal of computer vision, vol.115, issue.3, pp.211-252, 2015. ,
The perceptron: a probabilistic model for information storage and organization in the brain, Psychological review, vol.65, issue.6, p.386, 1958. ,
Random features for large-scale kernel machines, Advances in neural information processing systems, pp.1177-1184, 2008. ,
, Stationary random processes, 1967.
Fourier analysis on groups, vol.121967, 1962. ,
Principles of mathematical analysis, vol.3, 1964. ,
Are loss functions all the same?, Neural Computation, vol.16, issue.5, pp.1063-1076, 2004. ,
On the empirical distribution of eigenvalues of a class of large dimensional random matrices, Journal of Multivariate analysis, vol.54, issue.2, pp.175-192, 1995. ,
Analysis of the limiting spectral distribution of large dimensional random matrices, Journal of Multivariate Analysis, vol.54, issue.2, pp.295-309, 1995. ,
A modern maximumlikelihood theory for high-dimensional logistic regression, 2018. ,
Dropout: A simple way to prevent neural networks from overfitting, Deep learning in neural networks: An overview, vol.61, pp.1929-1958, 2014. ,
On random weights and unsupervised feature learning, ICML, vol.2, p.6, 2011. ,
Feed forward neural networks with random weights, International Conference on Pattern Recognition, pp.1-1, 1992. ,
Spectral clustering of graphs with the bethe hessian, Advances in Neural Information Processing Systems, pp.406-414, 2014. ,
URL : https://hal.archives-ouvertes.fr/cea-01140852
Exact solutions to the nonlinear dynamics of learning in deep linear neural networks, 2013. ,
Learning with kernels: support vector machines, regularization, optimization, and beyond, 2002. ,
A tutorial on support vector regression, Statistics and computing, vol.14, issue.3, pp.199-222, 2004. ,
Estimation of the mean of a multivariate normal distribution. The annals of Statistics, pp.1135-1151, 1981. ,
Least squares support vector machine classifiers. Neural processing letters, vol.9, pp.293-300, 1999. ,
Randomness in neural networks: an overview, Wiley Interdisciplinary Reviews: Data Mining and Knowledge Discovery, vol.7, issue.2, 2017. ,
Deep image prior, Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp.9446-9454, 2018. ,
Principles of risk minimization for learning theory, Advances in neural information processing systems, pp.831-838, 1992. ,
Cambridge university press, Asymptotic statistics, vol.3, 2000. ,
, A tutorial on spectral clustering. Statistics and computing, vol.17, issue.4, pp.395-416, 2007.
Coherency matrix estimation of heterogeneous clutter in high-resolution polarimetric sar images, IEEE Transactions on Geoscience and Remote Sensing, vol.48, issue.4, pp.1809-1826, 2009. ,
URL : https://hal.archives-ouvertes.fr/hal-00466647
Kernel recursive least-squares tracker for time-varying regression, IEEE transactions on neural networks and learning systems, vol.23, pp.1313-1326, 2012. ,
Efficient additive kernels via explicit feature maps. IEEE transactions on pattern analysis and machine intelligence, vol.34, pp.480-492, 2012. ,
Principal component analysis. Chemometrics and intelligent laboratory systems, vol.2, pp.37-52, 1987. ,
Computing with infinite networks, Advances in neural information processing systems, pp.295-301, 1997. ,
Data mining with big data, IEEE transactions on knowledge and data engineering, vol.26, issue.1, pp.97-107, 2013. ,
Fashion-mnist: a novel image dataset for benchmarking machine learning algorithms, 2017. ,
On early stopping in gradient descent learning, Constructive Approximation, vol.26, pp.289-315, 2007. ,
Convergence conditions for nonlinear programming algorithms, Management Science, vol.16, issue.1, pp.1-13, 1969. ,
Understanding deep learning requires rethinking generalization, 2016. ,
Semi-supervised learning literature survey, 2005. ,