Gaussian Processes for Machine Learning
by Carl E. Rasmussen, Christopher K. I. Williams
Publisher: The MIT Press 2005
Number of pages: 266
Gaussian processes (GPs) provide a principled, practical, probabilistic approach to learning in kernel machines. The book deals with the supervised-learning problem for both regression and classification, and includes detailed algorithms. A wide variety of covariance (kernel) functions are presented and their properties discussed. Model selection is discussed both from a Bayesian and a classical perspective. Many connections to other well-known techniques from machine learning and statistics are discussed, including support-vector machines, neural networks, splines, regularization networks, relevance vector machines and others.
Home page url
Download or read it online for free here:
(multiple PDF files)
by Abdelhamid Mellouk, Abdennacer Chebira - InTech
Neural machine learning approaches, Hamiltonian neural networks, similarity discriminant analysis, machine learning methods for spoken dialogue simulation and optimization, linear subspace learning for facial expression analysis, and more.
by David Beyer - O'Reilly Media
The series of interviews in this exclusive report unpack concepts and innovations that represent the frontiers of ever-smarter machines. You’ll get a rare glimpse into this exciting field through the eyes of some of its leading minds.
by Amnon Shashua - arXiv
Introduction to Machine learning covering Statistical Inference (Bayes, EM, ML/MaxEnt duality), algebraic and spectral methods (PCA, LDA, CCA, Clustering), and PAC learning (the Formal model, VC dimension, Double Sampling theorem).
by A. Goldenberg, A.X. Zheng, S.E. Fienberg, E.M. Airoldi - arXiv
We begin with the historical development of statistical network modeling and then we introduce some examples in the network literature. Our subsequent discussion focuses on prominent static and dynamic network models and their interconnections.