Advances in Kernel Methods: Support Vector Learning
MIT Press, 1999 - 376 pages
The Support Vector Machine is a powerful new learning algorithm for solving a variety of learning and function estimation problems, such as pattern recognition, regression estimation, and operator inversion. The impetus for this collection was a workshop on Support Vector Machines held at the 1997 NIPS conference. The contributors, both university researchers and engineers developing applications for the corporate world, form a Who's Who of this exciting new area.
Contributors: Peter Bartlett, Kristin P. Bennett, Christopher J. C. Burges, Nello Cristianini, Alex Gammerman, Federico Girosi, Simon Haykin, Thorsten Joachims, Linda Kaufman, Jens Kohlmorgen, Ulrich Kressel, Davide Mattera, Klaus-Robert Muller, Manfred Opper, Edgar E. Osuna, John C. Platt, Gunnar Ratsch, Bernhard Scholkopf, John Shawe-Taylor, Alexander J. Smola, Mark O. Stitson, Vladimir Vapnik, Volodya Vovk, Grace Wahba, Chris Watkins, Jason Weston, Robert C. Williamson.
What people are saying - Write a review
Three Remarks on the Support Vector Method of Function
Generalization Performance of Support Vector Machines
Bayesian Voting Schemes and Large Margin Classifiers
Support Vector Machines Reproducing Kernel Hilbert Spaces
Geometry and Invariance in Kernel Based Methods
Entropy Numbers Operators and Support Vector Kernels
Support Vector Machines for Dynamic Reconstruction of
Using Support Vector Machines for Time Series Prediction
Pairwise Classification and Support Vector Machines
Reducing the Runtime Complexity in Support Vector Machines
Support Vector Regression with ANOVA Decomposition Kernels
Support Vector Density Estimation
Combining Support Vector and Mathematical Programming
Kernel Principal Component Analysis
Making LargeScale Support Vector Machine Learning Practical
Fast Training of Support Vector Machines Using Sequential
algorithm ANOVA approach approximation approximation error benchmark bound chapter choice choose classifiers coefficients compute conjugate gradient consider constraints construct corresponding covering numbers CPU sec data set decision function decision tree decomposition defined density estimation dimensional distribution dot product e-insensitive eigenvalues eigenvectors entropy entropy numbers equation error fc(x feature space figure Gaussian given heuristic Hilbert space hyperplane input space invariant iteration kernel function kernel PCA Lagrange multipliers learning linear program loss function mapping margin matrix Mercer kernel method metric minimize multicategory nonlinear nonzero number of support obtained optimal hyperplane optimization problem parameters pattern recognition PCG chunking performance polynomial kernels principal components QP problem quadratic programming random reconstruction samples Scholkopf Smola solution solve sparse subset support vector machine SV machines SVM pairwise technique test set Theorem training data training examples training set Vapnik variables VC dimension Wahba zero
Page 368 - Golowich, and A. Smola. Support Vector Method for Function Approximation, Regression Estimation and Signal Processing.
Page 354 - Proc. of the 4th Midwest Artificial Intelligence and Cognitive Science Society Conference, pages 97-101, 1992.