Muutke küpsiste eelistusi

Semi-Supervised Learning [Pehme köide]

Edited by (Max Planck Institute for Intelligent Systems), Edited by , Edited by (Criteo)
  • Formaat: Paperback / softback, 528 pages, kõrgus x laius x paksus: 254x203x25 mm, kaal: 1043 g, 98 illus.
  • Sari: Adaptive Computation and Machine Learning series
  • Ilmumisaeg: 22-Jan-2010
  • Kirjastus: MIT Press
  • ISBN-10: 0262514125
  • ISBN-13: 9780262514125
Teised raamatud teemal:
  • Formaat: Paperback / softback, 528 pages, kõrgus x laius x paksus: 254x203x25 mm, kaal: 1043 g, 98 illus.
  • Sari: Adaptive Computation and Machine Learning series
  • Ilmumisaeg: 22-Jan-2010
  • Kirjastus: MIT Press
  • ISBN-10: 0262514125
  • ISBN-13: 9780262514125
Teised raamatud teemal:
A comprehensive review of an area of machine learning that deals with the use of unlabeled data in classification problems: state-of-the-art algorithms, a taxonomy of the field, applications, benchmark experiments, and directions for future research.

In the field of machine learning, semi-supervised learning (SSL) occupies the middle ground, between supervised learning (in which all training examples are labeled) and unsupervised learning (in which no label data are given). Interest in SSL has increased in recent years, particularly because of application domains in which unlabeled data are plentiful, such as images, text, and bioinformatics. This first comprehensive overview of SSL presents state-of-the-art algorithms, a taxonomy of the field, selected applications, benchmark experiments, and perspectives on ongoing and future research.

Semi-Supervised Learning first presents the key assumptions and ideas underlying the field: smoothness, cluster or low-density separation, manifold structure, and transduction. The core of the book is the presentation of SSL methods, organized according to algorithmic strategies. After an examination of generative models, the book describes algorithms that implement the low-density separation assumption, graph-based methods, and algorithms that perform two-step learning. The book then discusses SSL applications and offers guidelines for SSL practitioners by analyzing the results of extensive benchmark experiments. Finally, the book looks at interesting directions for SSL research. The book closes with a discussion of the relationship between semi-supervised learning and transduction.

Adaptive Computation and Machine Learning series
Series Foreword xi
Preface xiii
Introduction to Semi-Supervised Learning
1(12)
Supervised, Unsupervised, and Semi-Supervised Learning
1(3)
When Can Semi-Supervised Learning Work?
4(4)
Classes of Algorithms and Organization of This Book
8(5)
I Generative Models
13(90)
A Taxonomy for Semi-Supervised Learning Methods
15(18)
Matthias Seeger
The Semi-Supervised Learning Problem
15(2)
Paradigms for Semi-Supervised Learning
17(5)
Examples
22(9)
Conclusions
31(2)
Semi-Supervised Text Classification Using EM
33(24)
Kamal Nigam
Andrew McCallum
Tom Mitchell
Introduction
33(2)
A Generative Model for Text
35(6)
Experimental Results with Basic EM
41(2)
Using a More Expressive Generative Model
43(6)
Overcoming the Challenges of Local Maxima
49(5)
Conclusions and Summary
54(3)
Risks of Semi-Supervised Learning
57(16)
Fabio Cozman
Ira Cohen
Do Unlabeled Data Improve or Degrade Classification Performance?
57(2)
Understanding Unlabeled Data: Asymptotic Bias
59(4)
The Asymptotic Analysis of Generative Semi-Supervised Learning
63(4)
The Value of Labeled and Unlabeled Data
67(2)
Finite Sample Effects
69(1)
Model Search and Robustness
70(1)
Conclusion
71(2)
Probabilistic Semi-Supervised Clustering with Constraints
73(30)
Sugato Basu
Mikhail Bilenko
Arindam Banerjee
Raymond Mooney
Introduction
74(1)
HMRF Model for Semi-Supervised Clustering
75(6)
HMRF-KMeans Algorithm
81(12)
Active Learning for Constraint Acquisition
93(3)
Experimental Results
96(4)
Related Work
100(1)
Conclusions
101(2)
II Low-Density Separation
103(88)
Transductive Support Vector Machines
105(14)
Thorsten Joachims
Introduction
105(3)
Transductive Support Vector Machines
108(3)
Why Use Margin on the Test Set?
111(1)
Experiments and Applications of TSVMs
112(2)
Solving the TSVM Optimization Problem
114(2)
Connection to Related Approaches
116(1)
Summary and Conclusions
116(3)
Semi-Supervised Learning Using Semi-Definite Programming
119(18)
Tijl De Bie
Nello Cristianini
Relaxing SVM Transduction
119(7)
An Approximation for Speedup
126(2)
General Semi-Supervised Learning Settings
128(1)
Empirical Results
129(4)
Summary and Outlook
133(1)
Appendix: The Extended Schur Complement Lemma
134(3)
Gaussian Processes and the Null-Category Noise Model
137(14)
Neil D. Lawrence
Michael I. Jordan
Introduction
137(4)
The Noise Model
141(2)
Process Model and Effect of the Null-Category
143(2)
Posterior Inference and Prediction
145(2)
Results
147(2)
Discussion
149(2)
Entropy Regularization
151(18)
Yves Grandvalet
Yoshua Bengio
Introduction
151(1)
Derivation of the Criterion
152(3)
Optimization Algorithms
155(3)
Related Methods
158(2)
Experiments
160(6)
Conclusion
166(1)
Appendix: Proof of Theorem 9.1
166(3)
Data-Dependent Regularization
169(22)
Adrian Corduneanu
Tommi Jaakkola
Introduction
169(5)
Information Regularization on Metric Spaces
174(8)
Information Regularization and Relational Data
182(7)
Discussion
189(2)
III Graph-Based Methods
191(84)
Label Propagation and Quadratic Criterion
193(24)
Yoshua Bengio
Olivier Delalleau
Nicolas Le Roux
Introduction
193(1)
Label Propagation on a Similarity Graph
194(4)
Quadratic Cost Criterion
198(7)
From Transduction to Induction
205(1)
Incorporating Class Prior Knowledge
205(1)
Curse of Dimensionality for Semi-Supervised Learning
206(9)
Discussion
215(2)
The Geometric Basis of Semi-Supervised Learning
217(20)
Vikas Sindhwani
Misha Belkin
Partha Niyogi
Introduction
217(3)
Incorporating Geometry in Regularization
220(4)
Algorithms
224(5)
Data-Dependent Kernels for Semi-Supervised Learning
229(2)
Linear Methods for Large-Scale Semi-Supervised Learning
231(1)
Connections to Other Algorithms and Related Work
232(2)
Future Directions
234(3)
Discrete Regularization
237(14)
Dengyong Zhou
Bernhard Scholkopf
Introduction
237(2)
Discrete Analysis
239(6)
Discrete Regularization
245(4)
Conclusion
249(2)
Semi-Supervised Learning with Conditional Harmonic Mixing
251(24)
Christopher J. C. Burges
John C. Platt
Introduction
251(4)
Conditional Harmonic Mixing
255(1)
Learning in CHM Models
256(5)
Incorporating Prior Knowledge
261(1)
Learning the Conditionals
261(1)
Model Averaging
262(1)
Experiments
263(10)
Conclusions
273(2)
IV Change of Representation
275(8)
Graph Kernels by Spectral Transforms
277(6)
Xiaojin Zhu
Jaz Kandola
John Lafferty
Zoubin Ghahramani
The Graph Laplacian
278(2)
Kernels by Spectral Transforms
280(1)
Kernel Alignment
281(1)
Optimizing Alignment Using QCQP for Semi-Supervised Learning
282(1)
V Semi-Supervised Kernels with Order Constraints
283(112)
Experimental Results
285(4)
Conclusion
289(4)
Spectral Methods for Dimensionality Reduction
293(16)
Lawrence K. Saul
Kilian Q. Weinberger
Fei Sha
Jihun Ham
Daniel D. Lee
Introduction
293(2)
Linear Methods
295(2)
Graph-Based Methods
297(6)
Kernel Methods
303(3)
Discussion
306(3)
Modifying Distances
309(24)
Sajama, Alon Orlitsky
Introduction
309(3)
Estimating DBD Metrics
312(9)
Computing DBD Metrics
321(6)
Semi-Supervised Learning Using Density-Based Metrics
327(2)
Conclusions and Future Work
329(2)
Semi-Supervised Learning in Practice
331(2)
Large-Scale Algorithms
333(10)
Olivier Delalleau
Yoshua Bengio
Nicolas Le Roux
Introduction
333(1)
Cost Approximations
334(3)
Subset Selection
337(3)
Discussion
340(3)
Semi-Supervised Protein Classification Using Cluster Kernels
343(18)
Jason Weston
Christina Leslie
Eugene Ie
William Stafford Noble
Introduction
343(2)
Representations and Kernels for Protein Sequences
345(3)
Semi-Supervised Kernels for Protein Sequences
348(4)
Experiments
352(6)
Discussion
358(3)
Prediction of Protein Function from Networks
361(16)
Hyunjung Shin
Koji Tsuda
Introduction
361(3)
Graph-Based Semi-Supervised Learning
364(2)
Combining Multiple Graphs
366(3)
Experiments on Function Prediction of Proteins
369(5)
Conclusion and Outlook
374(3)
Analysis of Benchmarks
377(18)
The Benchmark
377(6)
Application of SSL Methods
383(7)
Results and Discussion
390(5)
VI Perspectives
395(114)
An Augmented PAC Model for Semi-Supervised Learning
397(24)
Maria-Florina Balcan
Avrim Blum
Introduction
398(2)
A Formal Framework
400(3)
Sample Complexity Results
403(9)
Algorithmic Results
412(4)
Related Models and Discussion
416(5)
Metric-Based Approaches for Semi-Supervised Regression and Classification
421(32)
Dale Schuurmans
Finnegan Southey
Dana Wilkinson
Yuhong Guo
Introduction
421(2)
Metric Structure of Supervised Learning
423(3)
Model Selection
426(10)
Regularization
436(9)
Classification
445(4)
Conclusion
449(4)
Transductive Inference and Semi-Supervised Learning
453(20)
Vladimir Vapnik
Problem Settings
453(2)
Problem of Generalization in Inductive and Transductive Inference
455(2)
Structure of the VC Bounds and Transductive Inference
457(1)
The Symmetrization Lemma and Transductive Inference
458(1)
Bounds for Transductive Inference
459(1)
The Structural Risk Minimization Principle for Induction and Transduction
460(2)
Combinatorics in Transductive Inference
462(1)
Measures of the Size of Equivalence Classes
463(2)
Algorithms for Inductive and Transductive SVMs
465(5)
Semi-Supervised Learning
470(1)
Conclusion: Transductive Inference and the New Problems of Inference
470(1)
Beyond Transduction: Selective Inference
471(2)
A Discussion of Semi-Supervised Learning and Transduction
473(36)
References
479(20)
Notation and Symbols
499(4)
Contributors
503(6)
Index 509