 Title Pages
 Series Foreword
 Preface

1 Introduction to SemiSupervised Learning 
1 A Taxonomy for SemiSupervised Learning Methods 
3 SemiSupervised Text Classification Using EM 
4 Risks of SemiSupervised Learning: How Unlabeled Data Can Degrade Performance of Generative Classifiers 
5 Probabilistic SemiSupervised Clustering with Constraints 
6 Transductive Support Vector Machines 
7 SemiSupervised Learning Using SemiDefinite Programming 
8 Gaussian Processes and the NullCategory Noise Model 
9 Entropy Regularization 
10 DataDependent Regularization 
11 Label Propagation and Quadratic Criterion 
12 The Geometric Basis of SemiSupervised Learning 
13 Discrete Regularization 
14 SemiSupervised Learning with Conditional Harmonic Mixing 
15 Graph Kernels by Spectral Transforms 
16 Spectral Methods for Dimensionality Reduction 
17 Modifying Distances 
18 LargeScale Algorithms 
19 SemiSupervised Protein Classification Using Cluster Kernels 
20 Prediction of Protein Function from Networks 
25 Analysis of Benchmarks 
22 An Augmented PAC Model for SemiSupervised Learning 
23 MetricBased Approaches for SemiSupervised Regression and Classification 
24 Transductive Inference and SemiSupervised Learning 
25 A Discussion of SemiSupervised Learning and Transduction  References
 Notation and Symbols
 Contributors
 Index
LargeScale Algorithms
LargeScale Algorithms
 Chapter:
 (p.332) (p.333) 18 LargeScale Algorithms
 Source:
 SemiSupervised Learning
 Author(s):
Delalleau Olivier
Bengio Yoshua
Le Roux Nicolas
 Publisher:
 The MIT Press
This chapter presents a subset selection method that can be used to reduce the original system to one of size m 〈〈 n. The idea is to solve for the labels of a subset S ⊂ X of only m points, while still retaining information from the rest of the data by approximating their label with a linear combination of the labels in S—using the induction formula presented in Chapter 11. This leads to an algorithm whose computational requirements scale as O(m^{2}n) and memory requirements as O(m^{2}), thus allowing one to take advantage of significantly bigger unlabeled data sets than with the original algorithms.
Keywords: subset selection method, linear combination, induction formula, algorithm, computational requirements, memory requirements, unlabeled data sets
MIT Press Scholarship Online requires a subscription or purchase to access the full text of books within the service. Public users can however freely search the site and view the abstracts and keywords for each book and chapter.
Please, subscribe or login to access full text content.
If you think you should have access to this title, please contact your librarian.
To troubleshoot, please check our FAQs, and if you can't find the answer there, please contact us.
 Title Pages
 Series Foreword
 Preface

1 Introduction to SemiSupervised Learning 
1 A Taxonomy for SemiSupervised Learning Methods 
3 SemiSupervised Text Classification Using EM 
4 Risks of SemiSupervised Learning: How Unlabeled Data Can Degrade Performance of Generative Classifiers 
5 Probabilistic SemiSupervised Clustering with Constraints 
6 Transductive Support Vector Machines 
7 SemiSupervised Learning Using SemiDefinite Programming 
8 Gaussian Processes and the NullCategory Noise Model 
9 Entropy Regularization 
10 DataDependent Regularization 
11 Label Propagation and Quadratic Criterion 
12 The Geometric Basis of SemiSupervised Learning 
13 Discrete Regularization 
14 SemiSupervised Learning with Conditional Harmonic Mixing 
15 Graph Kernels by Spectral Transforms 
16 Spectral Methods for Dimensionality Reduction 
17 Modifying Distances 
18 LargeScale Algorithms 
19 SemiSupervised Protein Classification Using Cluster Kernels 
20 Prediction of Protein Function from Networks 
25 Analysis of Benchmarks 
22 An Augmented PAC Model for SemiSupervised Learning 
23 MetricBased Approaches for SemiSupervised Regression and Classification 
24 Transductive Inference and SemiSupervised Learning 
25 A Discussion of SemiSupervised Learning and Transduction  References
 Notation and Symbols
 Contributors
 Index