Preface |
|
ix | |
Notation |
|
xiii | |
|
|
1 | (8) |
|
|
1 | (2) |
|
Learning and Generalisation |
|
|
3 | (1) |
|
|
4 | (2) |
|
Attractions and Drawbacks of Learning |
|
|
6 | (1) |
|
Support Vector Machines for Learning |
|
|
7 | (1) |
|
|
7 | (1) |
|
Further Reading and Advanced Topics |
|
|
8 | (1) |
|
|
9 | (17) |
|
|
9 | (11) |
|
|
11 | (8) |
|
|
19 | (1) |
|
Multi-class Discrimination |
|
|
20 | (1) |
|
|
20 | (4) |
|
|
21 | (1) |
|
|
22 | (2) |
|
Dual Representation of Linear Machines |
|
|
24 | (1) |
|
|
25 | (1) |
|
Further Reading and Advanced Topics |
|
|
25 | (1) |
|
Kernel-Induced Feature Spaces |
|
|
26 | (26) |
|
Learning in Feature Space |
|
|
27 | (3) |
|
The Implicit Mapping into Feature Space |
|
|
30 | (2) |
|
|
32 | (14) |
|
Characterisation of Kernels |
|
|
33 | (9) |
|
Making Kernels from Kernels |
|
|
42 | (2) |
|
Making Kernels from Features |
|
|
44 | (2) |
|
|
46 | (2) |
|
Kernels and Gaussian Processes |
|
|
48 | (1) |
|
|
49 | (1) |
|
Further Reading and Advanced Topics |
|
|
50 | (2) |
|
|
52 | (27) |
|
Probably Approximately Correct Learning |
|
|
52 | (2) |
|
Vapnik Chervonenkis (VC) Theory |
|
|
54 | (5) |
|
Margin-Based Bounds on Generalisation |
|
|
59 | (10) |
|
|
59 | (5) |
|
|
64 | (1) |
|
|
65 | (4) |
|
Other Bounds on Generalisation and Luckiness |
|
|
69 | (1) |
|
Generalisation for Regression |
|
|
70 | (4) |
|
Bayesian Analysis of Learning |
|
|
74 | (2) |
|
|
76 | (1) |
|
Further Reading and Advanced Topics |
|
|
76 | (3) |
|
|
79 | (14) |
|
|
79 | (2) |
|
|
81 | (6) |
|
|
87 | (2) |
|
|
89 | (1) |
|
Further Reading and Advanced Topics |
|
|
90 | (3) |
|
|
93 | (32) |
|
Support Vector Classification |
|
|
93 | (19) |
|
The Maximal Margin Classifier |
|
|
94 | (9) |
|
|
103 | (9) |
|
Linear Programming Support Vector Machines |
|
|
112 | (1) |
|
Support Vector Regression |
|
|
112 | (9) |
|
ε-Insensitive Loss Regression |
|
|
114 | (4) |
|
|
118 | (2) |
|
|
120 | (1) |
|
|
121 | (1) |
|
|
121 | (1) |
|
Further Reading and Advanced Topics |
|
|
122 | (3) |
|
Implementation Techniques |
|
|
125 | (24) |
|
|
125 | (4) |
|
The Native Solution: Gradient Ascent |
|
|
129 | (6) |
|
General Techniques and Packages |
|
|
135 | (1) |
|
Chunking and Decomposition |
|
|
136 | (1) |
|
Sequential Minimal Optimisation (SMO) |
|
|
137 | (7) |
|
Analytical Solution for Two Points |
|
|
138 | (2) |
|
|
140 | (4) |
|
Techniques for Gaussian Processes |
|
|
144 | (1) |
|
|
145 | (1) |
|
Further Reading and Advanced Topics |
|
|
146 | (3) |
|
Applications of Support Vector Machines |
|
|
149 | (13) |
|
|
150 | (2) |
|
A Kernel from IR Applied to Information Filtering |
|
|
150 | (2) |
|
|
152 | (4) |
|
Aspect Independent Classification |
|
|
153 | (1) |
|
Colour-Based Classification |
|
|
154 | (2) |
|
Hand-written Digit Recognition |
|
|
156 | (1) |
|
|
157 | (3) |
|
Protein Homology Detection |
|
|
157 | (2) |
|
|
159 | (1) |
|
Further Reading and Advanced Topics |
|
|
160 | (2) |
A Pseudocode for the SMO Algorithm |
|
162 | (3) |
B Background Mathematics |
|
165 | (8) |
|
|
165 | (2) |
|
|
167 | (2) |
|
|
169 | (2) |
|
Operators, Eigenvalues and Eigenvectors |
|
|
171 | (2) |
References |
|
173 | (14) |
Index |
|
187 | |