Next: 2-Norm Soft Margin
Up: Support Vector Machines (SVM)
Previous: Support Vector Machine
When the two classes are not linearly separable, the condition for the
optimal hyper-plane can be relaxed by including an extra term:
For minimum error, should be minimized as well as ,
and the objective function becomes:
Here is a regularization parameter that controls the trade-off between
maximizing the margin and minimizing the training error. Small C tends to
emphasize the margin while ignoring the outliers in the training data, while
large C may tend to overfit the training data.
When , it is called 2-norm soft margin problem:
Note that the condition is dropped, as if , we can
set it to zero and the objective function is further reduced.)
Alternatively, if we let , the problem can be formulated as
This is called 1-norm soft margin problem. The algorithm based on 1-norm
setup, when compared to 2-norm algorithm, is less sensitive to outliers in
training data. When the data is noisy, 1-norm method should be used to
ignore the outliers.
Subsections
Next: 2-Norm Soft Margin
Up: Support Vector Machines (SVM)
Previous: Support Vector Machine
Ruye Wang
2016-08-24