Support Vector Machine (SVM) :
You can see that in below diagram, there are two different classes of data( Blue and Red). You can separate them with several separation line such as A,B,C, and D. But there must be single separating line of hyper plane that classifies the two classes best. So, how to choose the best classifier?
Good vs Bad Classifier ?Data can be separated by a line or a simple hyper plane is known as linearly separable data. These hyper planes that can linearly separate the data are known as linear classifier.
If the training data is linearly separable, we can obtain two parallel hyper planes that separate the two classes of data, so that the distance between them is maximized. The region bounded by these two hyper planes is called the 'margin' and the maximum margin hyper plane is the hyper plane that lies half way between them.
We need to remember that a classifier is not worth when two classes seem to run too close to each other and when we run this same classifier through test data, there are high chances of misclassification. So the objective is to maximizes the margin, minimizes the risk involved in it and thus generalized it well on test data.
If you look at the data points below, you can't use linear classifier. To solve this we use non-linear classifier.
The non-linear classifier works well to classify them and more importantly without any misclassification.
Comments
Post a Comment