**Linear kernel**or

**Gaussian kernel**when working with

**Support Vector Machine (SVM)**. Please feel free to comment/suggest if I missed to mention one or more important points. Also, sorry for the typos.

Following are the key points described later in this article:

- When to Use Linear Kernel
- When to Use Gaussian Kernel

###### When to Use Linear Kernel

In case there are **large number of features and comparatively smaller number of training examples**, one would want to use linear kernel. As a matter of fact, it can also be called as **SVM with No Kernel**. One may recall that SVM with no kernel acts pretty much like **logistic regression** model where following holds true:

- Predict Y = 1 when W.X >= 0. Note that, in the prior equation, W is actually W transpose and also includes bias factor.
- Predict Y = 0 when W.X < 0.

Simply speaking, one may want to use SVM with linear kernel when data distribution is linearly separable.

###### When to Use Gaussian Kernel

In scenarios, where there are **smaller number of features and large number of training examples**, one may use what is called Gaussian Kernel. When working with Gaussian kernel, one may need to choose the value of variance (sigma square). The selection of variance would determine the bias-variance trade-offs. Higher value of variance would result in High bias, low variance classifier and, lower value of variance would result in low bias/high variance classifier.

- Beta Distribution Explained with Python Examples - September 24, 2020
- Bernoulli Distribution Explained with PythonExamples - September 23, 2020
- K-Nearest Neighbors Explained with Python Examples - September 22, 2020