In this section, we'll discuss the powerful linear classifier, Support Vector Machines. So the learning goals for this section will include gaining an understanding to the support vector machine approach to classification. Seeing how support vector machines compare with the logistic regression model that we introduced earlier in this course. We'll go over the cost function used to optimize our support vector machines. As we do with every model, we'll learn how to properly regularize to ensure that we find the right balance between bias and variance. Now, I'd like to start off by recalling what we learned earlier in the course in regards to logistic regression. The idea here was to transform our loss function for our linear regression so that we weigh the far-away points much less, so that our separation hyperplane does not try too hard to get those right. That was a measure taken in order not to be skewed by the outliers in our dataset. Now, let's try from there to build an intuition behind the support vector machine. With a support vector machine, here we're just looking at one dimension, that phone usage. We want to find the optimal point in the phone usage, in which we can split our data and create a decision boundary. Now, we see that here and with all values above this line, we are going to predict that we would not churn, and all values below this line, we'd predict that they did churn. Here we clearly see that is not a good choice as this specific boundary results in three misclassifications, right? We have three examples where we predict not churned given our decision boundary, when in actuality it should have been predicted churn. Next, we can move to the other gap in our phone usage. We see this gap here, and again, we have a bad misclassification. Because we have to misclassifications, it's a bad choice of decision boundary. We probably want to move somewhere in between. Now, this decision boundary has no misclassifications. But we want to think, is it really the best choice that we can make? In the close neighborhood of this choice and split, there are going to be other choices for our split that will also result in no misclassifications, right? Essentially, anything between those blue and pink dots will result in no misclassifications. So you want to start thinking, should we move this boundary a little bit to the left or a little bit to the right, in order to optimize where this decision boundary lays? We're also going to introduce here this dotted line, which will represent how far we are from the nearest point. So that nearest point here is going to be just to the left of the line, is going to be that blue dot, and we see that we have a very tight dotted lines across either side of our straight line, which is our decision boundary. Then we can think again for this points, we see that we're very close to the other side of our boundary and we end up with a very tight margin as well, with these dotted lines being very close to that decision boundary. So here's an idea. Let's maximize that region between the two classes. Let's draw the boundary at a place where it's equidistant from the closest blue and the closest red class, and this will ultimately be the goal of our support vector machine algorithm. In the next video, we'll continue to expand on this idea on how support vector machines will end up performing classifications on your dataset. See you there.