- In the previous article, we discussed Linear Regression, Polynomial Regression, and Logistic Regression.
- Today, we continue with a brief review of different supervised learning techniques.
- The simplest version of decision trees is something people use regularly in their daily lives.
- For example, deciding whether to purchase a specific car: If the car is broken — No. Otherwise, if the year is < 2020 — No. Otherwise...
- A Decision Tree is a tree-like model that makes decisions based on feature values. It splits data into branches based on conditions, eventually leading to predictions.
- Can be used for Classification (customer segmentation, recommendation systems) and Regression (predicting house prices, predicting employee salaries).
- Decision Trees are a great tool, but their usage is limited.
- Smart researchers found that, instead of using just one tree, it makes sense to use multiple decision trees and combine them to improve prediction accuracy.
- And so, the Random Forests method was created.
- In some cases, Random Forests can work better and faster than neural networks.
- This approach can be used in finance, healthcare, and bioinformatics.
- Primarily used for classification, but can also handle regression tasks.
- SVM acts like a smart "boundary drawer."
- It finds the best line (or boundary) to separate data points into different groups (classes).
- For example, if you have data on cats and dogs, SVM will try to draw a line that separates all the cats on one side and all the dogs on the other.
- Example use cases include text classification and image recognition.
- In k-NN, each data point is classified based on its nearest "neighbors" in the data.
- Imagine you have a new fruit, and you want to classify it as an apple or an orange based on its size and color. You measure the size and color of the nearest fruits (neighbors) around it. If most of them are apples, you classify the new fruit as an apple. If most are oranges, you label it an orange.
- k-NN is useful when you have well-separated data.
- Can be used for pattern recognition, recommendation systems.
- Imagine you’re learning a new skill, like cooking. The first time you try a recipe, it’s not perfect, but you learn from your mistakes. The next time, you fix some of those mistakes, getting a bit closer to a perfect dish. Over time, with each attempt, you keep getting better by learning from what went wrong before.
- GBM works similarly by "boosting" its accuracy in stages, learning from its errors at each step.
- GBM builds a series of simple models, often decision trees. Each weak learner on its own isn’t very accurate, but together, they build a stronger, more accurate model.
- Each new model focuses on correcting the mistakes made by the previous model.
- Can be used in predictive analytics, financial modeling.
- Some neural networks are also part of supervised learning. However, we don’t plan to discuss them in this article. Neural networks will be covered in a separate article.