From the course: Data Preparation, Feature Engineering, and Augmentation for AI Models
Unlock this course with a free trial
Join today to access over 25,500 courses taught by industry experts.
Feature engineering: Scaling and normalizing data
From the course: Data Preparation, Feature Engineering, and Augmentation for AI Models
Feature engineering: Scaling and normalizing data
- [Instructor] Sometimes we need to scale and normalize our data. This is because features with different ranges can distort model training and the predictions those models make. Now, many machine learning algorithms rely on distance calculations, so that requires that we work with comparable scales. And this is because we want to prevent large-magnitude features from really dominating the decisions that are made. Consider, for example, house sales. Now, the price of a house sale might be in the order of hundreds of thousands of dollars. That's one feature. Other features might be the number of bedrooms, which might be three, four, five, or so on. So the scales can really vary widely between different features. In addition, some algorithms will converge faster when we work with standardized and normalized data. And also, regularization techniques, which we use to prevent overfitting, work better when features are scaled. There are several coming scaling techniques. Min-max scaling…
Contents
-
-
-
(Locked)
Data exploration and initial quality assessment4m 49s
-
Detecting and managing missing data5m 13s
-
(Locked)
Detecting and managing outliers3m
-
(Locked)
Challenge: Assess data quality of a dataset18s
-
(Locked)
Solution: Assess data quality of a dataset23s
-
(Locked)
Feature engineering: Scaling and normalizing data4m 47s
-
(Locked)
Feature engineering: Categorical encodings4m 8s
-
(Locked)
Challenge: Apply feature engineering to a dataset18s
-
(Locked)
Solution: Apply feature engineering to a dataset16s
-
(Locked)
-
-
-
-
-
-