Customer Churn Prediction with ML Pipeline Improvements

Pipelines in ML really change how you build models!! So I rebuilt my Customer Churn Prediction project — this time using a proper ML pipeline. 🔧 What I improved: • Built an end-to-end Pipeline using ColumnTransformer • Switched from train-test split to 5-Fold Cross Validation • Removed unnecessary feature selection (Chi-Square) • Handled class imbalance using F1-score & class_weight • Tuned models like Random Forest & XGBoost 📊 Key Results (F1 Score): • Logistic Regression → ~0.62 • Decision Tree → ~0.60 • Random Forest → ⭐ ~0.63 • XGBoost → ~0.56 💡 Key Learnings: • My earlier results were slightly optimistic due to a single train-test split • Cross-validation gave me more honest and stable performance • Random Forest performed best → indicating non-linear patterns • Logistic Regression performed almost as well → dataset isn’t highly complex • XGBoost underperformed → showing advanced models need proper tuning Check out both codes here:- https://lnkd.in/gV5Cb5iC This project helped me move from “just building models” to actually understanding how ML systems should be structured and evaluated in practice. Would love to hear your feedback or suggestions! #MachineLearning #DataScience #Python #ScikitLearn #XGBoost #Analytics #LearningJourney

  • graphical user interface, text

To view or add a comment, sign in

Explore content categories