🚢 Titanic Survival Prediction – Machine Learning Project I recently worked on a beginner-level machine learning project using the Titanic dataset to predict passenger survival. In this project, I applied and compared multiple classification algorithms: • Logistic Regression • K-Nearest Neighbors (KNN) • Decision Tree • Naive Bayes 📊 Result: Logistic Regression performed slightly better compared to the other models in terms of accuracy on this dataset. 🔍 What I learned: • Data preprocessing and handling missing values • Feature selection and encoding • Training and evaluating multiple ML models • Comparing model performance This is a learning project as part of my machine learning practice, not a production-level system, but it helped me understand core concepts of classification and model evaluation. 💻 GitHub Project:https://lnkd.in/d9pDV4Dd I’m continuing to improve my skills in Machine Learning and Data Science, and more projects are coming soon. #MachineLearning #DataScience #Python #ScikitLearn #TitanicDataset #LearningJourney
More Relevant Posts
-
45 Days ML Journey — Day 15: Random Forest (Classifier & Regressor) Day 15 of my Machine Learning journey — exploring Random Forest, an ensemble learning technique used for both classification and regression tasks. Tools Used: Scikit-learn, NumPy, Pandas What is Random Forest? Random Forest is a supervised learning algorithm that builds multiple decision trees and combines their outputs to improve accuracy and reduce overfitting. Key concepts: Ensemble Learning : Combines multiple models to make better predictions Decision Trees : Individual models used as building blocks Bagging : Training trees on random subsets of data Feature Randomness : Random subset of features used for splitting RandomForestClassifier vs RandomForestRegressor: RandomForestClassifier : Used for classification tasks (predicting categories) RandomForestRegressor : Used for regression tasks (predicting continuous values) Why use Random Forest? Reduces overfitting compared to a single decision tree Handles large datasets with higher dimensionality Works well with both classification and regression problems Provides feature importance for better interpretability Code notebook: https://lnkd.in/gxsJwSmY Key takeaway: Random Forest leverages the power of multiple trees to deliver more accurate and stable predictions, making it one of the most reliable algorithms in machine learning. #MachineLearning #DataScience #RandomForest #Python #ScikitLearn #LearningInPublic #MLJourney
To view or add a comment, sign in
-
🚀 Excited to share my latest Machine Learning Project! 🌦️ Weather Prediction using Machine Learning In this project, I worked on a real-world weather dataset and performed: 🔹 Data Cleaning & Preprocessing 🔹 Exploratory Data Analysis (EDA) 🔹 Correlation Analysis (Heatmap) 🔹 Built Linear Regression model for temperature prediction 🔹 Developed Logistic Regression model to predict rain 🌧️ 🔹 Evaluated model using accuracy & confusion matrix 📊 Key Learning: Understanding data is more important than just building models. 🔗 GitHub Project: https://lnkd.in/geA8rz8v #MachineLearning #DataScience #Python #BeginnerProject #Learning #AI
To view or add a comment, sign in
-
Developed a Diabetes Prediction System using Machine Learning 🧠 Applied data preprocessing, handled missing values, and trained a Random Forest model to predict diabetes risk with good accuracy. Also implemented visualization techniques and a prediction system for real-time input. 📌 Tech Used: Python, Pandas, NumPy, Scikit-learn, Matplotlib, Seaborn This project helped me understand how ML can be applied in healthcare analytics. #MachineLearning #DataScience #Python #AI #Healthcare #StudentProject
To view or add a comment, sign in
-
🚀 Day 42 of My Data Science & Machine Learning Journey Support Vector Machine (SVM) Implementation 📌 What I focused on today? Instead of just theory, I worked on the implementation side of SVM using Scikit-learn 💻 I also explored how GridSearchCV helps in improving model performance 🔥 📊 What I learned: 🔹 How to train an SVM model using real data 🔹 Difference between kernels (linear vs RBF) 🔹 Importance of hyperparameters like C and gamma 🔹 How GridSearchCV automatically finds the best parameters 🔹 How SVM finds the optimal boundary between classes 🔥 Key Insight: SVM becomes much more powerful when combined with proper hyperparameter tuning instead of manual guessing. 🎥 Sharing a quick screen recording of my implementation (training + best parameters + accuracy) #MachineLearning #DataScience #SVM #Python #ScikitLearn #AI #LearningJourney #GridSearchCV
To view or add a comment, sign in
-
🚀 Day 47 of My Data Science & Machine Learning Journey K-Nearest Neighbors (KNN) Classification 👨💻 Instead of jumping straight into theory, I tried to understand it with a simple idea: 👉 “Your neighbors decide who you are.” Sounds funny, but that’s exactly how KNN works. 📌 What is KNN Classification? It classifies a data point based on the majority class of its nearest neighbors. Example: If most of your nearest neighbors are from Class A → you also belong to Class A ⚙️ How it works: 1️⃣ Choose value of K 2️⃣ Calculate distance (Euclidean) 3️⃣ Find K nearest neighbors 4️⃣ Majority voting → Final class 📊 Key Learnings: ✔ Simple and intuitive algorithm ✔ No training phase (lazy learning) ✔ Works well for small datasets ✔ Sensitive to value of K and scaling ⚠️ Challenges I faced: 🔸 Choosing the right K value 🔸 Understanding how distance impacts results #MachineLearning #DataScience #KNN #Classification #Python #LearningJourney #AI
To view or add a comment, sign in
-
-
Excited to share one of my recent Machine Learning mini projects: Student Grade Prediction using Linear Regression 🚀📊 This project was a part of my journey in Supervised Machine Learning, where I explored how models can learn from data to make predictions. In this project, I built a predictive model to analyze how factors like study hours and gaming hours can influence student academic performance. 🔹 What I used: • Python • Pandas & NumPy • Scikit-learn • Matplotlib 🔹 Project Workflow: • Imported dataset from kaggle and cleaned student dataset • Selected relevant features for prediction • Applied Train-Test Split on dataset • Built a Linear Regression Model to train and test model • Evaluated performance using MAE, MSE, RMSE, and R² Score • Visualized Actual vs Predicted Grades using scatter plots with regression line 🔹 What I Learned: This project helped me understand the ML workflow — from preprocessing data to training, testing, evaluating, and visualizing model results. Projects like these build strong fundamentals for solving real-world problems with data. Moving to next phase of Machine Learning journey. 🚀 #MachineLearning #SupervisedLearning #Python #DataScience #LinearRegression #ArtificialIntelligence #StudentProjects #ScikitLearn #Analytics #LearningByDoing #Tech
To view or add a comment, sign in
-
Maths and statistics aren’t just theory — they’re the backbone of every strong data science decision. From probability to linear algebra, from distributions to hypothesis testing… these are the tools that turn raw data into real insights. I made this quick cheat sheet to revise the fundamentals that actually matter when working on real-world problems. If you’re getting into data science, don’t skip this part. Strong basics = better models, better intuition, and better results. What topic do you find the most challenging in data science? ⸻ Hashtags: #DataScience #MachineLearning #Statistics #Mathematics #DataAnalytics #AI #DeepLearning #LearningInPublic #DataScienceJourney #Python #Analytics #BigData #StudentLife #TechSkills #CareerGrowth
To view or add a comment, sign in
-
-
45 Days ML Journey — Day 16: XGBoost (Classifier & Regressor) Day 16 of my Machine Learning journey — diving into XGBoost, a powerful and efficient gradient boosting algorithm used for both classification and regression tasks. Tools Used: Scikit-learn, NumPy, Pandas, XGBoost What is XGBoost? XGBoost (Extreme Gradient Boosting) is an advanced ensemble learning algorithm that builds models sequentially, where each new model corrects the errors of the previous ones. Key concepts: Boosting : Sequentially improving weak learners Gradient Descent : Minimizing errors using loss functions Decision Trees : Base learners used in boosting Regularization : Prevents overfitting and improves model generalization XGBoost Classifier vs Regressor: XGBClassifier : Used for classification tasks (predicting categories) XGBRegressor : Used for regression tasks (predicting continuous values) Why use XGBoost? High performance and speed compared to many algorithms Handles missing data efficiently Built-in regularization reduces overfitting Widely used in competitions and real-world applications Code notebook: https://lnkd.in/g7iSaTHR Key takeaway: XGBoost is a highly optimized boosting algorithm that delivers strong performance by continuously learning from errors, making it a go-to choice for structured data problems. #MachineLearning #DataScience #XGBoost #Python #ScikitLearn #LearningInPublic #MLJourney
To view or add a comment, sign in
-
𝐉𝐮𝐬𝐭 𝐜𝐨𝐦𝐩𝐥𝐞𝐭𝐞𝐝 𝐨𝐮𝐫 𝐃𝐫𝐲 𝐁𝐞𝐚𝐧𝐬 𝐂𝐥𝐚𝐬𝐬𝐢𝐟𝐢𝐜𝐚𝐭𝐢𝐨𝐧 𝐌𝐋 𝐏𝐫𝐨𝐣𝐞𝐜𝐭 🌱 Collaborated with Taimoor Tahir Satti 𝐃𝐚𝐭𝐚𝐬𝐞𝐭: 13,000+ records | 16 Features | 7 classes 𝐌𝐨𝐝𝐞𝐥 𝐏𝐞𝐫𝐟𝐨𝐫𝐦𝐚𝐧𝐜𝐞: Achieved 93%+ Accuracy, with Precision, Recall, and F1-Score all above 90%, ensuring balanced and reliable predictions across classes. 𝐖𝐡𝐚𝐭 𝐰𝐞 𝐝𝐢𝐝 𝐢𝐧 𝐭𝐡𝐢𝐬 𝐩𝐫𝐨𝐣𝐞𝐜𝐭: ● Exploratory Data Analysis (EDA) ● Outlier Detection & Handling ● SMOTE (handling class imbalance) ● Cross Validation ● Hyperparameter Tuning ● Trained & compared models (SVM, Random Forest, XGBoost) 𝐓𝐞𝐜𝐡 𝐒𝐭𝐚𝐜𝐤: Python, NumPy, Pandas, Matplotlib, Seaborn, Plotly, ydata-profiling, Scikit-learn, XGBoost, Streamlit 𝐏𝐫𝐨𝐣𝐞𝐜𝐭 𝐋𝐢𝐧𝐤𝐬: 🔗 Dataset: https://lnkd.in/dUPSMx_c 🔗 GitHub Repo: https://lnkd.in/dFSJq6zT 🔗 Live App: https://lnkd.in/d-E7kUjX We’ve been learning Machine Learning for around 1–1.5 months, mainly focusing on classical ML, and now moving towards Deep Learning and advanced topics. This is one of our first complete end-to-end + deployed ML projects, and a big step in our journey. Open to feedback and suggestions. #MachineLearning #DataScience #Python #AI #MLProjects #XGBoost #ScikitLearn #Streamlit #EDA #LearningJourney #F1Score #DataAnalytics #DeepLearning
To view or add a comment, sign in
-
🚀 Machine Learning Project: Pokémon Legendary Prediction Excited to share a project where I explored the Ultimate Pokémon Dataset 2025 and built a Machine Learning model to predict whether a Pokémon is Legendary or not. 🔍 Project Highlights: Performed data cleaning and preprocessing Selected relevant numerical features Trained a Random Forest Classifier Evaluated model performance using accuracy 📊 This project showed me how important data quality and preprocessing are in achieving good model performance. Even simple models can perform well with the right data preparation. 🛠 Tech Stack: Python | Pandas | NumPy | Scikit-learn 📁 GitHub Repository: 👉 https://lnkd.in/g2pjUHs3 💡 Next Steps: Apply feature engineering techniques Encode categorical variables instead of removing them Experiment with advanced models like XGBoost This was a great hands-on experience in building a complete machine learning pipeline from raw data to prediction. Fathima Murshida K #MachineLearning #DataScience #Python #AI #Kaggle #Projects #LearningJourney
To view or add a comment, sign in
Explore related topics
Explore content categories
- Career
- Productivity
- Finance
- Soft Skills & Emotional Intelligence
- Project Management
- Education
- Technology
- Leadership
- Ecommerce
- User Experience
- Recruitment & HR
- Customer Experience
- Real Estate
- Marketing
- Sales
- Retail & Merchandising
- Science
- Supply Chain Management
- Future Of Work
- Consulting
- Writing
- Economics
- Artificial Intelligence
- Employee Experience
- Workplace Trends
- Fundraising
- Networking
- Corporate Social Responsibility
- Negotiation
- Communication
- Engineering
- Hospitality & Tourism
- Business Strategy
- Change Management
- Organizational Culture
- Design
- Innovation
- Event Planning
- Training & Development
Wow, more grease to your elbow bro 💪💪💪