Day 7 / ∞ — Logistic Regression with Scikit-Learn Today's lab was all about classification basics: fitting a logistic regression model, making predictions, and calculating accuracy — all in just a few lines of Python. What stood out → scikit-learn abstracts away the math, but understanding what's happening under the hood (sigmoid function, decision boundaries) makes you a much better practitioner. The workflow is deceptively simple: → Prepare your feature matrix and labels → Fit the model → Predict and evaluate 100% accuracy on the training set sounds great until you remember that's 6 data points. Overfitting awareness starts early. One week in. The fundamentals are clicking. #MachineLearning #LogisticRegression #ScikitLearn #100DaysOfML
Logistic Regression with Scikit-Learn Basics
More Relevant Posts
-
🚀 Day 04 of My Machine Learning Journey: NumPy Data Types (dtypes) Today, I learned about NumPy data types (dtypes), which define the type of elements stored in an array. I explored: ✅ Different types like int, float, and bool ✅ How NumPy uses fixed data types for better performance ✅ Why choosing the right dtype helps optimize memory usage Understanding dtypes helps write more efficient and faster code — an important step for Machine Learning. 💡 #MachineLearning #NumPy #Python #LearningJourney #Day04
To view or add a comment, sign in
-
-
Excited to share my latest project: LinearRegression-ML This is a beginner-friendly Machine Learning project focused on understanding and implementing Linear Regression from scratch. It includes practical notebooks like profit analysis and medical data predictions, along with clear explanations of loss and cost functions. ???What I learned =>Fundamentals of Linear Regression =>Cost & loss function implementation =>Real-world dataset analysis using Python #https://lnkd.in/guCQQdNe #MachineLearning #Python_Jupyter_Notebook #DataScience
To view or add a comment, sign in
-
-
📊 Understanding Joint Distributions in Probability Ever wondered how to model the relationship between two random variables? A joint distribution is the key! It describes the probability of two (or more) events happening simultaneously, giving us a complete picture of their interaction. In my latest Python experiment, I created a simple joint distribution table for two discrete variables, X and Y, representing the number of heads and tails in two coin flips. Here’s what I learned: Joint distribution tells us the probability of both X and Y taking specific values. Marginal distributions help us understand each variable independently. Conditional distributions show how one variable behaves given a specific value of the other. This concept is foundational in statistics, machine learning, and data science. It’s amazing how much insight we can gain from just a few lines of code! 🔗 Check out the code snippet in the comments if you’re curious to try it yourself. #Probability #Statistics #DataScience #Python #MachineLearning #Coding
To view or add a comment, sign in
-
Just completed the Gradient Descent lab in Andrew Ng's ML Specialization — and it genuinely clicked for me here. The concept: instead of guessing the best values for w and b in a linear model, gradient descent finds them automatically by repeatedly moving in the direction that reduces error. What I built from scratch in Python: ✅ compute_cost() — measures how wrong the model is ✅ compute_gradient() — calculates which direction to move ✅ gradient_descent() — runs 10,000 iterations to find optimal parameters What surprised me most: → Starting from w=0, b=0, the algorithm found w≈200, b≈100 for a house price dataset → The cost dropped rapidly at first, then slowed as it approached the minimum — exactly like rolling a ball to the bottom of a bowl → Setting the learning rate too high (α = 0.8) caused the model to completely diverge — cost shot up instead of down That last point was the most valuable. Seeing divergence visually made the theory real. Building these functions line by line beats reading about them any day. #MachineLearning #Python #AndrewNg #LearningInPublic #DataScience #GradientDescent
To view or add a comment, sign in
-
-
Days 68-69 of the #three90challenge 📊 Today I explored NumPy operations — specifically indexing and slicing arrays. After understanding NumPy basics, this step made it easier to access and manipulate data efficiently. What I practiced today: • Accessing elements using indexing • Extracting subsets of data using slicing • Working with multi-dimensional arrays • Performing operations on selected data Example thinking: Instead of looping through data manually, I can directly select and operate on specific parts of an array. Example: import numpy as np arr = np.array([10, 20, 30, 40, 50]) print(arr[1:4]) # Output: [20 30 40] This makes data manipulation faster and more intuitive. From handling data → to controlling it efficiently 🚀 GeeksforGeeks #three90challenge #commitwithgfg #Python #NumPy #DataAnalytics #LearningInPublic #Consistency #Upskilling
To view or add a comment, sign in
-
Day 72. Spent time going deeper into XGBoost today. Covered classification and worked through the math: gradients & hessian leaf weights similarity score & gain Some questions I tried to answer while learning: Why do we need Taylor expansion here? Why can’t we directly differentiate the objective? What makes decision trees non-smooth / non-differentiable? The key realization: since trees produce piecewise constant outputs, the loss surface isn’t smooth — which is why second-order approximation becomes necessary. Still revising, but things are starting to connect. Notes: https://lnkd.in/gCqHUeK9 #MachineLearning #XGBoost #LearningInPublic #Python #DataScience
To view or add a comment, sign in
-
I trained a model last week. Got 97% accuracy. Showed it to the team. Everyone was thrilled. Then I changed one line of code. The accuracy dropped to 61%. Same model. Same data. Same algorithm. The only thing that changed? The random seed. I've spent the last few weeks building something to explain exactly why this happens — and how to make sure it never happens to you. Stay tuned. Spoiler: Had I used #skore library from :probabl. along with scikit-learn, this would have never happened.... https://lnkd.in/eMmwpj8a #skore #datascience #mlops #dataanalytics #crossvalidation #python #sklearn
To view or add a comment, sign in
-
-
Today, I focused on working with NumPy arrays. Building a solid foundation for data manipulation and analysis. Here’s what I practiced: 🔹 Created a 1D array with values from 1 to 15 🔹 Built a 2D array (3×4) filled with ones 🔹 Generated a 3×3 identity matrix 🔹 Explored key array properties like shape, type, and dimensions 🔹 Converted a regular Python list into a NumPy array This session helped me better understand how data is structured and handled in numerical computing. Getting comfortable with arrays is definitely a crucial step toward more advanced data analysis and machine learning tasks. Looking forward to building on this momentum 💡 #AI #MachineLearning #Python #NumPy #DataAnalysis #M4ACE
To view or add a comment, sign in
-
-
📊 Not everything in data science is a finished project most of it is exploration. This is a small snapshot from my Jupyter Notebook while working through a project. At this stage, it’s not about perfect results it’s about: • Understanding the data • Trying different approaches • Visualizing patterns • Making sense of what’s happening underneath What looks like simple code on the screen is actually a process of trial, error, and discovery. 💡 Key takeaway: Before insights come confusion. Before clarity comes experimentation. Every notebook is just a record of how thinking evolves through data. #DataScience #Python #JupyterNotebook #DataAnalytics #LearningInPublic
To view or add a comment, sign in
-
-
📊 Day 6 | K-Nearest Neighbors (KNN) 🤝📍 Today, I learned about K-Nearest Neighbors (KNN), a simple and intuitive Machine Learning algorithm. KNN works on the idea of distance — it classifies a data point based on the majority class of its nearest neighbors. 📌 In simple terms: “Similar data points are close to each other.” Example: ✔ Recommending products ✔ Classifying customers To understand this, I implemented KNN using Python and observed how it predicts based on nearby data points 💻 KNN is simple but powerful for many classification problems. #MachineLearning #KNN #DataScience #LearningInPublic #Python
To view or add a comment, sign in
-
More from this author
Explore related topics
Explore content categories
- Career
- Productivity
- Finance
- Soft Skills & Emotional Intelligence
- Project Management
- Education
- Technology
- Leadership
- Ecommerce
- User Experience
- Recruitment & HR
- Customer Experience
- Real Estate
- Marketing
- Sales
- Retail & Merchandising
- Science
- Supply Chain Management
- Future Of Work
- Consulting
- Writing
- Economics
- Artificial Intelligence
- Employee Experience
- Workplace Trends
- Fundraising
- Networking
- Corporate Social Responsibility
- Negotiation
- Communication
- Engineering
- Hospitality & Tourism
- Business Strategy
- Change Management
- Organizational Culture
- Design
- Innovation
- Event Planning
- Training & Development