Day 22 of 100 Completed Today continued with linked list fundamentals and took the first step into actual data analysis. • #876 - Middle of the Linked List (Easy) - solved • Started basics of EDA (Exploratory Data Analysis) 🔎 Focus Areas • Fast and slow pointer technique • Efficient traversal without extra space • Understanding the purpose of EDA in data workflows 💡 Key Takeaways (DSA) 📌 #876 Middle of the Linked List This is a classic pattern: use two pointers (slow and fast) slow moves 1 step, fast moves 2 steps when fast reaches the end, slow is at the middle Clean, efficient, and shows how smart traversal beats brute force. 🚀 Python + EDA Started basic Exploratory Data Analysis. This is where all the libraries finally start connecting. 💡 Key Takeaways (Python) • EDA is about understanding data before doing anything with it • Looking at distributions, missing values, and patterns • Visualization tools now actually have a purpose, not just syntax practice ⚡ Honest Reflection This was a meaningful shift. DSA is continuing steadily, but starting EDA makes things feel more real-world. Still early in EDA, so understanding is basic. Need to go deeper and work with actual datasets. Linked list patterns are becoming more intuitive now, which is a good sign. Consistency is strong. Direction is getting clearer. Patterns recognized: Fast-Slow Pointers | Linked List Traversal | Space Optimization | Data Understanding | EDA Basics #100DaysOfCode #DSA #Python #EDA #LinkedList #LeetCode #BuildInPublic #CodingJourney #Consistency
Linked List Fundamentals and EDA Basics
More Relevant Posts
-
Most pandas mistakes don’t come from complex logic. They come from confusing Series vs DataFrame. 🔹 Series = 1D (single column) 🔹 DataFrame = 2D (table of columns) Example: df['salary'] # Series (1D) df[['salary']] # DataFrame (2D) Why it matters: df['salary'].mean() # scalar df[['salary']].mean() # Series output Another common trap: df.mean() # column-wise df.mean(axis=1) # row-wise Real bug people make: df[df['salary'] > 100]['salary'] = 200 # ❌ SettingWithCopyWarning Fix: df.loc[df['salary'] > 100, 'salary'] = 200 Rule of thumb: Use Series → transformations Use DataFrame → analysis Small confusion → wrong output → bad business decisions. What’s a pandas bug that cost you time? #DataScience #Python #Pandas #DataAnalytics #MachineLearning #SQL #DataEngineering
To view or add a comment, sign in
-
-
“How do you actually deal with messy data in real projects?” Because the truth is most datasets are far from perfect. In one of my projects, I worked with thousands of records coming from different sources with missing values, inconsistent formats, duplicate entries… the usual chaos. At first, it felt overwhelming. But over time, I started following a simple approach: 1️⃣ Understand the data before touching it Instead of jumping into coding, I explore patterns, gaps, and inconsistencies. 2️⃣ Clean in layers, not all at once Handling missing values, standardizing formats, and removing duplicates step by step makes the process manageable. 3️⃣ Validate everything Even small errors can lead to wrong insights, so I always cross-check key metrics. 4️⃣ Automate what repeats If a task is done more than twice, it’s worth automating (Python/SQL saves a lot of time here). What I’ve learned is this: 👉 Data cleaning isn’t the “boring part” of analysis, it’s where most of the real work happens. A good model or dashboard is only as good as the data behind it. Curious to know what’s the messiest dataset you’ve worked with? #DataAnalytics #Python #SQL #DataCleaning #DataScience #Analytics
To view or add a comment, sign in
-
-
Today I explored User Defined Functions (UDF) in Python for Data Analysis 📊🐍 While working with datasets, I realized that sometimes built-in functions are not enough. That’s where UDFs become powerful. They allow us to create custom logic based on business requirements. Example: def profit_category(x): if x > 1000: return "High" elif x > 500: return "Medium" else: return "Low" df["Profit_Category"] = df["Profit"].apply(profit_category) 💡 What this helps with: • Custom data transformations • Business rule implementation • Feature engineering Real insight: In real-world analytics, the ability to create custom logic is what separates 👉 basic users from actual Data Analysts Step by step focusing on building practical and problem-solving skills in Data Analytics 🚀 #Python #Pandas #DataAnalytics #UDF #LearningJourney
To view or add a comment, sign in
-
Week 11: The Data Immersed Python Cohort This week was all about Exploratory Data Analysis (EDA), bringing together data cleaning, visualization, and analytical thinking into one structured workflow. I analyzed a supply chain dataset to: • Understand distributions of key variables (cost, quantity, delivery time) • Explore relationships between operational factors • Evaluate supplier performance and reliability • Detect anomalies and potential data quality issues EDA isn’t just about charts; it’s about asking the right questions and translating data into actionable insights. lLink to EDA Report: https://lnkd.in/effmSsYv Anne Nnamani TheData Immersed(TDI)
To view or add a comment, sign in
-
Stop wasting time on repetitive syntax. 🛑 When you’re in the middle of a data quality audit, the last thing you want to do is break your flow to look up how to fill a null or drop a duplicate. I’ve mapped out my "no-fluff" Pandas toolkit for Data Analysts. These aren't just functions, they are the exact commands I use daily to ensure data integrity at scale. Inside this guide: ✅ Inspection: Quick stats & null counts. ✅ Cleaning: Handling nulls & deduplication. ✅ Filtering: Advanced multi-condition logic. ✅ Aggregation: Summaries that stakeholders actually care about. Pro-tip: Don't just save it- apply it. Use the df.info() and df.duplicated() combo on your next raw dataset to spot red flags instantly. What’s your most-used Pandas function for data cleaning? 👇 #Python #Pandas #DataAnalytics #DataQuality #DataGovernance #WomenInData #SQL #BusinessIntelligence
To view or add a comment, sign in
-
-
Pandas Attributes vs Methods 🐼📊 (Explained Simply) If you're learning Pandas, understanding each attribute & method is a game changer 👇 🔸 ATTRIBUTES (Access Information – No parentheses) • shape → Shows the number of rows & columns 👉 Output: (rows, columns) • columns → Returns all column names 👉 Useful to understand dataset structure • index → Displays row labels / indexing • dtypes → Shows data type of each column (int, float, object, etc.) • size → Total number of elements (rows × columns) --- 🔹 METHODS (Perform Actions – Use parentheses) • head() → Displays first 5 rows 👉 Best for quick preview • describe() → Gives statistical summary 👉 mean, min, max, count, etc. • sum() → Calculates total of values • groupby() → Groups data for analysis 👉 Example: category-wise insights • fillna() → Fills missing values 👉 Important for data cleaning --- 📌 Golden Rule: Attributes = Information Methods = Action Master these basics and Pandas becomes much easier 💡 💬 Which one do you use the most in your projects? --- #Python #Pandas #DataAnalytics #DataScience #DataAnalyst #LearningPython #CodingTips #DataCleaning #Analytics
To view or add a comment, sign in
-
-
𝗜 𝗹𝗲𝗮𝗿𝗻𝗲𝗱 𝘀𝗼𝗺𝗲𝘁𝗵𝗶𝗻𝗴 𝘀𝗺𝗮𝗹𝗹 𝗯𝘂𝘁 𝘃𝗲𝗿𝘆 𝗽𝗼𝘄𝗲𝗿𝗳𝘂𝗹 𝘄𝗵𝗶𝗹𝗲 𝘄𝗼𝗿𝗸𝗶𝗻𝗴 𝘄𝗶𝘁𝗵 𝗣𝗮𝗻𝗱𝗮𝘀 𝗺𝗲𝗿𝗴𝗲𝘀 — 𝘂𝘀𝗶𝗻𝗴 𝗶𝗻𝗱𝗶𝗰𝗮𝘁𝗼𝗿=𝗧𝗿𝘂𝗲 At first, I used to merge DataFrames and just trust the result. If the output looked right, I would move on. But many times, hidden issues were there missing matches, unexpected duplicates, or extra rows. Then I discovered the indicator=True parameter. When you use it in a merge, Pandas adds a new column called "_merge". This column tells you exactly where each row came from: * "left_only" → present only in the left DataFrame * "right_only" → present only in the right DataFrame * "both" → matched in both This one column completely changed how I debug merges. Instead of guessing, I can now clearly see: * Which records didn’t match * If my join keys are correct * Whether I’m losing or gaining data unexpectedly For example, after a merge, I just do a quick check: df['_merge'].value_counts() In seconds, I know if something is wrong. This is especially useful in real-world data pipelines where data is messy and assumptions often fail. It’s a small trick, but it gives a lot of confidence in your data. #DataScience #Python #Pandas #DataEngineering #DataAnalytics
To view or add a comment, sign in
-
-
𝗗𝗮𝘁𝗮 𝗣𝗿𝗼𝗯𝗹𝗲𝗺𝘀 𝗢𝗳𝘁𝗲𝗻 𝗟𝗼𝗼𝗸 𝗧𝗲𝗰𝗵𝗻𝗶𝗰𝗮𝗹… 𝗕𝘂𝘁 𝗔𝗿𝗲𝗻’𝘁 At first, I thought most issues came from tools. SQL. Python. Dashboards. But over time, I noticed something different. Most problems come from: • Unclear definitions • Inconsistent processes • Lack of ownership • Poor communication The technical part is important. But it’s rarely the real problem. 〰️Kosi Clean data. Clear decisions. Better healthcare
To view or add a comment, sign in
-
-
Day 19 — Merging & Joining Data in Pandas As I continue deepening my understanding of pandas, today’s focus was on something very practical: combining datasets. In real-world scenarios, data rarely comes in a single clean table. You often have multiple datasets that need to be brought together before any meaningful analysis can happen. That’s where pandas functions like merge(), join(), and concat() come in. Here’s a quick breakdown of what I learned: 🔹 merge() This is similar to SQL joins. It allows you to combine datasets based on a common column. You can perform: Inner joins Left joins Right joins Outer joins Example: pd.merge(df1, df2, on="id", how="inner") 🔹 join() Used mainly for combining DataFrames based on their index. It’s a bit more concise when working with indexed data. 🔹 concat() Used to stack DataFrames either: Vertically (adding more rows) Horizontally (adding more columns) Example: pd.concat([df1, df2], axis=0) 💡 Key Insight: Understanding when to use each method is crucial. Use merge() when working with relational data Use concat() when stacking data Use join() for index-based alignment This concept is especially important in data cleaning and preprocessing, where datasets often come from different sources. Each day, pandas feels less like a tool and more like a language for working with data. #M4aceLearningChallenge #Day19 #DataScience #MachineLearning #Python #Pandas #DataAnalysis
To view or add a comment, sign in
-
🚀 Day 13 of My Pandas Journey - GroupBy + SQL-Style Joins Today I explored one of the most important concepts in Pandas - GroupBy, Aggregation, Merging & Joining. ✅ Learned: groupby() with single & multiple columns Aggregation using sum, mean, count, nunique Advanced agg() operations Custom functions with apply() Group-wise ranking & normalization DataFrame concatenation using pd.concat() SQL-style joins in Pandas: Inner Join Left Join Right Join Outer Join left_on & right_on np.intersect1d() and np.setdiff1d() 📌 Practiced data analysis concepts using sample sales, customer, and order datasets to better understand how real-world datasets are handled. 💡 Biggest realization today: Pandas feels like a perfect blend of Python + SQL for data analysis and data manipulation. Step by step, understanding how data is transformed, grouped, connected, and analyzed in real workflows. 📈 Github:- https://lnkd.in/g5qwr5Eu #Python #Pandas #SQL #DataAnalysis #DataEngineering #MachineLearning #DataScience #CodingJourney #LearnPython #Analytics
To view or add a comment, sign in
Explore content categories
- Career
- Productivity
- Finance
- Soft Skills & Emotional Intelligence
- Project Management
- Education
- Technology
- Leadership
- Ecommerce
- User Experience
- Recruitment & HR
- Customer Experience
- Real Estate
- Marketing
- Sales
- Retail & Merchandising
- Science
- Supply Chain Management
- Future Of Work
- Consulting
- Writing
- Economics
- Artificial Intelligence
- Employee Experience
- Workplace Trends
- Fundraising
- Networking
- Corporate Social Responsibility
- Negotiation
- Communication
- Engineering
- Hospitality & Tourism
- Business Strategy
- Change Management
- Organizational Culture
- Design
- Innovation
- Event Planning
- Training & Development