Day 24 of 100 Completed Today reinforced cycle detection patterns and continued working with real-world data through EDA. • #141 - Linked List Cycle (Easy) - solved • Continued EDA on dataset 🔎 Focus Areas • Fast-slow pointer technique for cycle detection • Recognizing repeated patterns across different problem types • Going deeper into data understanding and cleaning 💡 Key Takeaways (DSA) 📌 #141 Linked List Cycle This is a classic application of Floyd’s Cycle Detection: use slow and fast pointers if they meet → cycle exists no extra space needed, efficient and elegant Key insight: cycle detection isn’t limited to numbers - it applies to linked structures as well. 🚀 Python + EDA Continued working on EDA and exploring the dataset further. 💡 Key Takeaways (Python) • Better understanding of missing values and distributions • More confidence in using Pandas for exploration • Visualization is helping uncover patterns in data ⚡ Honest Reflection This was a steady day. Not very difficult, but important for reinforcing patterns. Cycle detection is now clearly a recurring concept across problems, which makes it easier to recognize. EDA still needs depth, especially in drawing meaningful insights instead of just running operations. Consistency is holding. Progress is gradual but real. Patterns recognized: Fast-Slow Pointers | Cycle Detection | Linked Lists | Data Cleaning | EDA | Pattern Recognition #100DaysOfCode #DSA #Python #EDA #LinkedList #LeetCode #BuildInPublic #CodingJourney #Consistency
Reinforced cycle detection patterns with linked lists and EDA
More Relevant Posts
-
Combining data from multiple sources is one of the most common tasks in data analysis and data engineering and in pandas, pd.concat() is the primary tool for getting it done. But there is more to it than just passing two DataFrames and getting one back. Understanding when to use axis=0 vs axis=1, how join handles mismatched columns, why concatenating inside a loop is a performance trap, and when to use concat vs merge. These are the details that separate clean, efficient data pipelines from slow, buggy ones. Get comfortable with pd.concat() and combining data from multiple sources becomes one of the fastest steps in your workflow. Read the full post here: https://lnkd.in/es7KJ7Y9 #Python #Pandas #DataScience #DataEngineering #Analytics #ETL
To view or add a comment, sign in
-
Raw data doesn’t become useful because you visualise it – it becomes useful because you model it properly. SQL for shaping logic. Python for cleaning and exploration. dbt for turning transformations into reliable, version-controlled data products. And GitHub is where all of it stops being “analysis” and starts becoming engineering. That’s the shift: from writing queries to building systems.
To view or add a comment, sign in
-
🚀 Stop looping through your DataFrames! I recently refactored a script processing 10 million rows. We were using a standard row-wise loop, which was choking our CI/CD pipeline and causing memory spikes. Before optimisation: for i, row in df.iterrows(): df.at[i, 'tax_total'] = row['price'] * 1.08 if row['state'] == 'NY' else row['price'] After optimisation: import numpy as np conditions = [df['state'] == 'NY'] choices = [df['price'] * 1.08] df['tax_total'] = np.select(conditions, choices, default=df['price']) Performance gain: 45x faster and 90% lower memory usage. By moving from row-wise iteration to NumPy’s vectorized selection, we eliminated the Python-level overhead entirely. The code is not only faster but cleaner and more readable for the rest of the team. Vectorization turns O(n) Python operations into high-performance C-level loops. It’s the single biggest quick win you can apply to any data pipeline. Have you ever seen a loop-heavy process that you successfully migrated to vectorized operations? #DataEngineering #Python #Pandas #PerformanceTuning #CodingTips
To view or add a comment, sign in
-
Outliers are one of the most misunderstood concepts in data analysis. Many analysts treat them as problems to be removed. But outliers can be data errors, extreme but valid values, or the most important signals in your entire dataset like a fraudulent transaction or a manufacturing defect. The right approach is never automatic. It requires understanding your data, your domain, and the impact of every decision you make. Master outlier detection and more importantly, master the judgment of knowing what to do with what you find. Read the full post here: https://lnkd.in/eQNyw8xG #DataScience #DataAnalysis #Python #MachineLearning #EDA #DataEngineering
To view or add a comment, sign in
-
The pipeline is live. The data is accumulating. Instead of overwriting state, it captures every meaningful change over time, allowing repository activity to be analyzed as a sequence rather than a snapshot. Core design decision: versioned state over snapshot storage All changes are tracked using SCD Type 2 modeling in dbt, preserving full historical state of repository attributes. This enables questions such as: -how repository popularity evolves over time -when growth begins or slows down -what distinguishes sustained momentum from short-term spikes Stack: Python · Prefect · Postgres (Supabase) · dbt · Streamlit The value isn’t in ingesting data, it’s in what becomes possible once data is treated as a record of change rather than a static snapshot. Live dashboard: https://lnkd.in/gU77tVF9
To view or add a comment, sign in
-
-
I used to think I was doing EDA the right way… Until I realized I was making some serious mistakes 😓 Here are the biggest EDA mistakes I made (and most beginners still do): ❌ Jumping to visualization without understanding data ❌ Ignoring missing values ❌ Not checking data types properly ❌ Trusting .describe() blindly ❌ Skipping outlier detection ❌ Creating too many useless charts ❌ Not asking “why” behind the data The truth is… EDA is not about making charts. It’s about understanding your data deeply. Now my approach is simple: 👉 First understand → Then visualize → Then analyze That one shift changed everything ⚡ If you're learning data analytics, Avoid these mistakes early… and you’ll grow 10x faster 🚀 #DataAnalytics #Python #EDA #DataScience #LearningInPublic #AnalyticsTips
To view or add a comment, sign in
-
Some problems are not about moving elements, but about understanding structure. Day 22/100 — Data Structures & Algorithms Journey Today’s Problem: Rotate List This problem helped me understand how linked lists behave when we manipulate their structure efficiently. Approach: Instead of rotating the list step by step (which is inefficient), I first calculated the length of the list. Then, I connected the tail to the head to form a circular linked list. By finding the correct breaking point using k % length, I was able to determine the new head and tail of the rotated list. At each step: Convert list into a circular structure Find the new tail position Break the circle to form the rotated list Key Takeaways: Understanding structure is more important than brute force Modulo operation helps avoid unnecessary rotations Linked list problems often become easier when visualized as cycles Efficient thinking leads to cleaner and faster solutions This problem strengthened my understanding of linked list manipulation and optimization techniques. #DSA #LeetCode #LinkedList #ProblemSolving #SoftwareEngineering #CodingJourney #100DaysOfCode #TechLearning #DeveloperJourney #Programming #Python #InterviewPreparation #CodingSkills #ComputerScience #FutureEngineer #TechCareers #SoftwareDeveloper #LearnInPublic #OpenToWork
To view or add a comment, sign in
-
-
One habit I’ve started building when working with data: Before writing any logic, I always run: df.head() df.info() df.describe() It sounds obvious. But early on, I skipped this step. I would immediately start writing transformations. And later realize things like: columns were strings instead of numbers values had unexpected formats missing data existed where I didn’t expect it Now I try to slow down and understand the data first. It saves a surprising amount of time later. 💡 Data engineering lesson I’m learning: Understanding the data is often more important than writing the code. #DataEngineering #Python #Pandas
To view or add a comment, sign in
-
Built an Event Scheduler Using Heaps and Hash Tables in Python Hi everyone, this week I implemented an Event Scheduler system focusing on algorithm efficiency and scalable data structures. Key Data Structures Used: Hash Table → O(1) event lookup by ID Min-Heap → O(log n) priority management Timestamp filtering → Efficient range queries The scheduler supports: ✔ Adding events ✔ Updating priorities ✔ Cancelling events ✔ Retrieving the next event ✔ Querying events within a time range This project reinforced how critical data structure selection is for system performance. Efficient design can transform potentially O(n) operations into O(1) or O(log n). Excited to continue building more algorithm-focused systems. #Python #DataStructures #Algorithms #ComputerScience #Heap #HashTable
To view or add a comment, sign in
-
Day 6/10 🚀 This is where your data starts to take shape. Collections — the backbone of every Python program. Without the right one? Slower code, messy logic. With the right one? Faster lookups, cleaner design. 📋 What I covered today: 01 → Lists — slicing & comprehensions 02 → Tuples — immutability & unpacking 03 → Dictionaries — CRUD & O(1) lookup 04 → Sets — unique values & operations 05 → Frozenset 06 → Advanced — defaultdict, Counter, namedtuple 07 → Iterators — iter() & next() 08 → Mini Project — Inventory Management System Built a simple system using dictionaries to manage stock & pricing — a real-world pattern used in inventory and data pipelines. Day 1 ✅ Day 2 ✅ Day 3 ✅ Day 4 ✅ Day 5 ✅ Day 6 ✅ 4 more to go. Drop a 🐍 if you’ve ever used a list when a set would’ve been better 😄 #Python #Collections #DataEngineering #LearningInPublic #CleanCode #10DaysOfPython #DataStructures
To view or add a comment, sign in
Explore content categories
- Career
- Productivity
- Finance
- Soft Skills & Emotional Intelligence
- Project Management
- Education
- Technology
- Leadership
- Ecommerce
- User Experience
- Recruitment & HR
- Customer Experience
- Real Estate
- Marketing
- Sales
- Retail & Merchandising
- Science
- Supply Chain Management
- Future Of Work
- Consulting
- Writing
- Economics
- Artificial Intelligence
- Employee Experience
- Workplace Trends
- Fundraising
- Networking
- Corporate Social Responsibility
- Negotiation
- Communication
- Engineering
- Hospitality & Tourism
- Business Strategy
- Change Management
- Organizational Culture
- Design
- Innovation
- Event Planning
- Training & Development