📘 Data Science Journey | Day 25 🔥 Day 50 of my #100daysofcodechallenge Today I learned about the Requests Library in Web Scraping. Here’s what I covered today: 📌 Introduction to Requests Library ▫ A Python library used to send HTTP requests easily ▫ Helps to fetch webpage content programmatically 📌 Making HTTP Requests ▫ Using requests.get() to retrieve webpage data 📌 Working with Response Object ▫ Accessing data using: ▪ .text → HTML content ▪ .status_code → request status (200 = success) 📌 Downloading Multiple Pages ▫ Using loops to scrape multiple pages ▫ Automating data collection from websites 👉 See you tomorrow for Day 51. #DataScience #Python #WebScraping #RequestsLibrary #DataCollection #LearningJourney #Consistency #CodeWithHarry #100daysofcode
Python Web Scraping with Requests Library
More Relevant Posts
-
🚀 Last month, I built and published my first Python package — Pristinizer I wanted to solve a simple but real problem in data science: 👉 Cleaning and understanding raw datasets takes way too much time. So I built Pristinizer, a lightweight Python package that helps streamline data cleaning + EDA in just a few lines of code. 🔍 What Pristinizer does: • Cleans messy datasets (duplicates, missing values, column formatting) • Generates structured dataset summaries • Visualizes missing data (heatmap, matrix, bar chart) ⚙️ Tech Stack: Python • pandas • matplotlib • seaborn 📦 Try it out: >> pip install pristinizer >> import pristinizer as ps df = ps.clean(df) ps.summarize(df) ps.missing_heatmap(df) 🧠 What I learned while building this: • Designing a clean and intuitive API • Structuring a real-world Python package • Publishing to PyPI • Writing proper documentation for users 📌 Next, I’m planning to add: • Outlier detection • Automated preprocessing pipelines • Advanced EDA reports Would love to hear your thoughts or feedback! #Python #DataScience #MachineLearning #OpenSource #Pandas #EDA #Projects
To view or add a comment, sign in
-
-
🐍 Working with data? Save this. Honest truth — I keep coming back to these commands more than I'd like to admit. In most data projects, cleaning takes up more time than the actual analysis, and having the right commands at hand makes a real difference. This Python Data Cleaning cheat sheet covers the 5 essentials I rely on constantly: ✅ Handling nulls and duplicates ✅ Quickly inspecting your dataset ✅ Renaming, converting & cleaning columns ✅ Filtering and slicing rows efficiently ✅ Merging and grouping data If you work with pandas regularly, this should always be within reach. Which of these do you use the most? 👇 #Python #DataScience #DataCleaning #Pandas #DataAnalytics
To view or add a comment, sign in
-
-
🚀 Day 29 – LeetCode Journey Today’s problem: Combine Two Tables ✔️ Used Pandas merge() to join datasets ✔️ Applied left join to retain all records from the primary table ✔️ Selected only required columns for clean output 💡 Key Insight: Understanding how to work with dataframes and joins is essential for real-world data analysis. Using merge() makes combining structured data simple and efficient. This problem strengthened my skills in Pandas, data manipulation, and SQL-like operations in Python. From algorithms to data handling — growing every day 📊🔥 #LeetCode #Day29 #Pandas #DataAnalysis #Python #ProblemSolving #CodingJourney #100DaysOfCode
To view or add a comment, sign in
-
-
📊 4 weeks. 100K+ Wikipedia edits. 1 key finding. I'm happy to share WikiPulse – my first end-to-end data analytics project. The question: Do Wikipedia edit spikes happen before or after real-world events? The finding: Most significant spikes occur 1–2 days before events, suggesting editors anticipate rather than just react. Strongest signal: Academy Awards (r = 0.977, p < 0.05) Tech stack: Python (pandas, NumPy, SciPy, statsmodels) Wikipedia API for data collection SQLite for local database storage Plotly for interactive visualizations Streamlit for dashboard & deployment Live demo: https://lnkd.in/g9bNc3jB GitHub: https://lnkd.in/ghTQfdng Open to feedback and suggestions. #DataAnalytics #Python #Streamlit #PortfolioProject
To view or add a comment, sign in
-
-
🚀 Project Setup (Logistic Regression) Setting up the right environment is the first step in building any Machine Learning project. This module explains how to prepare a Python project for Logistic Regression using essential tools and libraries. The process begins with installing Jupyter Notebook, one of the most widely used platforms for data science. As shown on page 1, using Anaconda Distribution simplifies installation by bundling Python and commonly used packages together. Next, the project setup involves installing required libraries like pandas, numpy, matplotlib, and scikit-learn using pip (page 2). These libraries are essential for data handling, visualization, and building machine learning models. The module also demonstrates how to import necessary packages (page 3), including preprocessing tools, LogisticRegression, and train_test_split from sklearn. Finally, as highlighted on page 4, running the code without errors confirms that the environment is successfully set up and ready for development. 💡 A crucial first step for anyone starting their journey in Machine Learning and data science projects. #Python #MachineLearning #LogisticRegression #DataScience #AshokIT
To view or add a comment, sign in
-
𝗠𝗔𝗖𝗛𝗜𝗡𝗘 𝗟𝗘𝗔𝗥𝗡𝗜𝗡𝗚 𝗙𝗢𝗥 𝗕𝗘𝗚𝗜𝗡𝗡𝗘𝗥𝗦 𝐃𝐚𝐭𝐚 𝐕𝐢𝐬𝐮𝐚𝐥𝐢𝐳𝐚𝐭𝐢𝐨𝐧: 𝐓𝐮𝐫𝐧𝐢𝐧𝐠 𝐃𝐚𝐭𝐚 𝐢𝐧𝐭𝐨 𝐃𝐞𝐜𝐢𝐬𝐢𝐨𝐧𝐬 Raw data is everywhere—but insights are rare. Data visualization is the bridge between numbers and understanding. It transforms complex datasets into clear, actionable insights that drive decisions. From identifying trends to uncovering hidden patterns , visualization is one of the most essential skills in data science. In this post, I’ll walk you through key visualization techniques using Python—designed especially for beginners to learn and apply. Let’s turn data into stories 🚀 #DataVisualization #Python #DataScience #EDA
To view or add a comment, sign in
-
Just wrapped up a simple, but insightful visualisation practice using Python 🐍🐼. I used a histogram to break down how many people passed vs failed in a dataset, and even with a small sample, the distribution already reveals something important. Clear labelling and readability made the difference in turning raw data into something meaningful. ✨ Something I'm focusing on more is not just analysing data, but presenting it in a way that makes insights easily recognisable. 🧠 Small steps, but each project sharpens my ability to communicate data effectively. 🔥📉📈 #DataAnalytics #Python #DataVisualization #LearningJourney Neo Matekane, your recent post "Changing Data into Insights 📊" was a wonderful resource! It gave me a fresh perspective on how to approach data visualisation and extract more meaningful insights from the process. 🥳✨✨ Shoutout to Shafiq Ahmed! His consistency in sharing data insights and breaking down projects in simple, easy-to-understand terms is something I truly look up to on my data journey. 🚀📊
To view or add a comment, sign in
-
-
Today I wrapped up Module 2 of my Python course, focusing on core data structures - lists, tuples, and dictionaries. These concepts are the foundation for real data work, and practicing them has already helped me think more clearly about how data is stored, accessed, and transformed. I’m building these fundamentals intentionally so that when I move into Pandas and SQL, the logic feels natural. A few things I strengthened today: • Understanding how lists and tuples behave differently • Accessing and slicing data efficiently • Using dictionaries to store structured information • Writing cleaner, more readable code Small steps, consistent progress - that’s the goal. On to Module 3. #PythonLearning #DataAnalytics #WomenInTech #CareerGrowth #TechSkills #DataStructures #LearningInPublic #AnalyticsJourney #HealthcareAnalytics #SQL #Pandas #DataAnalyst
To view or add a comment, sign in
-
🐍📊 Python + Data Science = A match made in heaven. If you're diving into data science (or leveling up your skills), mastering Python is non-negotiable. Here’s why: ✅ Simplicity – Clean syntax means you focus on solving problems, not fighting the language. ✅ Ecosystem – Pandas for wrangling, NumPy for numbers, Matplotlib/Seaborn for visuals, Scikit-learn for ML. ✅ Community – Thousands of free resources, libraries, and real-world projects to learn from. 🚀 3 Python tricks that saved me hours: df.query() instead of multiple slicing conditions in Pandas. seaborn.set_theme() for instantly better-looking plots. pd.to_datetime() with errors='coerce' to clean messy date columns fast. Whether you’re a beginner or a seasoned analyst, Python scales with you. 👇 What’s your go-to Python library for data work? #Python #DataScience #DataAnalytics #MachineLearning #Pandas #Coding
To view or add a comment, sign in
-
🚀 Day 46 of My 90-Day Data Science Challenge Today I worked on ROC Curve & AUC Score. 📊 Business Question: How can we evaluate a model’s performance across different classification thresholds? ROC Curve helps visualize the trade-off between True Positive Rate (Recall) and False Positive Rate. Using Python & scikit-learn: • Plotted ROC Curve • Calculated AUC Score • Compared model performance • Understood threshold impact • Evaluated classification quality 📈 Key Understanding: A higher AUC score indicates a better model at distinguishing between classes. 💡 Insight: ROC-AUC helps evaluate models beyond simple accuracy. 🎯 Takeaway: Good models maintain high true positives while minimizing false positives. Day 46 complete ✅ Advancing model evaluation techniques 🚀 #DataScience #MachineLearning #ROC #AUC #Python #LearningInPublic #90DaysChallenge
To view or add a comment, sign in
-
Explore content categories
- Career
- Productivity
- Finance
- Soft Skills & Emotional Intelligence
- Project Management
- Education
- Technology
- Leadership
- Ecommerce
- User Experience
- Recruitment & HR
- Customer Experience
- Real Estate
- Marketing
- Sales
- Retail & Merchandising
- Science
- Supply Chain Management
- Future Of Work
- Consulting
- Writing
- Economics
- Artificial Intelligence
- Employee Experience
- Workplace Trends
- Fundraising
- Networking
- Corporate Social Responsibility
- Negotiation
- Communication
- Engineering
- Hospitality & Tourism
- Business Strategy
- Change Management
- Organizational Culture
- Design
- Innovation
- Event Planning
- Training & Development