🚀 Built a Multi-Format PDF Data Extractor in Python I created a basic Python project that extracts structured data from different types of PDFs and raw text files, even when formats are inconsistent. 🔹 Handles multiple PDF layouts 🔹 Fallback extraction pipeline (regex → text → tables → OCR) 🔹 Extracts: PO, Brand, Size, Inseam, Quantity 🔹 Cleans and filters data automatically using pandas 🔹 Displays a clean table in terminal 🔹 Exports results to Excel 🔹 Works with messy and unstructured documents This is the first version. Next, I plan to add batch processing, logging, verification logic, and smarter format detection for higher accuracy. Learning by building real-world automation tools step by step. Feedback is welcome! #Python #PythonProgramming #PythonDeveloper #Automation #DataExtraction #PDFProcessing #Pandas #Regex #Camelot #pdfplumber #PyTesseract #OCR #DataEngineering #OpenPyXL #Tabulate #MachineLearning #AI #Developer #Coding #Tech #Programming #BuildInPublic #LearningByDoing
More Relevant Posts
-
Seaborn is a high-level Python visualization library designed for creating clear, attractive, and informative statistical graphics. Built on top of Matplotlib, it provides an intuitive interface for visualizing complex datasets with minimal code. Seaborn works seamlessly with Pandas DataFrames and supports a wide range of visualizations, including scatter plots, line plots, bar plots, box plots, violin plots, and heatmaps. It is particularly useful in Exploratory Data Analysis (EDA), where visualizing relationships between variables helps in understanding data behavior, detecting trends, and identifying outliers. Seaborn is widely used in Data Science and Machine Learning workflows, as effective visualization improves data understanding and supports better decision-making during model development.
To view or add a comment, sign in
-
-
🚆 Exploring & Understanding Training Data in Machine Learning I recently worked on a Jupyter Notebook project focused on analyzing a training dataset (Train.ipynb) as part of my data science journey. This project helped me understand how raw data is transformed into meaningful insights before feeding it into machine learning models. 🔍 What I worked on: • Data Exploration (EDA) • Data Cleaning & Handling Missing Values • Understanding feature relationships • Preparing structured training data 📊 Why Training Data Matters: Training data is the foundation of any machine learning model — the better the data quality, the better the predictions. 💡 Key Learnings: • Real-world datasets are messy and need preprocessing • Feature understanding is crucial before modeling • Data preparation directly impacts model accuracy • Practical exposure to ML workflow 🛠️ Tech Stack: Python | Pandas | NumPy | Jupyter Notebook 🚀 This project strengthened my understanding of data preprocessing and machine learning fundamentals 🔗 Check out the notebook here: https://lnkd.in/drXQ_7Rk 💬 Open to feedback, suggestions, and collaboration! #MachineLearning #DataScience #Python #EDA #AI #JupyterNotebook #StudentDeveloper #LearningJourney
To view or add a comment, sign in
-
🚀 Built a GUI-Based Data Analysis Tool while Learning Python with AI As part of my Python learning journey using AI-assisted development, I built a GUI-based data analysis tool that simplifies working with Excel and CSV data by helping users quickly explore datasets, generate summaries, and visualize insights without manual data processing. 🛠 Tech Stack: Python, Pandas, Tkinter, Matplotlib ✨ Key Features: ✅ Upload & analyze Excel/CSV files ✅ Automatic dataset profiling (rows, columns, headers) ✅ Smart detection of text & numeric columns ✅ GroupBy reports with multiple aggregations ✅ Built-in charts (Bar, Line, Column, Pie) ✅ Export reports (Excel/CSV) & charts (PNG) 🎯 This project helped me gain hands-on experience in Python development, data analysis workflows, and building practical business-focused tools with AI support. Excited to keep learning and building — feedback is welcome! #PythonLearning #DataAnalytics #AIAssistedDevelopment #Tkinter #Pandas #Automation #LearningByDoing
To view or add a comment, sign in
-
Today I'm giving a talk at PyCon DE & PyData about "Simplifying RAG Pipelines using Multimodal Embeddings". In Retrieval Augmented Generation systems – which enhance LLMs with external knowledge without retraining – PDFs are the dominant document format. The problem: they're display-oriented and not designed to make data easy to extract. Traditionally, OCR pipelines are used to pull out text, but they struggle with multi-column layouts, complex tables, and can't process charts or figures without dedicated extra steps. In my talk I look at a different approach: instead of fragile and complex OCR pipelines, you convert PDF pages to images and embed them using multimodal embedding models. This not only simplifies the setup and reduces costs – as I validated through benchmarks on real-world enterprise documents – it also improves retrieval performance. You can find the talk here: https://lnkd.in/d3ESkyc7 For anyone around, feel free to come say hi 👋 #PyCon #RAG #LLM #Python #MachineLearning
To view or add a comment, sign in
-
Day 9/60 Starting Chapter 2: Flow Control Topic l - Making Decisions 1. Conditions Smart programs use booleans to make decisions on whether to run lines of code or skip them. We use an if statement to write code that responds to different situations. We recognize it by the keyword if . The if statement runs code only if it is evaluated as true. It's like saying, if something is true, then do this. Let's make the evaluation true by simply using the boolean value True to display "Hello" in the console. 🧩Code if True: print("Hello!") 🖥️Output Hello! But what if we want to skip the code? In that case, we need the statement to evaluate as false. Let's easily make the statement evaluate to false by simply using the boolean value False . Nothing will print. 🧩Code if False: print("Hello!") 🖥️Output will be empty Values like True are called conditions. Statements relying on conditions are called conditionals. Conditions decide if the code runs or gets skipped. They come in between if and the colon : . 🧠Challenge of the day: Let’s see if you paid enough attention What is the output of following? 🧩Code: if true: print("3, 2, 1 GO") #python #ai #programming #bigtech
To view or add a comment, sign in
-
Polars — A Faster Alternative to Pandas for Data Processing While working with large datasets, performance becomes a real challenge. That’s where Polars is getting attention. What is Polars? Polars is a high-performance DataFrame library designed for fast and efficient data processing. It is built using Rust and provides a Python API similar to Pandas. Why developers are switching to it: Faster execution on large datasets Lower memory usage Parallel processing support Cleaner and modern API design What makes it interesting: Instead of processing data in a single-threaded way like traditional workflows… Polars is optimized for speed from the ground up. Real use cases: Data analytics pipelines Large CSV and parquet processing Machine learning preprocessing High-performance data engineering tasks Why it matters: As datasets continue to grow, performance and scalability become more important than ever. Tools like Polars show how modern data processing is evolving. Final thought: Pandas changed how developers work with data. Polars is pushing that experience toward speed and scalability. Follow Saif Modan #Python #DataScience #Polars #MachineLearning #Analytics #Tech
To view or add a comment, sign in
-
-
𝐏𝐂𝐀 (𝐏𝐫𝐢𝐧𝐜𝐢𝐩𝐚𝐥 𝐂𝐨𝐦𝐩𝐨𝐧𝐞𝐧𝐭 𝐀𝐧𝐚𝐥𝐲𝐬𝐢𝐬)- 𝐖𝐡𝐞𝐧 𝐭𝐨𝐨 𝐦𝐚𝐧𝐲 𝐟𝐞𝐚𝐭𝐮𝐫𝐞𝐬 𝐬𝐭𝐚𝐫𝐭 𝐛𝐞𝐜𝐨𝐦𝐢𝐧𝐠 𝐚 𝐩𝐫𝐨𝐛𝐥𝐞𝐦… While working on datasets with a large number of features, I realized something important: 𝐌𝐨𝐫𝐞 𝐟𝐞𝐚𝐭𝐮𝐫𝐞𝐬 ≠ 𝐛𝐞𝐭𝐭𝐞𝐫 𝐦𝐨𝐝𝐞𝐥 In fact, too many features can lead to a problem called: - Curse of Dimensionality - Models become slow - Computation increases - Noise increases - Visualization becomes difficult 𝐒𝐨𝐥𝐮𝐭𝐢𝐨𝐧 → 𝐃𝐢𝐦𝐞𝐧𝐬𝐢𝐨𝐧𝐚𝐥𝐢𝐭𝐲 𝐑𝐞𝐝𝐮𝐜𝐭𝐢𝐨𝐧 𝐏𝐂𝐀 is an 𝐮𝐧𝐬𝐮𝐩𝐞𝐫𝐯𝐢𝐬𝐞𝐝 𝐥𝐞𝐚𝐫𝐧𝐢𝐧𝐠 technique used when we only have input features (no target/output). It is a 𝐟𝐞𝐚𝐭𝐮𝐫𝐞 𝐞𝐱𝐭𝐫𝐚𝐜𝐭𝐢𝐨𝐧 𝐭𝐞𝐜𝐡𝐧𝐢𝐪𝐮𝐞 𝐭𝐡𝐚𝐭 𝐭𝐫𝐚𝐧𝐬𝐟𝐨𝐫𝐦𝐬 𝐡𝐢𝐠𝐡-𝐝𝐢𝐦𝐞𝐧𝐬𝐢𝐨𝐧𝐚𝐥 𝐝𝐚𝐭𝐚 𝐢𝐧𝐭𝐨 𝐥𝐨𝐰𝐞𝐫 𝐝𝐢𝐦𝐞𝐧𝐬𝐢𝐨𝐧𝐬 while preserving most of the important information. " In simple words: It keeps the essence of data but reduces complexity." 𝐔𝐬𝐢𝐧𝐠 𝐏𝐂𝐀 𝐡𝐞𝐥𝐩𝐬:- Reduce number of features - Improve model performance - Reduce computation cost - Speed up training - Make data easier to visualize 𝐇𝐨𝐰 𝐏𝐂𝐀 𝐖𝐨𝐫𝐤𝐬 (𝐒𝐭𝐞𝐩𝐬 𝐈 𝐏𝐫𝐚𝐜𝐭𝐢𝐜𝐞𝐝) 𝐒𝐭𝐞𝐩 1️⃣: 𝐒𝐭𝐚𝐧𝐝𝐚𝐫𝐝𝐢𝐳𝐞 𝐭𝐡𝐞 𝐝𝐚𝐭𝐚 Because PCA is scale-sensitive 𝐒𝐭𝐞𝐩 2️⃣: 𝐂𝐨𝐦𝐩𝐮𝐭𝐞 𝐂𝐨𝐯𝐚𝐫𝐢𝐚𝐧𝐜𝐞 𝐌𝐚𝐭𝐫𝐢𝐱 To understand relationships between features 𝐒𝐭𝐞𝐩 3️⃣: 𝐅𝐢𝐧𝐝 𝐄𝐢𝐠𝐞𝐧𝐯𝐚𝐥𝐮𝐞𝐬 & 𝐄𝐢𝐠𝐞𝐧𝐯𝐞𝐜𝐭𝐨𝐫𝐬 import numpy as np eigen_values, eigen_vectors=np.linalg.eig(cov_matrix) 𝐒𝐭𝐞𝐩 4️⃣: 𝐒𝐞𝐥𝐞𝐜𝐭 𝐏𝐫𝐢𝐧𝐜𝐢𝐩𝐚𝐥 𝐂𝐨𝐦𝐩𝐨𝐧𝐞𝐧𝐭𝐬 Choose top components with highest variance 𝘗𝘊𝘈 𝘪𝘴 𝘯𝘰𝘵 𝘫𝘶𝘴𝘵 𝘳𝘦𝘥𝘶𝘤𝘪𝘯𝘨 𝘤𝘰𝘭𝘶𝘮𝘯𝘴… 𝘐𝘵’𝘴 𝘢𝘣𝘰𝘶𝘵 𝘬𝘦𝘦𝘱𝘪𝘯𝘨 𝘵𝘩𝘦 𝘮𝘰𝘴𝘵 𝘪𝘮𝘱𝘰𝘳𝘵𝘢𝘯𝘵 𝘪𝘯𝘧𝘰𝘳𝘮𝘢𝘵𝘪𝘰𝘯 𝘸𝘩𝘪𝘭𝘦 𝘳𝘦𝘮𝘰𝘷𝘪𝘯𝘨 𝘳𝘦𝘥𝘶𝘯𝘥𝘢𝘯𝘤𝘺 #Datascience #Dataanalyst #Machinelearning #curseofdimensionality #featureextraction #python #numpy
To view or add a comment, sign in
-
🚀 Automated My Email Workflow with Python! The Problem: Manually replying to dozens of emails daily with document attachments - 2-3 minutes each, that’s hours of repetitive work every week! The Solution: Built a Python script that: ✅ Monitors a folder for new files ✅ Finds matching emails by reference number ✅ Auto-replies with personalized greetings & attachments ✅ Runs 24/7 in the background How I Did It: Used AI (Claude) to generate the code, then customized it for my needs - Reply All, multi-folder search, duplicate prevention, and HTML formatting. The Impact: ⏱️ Saves 1-2 hours daily ✅ Zero errors 📈 Consistent responses Key Takeaway: You don’t need to be a developer to automate. With AI assistance, anyone can build practical solutions. Tools: Python | pywin32 | Claude AI #Automation #Python #ProductivityHacks #AI #WorkSmarter
To view or add a comment, sign in
-
Getting the "plumbing" right before the ML takes over. I’m currently building a House Price Valuation System, and if there’s one thing my CS background has taught me, it’s that a model is only as good as the data pipeline behind it. This screenshot is from the Data Preprocessing phase. I’m using Python (Pandas/NumPy) to handle the messy reality of raw data—things like categorical imputation and logical defaults—so the data is actually structured and ready for testing in the ML models. Whether it’s an ML project or a business dashboard, I’ve found that the real engineering happens in the "boring" parts: the cleaning, the logic, and the automated pipelines. Once the technical foundation is solid, the rest usually falls into place. #CSEngineer #Python #MachineLearning #SystemArchitecture #BuildingInPublic
To view or add a comment, sign in
-
-
Ever run a Python script and get a frustrating “file not found” error? 😤 This simple snippet can save you hours 👇 import os # Check if we're in the right place print("Current directory: ", os.getcwd()) # Check if our data file exists data_path = "data/sales.csv" if os.path.exists(data_path): print(f"Found {data_path}") else: print(f"X Cannot find {data_path}") print("Make sure you're running from the sales-analysis folder!") 💡 What’s happening here? 🔹 os.getcwd() Prints your current working directory — this tells you where your script is running from. Many errors happen because you're in the wrong folder. 🔹 data_path = "data/sales.csv" Defines the relative path to your dataset. 🔹 os.path.exists(data_path) Checks if the file actually exists before trying to use it. 🔹 Conditional check (if / else) Gives clear feedback: ✔ Found the file ❌ Or tells you it’s missing 🚀 Why this matters Prevents runtime errors Helps debug file path issues quickly Makes your scripts more reliable Essential habit for data analysis projects 📊 Whether you're working on data science, automation, or AI — always verify your file paths before processing data. Small habit. Big impact. #Python #Programming #DataScience #AI #CodingTips #Debugging
To view or add a comment, sign in
Explore content categories
- Career
- Productivity
- Finance
- Soft Skills & Emotional Intelligence
- Project Management
- Education
- Technology
- Leadership
- Ecommerce
- User Experience
- Recruitment & HR
- Customer Experience
- Real Estate
- Marketing
- Sales
- Retail & Merchandising
- Science
- Supply Chain Management
- Future Of Work
- Consulting
- Writing
- Economics
- Artificial Intelligence
- Employee Experience
- Workplace Trends
- Fundraising
- Networking
- Corporate Social Responsibility
- Negotiation
- Communication
- Engineering
- Hospitality & Tourism
- Business Strategy
- Change Management
- Organizational Culture
- Design
- Innovation
- Event Planning
- Training & Development