7,250 downloads. 1,880 clones in 14 days. 404 developers using it . When we started building SynapseKit, we made one rule: Don't ship the framework without shipping the documentation. Because I've used too many "promising" Python libraries that had great internals and zero explanation of how to actually use them. You'd clone it, stare at the source code for 20 minutes, and give up. SynapseKit was built to be the opposite of that. What is SynapseKit? An async-native Python framework for building LLM applications — RAG pipelines, AI agents, and graph workflows — across 27 providers with one interface. Swap OpenAI for Anthropic[Anthropic]. Swap Anthropic for Ollama[Ollama]. Zero rewrites. Streaming-first. Async by default. Two hard dependencies. But here's what actually makes me proud: The 7,250 downloads aren't from a viral post or a Product Hunt launch. They came from developers finding it on GitHub, engineers discovering it on PyPI while searching for tools, and people landing on the docs and actually understanding what they found. That last one is everything. Good documentation doesn't just explain your code. It builds trust. It tells engineers — "this project is maintained, this project respects your time, this project will still work six months from now." 105 open issues. 30 pull requests in March alone. People aren't just downloading SynapseKit — they're contributing to it. What's inside: → RAG Pipelines — streaming, BM25 reranking, memory, token tracing → Agents — ReAct loop, native function calling for OpenAI / Anthropic / Gemini / Mistral → Graph Workflows — DAG async, parallel routing, human-in-the-loop → Observability — CostTracker, BudgetGuard, OpenTelemetry — no SaaS required → Vector Stores — ChromaDB, FAISS, Qdrant, Pinecone behind one interface All of it documented. All of it referenced. All of it open source. If you're building LLM applications in Python, I'd genuinely love for you to take it for a spin. 📖 https://lnkd.in/dvr6Nyhx ⭐ https://lnkd.in/d2fGSPkX And if you find something broken, missing, or confusing - open an issue. That's exactly how 105 conversations started. No framework survives bad documentation. We're building both. #Python #OpenSource #LLMFramework #SynapseKit #AIEngineering #RAG #AIAgents #BuildInPublic #MachineLearning #LLM
SynapseKit: Async Python LLM Framework with Open Source Docs
More Relevant Posts
-
I thought knowing Next.js, GraphQL, Docker, Python, and every AI tool made me untouchable. Until a client asked a simple question about code I’d shipped just a week earlier and I froze. I wasn’t thinking. I had just been copying and pasting answers from AI. The trap isn’t using AI. The trap is feeling productive while your real understanding quietly disappears. I wrote the full breakdown of what I call the Copy‑Paste Trap and the uncomfortable practices that help you escape it. If you’ve ever shipped code you couldn’t fully explain, this one’s for you.
To view or add a comment, sign in
-
There’s a small change coming to Python that looks simple on the surface — but has real impact once you think in terms of systems. PEP 810 introduces explicit lazy imports - modules don’t load at startup - they load only when actually used At first glance, this sounds like a minor optimization. It’s not. Every engineer has seen this pattern: You run a CLI with -help - and it still takes seconds to respond Why? Because the runtime eagerly loads everything - even code paths you’ll never touch in that execution That startup cost adds up - especially in services, scripts, and short-lived jobs Lazy imports change that behavior. Instead of front-loading everything at startup - the runtime defers work until it’s actually needed So now: - unused dependencies don’t slow you down - cold starts improve - CLI tools feel instant again It’s a small shift in syntax - but a meaningful shift in execution model What’s interesting is not the idea itself. Lazy loading has existed for years - across languages, frameworks, and runtimes But Python never had a standard way to do it - teams built custom wrappers - some even forked the runtime That fragmentation was the real problem. PEP 810 fixes that - by making it opt-in - preserving backward compatibility - while finally standardizing the pattern That decision matters more than the feature. Earlier attempts tried to make lazy imports the default - and ran straight into compatibility risks This time, the approach is pragmatic: - no breaking changes - no surprises in existing systems - but a clear path for teams that need performance gains That’s how ecosystem-level changes actually stick. From a systems perspective, this connects to a broader principle: Startup time is part of user experience. Whether it’s: - a CLI tool - a containerized service - a serverless function Cold start latency directly impacts usability and cost And most of that latency isn’t business logic - it’s initialization overhead Lazy imports attack that overhead at the root. Not by optimizing logic - but by avoiding unnecessary work entirely Which is often the highest-leverage optimization you can make. The bigger takeaway isn’t just about Python. It’s this: Modern systems are moving toward just-in-time execution - load less upfront - execute only what’s needed - keep everything else deferred You see it in: - class loading strategies - dependency injection frameworks - container startup tuning Now it’s becoming part of the language itself. It’ll take time before this shows up in everyday workflows. But once it does, expect a shift in how people structure imports - especially in performance-sensitive paths Explore more : https://lnkd.in/gP-SeCMD #SoftwareEngineering #Python #Java #Backend #Data #DevOps #AWS #C2C #W2 #Azure #Hiring #BackendEngineering Boston Consulting Group (BCG) Kforce Inc Motion Recruitment Huxley Randstad Digital UST CyberCoders Insight Global
To view or add a comment, sign in
-
Everyone said AI would make Python unstoppable. GitHub's data says the opposite happened. Please note, I’m as proficient in Python as I am in TypeScript. TypeScript just became the #1 language on GitHub. Not Python. Not JavaScript. The typed version of JavaScript that some developers still avoid. And AI coding tools are the main reason. Here are five things the numbers show: 1. 94% of LLM-generated compilation errors are type-check failures. A 2025 academic study confirmed this. AI models produce code that compiles and runs but fails type checks. TypeScript catches those errors before they reach production. Python does not. 2. TypeScript contributors on GitHub grew 66% year-over-year. That is 2.6 million monthly contributors, more than any other language on the platform. The growth accelerated after Copilot went mainstream. 3. 80% of new GitHub developers use Copilot in their first week. These developers do not choose languages based on tradition. They choose whatever the AI writes best. And AI arguably writes TypeScript better than almost anything else. 4. Every major framework now defaults to TypeScript. Start a new project today and you get TypeScript whether you asked for it or not. 5. 1.1 million public repos now use an LLM SDK. That is up 178% in one year. The tools developers build with AI are being built in TypeScript. The language and the tooling are converging. The takeaway for builders: if you are still writing untyped JavaScript or betting everything on Python for web products, the industry moved while you were deciding. Types are not a preference anymore. They are a production requirement in the age of AI-generated code. Don’t get me wrong, we still build some services using Python and FastAPI in Deveote. Especially ML-rich services. But TypeScript is our major language, and will continue to be. What language are you betting on for web applications?
To view or add a comment, sign in
-
Building a Multimodal Agent with the ADK, Amazon Lightsail, and Gemini Flash Live 3.1: Leveraging the Google Agent Development Kit (ADK) and the underlying Gemini LLM to build Agentic apps using the Gemini Live API with the Python programming language deployed to Amazon Lightsail. Aren’t There a Billion Python ADK Demos? Yes there are. Python has traditionally been the main coding language for ML and AI tools. The goal of this article is to provide a minimal viable basic working ADK streaming multi-modal agent using the latest Gemini Live Models. In the Spirit of Mr. McConaughey’s “alright, alright, alright” So what is different about this lab compared to all the others out there? This is one of the first implementations of the latest Gemini 3.1 Flash Live Model with the Agent Development Kit (ADK). The starting point for the demo was an existing Code lab- which was updated and re-engineered with Gemini CLI. The original Codelab- is here: Way Back Home - Building an ADK Bi-Directional Streaming Agent | Google Codelabs What Is Python? Python is an interpreted language that allows for rapid development and testing and has deep libraries for working with ML and AI: Welcome to Python.org Python Version Management One of the downsides of the wide deployment of Python has been managing the language versions across platforms and maintaining a supported version. The pyenv tool enables deploying consistent versions of Python: GitHub - pyenv/pyenv: Simple Python version management As of writing — the mainstream python version is 3.13. To validate your current Python:python --version Python 3.13.12 Amazon Lightsail Amazon Lightsail is an easy-to-use virtual private server (VPS) provider and cloud platform designed by AWS for simpler workloads, offering developers pre-configured compute, storage, and networking for a low, predictable monthly price. It is ideal for hosting small websites, simple web apps, or creating development environments. More information is available on the official site here: Amazon's Simple Cloud Server | Amazon Lightsail And this is the direct URL to the console:https://lnkd.in/eV7DaV8y Gemini Live Models Gemini Live is a conversational AI feature from Google that enables free-flowing, real-time voice, video, and screen-sharing interactions, allowing you to brainstorm, learn, or problem-solve through natural dialogue. Powered by the Gemini 3.1 Flash Live model, it provides low-latency, human-like, and emotionally aware speech in over 200 countries. More details are available here: Gemini 3.1 Flash Live Preview | Gemini API | Google AI for Developers The Gemini Live Models bring unique real-time capabilities than can be used directly from an Agent. A summary of the model is also available here:https://lnkd.in/ekCsUE3q Gemini CLI If not pre-installed… #genai #shared #ai
To view or add a comment, sign in
-
I replaced my AI assistant platform with 300 lines of Python. For a few months I ran OpenClaw, a self-hosted AI assistant with multi-agent routing and sandbox execution. It worked. But the overhead was real: → Every design decision filtered through "how many tokens does this cost?" → 6-12 hours to deploy across gateway config, sandbox hardening, networking, and agent tuning → Memory leaks and provider deprecations on my Mac Mini → A known-issues doc that kept growing After attending an Anthropic workshop at their headquarters about the Claude Agent SDK, I figured I'd give it a shot. It's a Python package that spawns the Claude CLI as a subprocess. It inherits your existing auth session, so my Claude Pro subscription covers everything. No API keys, no per-token billing. 𝗪𝗵𝗮𝘁 𝗰𝗵𝗮𝗻𝗴𝗲𝗱: The entire Telegram bot is ~300 lines. MCP server support, multi-turn sessions, tool use, skills, all inherited from my Claude Code config. I added Python hook callbacks for guardrails (hard-block destructive commands, log every tool call) and built custom tools like a subscription usage tracker and semantic memory retrieval. The MCP integrations are where it gets fun. Through a single MetaMCP aggregator, the bot can pull YouTube transcripts and comments, manage my Unraid NAS, check Plex library status, handle media requests through Overseerr, browse Reddit, and manage my DNS records on Porkbun. I can ask it to look up what was said in a YouTube video, check if a movie is available on Plex, or see which Docker containers are unhealthy on my server — all from Telegram. Scheduled tasks are even simpler. A daily homelab health check is two lines: from task_common import run run("Check all Docker stacks and container health", max_turns=15) Runs via cron, sends results to Telegram. 𝗧𝗵𝗲 𝗯𝗶𝗴𝗴𝗲𝘀𝘁 𝘀𝗵𝗶𝗳𝘁 𝘄𝗮𝘀 𝗺𝗲𝗻𝘁𝗮𝗹. With OpenClaw I picked Gemini Flash because it was cheap. With the SDK I use Sonnet because it's better, and the cost is the same either way. I stopped optimizing token budgets and started thinking about what I actually want the bot to do. OpenClaw is a solid project if you need multi-channel support or provider failover, or if you want something turnkey without writing code. But for a single user comfortable with Python who's already paying for Claude Pro, the SDK won on every axis I care about. The bot runs under launchd on a Mac Mini. Starts on boot, restarts on crash, idles at ~50MB. I haven't touched it in weeks. #ClaudeCode #AgentSDK #AI #Python #BuildInPublic #Homelab
To view or add a comment, sign in
-
Python developers in 2026 are sitting on a goldmine and not using it. You already know FastAPI. You already know Django. Your CRUD is clean. Your endpoints are solid. Your logic is tight. But here's the thing That's the baseline now. Not the advantage. Every developer ships CRUD. Not every developer ships a product that thinks. And the good news? If you're already in Python you're one integration away. Python is the only language where the gap between "CRUD app" and "AI-powered product" is measured in hours, not months. Here's what that gap looks like in practice: → Add openai or anthropic SDK — your app now understands user input, not just stores it → Plug in LangChain — your endpoints start making decisions, not just returning rows → Use scikit-learn or Prophet — your FastAPI routes now predict, not just fetch → Connect Celery + an AI model — your background tasks now act intelligently on patterns → Drop in pgvector with PostgreSQL — your database now does semantic search, not just SQL filters This is not a rewrite. This is an upgrade. What CRUD alone gives your users in 2026: ❌ The same experience on day 1 and day 500 ❌ Manual decisions they have to make themselves ❌ A product that stores their data but never understands it ❌ A reason to switch the moment something smarter appears What Python + AI gives your users in 2026: ✅ An app that learns their behavior and adapts ✅ Recommendations, predictions and alerts automatically ✅ A product that gets more valuable the more they use it ✅ A reason to stay and a reason to tell others The architecture stays familiar. FastAPI route → AI layer → response. You're not rebuilding anything. You're making what you already built actually intelligent. Python developers have transformers, LangChain, OpenAI SDK, Hugging Face all production-ready, all pip-installable, and all designed to sit right next to your existing FastAPI or Django project. No other ecosystem makes this this accessible. CRUD was the foundation. AI is the product. And if you're already writing Python you're already holding the tools. The only move left is using them. Which Python AI library are you integrating into your stack this year? 👇 #Python #FastAPI #Django #AIIntegration #SoftwareDevelopment #LangChain #MachineLearning #BackendDevelopment #TechIn2026 #BuildInPublic
To view or add a comment, sign in
-
-
A Python script answers questions. Nobody else can use it. A FastAPI endpoint answers questions. Everyone can. That gap is 10 lines of code. I closed it on Day 17 — here is everything I measured. —— I spent 20 days building an AI system from scratch. No LangChain. No frameworks. Pure Python. Phase 5 was wrapping it in FastAPI and measuring everything honestly. —— Day 17 — two endpoints, full pipeline behind HTTP POST /ask runs the full multi-agent pipeline. GET /health reports server status and tool count. Swagger UI at /docs — interactive docs, zero extra code. First real response: 60,329ms. Day 18 — one log file changed everything Per-stage timing showed this: mcp_init: 31,121ms planner: 748ms orchestrator: 3,127ms synthesizer: 1,331ms 31 of 60 seconds was initialization. Not the model. Not retrieval. The setup — running fresh every request. Two fixes. No model change. Fix 1: direct Python calls instead of subprocess per tool. Fix 2: MCP init moved to server startup — paid once, never again. Result: 60s → 5.7s. 83% faster. Day 19 — RAGAS on the live API Same 6 questions from Phase 2. Real HTTP calls. Honest numbers. Faithfulness: 0.638 → 1.000 Answer relevancy: 0.638 → 0.959 Context recall: went down — keeping that in. Explained in the post. —— The number that reframes the whole journey: 54 seconds saved by initializing in the right place. Not a faster model. Not more compute. Just knowing what to load at startup and what to create per request. Expensive + stateless → load once at startup. Stateful or cheap → create fresh per request. That one decision is the difference between a demo and a production system. —— The full score progression — all 20 days: Phase 2 baseline: 0.638 Phase 2 hybrid retrieval: 0.807 Phase 2 selective expansion: 0.827 Phase 5 answer relevancy: 0.959 Phase 5 faithfulness: 1.000 —— 20 days. Pure Python. No frameworks. Every number real. Every failure documented. Full writeup with code, RAGAS setup, and the FastAPI tutorial: https://lnkd.in/eBDdAMiY GitHub — everything is open source: https://lnkd.in/es7ShuJr If you have built something with FastAPI — what was the first thing you wished someone had told you? #AIEngineering #FastAPI #Python #BuildInPublic #LearningInPublic
To view or add a comment, sign in
-
-
Story time. There was a phase when Python quietly stopped getting picked. Not because it disappeared. Not because people didn’t love it. But when the question was “what should we use for a serious backend?” — the answers were predictable. Node for async. Go for concurrency. Java for scale. Python? “Too slow.” “GIL issues.” “Not for production.” And to be fair — those criticisms weren’t wrong. The GIL wasn’t a bug. It was a design choice for safety. It ensured: memory consistency simpler garbage collection a stable C-extension ecosystem But the tradeoff was brutal: Only one thread could execute Python bytecode at a time. No true parallelism. People tried to “fix” it: joblib, threads, thread pools… But none of them actually removed the constraint. They just worked around it. Meanwhile, Go was doing real concurrency out of the box. Lightweight goroutines. Multi-core efficiency. If this was a race — Python wasn’t winning. But here’s the part most people miss: There was no rivalry. No “Python vs Go” war. Just a quiet shift in what the industry valued. While everyone was optimizing for speed, Python went somewhere else entirely. Data. Machine learning. AI. It didn’t try to win the same game. Then… the stack evolved. Async became usable. And a big unlock came in quietly: uvloop. A faster event loop that made Python’s async actually fast Lower latency. Better throughput. Real gains. But speed alone wasn’t enough. Enter FastAPI Not just a framework — but the missing piece that made everything click: Async-first by design Type-driven development Automatic docs Clean, production-ready APIs Now the stack looked like: async + uvloop + ASGI + FastAPI Not true parallelism. But extremely efficient I/O concurrency. And something shifted. Python didn’t need to beat Go at concurrency. It just needed to be good enough for the systems people were actually building. Then the real change happened. Backends stopped being just CRUD layers. They became: model serving systems Data pipelines AI-native applications And now the question wasn’t: “What’s the fastest language?” It was: “What fits the system end-to-end?” That’s when Python walked back in. Not as the fastest. Not as the best at concurrency. But as the most aligned. So no — Python didn’t beat Go. It just stopped playing the same game… and won a bigger one. Funny how a design choice made for safety… was once seen as a limitation — and later became irrelevant to the problems that mattered. #Python #FastAPI #uvloop #AI #Backend #SystemDesign
To view or add a comment, sign in
-
📨 AI-Driven Micro-Loan Platform 5/10: The gRPC Fast-Lane – When "Async" isn't fast enough ⚡ In Post 3, we talked about the "Deep-Dive" via Service Bus. But what if you need an answer now? For the initial "Pre-Score" (the 5-second decision that keeps a user in the app), we can't wait for a message queue. We need a direct, high-speed connection between .NET 10 and Python. Enter gRPC. 🏎️ 1. Why gRPC over REST? ⚖️ Most teams default to JSON over HTTP. In a high-volume microservices environment, that's "death by a thousand cuts." Protobuf vs. JSON: gRPC uses Protocol Buffers (binary). It’s smaller, faster to serialize, and strictly typed. Multiplexing: Using HTTP/2, we keep a single connection open for multiple requests, reducing the overhead of constant "handshakes." 2. The "Pre-Score" Flow 🟢 When the user hits "Check My Limit": .NET API calls the Python ML Service via a gRPC client. Python pulls the "Lightweight" features from Redis. Inference happens in <20ms. The result returns to .NET, and the user sees a "Preliminary Offer" immediately. 3. The Contract-First Advantage 📝 One of the biggest headaches in Polyglot teams (.NET + Python) is API breaking changes. The .proto file: This is our "Single Source of Truth." Both teams agree on the input and output types. Auto-Generation: .NET generates its client, and Python generates its server from the same file. No more "Expected an integer but got a string" bugs in production. 4. Handling the "Timeout" Trap 🛡️ Direct calls are risky. If Python is slow, .NET hangs. The Strategy: We implement strict Deadlines. If Python doesn't answer in 100ms, the gRPC call cuts off. The Fallback: If the Fast-Lane fails, the system gracefully falls back to the "Deep-Dive" Async flow we discussed earlier. The user gets a "We're processing your request" message instead of a crash. 📈 The Results: ✅ Real-Time UX: Users get instant gratification. ✅ Polyglot Harmony: .NET and Python talk as if they were in the same project. ✅ Efficiency: Reduced CPU overhead on both sides compared to REST/JSON. 🧠 Post 6: The Watchtower – Real-time Observability with OpenTelemetry & Dashboards. #gRPC #Microservices #DotNet #Python #SystemDesign #FinTech #MLOps #API #SoftwareEngineering #PerformanceOptimization
To view or add a comment, sign in
-
More from this author
Explore related topics
- Open Source Frameworks for Building Autonomous Agents
- Building AI Applications with Open Source LLM Models
- Using LLMs as Microservices in Application Development
- Open Source AI Tools and Frameworks
- Building Scalable Applications With AI Frameworks
- How to Build Reliable LLM Systems for Production
- How to Streamline AI Agent Deployment Infrastructure
Explore content categories
- Career
- Productivity
- Finance
- Soft Skills & Emotional Intelligence
- Project Management
- Education
- Technology
- Leadership
- Ecommerce
- User Experience
- Recruitment & HR
- Customer Experience
- Real Estate
- Marketing
- Sales
- Retail & Merchandising
- Science
- Supply Chain Management
- Future Of Work
- Consulting
- Writing
- Economics
- Artificial Intelligence
- Employee Experience
- Workplace Trends
- Fundraising
- Networking
- Corporate Social Responsibility
- Negotiation
- Communication
- Engineering
- Hospitality & Tourism
- Business Strategy
- Change Management
- Organizational Culture
- Design
- Innovation
- Event Planning
- Training & Development