Why the Future of Data Engineering is About Orchestration, Not Just Code
AI generated image

Why the Future of Data Engineering is About Orchestration, Not Just Code

Over my 30+ years in enterprise data, I’ve navigated multiple paradigm shifts—from on-prem mainframes to the cloud, and from fragmented databases to unified Customer Data Platforms. But to truly understand how Generative AI is shifting our day-to-day workflows, I knew I needed to get into the weeds.

Today is Good Friday, and being a flexible holiday, I decided to go to office and sit with three of my engineering colleagues and learn vibe coding and how it can enable people like me who don't code. I sat down with my engineering team for a hands-on "vibe coding" session.

We purposefully chose a simple, internal use case as our sandbox: taking a static 2026 KPI spreadsheet and converting it into a live, automated HTML page and Power BI dashboard. I didn't just want to see it work; I wanted to see exactly where the AI excelled, and where it broke down in an enterprise context.

Initially, we thought we could just feed the raw data into a BI copilot and watch the magic happen. But we quickly hit the reality of current enterprise AI limits—whether it was dataset size constraints or the AI struggling to parse the unstructured schema of our raw metrics.

So, we pivoted. We took our basic wireframes and fed them into GitHub Copilot using natural English prompts. We iteratively prompted the AI to help us build a Markdown (MD) file that explicitly defined the semantic data model we needed. The AI generated the code, structured the relationships, and gave us the exact steps to import this newly minted semantic model into Power BI.

Watching the AI generate the boilerplate code in minutes validated a strategic thesis I’ve been developing: If AI can write the queries, build the HTML, and generate the dashboards, the role of the data professional isn't disappearing—it is scaling.

The barrier to entry for building data products has never been lower, which means our capacity for strategic innovation has never been higher. Based on our session today, here is how AI is fundamentally transforming what we do:

  • From Syntax Writers to Semantic Architects: The AI easily generated the code, but it still needed us to tell it what the data meant. During our session, we realized that while a Copilot can tell you how to connect BigQuery to Power BI, it doesn’t know which tables matter. It doesn’t inherently know the business logic behind a complex B2B fleet hierarchy, or how a "disposed vehicle" differs from an "active" one in our CDP.

The value of a data professional is shifting entirely away from memorizing SQL syntax. Instead, we are becoming Semantic Architects. Our job is to deeply understand the business context, structure the semantic layer, and ensure the AI has a clean, logical foundation to build upon.

  • From Builders to Orchestrators: We are no longer just building static data pipelines; we are orchestrating intelligent agents. During our vibe coding session, we didn't just want a one-off HTML page; we discussed creating an AI agent that automatically monitors a SharePoint folder, detects when a product manager updates the KPI sheet, and dynamically rewrites the code to update the live dashboard without human intervention.

In the near future, data engineers will manage fleets of AI agents that handle ingestion, transformation, and visualization, stepping in only to orchestrate the workflow and handle edge cases.

  • The Guardians of Context and Quality: Later in our session, we explored BigQuery’s new natural language Data plex Universal data catalog. The promise is incredible: a business user can type, "Find the table where customer and associated VINs are mapped," and the AI will fetch it.

But we immediately spotted a catch. Without human oversight, the AI might rank a deprecated backup table higher than the actual production table simply because it was modified more recently. AI lacks intuition. It needs data professionals to build the guardrails, enforce "Policy as Code," and curate the knowledge base so that when the AI answers a business question like, "How many vehicles does this enterprise customer have?" it delivers a trusted, accurate result.

  • The Death of the "Blank Page: The hardest part of analytics used to be the initial build—staring at a blank IDE or an empty canvas. Today, the time-to-value is shrinking to near zero. We can take a rough concept, feed it to an LLM, and get a working prototype in minutes.

Our job is now to refine, validate, and scale those prototypes. We are moving from a world of manual execution to a world of strategic curation. In the AI era, curiosity, business acumen, and problem-solving are the new programming languages.

I walked away from today’s session incredibly energized about the future of our industry. Sincere thanks to Vignesh Kumar G Anbu Ananada and Amridha Venkat for spending time with me today.

#ArtificialIntelligence #DataEngineering #CustomerDataPlatform #PowerBI #GenerativeAI #FutureOfWork #DataAnalytics #VibeCoding #Leadership #FordPro

Can't agree more with this line "In the AI era, curiosity, business acumen, and problem-solving are the new programming languages."

Like
Reply

This resonates hard. The best engineers I work with now spend 80% of their time orchestrating, not coding.

Like
Reply

We are building a few chatbots over copilot in power BI and discovering similar constraints - once the semantic model is ready, the time to market takes off! I don’t think you need too many dashboards though. Wouldn’t you just give a chatbot to your end user and let them ask more questions than they ever can to a dashboard?

Good insights, Venky! We’ve faced similar constraints with coiplot in Power BI as well. Really liked how your team pivoted to find an effective solution

Like
Reply

Great write up Venky. You always inspire us with your passion to learn and lead!

Like
Reply

To view or add a comment, sign in

More articles by Srinivasan Venkatesan

Others also viewed

Explore content categories