MCPs and location data: Why reference data is the hardest context for AI
TL;DR
The Model Context Protocol (MCP) is an open-source standard that enables Large Language Models (LLMs) to access files, databases, and APIs.
I think of MCPs as a bridge between language models and the external world. MCP clients such as Claude, Cursor, and ChatGPT now integrate with MCP servers, enabling AI to interact with everything from calendars to APIs.
But early implementations have exposed challenges that I've been watching closely. As Kent C. Dodds (world-renowned coding expert and speaker) observes, MCP has entered its natural critique phase, similar to early web browsers that were slow and inefficient but evolved into today's dominant platform. The protocol works; people use it extensively every day. We're now in the phase of making it right, then making it fast.
In my experience working with location data, I've seen a specific tension emerge: while LLMs and MCP handle data reasonably well, they struggle to grasp the difference between reference data and other data. And within reference data, location information is the hardest test case.
Understanding MCP
The Model Context Protocol defines two key components: MCP clients and MCP servers. Clients are embedded in AI tools like Claude, Cursor, or custom applications. Servers are hosted by organizations providing data or tools for LLM consumption.
When an LLM needs external capabilities, it uses its client to connect to an MCP server, discover available tools and resources, and determine how to leverage them to complete its task. The protocol standardizes this discovery and invocation process, making it theoretically simple to connect AI to any service.
The elegance of this design explains MCP's rapid adoption. But the protocol itself is only infrastructure. What matters — and what I care most about — is what flows through it.
The challenge of using MCPs with location data
Reference data is structured, authoritative, and foundational. Location reference data — postal codes, administrative boundaries, time zones, and coordinates — is a clear example. Unlike conversational text, reference data must be correct.
Geographic information also involves complex hierarchies: cities contain neighborhoods, regions contain cities, and countries contain regions.
What makes this particularly difficult, in my view, is that location data is constantly evolving. Postal codes are introduced and retired. Municipalities merge or split. Administrative divisions are redrawn. Maintaining both accuracy and backward compatibility requires careful structuring.
These decisions require domain expertise. Without that expert-level context, LLMs are likely to provide incorrect answers confidently, what the field calls hallucinations.
The hallucination problem with location data in LLMs
LLMs generate responses through statistical analysis of their training data. They identify patterns and generate text that follows them. For most conversational tasks, this works well. For reference location data, it creates a fundamental problem.
The location data in LLM training sets is often outdated, inconsistent, or incomplete. Models have no mechanism to recognize this; they don't know which information is current and which is stale. They can't distinguish authoritative sources from casual mentions. They simply generate responses that match statistical patterns in what they've seen.
Recommended by LinkedIn
I'll give a concrete example we've encountered directly. Consider postal codes in China. The country maintains alternative postcodes for certain administrative divisions, but information about these is intentionally opaque. When we cross-referenced available sources and compared them to AI-generated responses, we found significant discrepancies. The AI confidently provided postal codes that didn't match any authoritative source. It wasn't lying; it was generating statistically plausible patterns based on incomplete training data.
The downstream consequences are real. A customer service system that routes packages based on AI-provided postal codes will send them to the wrong place. A logistics platform that validates addresses against hallucinated data will reject legitimate locations. An analytics system that aggregates data by administrative division will produce meaningless results when the boundaries are wrong.
The challenge of choosing a reliable data source
Even outside of MCP and AI systems, this is a problem I see organizations struggle with constantly. Maintaining reliable location data is difficult. Postal codes, administrative divisions, boundaries, and coordinates form one of the most critical reference data domains because they underpin logistics, taxation, reporting, and regulatory processes. Maintaining this data requires significant effort and expertise.
Postal systems evolve constantly: municipalities merge or split, administrative boundaries change, and postal codes are introduced or retired. Ensuring accuracy requires continuous validation against authoritative sources.
Many organizations underestimate this effort. When reference data becomes fragmented, outdated, or inconsistent across systems, downstream platforms inherit those inconsistencies. Because maintaining this layer requires continuous reconciliation of multiple authoritative sources, many organizations rely on specialized data providers that curate and maintain global location reference datasets.
Conclusion: Use-case-driven MCP design with quality location data
From where I sit, your MCP's maturity depends on parallel evolution in two areas: server design and data quality.
Well-designed MCP servers are tailored to specific use cases, not generic wrappers around existing non-MCP optimized APIs. They minimize token consumption, return actionable responses, and reduce errors. They understand that serving an LLM is fundamentally different from serving a programmer.
MCP amplifies the consequences of poor data. When LLMs interact with APIs, stale or inconsistent data produces wrong answers and propagates errors at scale through hallucinations and misinterpretation.
This is precisely why curated datasets with active governance are becoming increasingly critical in AI workflows. An expert provider needs to cross-reference sources, track changes over time, and maintain data freshness. AI can't do this work reliably because it requires knowing which sources are authoritative, knowledge that comes from domain expertise, not statistical patterns.
That distinction is one I've spent fifteen years thinking about. It doesn't get simpler as the tooling around it becomes more powerful.
Looking for self-hosted ZIP code or address data? At GeoPostcodes, we provide high-quality data to integrate as a global reference layer.
👉 Discover GeoPostcodes' self-hosted data offering.
Until next time,
Jérôme Urbain
Head of Products at GeoPostcodes
Really interesting piece. Couldn’t agree more that reference data is becoming the critical limiter in any AI‑led location or address solution. What we’re seeing both across the industry and from our own partners is that the real challenge isn’t just accuracy, it’s the speed of change, the layers of context needed for automation, and the operational knock‑on effects when any part of that foundation slips. Good to see more focus on treating reference data as core infrastructure rather than an afterthought. It’s a conversation the industry definitely needs more of.
Bachelor in Geography with special interest in GIS and Remote Sensing based on open source systems
1moSofia Di Croce