Europe is sitting on a wealth of public data – but much of its potential remains untapped. The challenges are well known: fragmented portals, incompatible interfaces, and growing reliance on non-European platforms that slow innovation. While new industrial data spaces are emerging – enabling secure and sovereign exchange of sensitive information – public and industrial data ecosystems remain largely siloed. This article explores how Artificial Intelligence (AI) and the Model Context Protocol (MCP) can help bridge that gap and accelerate Europe’s shift from Open Data to Open Knowledge—supporting digital sovereignty and delivering greater value to society.


Europe is sitting on vast amounts of underused data: Germany alone has more than 500 open data portals, yet fewer than 5% of that data is put to productive use.
Making data available is not enough: it has to be contextualized and connected to create real value.
AI and MCP can serve as a bridge: a federated knowledge architecture connects open data with industrial data spaces and makes both usable through a local LLM.
The shift from collecting data to creating knowledge is technically achievable and also strengthens digital sovereignty.
Europe possesses enormous public datasets, yet their economic and societal value has consistently fallen short of expectations. The causes are fragmented data portals, incompatible interfaces, and increasing dependencies on non-European platforms – all factors that hinder genuine innovation.
In parallel, new data spaces are emerging in industry, for example within the framework of Gaia-X, Catena-X, or the International Data Spaces (IDS) reference model. Unlike traditional open data portals, these data spaces enable secure, structured, and cross-industry exchange of sensitive information – based on standardized contracts, identity, and access mechanisms. Each participant retains control over their data while creating a trustworthy ecosystem for collaborative value creation.
As a result, two extensive but largely separate data worlds exist today:
Both areas contain vast knowledge – yet as long as there is no connecting, easily usable bridge, this potential remains untapped. In this article, you will learn how artificial intelligence (AI) and the Model Context Protocol (MCP) together pave the way from open data to open knowledge – and how you as a decision-maker can advance seamless networking of public and industrial datasets without endangering your company’s digital sovereignty.
Open data stands for the principle of providing administrative and research data openly, machine-readable, and without legal or technical access barriers. This concept is by no means new: as early as 2003, the PSI Directive (Public Sector Information) committed EU member states to make publicly funded data as accessible as possible. In 2019, the framework was strengthened again with the Open Data Directive – with the goal of boosting innovation, transparency, and economic growth. As a result, authorities provide a wide variety of datasets, from geodata and weather information to budget figures and traffic data – with the aspiration of enabling new services and insights from them.
However, reality looks different: there is a significant gap between aspiration and actual added value.
Open data is abundant, but its societal impact remains marginal.
Despite considerable investment, only a fraction of this potential is being exploited. In Germany, for example, there are over 500 open data portals – yet each works with its own metadata structures, formats, and interfaces. For developers, this means: complex integration, inconsistent formats, and poorly understandable documentation make usage difficult. On average, a portal counts fewer than 100 accesses per month, some datasets are virtually never used. Since 2010, over 250 million euros have been invested in such infrastructures, yet fewer than five percent of the provided data finds productive use.
The result: expensive isolated solutions emerge, unnecessary duplication of work, and frustration among all stakeholders. Additionally, dependency on global cloud providers grows – particularly for analysis, hosting, or processing services. This, in turn, contradicts the goal of strengthening Europe’s digital sovereignty.
In the following sections, this article explains how modern technologies like artificial intelligence (AI) and the Model Context Protocol (MCP) can help overcome these blockages and bring the vision of open knowledge within reach.
While open data portals primarily target the general public, so-called data spaces emerged in industry – especially since 2015 with the International Data Spaces (IDS) initiative by the Fraunhofer Institute. Their goal is to enable a trustworthy, decentralized data marketplace where companies can exchange sensitive information without losing sovereignty over this data. Projects like Gaia-X or its industry-specific implementations (Catena-X for the automotive industry, Manufacturing-X for production, Agricultural Data Space, etc.) build on this concept.
Key features of a data space:
This enables digital twins to be built along the entire value chain – from raw material procurement to recycling – providing real-time information about the condition, usage, and carbon footprint of a product. Machine suppliers, logistics service providers, suppliers, and operators thus receive a common but finely granulated situational picture without having to disclose their proprietary databases.
Data spaces thus complement the public open data ecosystem with high-resolution, domain-specific knowledge treasures. When both worlds are connected and made easily accessible for AI systems like large language models, a solid foundation for truly data-driven innovations emerges – from preventive maintenance via resilient supply chains to sustainable product cycles.
Digital sovereignty means control over data, infrastructure, and value creation. Three misconceptions stand in the way of genuine sovereignty:
Only when data is contextualized, accessible, and processable does value emerge. Wouldn’t it be easier to transfer the open data portals into data spaces?
Data space architectures – whether Gaia-X, Catena-X, Manufacturing-X, or the cross-sectional IDS reference – promise the holy grail of data sovereignty. Technically, they deliver:
Thus, the question Who may see which data under which conditions? is now well answered. However, the much more important question remains open: And for what purpose?
In short: Data spaces lay a secure pipeline, but the water still needs to be refined into drinking water. When AI-based services like LLMs and lightweight protocols like MCP make data automatically discoverable, semantically harmonized, and translatable into natural language, the “sovereignty-value creation gap” closes. Then genuinely usable knowledge emerges from sovereignly shared raw data – from supply chain resilience via digital twins to circular economy.
This section describes a concrete use case that shows how, through the combination of open data and data spaces using AI and MCP, a sovereign architectural pattern – the Federated Knowledge Architecture (FKA) – can emerge that creates genuine added value. Using the example of the planned construction of a manufacturing hall, it illustrates how this architectural pattern bridges open and domain-specific data spaces, enabling innovative knowledge landscapes.
Federated Knowledge Architecture (FKA) refers to an architectural pattern in which distributed knowledge and data services are federated via MCP. It connects open data sources and domain-specific data spaces into a sovereign knowledge layer for AI-supported analysis – with clear governance and without central data storage.
The Federated Knowledge System (FKS) ) is the concrete implementation of the FKA in an organization or ecosystem – including MCP servers, LLM orchestration, and domain adapters.
Planning and building a modern manufacturing facility is inherently complex. It involves pulling in data from a wide range of sources: public open data portals on soil quality, drinking water protection zones, and flood risks, as well as scientific and ecological considerations for sustainable construction to minimize environmental impact. At the same time, industry-specific data spaces need to be tapped to identify low-emission materials—whether construction materials or electrical components. Ideally, sustainability metrics should be calculated across the full lifecycle of the facility, not just during construction.
Today, this process is largely manual: consulting with multiple agencies, filling out forms, and navigating time-consuming coordination loops. Each step introduces potential delays and errors.
This is where the concept of a Federated Knowledge Architecture comes in. Using the Model Context Protocol (MCP) as a federated data translator – combined with locally hosted Large Language Models (LLMs)—this architectural pattern simplifies complexity and unifies access to all relevant data through a single intelligent layer.
How it works:
The Outcome: Rather than contacting agencies individually or relying on expert input, a project manager could submit a query like: “How do I design an environmentally optimized factory in a drinking water protection zone?” The system would return relevant regulations, environmental assessments, and supporting documentation – within seconds. These resources remain accessible throughout the entire construction lifecycle and can be continuously updated as the project evolves.
Technically, this setup relies on local LLMs enriched in real time via MCP servers, connected to public and private data sources. No migration of legacy systems is required, making implementation straightforward.
This approach fundamentally transforms how industrial building projects are planned and executed:
This is more than a technical innovation – it’s a blueprint for AI-augmented infrastructure that leverages Open Data and Data Spaces to move from fragmented silos to actionable Open Knowledge.
What does such an architecture look like in practice? The diagram below outlines a representative setup: it shows how LLMs, MCP servers, and data sources interact to deliver actionable insights to project stakeholders. For clarity, not all data sources from the example are shown. The MCP servers illustrated can be deployed either within the organization’s own Federated Knowledge Architecture or in an external system, such as one operated by an Open Data Portal provider.

Use Cases:
The example above demonstrates how different systems can interoperate effectively. The more MCP servers are connected, the richer and more complete the responses become. This spans use cases from generating summaries and recommending actions to pre-filling applications – or even controlling infrastructure components directly. The potential of this architectural pattern is significant: it enables organizations to turn existing data into actionable innovation. But how do you make it real?
Before implementing a Federated Knowledge Architecture, you need to define its core structure – specifically, how the MCP server and LLM will fit into your overall system landscape.
In early-stage projects, chat-based interfaces offer a lightweight and flexible way to get started. They allow teams to quickly prototype with LLMs, experiment with prompts, and develop MCP features tailored to specific domains. If needed, you can integrate a custom MCP client as a plugin within the chat environment.
As your proof of concept matures, and the complexity of domain logic grows, a structured user interface becomes more practical. Advanced use cases often require collecting more detailed inputs—something that’s much easier to manage through structured fields than free-text prompts. Structured interfaces also help scope your system appropriately. A construction-focused knowledge architecture doesn’t need access to zoological data – even if some environmental datasets touch on those areas.
Most organizations already have what they need: existing data silos, many of which are accessible via APIs. Even a basic database can be connected to an MCP server with minimal effort. The open-source MCP community continues to release ready-to-use integrations. For example, APIs described using OpenAPI can be automatically translated into MCP features. Prebuilt adapters also exist for many common databases.
In its initial form, an MCP server doesn’t need to do much. It simply acts as a bridge between the LLM and the data source. The logic for processing and interpreting the data lives in the prompt itself and is executed by the LLM. Over time, as more use cases are implemented, your MCP features can evolve to handle more complex logic. Selecting an LLM isn’t just about accuracy or benchmark scores. Deployment strategy, data governance, and regulatory fit all play a role. Here are key considerations:
What About gpt-oss (Open-Weight Model)?: The release of gpt-oss-120b and gpt-oss-20b, both available under Apache 2.0 licensing, has shifted the equation. Local deployment is now more affordable, more powerful, and more flexible. A Federated Knowledge Architecture can remain model-agnostic: switching to an open-weight model like gpt-oss is as simple as changing the model reference in your MCP layer—no architectural changes required. That said, governance remains critical. Prompt logging, evaluation pipelines, and access controls should be in place from the start.
Your LLM should be integrated in a way that allows for easy replacement down the line. That’s not just smart architecture – it’s a necessity. The LLM ecosystem is evolving fast, and models will almost certainly need to be swapped out over time. Reasons include cost optimization, improved performance, changing business needs, or new regulatory requirements.
Once your technical foundation is in place, the most important step is: get started. Thanks to years of investment in Big Data, Open Data, and Data Spaces, most enterprises and public-sector organizations already have data silos in place. Many of them are API-ready and just waiting to be connected. Chances are, there’s already a perfect starting point in your environment.
The discussion about open data and semantic interoperability culminates in a clear appeal: data must no longer be merely collected and provided; it must be strategically deployed to drive innovation. The vision of a Federated Knowledge Architecture, based on the Model Context Protocol (MCP) and the intelligent linking of open data with domain-specific industry data spaces, is pioneering in this regard.
Paradigm shift toward targeted data utilization: Instead of merely publishing data, open interfaces should be created that promote their semantic integration and AI-supported utilizability. The implementation of lightweight MCP servers enables rapid prototyping and supports connection to existing APIs and data sources.
Recommendations:
Final thought: It is up to Europe’s stakeholders to take the next step. From data collection to knowledge

We support you on your path to digital sovereignty, wherever you are today.