Data Strategy Only in God may we trust The rest bring Data
Data Strategy Only in God may we trust The rest bring Data
4. SAP gets the point and releases BDC with Fabric, Mesh, Dashboards
and Gen AI
Traditionally, building analytics in SAP primarily built around back end and
presentation layers, Business Warehouse (BW) and Business Objects (BO)
gave enterprise reporting and data analysis for many organizations. SAP BW
provided the data warehousing platform, centralizing data from mostly SAP
sources into a structured environment optimized for reporting. On the
presentation layer, Business Objects offered a suite of tools for creating
reports, dashboards, and ad-hoc queries, it has given business users with
self-service analytics.
The rise of cloud computing, big data, and real-time analytics brought
different architectures more flexible, scalable. The focus shifted towards
cloud-based platforms like SAP Analytics Cloud and embedded analytics
within S/4HANA, the raise of cloud hyperscalers and modern analytics
platforms, first Snowflake and then Databricks, got the idea perfectly and
leveraged a nice combination of technological improvements and platform
modernity. The medallion architecture was born.
In this blog, I will go through the new offering from SAP, Business Data Cloud
(BDC from now on), the vision to unify applications, data, and AI, with a core
focus on Data Products.
Spoiler
In the world of data management, the Medallion architecture, also known as
multi-hop architecture, is an approach to data model design that encourages
the logical organisation of data within a data lakehouse.
This concept, called Data Mesh is perfect for LLM consumption, but came
before LLMs, its accidental feature. Introduced here in 2020 describes that
data monolith (being a data lake or warehouse), is many times a bottleneck
when organizations grow and data complexity increases. Data Lakes, coming
from the old days of Big Data, lack agility and scaling, because data needs to
move over and offer to the next phase of "curation".
By Author: Data mesh conceptual diagram showing decentralized data
ownership
Data becomes a DaaS product at the point of consumption, not merely at the
point of collection or storage. It's like preparing a meal, you collect
ingredients which are not consumed until you have a final dish, and you
collect the ingredients you want, not the other way around, being given
ingredients "what can I do with all this".
Also, the lifecycle of the data is critical in its design. Adopting a product-
centric mindset within data teams. Data teams should possess a strong
understanding of business needs and user workflows. This requires a shift in
perspective, viewing data not just as a technical asset but as a product that
needs to be carefully crafted, marketed, and supported to ensure user
adoption and satisfaction.
SAP has long been player in Data since the introduction of BW in 1998. This
was the dawn of an era where organizations sought to consolidate and
analyze their business data. SAP BW evolved through capabilities in ETL,
data modeling, and integration, ultimately the in-memory HANA around
2015.
The introduction of SAP Data Warehouse Cloud, around 2019, which was later
SAP Datasphere in 2023, and built on the BTP, has been the first business
data fabric architecture, because it aimed the consolidation of critical data
from various sources, on HANA Cloud. This architecture, with its focus on
integrating diverse data landscapes, aligns with the decentralized and
interconnected nature of a Data Mesh. SAP Datasphere provides a
centralized catalog for data discovery and governance across these
interconnected sources. The Catalog is important, and I will discuss later on.
This 2025, the introduction of the SAP Business Data Cloud (BDC) is now the
future of SAP's data and analytics strategy. It integrates the strengths of SAP
BW, SAP Datasphere, and SAP Analytics Cloud (SAC) on a single platform.
What is really new is the concept of data products, central to SAP BDC, with
SAP aiming to deliver out-of-the-box data products following a harmonized
data model. This strong emphasis on data products as fundamental building
blocks clearly echoes the core tenets of a Data Mesh.
LLMs are trained on vast amounts of unstructured data; they learn about the
data and store this information as part of weights and biases in their neural
network; this information includes the language understanding and the
general knowledge provided in the training data.
To date, off-the-shelf LLMs are not prepared with structured and relevant
data for enterprise use cases. The most popular use case for enterprises is to
be able to query their extensive set of tables and data lakes using LLMs.
Here are two broad ways in which LLMs are being used in enterprises today:
Organizations may choose to organize their data with defined schemas and
entities in catalogs before using LLMs, which helps the LLMs understand the
data and improves accuracy and efficiency. However, this method requires
ongoing updates, involves data movement, and has high upfront costs for
organizing and cataloging large datasets. Additionally, even with this
structure, LLMs may still not fully comprehend the data's context and
semantics, which results in inaccuracies.
By Author; LLM
with data catalogs diagram showing the addition of data catalog between
query engine and data lakes
Enter the data product era! A data product is a comprehensive solution that
integrates Data, Metadata (including semantics), Code (transformations,
workflows, policies, and SLAs), and Infrastructure (storage and compute). It is
specifically designed to address various data and analytics (D&A) and AI
scenarios, such as data sharing, LLM training, data monetization, analytics,
and application integration. Across various industries, organizations are
increasingly turning to sophisticated data products to transform raw data
into actionable insights, enhancing decision-making and operational
efficiency. Data products are not merely a solution; they are transformative
tools that prepare and present data effectively for AI consumption—trusted
by its users, up-to-date, and governed appropriately.
B
y Author; LLM with data product layer diagram showing how data products
sit between LLM and data lakes
At the heart of SAP BDC lies the concept of Data Products. SAP recognizes
that data is only valuable if it's accessible, understandable, and trustworthy.
Data Products in BDC are not just raw data; they are curated, enriched, and
contextualized data assets designed for specific business purposes. They are
the core components of the Business Data Cloud.
Within SAP, a Data Product is a data set made available for use outside its
original application through APIs. It comes with detailed, high-quality
descriptions accessible via a Data Product Catalog. It's important to note
that "Data Product" doesn't mean something you purchase isolated; it
simply refers to data that is "packaged" for straightforward use.
Cash Flow Data Product example showing interface elements and properties
Consumable: via APIs or via Events. Supported API types are SQL (incl.
SQL interface views), Delta Sharing, Event, REST, (oData)
Described: with metadata that is of high quality and provided via Open
Resource Discovery (ORD), following ORD schema for Data Product*
ORD will be explained in a minute
The core components of SAP BDC architecture work together to support the
Data Product lifecycle:
SAP Datasphere
Datasphere serves as the central hub for integrating data from various
sources, both SAP and non-SAP. It plays a crucial role in harmonizing data
across different formats and structures, ensuring consistency and
compatibility. Through its data modeling and transformation capabilities,
Datasphere creates the foundation for high-quality Data Products by
providing a unified semantic layer that bridges technical data storage and
business meaning.
SAP Databricks
The journey of creating and consuming Data Products in SAP BDC follows a
structured process that ensures quality, governance, and accessibility. This
process encompasses multiple steps, from initial data package activation to
the creation of business value through Data Product consumption.
SAP BDC offers seamless integration with SAP Business Warehouse, allowing
organizations to leverage their existing BW investments while moving toward
a modern data product approach. With the introduction of SAP BDC,
customer-managed BW data products can be transitioned to SAP-managed
data products.
Customers will first work with BW data products based on their BW data
models from SAP BW, starting use cases in SAP Databricks and SAP
Datasphere, with exposure via Delta Share. As they begin exploring what's
possible with SAP-managed data products and insight apps, customers can
gradually replace BW Data Products with SAP-managed Data Products as
they move into a clean data approach.
BW can publish "BW data products" into the SAP BDC object store. We could
use SAP Databricks delta share capabilities to expose those data products to
3rd party data lakes. We could also access 3rd party data lakes from BDC via
delta share, and for example, join data from BW with data from Azure Data
Lake in a Datasphere data layer.
The Role of Open Resource Discovery (ORD)
A real-world impact of Data Products within SAP BDC, let's consider a finance
department struggling with cash flow visibility and forecasting accuracy.
Before implementing a Data Product approach, the finance team dealt with
fragmented data sources: bank transaction data in SAP S/4HANA, accounts
receivable in an Ariba system, and forecasting data in various Excel
spreadsheets. Analysts spent days each month reconciling and consolidating
this data, often discovering discrepancies too late to inform decision-making.
Finance analysts now access this Data Product through SAP Analytics Cloud,
where they can immediately visualize current cash positions, analyze trends,
and generate accurate forecasts. The Data Product ensures that all data is
current, consistent, and properly contextualized with business meaning.
By Author
The Data Product approach within SAP Business Data Cloud delivers several
key advantages that collectively transform how organizations leverage their
data assets:
TL:DR
In the age of AI, the ability to link business understanding with strategic
choices is vital. As every company becomes an AI data creator, strong
business context becomes necessary to get AI ready for real-world use.
Building a solid base for Business AI means bringing together data systems
and business operations.
The data product approach fundamentally shifts how we think about data,
moving from a technical resource to be managed to a strategic asset to be
leveraged. It brings business context to the forefront, ensuring that data is
not just stored but is truly useful and actionable.
In SAP, the new Business Data Cloud, with Data Products, moves into how
organizations manage and leverage their data in this new paradigm by
treating data as a product—discoverable, accessible, trustworthy, and
natively usable.