Building Trust into the Data Lakehouse

Author: Denis Avetisyan


A new architectural approach leverages database principles to ensure safe concurrency and robust governance in complex, agent-driven data pipelines.

A self-correcting system leverages a ReACT loop within an agentic lakehouse, employing an initial verification step before deferring to human confirmation through a branch-then-merge workflow, ensuring robust and reliable outcomes.
A self-correcting system leverages a ReACT loop within an agentic lakehouse, employing an initial verification step before deferring to human confirmation through a branch-then-merge workflow, ensuring robust and reliable outcomes.

This review proposes the ‘Agentic Lakehouse’, adapting Multi-Version Concurrency Control for data and compute isolation in modern data architectures.

Despite rapid advances in artificial intelligence, enterprise adoption remains hampered by concerns regarding the trustworthiness of agents accessing production data. This paper, ‘Trustworthy AI in the Agentic Lakehouse: from Concurrency to Governance’, argues that building reliable agentic workflows necessitates a fundamental rethinking of data lakehouse architecture. We propose ‘Bauplan’, an agent-first design that adapts database principles-specifically, multi-version concurrency control-to address the unique challenges of decoupled, multi-language data pipelines. Can this approach unlock the full potential of AI agents while ensuring data integrity and robust governance in the modern data landscape?


The Concurrency Challenge in Modern Data Lakehouses

Modern data lakehouses, designed for expansive data storage and analytical processing, often encounter significant challenges when multiple users or applications attempt to access and modify data simultaneously. This concurrency issue stems from the lakehouse architecture’s reliance on object storage, which doesn’t inherently provide the fine-grained locking mechanisms found in traditional databases. Consequently, concurrent operations can lead to data corruption, inconsistent results, and substantial performance degradation as the system struggles to manage conflicting requests. While lakehouses excel at handling large datasets, the lack of robust concurrency control necessitates careful design considerations and often requires implementing complex, custom solutions to ensure data integrity and reliable performance under heavy load. These challenges are particularly acute in real-time analytics and operational applications where low latency and consistent data are critical.

Historically, data management relied on robust, yet rigid, transactional systems – monolithic databases designed to guarantee data consistency through strict rules and immediate confirmations. These systems excel at preventing data corruption and ensuring accuracy in scenarios like financial transactions, but their architecture inherently limits scalability and adaptability. Unlike the distributed, open formats favored by modern lakehouses, monolithic databases struggle to efficiently handle the volume, velocity, and variety of data characteristic of contemporary applications. This inflexibility hinders their ability to support the diverse analytical workloads and evolving data schemas demanded by data science and machine learning initiatives, creating a bottleneck as organizations seek to unlock the full potential of their data assets.

Utilizing a Bauplan run API ensures atomic updates of interdependent components, guaranteeing consistency and isolation during pipeline execution, unlike traditional methods where version mismatches can occur.
Utilizing a Bauplan run API ensures atomic updates of interdependent components, guaranteeing consistency and isolation during pipeline execution, unlike traditional methods where version mismatches can occur.

Bauplan: A Git-Inspired Approach to Data Consistency

Bauplan utilizes a data architecture inspired by Git version control, addressing concurrency challenges in lakehouse environments. Traditional lakehouses often struggle with simultaneous data modifications, leading to conflicts and data inconsistency. Bauplan introduces the concept of temporary, isolated branches for data manipulation, mirroring Git’s branching strategy. This allows multiple users or processes to operate on data independently without impacting the core dataset or other ongoing operations. Changes made within these branches are then integrated using an atomic merge process, ensuring data integrity and a complete audit trail of all modifications, similar to Git’s commit history. This approach enables a more collaborative and robust data engineering workflow by facilitating parallel data transformations and analysis.

Bauplan utilizes temporary branches as isolated environments for data manipulation, allowing multiple concurrent operations on a lakehouse without impacting the established data lineage. These branches function as independent workspaces where transformations, cleaning, or enrichment can occur. Changes within a branch are not visible or applied to the main data set until explicitly merged. This isolation prevents conflicting modifications and ensures that the core data remains consistent during ongoing analytical or processing tasks. The branching model facilitates parallel data workstreams, improving efficiency and reducing contention for resources within the data lakehouse.

Atomic merge, within the Bauplan architecture, functions by treating data transformations as discrete, versioned commits. Each change is isolated until explicitly merged into the main data lineage, and the merge operation itself is all-or-nothing; it either completes successfully, integrating the changes as a single unit, or fails entirely, leaving the lakehouse in its original state. This approach prevents partial updates and ensures data consistency by maintaining a verifiable history of all changes. The system utilizes techniques such as optimistic locking and conflict detection during the merge process to guarantee that concurrent modifications do not corrupt the data, effectively providing ACID properties for data transformations within the lakehouse.

Self-Healing Pipelines Through Autonomous Agents and Verification

Bauplan enables the construction of self-healing data pipelines by combining autonomous agents with computational verification. These agents continuously monitor data flow and pipeline performance, identifying inconsistencies or failures. The system employs a layered approach where agents, upon detecting an issue, propose and implement corrections. Crucially, all agent-driven modifications are subject to validation by dedicated verifiers. These verifiers utilize pre-defined acceptance criteria – essentially, computational rules – to assess the quality and validity of the proposed fix before it is permanently applied, ensuring data integrity and pipeline stability. This integration of proactive agents and rigorous verification constitutes the core mechanism for automated self-healing within the Bauplan framework.

Agents within Bauplan utilize ReAct (Reason + Act) reasoning to address data inconsistencies in pipelines autonomously. This process involves the agent first observing a discrepancy, then formulating a plan of action based on the observed state and pre-defined knowledge. Following the reasoning step, the agent executes the plan by initiating repair actions, such as data transformation, correction, or re-processing. The ReAct cycle allows agents to dynamically respond to issues without human intervention, improving pipeline resilience and reducing manual maintenance efforts. The system logs each reasoning step and action for auditability and further refinement of agent behavior.

Verifiers within Bauplan pipelines function as computational acceptance criteria, rigorously evaluating the corrections proposed by agents. These verifiers are not simply boolean pass/fail checks; they execute defined computations on the modified data to assess validity against pre-established standards. This process ensures that agent-driven repairs not only address identified inconsistencies but also maintain data integrity and adhere to specified business rules. The computational nature of these verifiers allows for nuanced validation beyond simple data type or range checks, incorporating complex logic and dependencies to confirm the overall quality of the corrected data before it propagates further down the pipeline.

Bauplan’s Architectural Foundation: Scalability, Security, and Governance

Bauplan’s core architecture strategically employs Function-as-a-Service (FaaS) to establish granular network isolation between its various components. This approach moves away from traditional, monolithic network configurations and instead distributes functionality into independently deployable, stateless functions. Each function operates within a strictly defined security perimeter, significantly reducing the potential attack surface and limiting the blast radius of any security breaches. Beyond security, this FaaS implementation enables exceptional scalability; individual functions can be automatically scaled up or down based on demand, optimizing resource utilization and ensuring consistent performance even during peak loads. By embracing a serverless, event-driven model, Bauplan achieves a highly resilient and adaptable infrastructure capable of handling dynamic workloads with efficiency and security.

Bauplan’s data handling is streamlined through a declarative I/O approach, fundamentally altering how applications interact with storage. Rather than specifying how to retrieve or process data, applications simply declare what data is needed, allowing the system to optimize the access patterns automatically. This shift dramatically reduces the complexity for developers, eliminating the need to write intricate data access code and minimizing potential errors. Consequently, performance is significantly improved as the system can leverage efficient storage mechanisms and parallel processing techniques, tailoring data delivery to application requirements. The declarative model also fosters greater flexibility and scalability, enabling Bauplan to seamlessly adapt to evolving data volumes and diverse application workloads without requiring extensive code modifications.

Bauplan’s foundational architecture deliberately integrates established industry standards to ensure data integrity and secure access. Specifically, the system employs Apache Iceberg, an open table format, to guarantee transactional consistency – meaning all data operations either fully succeed or fully fail, preventing partial writes and ensuring reliable data states. Complementing this is a robust Role-Based Access Control (RBAC) system, which meticulously defines and enforces permissions based on user roles, rather than individual identities. This approach not only simplifies user management but also significantly enhances security by limiting data access to authorized personnel and processes, creating a highly governed and auditable data environment. By building upon these proven technologies, Bauplan minimizes risk and maximizes interoperability with existing data infrastructure.

Envisioning the Future of Data Engineering: Trustworthy and Autonomous Systems

Bauplan introduces a novel approach to data pipeline construction centered around autonomous “agents” that proactively monitor and remediate issues, marking a departure from traditional, static configurations. These agents operate within existing data lakehouse ecosystems – leveraging technologies like Delta Lake and Apache Iceberg – to dynamically adapt to changing data schemas, quality drifts, and system failures. This integration isn’t about replacing current infrastructure, but rather augmenting it with a layer of intelligent automation; agents can automatically detect and correct data inconsistencies, optimize pipeline performance, and even self-heal from errors without human intervention. The result is a data infrastructure that isn’t merely reactive to problems, but actively anticipates and resolves them, promising increased reliability, reduced operational overhead, and ultimately, a more trustworthy data foundation for data-driven decision-making.

The evolving architecture of modern data engineering, particularly with systems like Bauplan, promises substantial gains in data quality, reliability, and scalability-critical needs in the current era of big data. Traditional pipelines often struggle with the velocity and variety of incoming information, leading to data inconsistencies and system failures; however, agent-first approaches enable automated monitoring, self-healing capabilities, and dynamic adjustments to pipeline configurations. This proactive stance minimizes errors and ensures data integrity, while the ability to dynamically scale resources in response to fluctuating data volumes guarantees consistent performance even under peak loads. Ultimately, this architectural shift moves beyond reactive problem-solving toward a predictive and resilient data infrastructure, unlocking the full potential of big data analytics and driving more informed decision-making.

Bauplan proposes a fundamental shift in data engineering, focusing on the creation of a demonstrably trustworthy infrastructure. The system doesn’t merely process data; it actively verifies and validates each stage of the pipeline, building confidence in the data’s integrity from source to consumption. This is achieved through automated testing, continuous monitoring, and a self-healing architecture that proactively addresses potential issues before they impact downstream applications. By prioritizing data quality and reliability as core design principles, Bauplan aims to move beyond simply having data to possessing data that is consistently accurate, complete, and dependable – a critical need in an era increasingly driven by data-informed decision-making.

The Agentic Lakehouse, as proposed, necessitates a fundamental shift in how data systems are conceived. It moves beyond merely storing and processing data to actively governing interactions within the lakehouse itself. This architecture implicitly acknowledges that systems break along invisible boundaries – if one cannot clearly define and enforce isolation between agents and data versions, pain is coming. John McCarthy famously stated, “The best way to predict the future is to invent it.” This rings true here; the Agentic Lakehouse isn’t simply reacting to the challenges of concurrency and governance, but proactively inventing a future where these concerns are addressed through declarative programming and robust MVCC principles, effectively building a more predictable and trustworthy data environment.

The Horizon Beckons

The Agentic Lakehouse, as presented, addresses a critical, if often overlooked, truth: scaling compute without commensurate attention to data integrity is a recipe for systemic fragility. The adaptation of Multi-Version Concurrency Control to this distributed landscape is a logical, even inevitable, step. However, the architecture’s true test will not be in demonstrating isolation – that much is established database practice – but in revealing the performance trade-offs inherent in maintaining such guarantees across increasingly heterogeneous pipelines. The devil, predictably, resides in the details of implementation and the cost of coordination.

A particularly intriguing, and currently underexplored, aspect concerns the interplay between declarative programming and agent autonomy. While declarative definitions promise a degree of predictability, true agency implies a capacity for improvisation. Resolving this tension – ensuring agents operate within defined boundaries while still exhibiting intelligent behavior – demands a deeper understanding of how to formalize trust and accountability within the lakehouse itself.

Ultimately, the success of this paradigm will hinge not on technological novelty, but on a philosophical shift. The lakehouse is not merely a repository of data; it is an ecosystem. Treating its components as isolated entities, solvable through independent optimization, will inevitably lead to emergent failures. A holistic view, recognizing that modifying one part of the system triggers a domino effect, is paramount. The question is not whether the Agentic Lakehouse can be built, but whether the field is prepared to embrace the systemic thinking required to build it well.


Original article: https://arxiv.org/pdf/2511.16402.pdf

Contact the author: https://www.linkedin.com/in/avetisyan/

See also:

2025-11-23 22:11