Score: 1

Safe, Untrusted, "Proof-Carrying" AI Agents: toward the agentic lakehouse

Published: October 10, 2025 | arXiv ID: 2510.09567v1

By: Jacopo Tagliabue, Ciro Greco

Potential Business Impact:

Lets AI fix data without breaking it.

Business Areas:
Machine Learning Artificial Intelligence, Data and Analytics, Software

Data lakehouses run sensitive workloads, where AI-driven automation raises concerns about trust, correctness, and governance. We argue that API-first, programmable lakehouses provide the right abstractions for safe-by-design, agentic workflows. Using Bauplan as a case study, we show how data branching and declarative environments extend naturally to agents, enabling reproducibility and observability while reducing the attack surface. We present a proof-of-concept in which agents repair data pipelines using correctness checks inspired by proof-carrying code. Our prototype demonstrates that untrusted AI agents can operate safely on production data and outlines a path toward a fully agentic lakehouse.

Repos / Data Links

Page Count
5 pages

Category
Computer Science:
Artificial Intelligence