LLM Agents for Interactive Workflow Provenance: Reference Architecture and Evaluation Methodology
By: Renan Souza , Timothy Poteet , Brian Etz and more
Potential Business Impact:
Lets scientists ask questions about data using normal talk.
Modern scientific discovery increasingly relies on workflows that process data across the Edge, Cloud, and High Performance Computing (HPC) continuum. Comprehensive and in-depth analyses of these data are critical for hypothesis validation, anomaly detection, reproducibility, and impactful findings. Although workflow provenance techniques support such analyses, at large scale, the provenance data become complex and difficult to analyze. Existing systems depend on custom scripts, structured queries, or static dashboards, limiting data interaction. In this work, we introduce an evaluation methodology, reference architecture, and open-source implementation that leverages interactive Large Language Model (LLM) agents for runtime data analysis. Our approach uses a lightweight, metadata-driven design that translates natural language into structured provenance queries. Evaluations across LLaMA, GPT, Gemini, and Claude, covering diverse query classes and a real-world chemistry workflow, show that modular design, prompt tuning, and Retrieval-Augmented Generation (RAG) enable accurate and insightful LLM agent responses beyond recorded provenance.
Similar Papers
LLM Agents for Interactive Workflow Provenance: Reference Architecture and Evaluation Methodology
Distributed, Parallel, and Cluster Computing
Lets scientists ask questions about data using normal words.
PROV-AGENT: Unified Provenance for Tracking AI Agent Interactions in Agentic Workflows
Distributed, Parallel, and Cluster Computing
Tracks AI agent mistakes to make them more reliable.
PROV-AGENT: Unified Provenance for Tracking AI Agent Interactions in Agentic Workflows
Distributed, Parallel, and Cluster Computing
Makes AI agents trustworthy and their work reliable.