Discovering and Learning Probabilistic Models of Black-Box AI Capabilities
By: Daniel Bramblett , Rushang Karia , Adrian Ciotinga and more
Black-box AI (BBAI) systems such as foundational models are increasingly being used for sequential decision making. To ensure that such systems are safe to operate and deploy, it is imperative to develop efficient methods that can provide a sound and interpretable representation of the BBAI's capabilities. This paper shows that PDDL-style representations can be used to efficiently learn and model an input BBAI's planning capabilities. It uses the Monte-Carlo tree search paradigm to systematically create test tasks, acquire data, and prune the hypothesis space of possible symbolic models. Learned models describe a BBAI's capabilities, the conditions under which they can be executed, and the possible outcomes of executing them along with their associated probabilities. Theoretical results show soundness, completeness and convergence of the learned models. Empirical results with multiple BBAI systems illustrate the scope, efficiency, and accuracy of the presented methods.
Similar Papers
Investigating Advanced Reasoning of Large Language Models via Black-Box Interaction
Artificial Intelligence
Teaches computers to figure out hidden rules.
Integrating Machine Learning into Belief-Desire-Intention Agents: Current Advances and Open Challenges
Artificial Intelligence
Helps smart robots make better decisions.
Agential AI for Integrated Continual Learning, Deliberative Behavior, and Comprehensible Models
Artificial Intelligence
AI learns and plans like a person.