A General Framework for Property-Driven Machine Learning
By: Thomas Flinkow , Marco Casadio , Colin Kessler and more
Potential Business Impact:
Teaches computers to follow rules, not just guess.
Neural networks have been shown to frequently fail to learn critical safety and correctness properties purely from data, highlighting the need for training methods that directly integrate logical specifications. While adversarial training can be used to improve robustness to small perturbations within $\epsilon$-cubes, domains other than computer vision -- such as control systems and natural language processing -- may require more flexible input region specifications via generalised hyper-rectangles. Differentiable logics offer a way to encode arbitrary logical constraints as additional loss terms that guide the learning process towards satisfying these constraints. In this paper, we investigate how these two complementary approaches can be unified within a single framework for property-driven machine learning, as a step toward effective formal verification of neural networks. We show that well-known properties from the literature are subcases of this general approach, and we demonstrate its practical effectiveness on a case study involving a neural network controller for a drone system. Our framework is made publicly available at https://github.com/tflinkow/property-driven-ml.
Similar Papers
Attack logics, not outputs: Towards efficient robustification of deep neural networks by falsifying concept-based properties
Cryptography and Security
Makes AI understand things more logically and safely.
Propositional Logic for Probing Generalization in Neural Networks
Machine Learning (CS)
Computers struggle to learn logic rules.
Categorical Construction of Logically Verifiable Neural Architectures
Logic in Computer Science
Makes AI think logically, preventing mistakes.