Light Differentiable Logic Gate Networks
By: Lukas Rüttgers , Till Aczel , Andreas Plesner and more
Potential Business Impact:
Makes AI learn faster and use less memory.
Differentiable logic gate networks (DLGNs) exhibit extraordinary efficiency at inference while sustaining competitive accuracy. But vanishing gradients, discretization errors, and high training cost impede scaling these networks. Even with dedicated parameter initialization schemes from subsequent works, increasing depth still harms accuracy. We show that the root cause of these issues lies in the underlying parametrization of logic gate neurons themselves. To overcome this issue, we propose a reparametrization that also shrinks the parameter size logarithmically in the number of inputs per gate. For binary inputs, this already reduces the model size by 4x, speeds up the backward pass by up to 1.86x, and converges in 8.5x fewer training steps. On top of that, we show that the accuracy on CIFAR-100 remains stable and sometimes superior to the original parametrization.
Similar Papers
Mind the Gap: Removing the Discretization Gap in Differentiable Logic Gate Networks
Machine Learning (CS)
Makes smart computer pictures learn much faster.
Recurrent Deep Differentiable Logic Gate Networks
Machine Learning (CS)
Makes computers learn by thinking like simple switches.
eXpLogic: Explaining Logic Types and Patterns in DiffLogic Networks
Machine Learning (CS)
Explains how computer brains make decisions.