Light differentiable logic gate networks

Illustration of reparametrization

Abstract

Differentiable logic gate networks (DLGNs) exhibit extraordinary efficiency at inference while sustaining competitive accuracy. But vanishing gradients, discretization errors, and high training cost impede scaling these networks. With our reparametrization, we tackle these shortcoming and make DLGNs more scalable in depth and complexity. The implementation of our reparametrization is available at GitHub.

Presentation

To navigate through the slides yourself, open the presentation here.