ELLIS header
University of Stuttgart Logo
Max Planck Institute for Intelligent Systems Logo

State-Regularized Recurrent Neural Networks

Cheng Wang, Mathias Niepert

Proc. of the 36th International Conference on Machine Learning (ICML), pp. 6596–6606, 2019.


Abstract

Recurrent neural networks are a widely used class of neural architectures with two shortcomings. First, it is difficult to understand what exactly they learn. Second, they tend to work poorly on sequences requiring long-term memorization, despite having this capacity in principle. We aim to address both shortcomings with a class of recurrent networks that use a stochastic state transition mechanism between cell applications. This mechanism, which we term state-regularization, makes RNNs transition between a finite set of learnable states. We evaluate state-regularized RNNs on (1) regular languages for the purpose of automata extraction; (2) nonregular languages such as balanced parentheses, palindromes, and the copy task where external memory is required; and (3) real-word sequence learning tasks for sentiment analysis, visual object recognition, and language modeling. We show that state-regularization simplifies the extraction of finite state automata from the RNN’s state transition dynamics; forces RNNs to operate more like automata with external memory and less like finite state machines; and makes RNNs more interpretable.

Links


BibTeX

@inproceedings{wang19_icml, title = {State-Regularized Recurrent Neural Networks}, author = {Wang, Cheng and Niepert, Mathias}, year = {2019}, booktitle = {Proc. of the 36th International Conference on Machine Learning (ICML)}, pages = {6596--6606}, url = {https://proceedings.mlr.press/v97/wang19j.html} }