Surprisal-Based Activation in Recurrent Neural Networks

Proceedings of the 26th European Symposium on Artificial Neural Networks, Computational Intelligence and Machine Learning (ESANN 2018), pages 597--602 - Apr 2018.
Associated documents : es2018-120.pdf [1.4Mo]  
Learning hierarchical abstractions from sequences is a challenging and open problem for Recurrent Neural Networks (RNNs). This is mainly due to the difficulty of detecting features that span over long distances with also different frequencies. In this paper, we address this challenge by introducing surprisal-based activation, a novel method to preserve activations contingent on encoding-based self-information. The preserved activations can be considered as temporal shortcuts with perfect memory. We evaluate surprisal-based activation on language modelling by testing it on the Penn Treebank corpus and find that it can improve performance when compared to a baseline RNN.

 

@InProceedings{AAW18,
  author       = "Alpay, Tayfun and Abawi, Fares and Wermter, Stefan",
  title        = "Surprisal-Based Activation in Recurrent Neural Networks",
  booktitle    = "Proceedings of the 26th European Symposium on Artificial Neural Networks, Computational Intelligence and Machine Learning (ESANN 2018)",
  pages        = "597--602",
  month        = "Apr",
  year         = "2018",
  url          = "https://www2.informatik.uni-hamburg.de/wtm/publications/2018/AAW18/es2018-120.pdf"
}

» Tayfun Alpay
» Fares Abawi
» Stefan Wermter