TR2019-049

Analysis of the contribution and temporal dependency of LSTM layers for reinforcement learning tasks


    •  Lee, T.-Y., van Baar, J., Wittenburg, K.B., Sullivan, A., "Analysis of the contribution and temporal dependency of LSTM layers for reinforcement learning tasks", IEEE Conference on Computer Vision and Pattern Recognition (CVPR) Explanable AI Workshop, June 2019, pp. 99-102.
      BibTeX TR2019-049 PDF
      • @inproceedings{Lee2019jun2,
      • author = {Lee, Teng-Yok and van Baar, Jeroen and Wittenburg, Kent B. and Sullivan, Alan},
      • title = {Analysis of the contribution and temporal dependency of LSTM layers for reinforcement learning tasks},
      • booktitle = {IEEE Conference on Computer Vision and Pattern Recognition (CVPR) Explanable AI Workshop},
      • year = 2019,
      • pages = {99--102},
      • month = jun,
      • url = {https://www.merl.com/publications/TR2019-049}
      • }
  • Research Areas:

    Artificial Intelligence, Computer Vision

Abstract:

Long short-term memory (LSTM) architectures are widely used in deep neural networks (DNN) when the input data is time-varying, because of their ability to capture (often unknown) long-term dependencies of sequential data. In this paper, we present an approach to analyze the temporal dependencies needed by an LSTM layer. Our approach first locates so-called salient LSTM cells that contribute most to the neural network output, by combining both forward and backward propagation. For these salient cells, we compare their output contributions and the internal gates of LSTM to see whether the activation of gates precedes the increasing of contribution, and how far beforehand the precedence occurs. We apply our analysis in the context of reinforcement learning (RL) for robot control to understand how the LSTM layer reacts under different circumstances.