(344e) Attention-Based Recurrent Neural Network for Multi-Step-Ahead Prediction
AIChE Annual Meeting
2019
2019 AIChE Annual Meeting
Computing and Systems Technology Division
Machine Learning Applications and Intelligent Systems
Tuesday, November 12, 2019 - 1:46pm to 2:05pm
Motivated by [1, 2], we introduce an attention-based RNN for multi-step-ahead prediction that can have applications in model predictive control, fault diagnosis, etc. This model consists of an RNN that encodes a sequence of input time series data into a new representation (called context vector) and another RNN that decodes the representation into output target sequence. An attention model integrated to the encoder-decoder RNN model allows the network to focus on parts of the input sequence that are relevant to predicting the target sequence. The attention model is jointly trained with all other components of the model. By having a deep architecture, the model can learn a very complex dynamic system, and it is robust to noise. In order to show the effectiveness of the proposed approach, we perform a comparative study on the problem of catalyst activation prediction, by using Nonlinear AutoRegressive with eXogeneous (NARX) predictor, and conventional machine learning techniques such as Support Vector Regression (SVR).
[1] Cho, K., van Merrienboer, B., Gulcehre, C., Bougares, F., Schwenk, H., and Bengio, Y. Learning phrase representations using RNN encoder-decoder for statistical machine translation. In Proceedings of the Empiricial Methods in Natural Language Processing (EMNLP 2014).
[2] Dzmitry Bahdanau, Kyunghyun Cho, and Yoshua Bengio. 2015. Neural machine translation by jointly learning to align and translate. In Proceeding of the International Conference on Learning Representations (ICLR 2015).
_________________________________________________________________
* All correspondence should be addressed to Prof. Yinlun Huang (Phone: 313-577-3771; E-mail: yhuang@wayne.edu).