Data Science and Machine Learning Series- Recurrent Neural Networks Free Tutorial Download
Recurrent neural networks are very famous deep learning networks which are applied to sequence data: time series forecasting, speech recognition, sentiment classification, machine translation, Named Entity Recognition, etc..
The use of feedforward neural networks on sequence data raises two majors problems:
- Input & outputs can have different lengths in different examples
- MLPs do not share features learned across different positions of the data sample
In this article, we will discover the mathematics behind the success of RNNs as well as some special types of cells such as LSTMs and GRUs. We will finally dig into the encoder-decoder architectures combined with attention mechanisms.
NB: Since Medium does not support LaTeX, the mathematical expressions are inserted as images. Hence, I advise you to turn the dark mode off for a better reading experience.
Table of content
- RNN model
- Different types of RNNs
- Advanced types of cells
- Encoder & Decoder architecture
- Attention mechanisms