Back to GlossaryArchitecture

RNN (Recurrent Neural Network)

Definition

A neural network architecture designed for sequential data, where the output of each step is fed back as input to the next step, maintaining a hidden state that captures information from previous inputs.

Recurrent Neural Networks process sequences one element at a time, maintaining a hidden state that acts as memory of past inputs. This makes them naturally suited for tasks like language modeling, speech recognition, and time series prediction. However, basic RNNs struggle with long sequences due to the vanishing gradient problem, where gradients diminish as they propagate back through many time steps. This limitation led to improved architectures like LSTM and GRU, which use gating mechanisms to better preserve long-term information. Despite these improvements, RNNs have been largely replaced by transformers for most sequence tasks, as transformers can process all positions in parallel and handle long-range dependencies more effectively.

Companies in Architecture

View Architecture companies →