Fact-checked by Grok 2 weeks ago

Recurrent neural network

A recurrent neural network (RNN) is a class of artificial neural networks designed to process sequential by incorporating loops that allow to persist across time steps through a hidden state, enabling the modeling of temporal dependencies unlike traditional networks. These networks share weights across time steps, making them suitable for variable-length sequences such as text, speech, or time-series . The foundations of RNNs trace back to early work on feedback in neural models, with the in 1982 introducing recurrent connections for associative memory and in physical systems inspired by biological computation. Modern RNN architectures emerged in the late 1980s and 1990s, notably with Jordan's 1986 proposal for context units and Elman's 1990 simple recurrent network, which demonstrated the ability to learn temporal structures in sequences like . Training RNNs relies on backpropagation through time (BPTT), an extension of that unrolls the network across time steps, though it was formalized earlier by Werbos in the 1970s and popularized in the 1980s. Despite their promise, vanilla RNNs suffer from vanishing or exploding gradients during training, which hinder learning long-term dependencies; this issue was analyzed in the and addressed by gated variants. The unit, introduced by Hochreiter and Schmidhuber in 1997, incorporates input, forget, and output gates to regulate information flow and mitigate gradient problems, enabling effective handling of extended sequences. Similarly, the , proposed by Cho et al. in 2014, offers a simpler alternative with update and reset gates, achieving comparable performance with fewer parameters and faster computation. Other variants include bidirectional RNNs for processing sequences in both directions and echo state networks for . RNNs have found widespread applications in domains requiring sequential modeling, including tasks like and , systems such as DeepSpeech, and time-series for stock prices or weather patterns. In , they support video analysis and image captioning by integrating with convolutional networks, while in bioinformatics, they aid protein sequence prediction. Although transformers have recently overshadowed RNNs in many large-scale tasks due to parallelization advantages, RNNs remain valuable for resource-constrained environments and as components in hybrid models.

Overview

Definition and purpose

A recurrent neural network (RNN) is a class of artificial neural networks in which connections between nodes form directed cycles, creating an internal state that allows the network to retain and utilize information from previous inputs across multiple time steps. This recurrent structure enables the processing of sequential data by feeding the output of hidden units back as input, effectively providing a form of dynamic that persists over time. The primary purpose of RNNs is to model temporal dependencies and patterns in sequences, such as those found in text, speech, or data, where the order of elements influences the overall meaning or prediction. Unlike neural networks, which process inputs independently without regard to sequence order, RNNs capture context by incorporating prior information into current computations, making them suitable for tasks requiring ongoing awareness of historical data. For example, an RNN processing a can analyze it word by word, maintaining a of earlier words to inform interpretations or predictions for later ones, such as anticipating the next word based on accumulated context. This capability stems from the historical motivation to simulate human-like sequential processing and in cognitive tasks, addressing challenges in representing time implicitly through network dynamics rather than explicit input features.

Key characteristics

Recurrent neural networks (RNNs) are distinguished by their use of a hidden state, which serves as a form of internal that persists across time steps. This hidden state is a updated at each step based on the current input and the previous hidden state, allowing the network to incorporate information from prior elements in a . By feeding the hidden state back into the network, RNNs can capture temporal dependencies, making them suitable for tasks involving sequential data such as modeling or prediction. A core feature of RNNs is parameter sharing, where the same set of weights is applied at every time step rather than having distinct for each position in the sequence. This reuse of weights reduces the total number of parameters compared to a fully unrolled network of equivalent depth, promoting efficiency and enforcing a consistent mechanism across the sequence. Parameter sharing also enables RNNs to generalize to sequences of varying lengths without requiring architecture changes, as the model applies the identical transformation repeatedly. RNNs are often visualized in an unfolded or unrolled representation, which depicts the recurrent structure as a deep feedforward where each layer corresponds to a time step, with shared weights connecting identical modules horizontally. This unfolding aids in understanding training dynamics, such as through time, but highlights challenges like to weight initialization, where poor starting values can lead to unstable dynamics or poor during learning.

History

Early foundations

The early foundations of recurrent neural networks emerged from the interdisciplinary field of in the and , which emphasized loops as fundamental mechanisms for control and adaptation in both biological and engineered systems. Norbert Wiener's 1948 book Cybernetics: Or Control and Communication in the Animal and the Machine formalized as a process where system outputs influence future inputs, providing a theoretical basis for recurrent dynamics in neural modeling. During the , ideas extended to systems, influencing early explorations of looped interactions in computational models of and . A pivotal contribution came in 1943 when Warren S. McCulloch and introduced a logical calculus for neural activity, modeling neurons as binary threshold units capable of forming networks with feedback loops to simulate complex propositional functions over time. Their framework demonstrated that simple recurrent connections could realize any finite logical expression, establishing recurrent structures as a means to handle temporal and sequential processing in artificial neural systems. In the 1980s, these ideas advanced with John J. Hopfield's model of a recurrent neural network for associative memory, where fully connected units with symmetric weights settled into stable states via an energy function, enabling pattern storage and retrieval through dynamic feedback. Hopfield's network highlighted the computational power of recurrent connections in mimicking physical systems with emergent collective behaviors. Building on this, proposed in 1986 a simple recurrent architecture known as the Jordan network, which incorporated feedback from output activations to the hidden layer, allowing the model to learn and generate ordered sequences such as motor actions or speech patterns. This design emphasized parallel distributed processing for serial order, with context maintained through recurrent loops. By 1990, Jeffrey L. Elman extended these concepts in the Elman network, introducing recurrent connections from hidden layer outputs to a dedicated layer that fed back as additional inputs, facilitating the discovery of temporal structures in data like language sequences. This input-to-hidden recurrency enabled simple networks to capture dependencies over time without explicit programming. Key figures in these early developments include for cybernetic theory, McCulloch and for foundational neural models, Hopfield for associative recurrent systems, for output-feedback mechanisms, and Elman for context-learning architectures.

Modern developments

The 1990s marked a pivotal era for recurrent neural networks (RNNs) with the popularization of through time (BPTT), a algorithm that extends gradient-based learning to temporal sequences. David Rumelhart and colleagues popularized in 1986, laying the groundwork for its extension to recurrent structures. Paul Werbos formalized BPTT in 1990, providing a practical framework for computing gradients in recurrent structures, which addressed the challenges of networks on sequential data. A major breakthrough came in 1997 with the introduction of (LSTM) units by and , designed specifically to mitigate the vanishing and exploding gradient problems that plagued vanilla RNNs during long-sequence training. This architecture incorporated gating mechanisms to selectively retain or discard information over extended time lags, significantly improving RNN performance on tasks requiring memory of distant inputs. In 2006, and colleagues advanced RNN applications in sequence labeling with (CTC), a that allows training without explicit alignment between inputs and outputs, proving particularly effective for tasks like . The 2010s saw an explosion in RNN adoption, driven by innovations in simplified architectures and sequence-to-sequence (seq2seq) modeling. In 2014, Kyunghyun Cho and co-authors proposed the gated recurrent unit (GRU) as a lightweight alternative to LSTM, featuring fewer parameters while maintaining comparable performance on capturing dependencies in sequences. That same year, Ilya Sutskever, Oriol Vinyals, and Quoc Le introduced the seq2seq framework using LSTMs, which encoded input sequences into fixed representations and decoded them into outputs, catalyzing widespread use of RNNs in natural language processing tasks such as machine translation. Extending RNNs beyond text, Aaron van den Oord and colleagues developed PixelRNN in 2016, a generative model that autoregressively predicts image pixels using row-wise and column-wise recurrent connections, achieving state-of-the-art density estimation on datasets like CIFAR-10. Entering the 2020s, RNNs have evolved through hybrid integrations with architectures to enhance efficiency, combining recurrent state management with mechanisms for better handling of long-range dependencies at reduced computational cost. Despite the dominance of transformers in large-scale models, RNN variants like LSTMs and GRUs persist in edge devices due to their sequential processing and lower , enabling on resource-constrained hardware such as microcontrollers for applications in and wearables.

Fundamentals

Mathematical formulation

A recurrent neural network (RNN) processes sequential data by maintaining a hidden state that captures information from previous time steps. Consider an input sequence \mathbf{x}_{1:T} = (\mathbf{x}_1, \mathbf{x}_2, \dots, \mathbf{x}_T), where each \mathbf{x}_t \in \mathbb{R}^{d_x} is the input vector at time step t, and T is the sequence length. The RNN uses shared parameters \theta across all time steps, including weight matrices U \in \mathbb{R}^{d_h \times d_x} (input-to-hidden), W \in \mathbb{R}^{d_h \times d_h} (hidden-to-hidden), and V \in \mathbb{R}^{d_y \times d_h} (hidden-to-output), along with bias vectors \mathbf{b} \in \mathbb{R}^{d_h} and \mathbf{c} \in \mathbb{R}^{d_y}, where d_h is the hidden state dimension and d_y is the output dimension. The core computation is the deterministic update of the hidden state \mathbf{h}_t \in \mathbb{R}^{d_h} at each time step t: \mathbf{a}^{(t)} = \mathbf{b} + W \mathbf{h}^{(t-1)} + U \mathbf{x}^{(t)} \mathbf{h}^{(t)} = \tanh(\mathbf{a}^{(t)}) Here, \tanh is the hyperbolic tangent activation function, applied elementwise, which ensures the hidden state remains bounded. The initial hidden state is typically \mathbf{h}_0 = \mathbf{0}. This recurrence allows to incorporate temporal dependencies through the persistence of the hidden state. The output \mathbf{o}_t \in \mathbb{R}^{d_y} at time t is computed linearly from the current hidden state: \mathbf{o}^{(t)} = \mathbf{c} + V \mathbf{h}^{(t)} For tasks like sequence classification or labeling, the predicted output \hat{\mathbf{y}}_t is often obtained via a softmax transformation: \hat{\mathbf{y}}^{(t)} = \softmax(\mathbf{o}^{(t)}) where \softmax(\mathbf{z})_i = \frac{\exp(z_i)}{\sum_j \exp(z_j)}. The objective during training is to minimize a loss function over the sequence. For probabilistic prediction tasks, the total loss L is the sum of per-time-step losses L_t, such as cross-entropy: L(\mathbf{y}_{1:T}, \hat{\mathbf{y}}_{1:T}) = \sum_{t=1}^T L_t(\mathbf{y}_t, \hat{\mathbf{y}}_t) where L_t(\mathbf{y}_t, \hat{\mathbf{y}}_t) = -\sum_i y_{t,i} \log \hat{y}_{t,i}, and \mathbf{y}_t is the target distribution at time t. This formulation enables end-to-end optimization of the shared parameters \theta. Computationally, the recurrent structure is unfolded into a deep network with T layers, where each layer corresponds to a time step and reuses the same weights. This unrolled representation, \mathbf{h}^{(t)} = g^{(t)}(\mathbf{x}^{(t)}, \dots, \mathbf{x}^{(1)}; \theta), facilitates gradient-based training while preserving the sequential dependencies.

Backpropagation through time

Backpropagation through time (BPTT) is the standard gradient-based algorithm for training recurrent neural networks (RNNs) by unfolding the recurrent computation graph across time steps and applying the chain rule to compute derivatives of the loss with respect to the network parameters. This approach treats the RNN as a deep network where each layer corresponds to a time step, allowing standard to propagate errors backward through the sequence. In BPTT, the network is unrolled for the entire sequence length T, creating a with T layers, and are computed by summing contributions from each time step while accounting for dependencies across time. For the hidden-to-hidden weight matrix W, the of the loss L with respect to W is given by \frac{\partial L}{\partial W} = \sum_{t=1}^T \delta_t \mathbf{h}_{t-1}^\top, where \delta_t = \frac{\partial L}{\partial \mathbf{h}_t} is the backpropagated error at time t, computed recursively in the backward pass starting from \delta_T = \frac{\partial L_T}{\partial \mathbf{h}_T} and for t = T-1, \dots, 1, \delta_t = \frac{\partial L_t}{\partial \mathbf{h}_t} + (1 - \mathbf{h}_t \odot \mathbf{h}_t) \odot (W^\top \delta_{t+1}), with \odot denoting elementwise (reflecting the of the tanh ). This formulation highlights how errors from future time steps influence earlier hidden states, enabling the RNN to learn temporal patterns. To address computational and memory demands for long sequences, truncated BPTT limits the unrolling to a fixed number of steps \tau \ll T, approximating gradients by ignoring dependencies beyond \tau time steps and reducing complexity from O(T N^2) to O(\tau N^2), where N is the hidden dimension. This truncation mitigates issues like vanishing gradients but may weaken learning of long-range dependencies. An alternative to BPTT is real-time recurrent learning (RTRL), which computes gradients incrementally during the forward pass without unrolling, allowing online updates but at higher computational cost of O(N^4) per step due to maintaining full matrices. RTRL is less efficient than BPTT for most applications and is typically used only when adaptation is critical.

Configurations

Unidirectional configuration

In the unidirectional configuration, information in a recurrent neural network (RNN) propagates solely in the forward direction through time, with the state at each time step t being updated based exclusively on the input at t and the state from the previous step t-1. This sequential forms a linear across time steps, where each in the receives input from the prior and passes its output forward, enabling the network to maintain a of past observations without access to future data. The structure resembles a network unrolled over time, with recurrent connections linking consecutive states in a one-way manner. This setup is particularly suited for tasks requiring causal processing, such as prediction and , where decisions must be made incrementally as new data arrives without lookahead. For instance, in automatic , unidirectional RNNs process audio streams sequentially to generate transcriptions on-the-fly, supporting applications like live captioning or voice assistants. The configuration's ensures that predictions at any point depend only on preceding context, making it ideal for autoregressive generation tasks, such as next-word prediction in language modeling, where the model samples outputs step-by-step to build sequences. The advantages of this unidirectional approach include computational efficiency in streaming scenarios, as it avoids the need to buffer inputs, and its inherent suitability for generative processes that mimic temporal in real-world . By enforcing a strict left-to-right update of the hidden state, the model captures temporal dependencies in a manner that aligns with sequential , though it limits context to historical information alone.

Bidirectional configuration

Bidirectional recurrent neural networks (BRNNs) extend the standard unidirectional RNN by processing input sequences in both forward and backward directions, enabling the model to capture contextual information from the entire sequence rather than just the past. This configuration employs two separate unidirectional RNNs: one that propagates information forward from the beginning of the sequence to produce hidden states \overrightarrow{h_t}, and another that processes the sequence in reverse to generate backward hidden states \overleftarrow{h_t}. At each time step t, these states are concatenated to form the final hidden state h_t = [\overrightarrow{h_t}; \overleftarrow{h_t}], which incorporates dependencies from both preceding and succeeding elements in the sequence. The output at each time step is typically derived from the concatenated hidden state through a layer, often using a weighted or direct to produce predictions that the bidirectional context. This integration allows BRNNs to achieve higher accuracy in tasks where future context is informative, though it doubles the number of parameters compared to a unidirectional RNN of equivalent size due to the parallel processing layers. Computationally, BRNNs require the full sequence to be available upfront, making them suitable for offline processing but incompatible with streaming applications. BRNNs have proven particularly effective in offline sequence labeling tasks, such as , where access to the complete input is feasible. For instance, in offline systems, bidirectional configurations process image-derived s to model both left-to-right and right-to-left dependencies, leading to improved transcription accuracy on datasets like the IAM handwriting database.

Stacked and deep configurations

recurrent neural networks extend the by layering multiple recurrent units vertically, where the output hidden state of one layer serves as the input to the subsequent layer at each time step. This configuration allows for hierarchical processing of sequential , with lower layers capturing lower-level patterns and higher layers abstracting more complex representations. The hidden state update for the l-th layer can be formulated as h_t^{(l)} = \sigma \left( W_{hh}^{(l)} h_{t-1}^{(l)} + W_{xh}^{(l)} h_t^{(l-1)} \right), where \sigma is a nonlinearity such as the hyperbolic tangent, h_t^{(l)} is the hidden state at time t for layer l, W_{hh}^{(l)} are the recurrent weights, and W_{xh}^{(l)} connect the previous layer's output h_t^{(l-1)} as input. The primary benefit of stacking is enhanced representational power through hierarchical feature extraction, enabling the model to handle longer sequences and more intricate dependencies more effectively than single-layer RNNs. For instance, deeper configurations have demonstrated superior performance in tasks requiring multi-scale temporal modeling, such as language processing, by allowing progressive refinement of features across layers. However, increasing depth exacerbates the during training, as gradients propagate through multiple layers and time steps, leading to exponentially diminishing updates for early layers or initial time steps. This challenge is particularly pronounced in standard RNNs without specialized mitigations like gradient clipping. In practice, stacked RNNs often incorporate bidirectional processing from lower layers to capture context in both directions before unidirectional refinement in upper layers. A notable application is in , where deep stacks of LSTM units—typically 4 to 8 layers—form the encoder and decoder, achieving significant improvements in translation quality by modeling source and target sequences hierarchically; for example, the sequence-to-sequence framework used 4-layer LSTMs to outperform shallow models on English-to-French translation benchmarks. Similarly, production-scale systems employed 8-layer stacked LSTMs with residual connections to mitigate issues, yielding up to 60% relative error reduction compared to phrase-based systems on language pairs such as English-French, English-Spanish, and English-Chinese.

Architectures

Vanilla recurrent networks

Vanilla recurrent neural networks (RNNs) represent the foundational architecture for processing sequential data, characterized by their simplicity and lack of gating mechanisms. In this basic form, the network maintains a hidden state h_t that is updated at each time step t based on the current input x_t and the previous hidden state h_{t-1}. The core update equation is h_t = \tanh(W_{hh} h_{t-1} + W_{xh} x_t + b_h), where W_{hh} denotes the recurrent weight matrix, W_{xh} the input-to-hidden weight matrix, b_h the hidden bias, and \tanh the , which bounds the hidden state values between -1 and 1. The output y_t at each step is typically obtained via a linear transformation of the hidden state, y_t = W_{hy} h_t + b_y, often followed by a softmax for probabilistic predictions in tasks like sequence classification. This structure allows the network to share weights across time steps, enabling it to model temporal dependencies through the recurrent connection. A key variant of the vanilla RNN is the simple recurrent network (SRN), proposed by Elman in as a tool for discovering structure in temporal sequences. In the SRN, the hidden layer activations are copied to a dedicated context layer at each time step, which then provides feedback to the hidden layer in the next step, initially with unity weights to preserve information flow. This design facilitates retention without additional complexity, making it suitable for modeling phenomena like language processing where local dependencies predominate. Elman's work demonstrated the SRN's ability to learn hierarchical representations from sequential inputs, influencing early applications in cognitive modeling. Despite their elegance, vanilla RNNs face significant challenges during training, particularly the vanishing and exploding problems when handling long sequences. As are propagated backward through time via through time (BPTT), they are repeatedly multiplied by the of the \tanh , which has a less than 1; this leads to (vanishing) over extended timesteps, hindering the learning of long-range dependencies. Conversely, if the recurrent weights have eigenvalues greater than 1, can explode, causing unstable training. These issues were systematically analyzed in early work, showing that RNNs struggle with tasks requiring beyond 5-10 steps. Vanilla RNNs, including variants like the SRN and earlier models such as Jordan's 1986 network with output feedback to hidden units, were pivotal in early sequence modeling efforts from the mid-1980s to the mid-1990s, applied to problems in , time series prediction, and grammatical inference before the advent of gated architectures addressed their limitations.

Long short-term memory (LSTM)

Long short-term memory (LSTM) networks address the vanishing and exploding problems in vanilla recurrent neural networks by incorporating a cell state that acts as a conveyor belt, allowing information to flow across time steps with minimal alteration, and using gating mechanisms to regulate the flow of information. Introduced by Hochreiter and Schmidhuber in 1997, LSTMs enable the learning of long-term dependencies in sequential data, which traditional RNNs struggle with due to gradient truncation during . The core of an LSTM unit consists of three gates—forget, input, and output—along with a , all computed using (\sigma) and hyperbolic tangent (\tanh) activations. The forget determines what information to discard from the previous : f_t = \sigma(W_f \cdot [h_{t-1}, x_t] + b_f) where W_f is the weight matrix, b_f the , h_{t-1} the previous , and x_t the input. The input i_t decides which new information to store: i_t = \sigma(W_i \cdot [h_{t-1}, x_t] + b_i) A candidate cell state \tilde{c_t} is proposed using: \tilde{c_t} = \tanh(W_c \cdot [h_{t-1}, x_t] + b_c) The cell state is then updated as: c_t = f_t \odot c_{t-1} + i_t \odot \tilde{c_t} where \odot denotes element-wise multiplication. Finally, the output gate o_t filters the cell state to produce the hidden state: o_t = \sigma(W_o \cdot [h_{t-1}, x_t] + b_o), \quad h_t = o_t \odot \tanh(c_t) These gates collectively mitigate gradient issues by selectively preserving or updating relevant information over extended sequences. A variant of LSTM incorporates peephole connections, allowing the gates to directly access the cell state from the previous time step, which enhances timing precision and performance on tasks requiring exact sequence lengths. In this setup, the forget and input gates include terms like \sigma(W_{f,cell} \cdot c_{t-1} + \dots) and similarly for the output gate, as proposed by Gers and Schmidhuber in 2000. This modification has been shown to improve LSTM's ability to learn context-sensitive languages and precise timing patterns compared to the standard architecture. LSTMs excel at capturing long-term dependencies, enabling applications in , , and time-series forecasting where context spans many time steps, often outperforming vanilla RNNs by orders of magnitude in tasks involving lags of 1000+ steps.

Gated recurrent unit (GRU)

The (GRU) is a type of recurrent neural network unit designed to address the while offering a simpler architecture than (LSTM) units. Introduced by Cho et al. in 2014 for use in tasks, the GRU incorporates gating mechanisms to selectively update and reset the hidden state, enabling effective capture of dependencies in sequential data. Unlike the LSTM, which maintains a separate cell state and employs distinct input, forget, and output gates, the GRU merges the forget and input gates into a single update gate and eliminates the cell state, relying instead on a direct between the previous and candidate hidden states. This design reduces the number of parameters, making the more computationally efficient. Empirical evaluations have shown that GRUs achieve performance comparable to LSTMs on sequence modeling tasks while requiring fewer parameters, leading to faster training times. The core computation in a GRU at time step t involves two sigmoid-activated gates: the update gate z_t, which determines the extent to which the hidden state is updated, and the reset gate r_t, which decides how much of the previous hidden state to forget when computing the candidate activation. These are defined as: z_t = \sigma(W_z [h_{t-1}, x_t]) r_t = \sigma(W_r [h_{t-1}, x_t]) where \sigma is the , W_z and W_r are learnable weight matrices, h_{t-1} is the previous hidden state, and x_t is the input at time t. The candidate hidden state \tilde{h}_t is then computed as: \tilde{h}_t = \tanh(W_h [r_t \odot h_{t-1}, x_t]) Finally, the hidden state h_t is updated via linear interpolation: h_t = (1 - z_t) \odot h_{t-1} + z_t \odot \tilde{h}_t where \odot denotes element-wise multiplication. This formulation allows the GRU to adaptively control information flow with reduced complexity compared to more elaborate gated architectures.

Reservoir computing variants

Reservoir computing variants represent a class of recurrent neural networks where the recurrent weights are fixed and randomly initialized, with training confined to the output readout layer, offering an efficient alternative to full backpropagation through time. These approaches leverage the inherent dynamics of a high-dimensional "reservoir" to process temporal data, enabling rapid training and deployment in resource-constrained settings. The Echo State Network (ESN), introduced by Jaeger in , consists of a sparsely connected recurrent layer with random, fixed weights that generates a rich set of transient states in response to inputs. Central to the ESN is the echo state property, which ensures that the reservoir's state at any time depends asymptotically only on the input history, providing a form of fading memory where older inputs influence the current state less strongly. This property allows the network to capture temporal dependencies without requiring the recurrent layer to be trained, making ESNs particularly effective for tasks like time-series prediction and . The Liquid State Machine (LSM), proposed by Maass et al. in 2002, extends similar principles to biologically inspired , using a random recurrent of integrate-and-fire neurons as the to process continuous or spike-train inputs in . Unlike traditional models reliant on stable attractors, the LSM emphasizes computation through transient perturbations in the "liquid" state of the , achieving universal computational power for fading memory tasks on time-varying inputs. This spiking-based structure draws from , modeling cortical microcircuits where generic connectivity suffices for diverse computations. In both ESN and LSM, training involves collecting reservoir states during input presentation and applying —often via pseudoinverse or perceptron-like rules—to learn a memoryless readout from these states to desired outputs, bypassing the computational overhead of gradient-based optimization. This readout-only approach yields fast training times, often orders of magnitude quicker than fully trainable recurrent networks, and facilitates hardware implementations on non-digital substrates like optical systems or arrays due to the fixed recurrent core.

Advanced Architectures

Encoder-decoder frameworks

Encoder-decoder frameworks, also known as sequence-to-sequence (seq2seq) models, utilize two recurrent neural networks (RNNs) to map variable-length input sequences to variable-length output sequences. The encoder RNN processes the input sequence step by step, transforming each element into a hidden state that captures contextual information, ultimately compressing the entire input into a fixed-dimensional context vector. This vector serves as the initial hidden state for the decoder RNN, which generates the output sequence autoregressively, predicting one element at a time conditioned on the previous outputs and the context vector. This architecture was introduced independently in two seminal works: one proposing an RNN encoder-decoder for statistical machine translation using multilayer RNNs, and another applying long short-term memory (LSTM) units in a seq2seq setup for neural machine translation. A key limitation of the fixed context vector is its inability to fully represent long input sequences, leading to information bottlenecks. To address this, attention mechanisms were integrated, allowing the to dynamically weigh different parts of the input sequence at each output step rather than relying on a single vector. In the additive attention model, alignment scores are computed as the between the 's current hidden state and each encoder hidden state, passed through a network, and normalized via softmax to produce a context vector as a weighted sum of encoder states. This approach, pioneered in , significantly improved performance on longer sentences by enabling soft alignment between input and output elements. These frameworks have been widely applied to tasks involving sequence transduction, such as , where they achieved state-of-the-art results on benchmarks like WMT English-to-French translation, outperforming traditional phrase-based systems. They also extend to abstractive text summarization, where the encoder processes a and the generates a concise summary, demonstrating superior scores compared to extractive methods on datasets like /. During inference, variants like are employed to explore multiple output hypotheses, maintaining a fixed number of high-probability partial sequences to mitigate error propagation in autoregressive generation and improve overall sequence quality.

Hierarchical and recursive variants

Hierarchical recurrent neural networks (HRNNs) and recursive variants extend standard RNNs by incorporating multi-level or tree-structured recurrence, enabling the modeling of complex hierarchical data such as parse trees in or compositional scenes in images. Unlike unidirectional or bidirectional RNNs that process sequences linearly, these architectures recursively apply the same weights across levels of a , building representations bottom-up from atomic elements like words or image segments to higher-level phrases or objects. This approach captures compositional semantics, where meaning emerges from structured combinations rather than flat sequences. Socher et al. (2013) introduced such layered structures in HRNNs for , using recursive layers to represent phrases within sentences, achieving improved handling of syntactic dependencies. Recursive neural networks (RvNNs), a foundational variant, operate on tree-structured inputs by merging child representations into parent nodes through a shared composition function, propagating information hierarchically without fixed sequence order. In Tree-RNNs, input units—such as words in sentences or segments in images—are initialized as vectors, then greedily merged based on learned scores to form a , with each non-terminal node representing a compositional vector derived from its children. This bottom-up pooling of child representations allows for flexible of variable-length structures, emphasizing semantic compositionality over sequential order. Socher et al. (2011) demonstrated Tree-RNNs' efficacy in on the Penn Treebank , attaining a 90.29% F1 score, and in scene understanding on the Stanford Background , reaching 78.1% pixel-level accuracy by hierarchically labeling image regions as objects or scenes. A key advancement, the Recursive Neural Tensor Network (RNTN), enhances recursive variants by employing a function that directly models multi-way interactions between child vectors, addressing limitations in bilinear merging of earlier RvNNs. Applied to over full parse trees, RNTN processes the Stanford Sentiment Treebank by computing phrase-level sentiments hierarchically, outperforming prior models with 80.7% accuracy in fine-grained (5-class) classification across 215,154 phrases. Socher et al. (2013) highlighted its strength in capturing and intensification through tree structures, such as distinguishing "not bad" from "bad" at intermediate nodes. These variants have been pivotal in applications like dependency tree parsing for image-sentence , where hierarchical representations map visual scenes to descriptive text, improving retrieval mean rank to 13.6 on benchmark datasets.

Neural Turing machines and extensions

The Neural Turing Machine (NTM) is a memory-augmented recurrent neural network architecture that incorporates an external memory matrix and differentiable read/write heads to emulate the computational capabilities of a . Introduced by Graves, Wayne, and Danihelka in 2014, the NTM extends traditional RNNs by allowing the network to store and retrieve information from a fixed-size memory matrix M_t \in \mathbb{R}^{N \times W}, where N is the number of memory locations and W is the width of each location. This design addresses limitations in vanilla RNNs and LSTMs, where internal state capacity is constrained, by providing an addressable external memory that supports variable-length sequence processing and algorithmic learning. The read and write operations in an NTM rely on attention-based addressing mechanisms, enabling content-based and location-based access to the . For content-based addressing, the read/write weights w_t are computed as w_t = \softmax(K(k_t, M_{t-1})), where k_t is a produced by the controller network, and K measures the similarity (e.g., via cosine or ) between k_t and the memory rows M_{t-1}. Writing involves erasing and adding to selected memory locations, while reading retrieves vectors by weighting memory contents with w_t. These operations are fully differentiable, allowing end-to-end training via , and the controller—a or recurrent neural network—learns to generate keys, strengths, and shifts for addressing. Building on the NTM, the (DNC) introduces dynamic external management to handle growing and sparse memory usage more efficiently. Proposed by Graves et al. in , the DNC augments the memory matrix with a usage and a temporal linkage that tracks allocation and writing order, enabling the to differentiate between free and occupied locations and to perform content- and temporal-based reads. This allows the DNC to maintain long-term dependencies over extended sequences without fixed-size limitations degrading performance. NTMs and DNCs demonstrate superior performance on algorithmic tasks that challenge standard RNNs, such as , , and associative recall, where they learn simple programs from input-output examples with minimal supervision. For instance, on the task with sequences up to length 20, NTMs achieve near-perfect accuracy after , inferring reversal and comparison operations implicitly. These architectures highlight the potential of memory-augmented RNNs for and reasoning, influencing subsequent work in neural program induction.

Training Methods

Gradient-based optimization

Gradient-based optimization is a cornerstone of training recurrent neural networks (RNNs), where parameters are updated iteratively using gradients computed via through time (BPTT). In the vanilla approach, the weights W are updated according to the rule W \leftarrow W - \eta \frac{\partial L}{\partial W}, where \eta is the and L is function, with gradients derived from unfolding the RNN over length. To mitigate exploding gradients, which can cause unstable training due to repeated multiplication of large values in recurrent connections, clipping is applied by rescaling the gradient vector if its norm exceeds a threshold, typically between 1 and 5. Advanced optimizers adapt the dynamically to handle the non-stationary gradients prevalent in RNN training. RMSprop maintains a of the squared gradients to normalize updates, effectively coordinate-wise adaptation from manual rate scheduling and improving on sequence data. Similarly, combines with adaptive estimates of first- and second-order moments of the gradients, making it particularly effective for RNNs by accelerating initial progress and stabilizing updates in noisy, high-dimensional spaces. Regularization techniques complement these optimizers to prevent in RNNs. Dropout, when applied selectively to non-recurrent connections (e.g., input-to-hidden weights but not hidden-to-hidden), acts as a Bayesian that maintains temporal while introducing stochasticity for better generalization. Hyperparameter tuning, such as scheduling, is crucial for sequence tasks where long dependencies amplify optimization challenges. A common strategy involves starting with a higher initial rate and halving it periodically after initial epochs to refine without stalling.

Teacher forcing and scheduled sampling

In autoregressive recurrent neural networks (RNNs), is a training strategy that feeds the ground-truth previous outputs as inputs to the model at each time step, rather than the model's own predictions. This technique, introduced by Williams and Zipser in , enables faster convergence during gradient-based optimization by preventing the immediate accumulation of prediction errors that would otherwise propagate through the sequence. Despite its benefits, teacher forcing introduces exposure bias, a mismatch between the training regime—where the model relies on perfect previous tokens—and inference, where it must generate sequences using its own potentially erroneous outputs. This discrepancy causes error accumulation at test time, degrading performance in tasks like and text generation, as the model becomes unaccustomed to correcting its mistakes. To address exposure bias, scheduled sampling was proposed by Bengio et al. in 2015 as a curriculum learning approach that progressively replaces ground-truth inputs with the model's predictions during training. The method uses a time-dependent probability schedule to decide at each step whether to sample from the true previous token or the model's output distribution, starting with high reliance on ground truth and gradually increasing self-reliance. This gradual exposure improves the model's robustness to its own errors, leading to better generalization in sequence prediction tasks. An alternative to scheduled sampling is Professor Forcing, introduced by Lamb et al. in 2016, which employs adversarial domain adaptation to align the hidden state dynamics between teacher-forced training and free-running generation modes. By training a discriminator to distinguish between the two regimes and minimizing the discrepancy via a secondary loss, Professor Forcing achieves faster convergence and reduced exposure bias compared to pure teacher forcing, particularly in complex sequence modeling scenarios.

Specialized loss functions

Specialized loss functions in recurrent neural networks (RNNs) address challenges in training on sequential data where input and output alignments are unknown or variable, enabling end-to-end learning without explicit segmentation. One prominent example is (CTC), which marginalizes over all possible alignments between input sequences and target labels, allowing RNNs to directly optimize the probability of the output sequence given the input. CTC defines the probability of a target y given an input x as the sum over all possible \pi that map to y, formulated as: P(y|x) = \sum_{\pi \in \mathcal{B}^{-1}(y)} P(\pi|x) where \mathcal{B}^{-1}(y) denotes the set of all paths that collapse to y via a function \mathcal{B} that handles blank symbols and repetitions. This approach, introduced by Graves et al. in 2006, uses dynamic programming to compute the forward-backward efficiently, making it tractable for long and eliminating the need for supervised during . CTC has been widely applied in , where it enables end-to-end models to transcribe audio without phonetic alignments; for instance, the Deep Speech system achieved a 16.0% on the Switchboard Hub5'00 using CTC with deep RNNs. In , CTC facilitated direct labeling of unsegmented stroke sequences. Variants of CTC, such as hybrid CTC- architectures, combine CTC's alignment-free properties with mechanisms to improve decoding and handle non-monotonic alignments, as seen in end-to-end automatic systems that reduced by up to 10% relative to pure CTC models. Another specialized loss is the objective in sequence-to-sequence frameworks, adapted with masking to ignore in variable-length outputs during training, often paired with to provide ground-truth previous tokens.

Limitations and Extensions

Vanishing and exploding gradients

One of the primary challenges in training recurrent neural networks (RNNs) arises from the , where gradients with respect to the initial hidden states or early weights diminish exponentially during through time (BPTT). This occurs because the gradients are computed as products of the matrices over multiple time steps, and if the recurrent weight matrix W_{hh} has eigenvalues with magnitudes less than 1, repeated multiplications lead to , making it difficult to learn long-term dependencies. Conversely, the exploding problem manifests when the eigenvalues of W_{hh} have magnitudes greater than 1, causing to grow exponentially and potentially leading to numerical instability, such as overflow to values during training. This instability is typically detected by monitoring the norm of the ; if it exceeds a predefined (e.g., 1 or 5, depending on the ), it signals potential . To mitigate these issues within traditional RNN frameworks, clipping is a widely adopted that rescales the entire if its norm surpasses the , preventing explosions without altering the direction of updates. Additionally, orthogonal initialization of the recurrent weights ensures that the singular values (and thus eigenvalues for square matrices) lie on the unit circle, promoting stability by avoiding both shrinkage and amplification in the early training phases. The stability of RNN training can be analyzed through the eigenvalue spectrum of W_{hh}: for gradients to propagate effectively over long sequences without vanishing or exploding, the (the largest absolute eigenvalue) should be close to 1, balancing the trade-off between short-term and long-term learning dynamics. This highlights why random initializations often fail, as they rarely yield such balanced spectra, underscoring the need for targeted initialization strategies.

Modern integrations and alternatives

In the era dominated by Transformer architectures, recurrent neural networks (RNNs) exhibit limitations stemming from their inherent sequential processing, which imposes linear time complexity O(n) but precludes efficient parallelization during and . In contrast, Transformers leverage self-attention mechanisms with O(n²) in sequence length, yet achieve superior wall-clock performance through massive parallelism, particularly on modern hardware. This disparity has driven a significant shift in sequence modeling away from RNNs toward s for tasks involving long-range dependencies. Hybrid models have emerged to mitigate these issues by integrating recurrent elements with attention-based frameworks. Transformer-XL, for instance, introduces segment-level recurrence to reuse hidden states across segments, enabling the capture of dependencies up to 450% longer than vanilla Transformers while preserving temporal coherence. Similarly, the Simple Recurrent Unit (SRU), proposed in 2017, reformulates recurrence to allow full parallelization within each time step, achieving linear scaling and up to 5-9 times faster training than LSTM variants on language modeling tasks without accuracy loss. Amid 2020s advancements, RNNs retain relevance in low-resource environments, such as for low-data languages like English-to-Igbo, where their modest parameter counts and efficiency yield competitive scores of approximately 35-38 with . In , recurrent policy networks employ RNNs to maintain evolving hidden states, facilitating and in partially Markov decision processes, as demonstrated in real-time setups solving complex games like . Recent extensions include state space models like (2023), which provide selective scanning mechanisms for efficient long-sequence modeling, combining RNN-style recurrence with linear-time computation as an alternative to both traditional RNNs and Transformers. Although alternatives like Transformers and convolutional neural networks have supplanted RNNs in many offline sequence tasks due to better scalability, RNN architectures endure in streaming scenarios, including time-series forecasting, where their incremental processing supports low-latency applications without buffering entire sequences.

Applications

Sequence modeling tasks

Recurrent neural networks (RNNs) have been pivotal in (NLP) tasks that involve modeling sequential dependencies in text data. In language modeling, RNNs predict the probability of the next word in a sequence given previous context, enabling applications like and text completion. A seminal work demonstrated that RNN-based language models significantly outperform traditional n-gram models by capturing long-range dependencies, achieving up to 50% reduction in perplexity on benchmark corpora. Bidirectional long short-term memory (LSTM) networks, a variant of RNNs that process sequences in both forward and backward directions, have been effectively applied to . This allows the model to incorporate contextual from the entire , improving of positive, negative, or neutral sentiments in reviews and social media text. For instance, bidirectional LSTMs have achieved high accuracy on datasets like the Stanford Sentiment Treebank by jointly modeling word embeddings and . In , RNNs form the basis of frameworks, where an encoder RNN compresses the input sentence into a fixed representation, and a RNN generates the output . Integrating mechanisms into these models allows the decoder to focus on relevant parts of the input, substantially improving translation quality over fixed representations, as detailed in advanced architectures. RNNs also enable creative text generation by modeling character-level or word-level sequences autoregressively. Early demonstrations trained RNNs on large corpora to generate coherent , such as Shakespearean text, by optimizing the model with advanced techniques like Hessian-free optimization to handle the non-convex loss landscape. These models produce novel sequences that capture stylistic patterns, paving the way for applications in story writing and dialogue systems. For , RNNs support sequence transduction tasks like streaming transcription through models such as the RNN Transducer (RNN-T). This end-to-end approach aligns input audio frames with output text labels without requiring predefined alignments, enabling processing and achieving competitive word error rates on large-vocabulary datasets. RNN-T combines a prediction network for language modeling with an encoder for acoustic features, outputting a distribution over sequences via a joint network.

Time series and signal processing

Recurrent neural networks (RNNs), especially (LSTM) variants, excel in by maintaining memory of past observations to predict future values in sequential data. LSTMs address the limitations of vanilla RNNs in handling long-range dependencies, making them suitable for multivariate where multiple interrelated variables evolve over time. In stock price , LSTM models process historical prices, trading volumes, and economic indicators as input sequences, often achieving lower mean absolute errors compared to (ARIMA) models on benchmarks like the S&P 500 index. Similarly, for weather prediction, LSTMs integrate diverse inputs such as , , and data to forecast short- to medium-term conditions, demonstrating improved accuracy over statistical baselines in multivariate scenarios. Anomaly detection in sensor data leverages RNN-based autoencoders, which compress sequential inputs into latent representations and reconstruct them, identifying outliers where reconstruction errors exceed thresholds indicative of abnormal behavior. These models learn temporal patterns in streams from industrial sensors, such as vibration or temperature readings, enabling unsupervised detection of faults in predictive maintenance systems with high precision on datasets like those from power plants. In , RNNs facilitate audio by autoregressively generating raw waveforms, capturing fine-grained temporal structures. , a single-layer RNN architecture, produces high-fidelity speech and music samples at speeds suitable for real-time applications, rivaling convolutional models like in perceptual quality while reducing computational demands. Representative applications include electrocardiogram (ECG) analysis, where bidirectional LSTMs process heartbeat sequences to classify arrhythmias, achieving over 99% accuracy on MIT-BIH datasets by modeling subtle temporal variations in QRS complexes. In traffic prediction, spatio-temporal RNNs forecast vehicle flows using historical speed and volume data from road networks, outperforming graph-based alternatives in urban settings like by capturing dynamic dependencies.

Other domains

Recurrent neural networks have been integrated into frameworks to handle partially observable Markov decision processes (POMDPs), where s must maintain internal state representations to compensate for incomplete observations. In such settings, policy RNNs extend value-based methods like Deep Q-Networks (DQNs) by incorporating recurrent layers, such as LSTMs, to process sequential observations and produce state-dependent actions. For instance, the Deep Recurrent Q-Network (DRQN) replaces the fully connected layers in a standard DQN with recurrent units, enabling the to learn temporal dependencies and achieve superior in tasks like partially observable , where non-recurrent DQNs fail due to limitations. In , RNNs facilitate the analysis of spatiotemporal data in video sequences, particularly for action recognition tasks that require modeling both spatial features from frames and temporal dynamics across them. A prominent approach combines convolutional neural networks (CNNs) for feature extraction with recurrent layers for sequence modeling, as exemplified by Long-term Recurrent Convolutional Networks (LRCNs). These models process variable-length video inputs end-to-end, capturing long-term dependencies to classify actions with accuracies exceeding 80% on benchmarks like the UCF-101 dataset, outperforming purely convolutional alternatives by integrating temporal context effectively. RNNs also play a key role in for , where recurrent policies enable in dynamic environments with partial . By parameterizing policies as recurrent functions, these networks can generate sequential action trajectories that account for historical states and uncertainties, facilitating tasks like or in POMDPs. Seminal work on recurrent policy gradients demonstrates how such models solve deep POMDPs by learning policies that maintain states for , achieving efficient and in robotic simulations with high-dimensional state spaces. In healthcare, RNNs model electronic health records (EHRs) to predict patient trajectories by capturing the irregular, sequential nature of clinical events like diagnoses, procedures, and medications. These models treat patient histories as , using gated recurrent units to handle variable visit intervals and predict future outcomes such as disease progression or readmissions. For example, the Doctor framework employs RNNs to forecast clinical events from longitudinal EHR data of approximately 250,000 patients, achieving AUROC scores above 0.75 for multi-label predictions and demonstrating the ability to learn patient-specific representations that mimic reasoning patterns. In bioinformatics, RNNs, particularly LSTMs, are used for protein tasks such as secondary and . These models sequences to capture long-range dependencies, enabling accurate predictions of patterns and functional annotations from primary sequence data alone. More recently, RNNs have been deployed in AI applications for IoT , enabling real-time ing of sensor data on resource-constrained devices. In 2025, implementations combining RNNs with autoencoders have shown promise for fault detection in industrial IoT networks, where LSTMs analyze streaming time-series data to identify deviations with low and minimal computational overhead. These edge-based systems reduce reliance on , achieving detection accuracies over 95% in simulated environments while preserving for applications like and environmental .

References

  1. [1]
    Recurrent Neural Networks (RNNs): Architectures, Training Tricks ...
    Jul 23, 2023 · Recurrent neural network (RNN) is a specialized neural network with feedback connection for processing sequential data or time-series data in ...
  2. [2]
    Neural networks and physical systems with emergent collective ...
    Neural networks and physical systems with emergent collective computational abilities. J J Hopfield ... ArticleApril 15, 1982. Sequence-specific ...
  3. [3]
    Finding Structure in Time - Elman - 1990 - Cognitive Science
    The current report develops a proposal along these lines first described by Jordan (1986) which involves the use of recurrent links in order to provide networks ...
  4. [4]
    A Comprehensive Review of Architectures, Variants, and Applications
    Recurrent neural networks (RNNs) are a class of deep learning models that are fundamentally designed to handle sequential data [10,11]. Unlike feedforward ...
  5. [5]
    Long Short-Term Memory | Neural Computation - MIT Press Direct
    Nov 15, 1997 · We briefly review Hochreiter's (1991) analysis of this problem, then address it by introducing a novel, efficient, gradient based method called ...Missing: original | Show results with:original
  6. [6]
    Learning Phrase Representations using RNN Encoder-Decoder for ...
    Jun 3, 2014 · In this paper, we propose a novel neural network model called RNN Encoder-Decoder that consists of two recurrent neural networks (RNN).
  7. [7]
    [PDF] A PARALLEL DISTRmUTED PROCESSING APPROACH - UCSD CSE
    The approach would therefore seem to have some potential for reconciling problems of serial order with problems relating to the continuous nature of behavior.
  8. [8]
    Recurrent Neural Networks (RNNs): A gentle Introduction and ...
    Nov 23, 2019 · In this work we give a short overview over some of the most important concepts in the realm of Recurrent Neural Networks.
  9. [9]
    On the difficulty of training recurrent neural networks
    There are two widely known issues with properly training recurrent neural networks, the vanishing and the exploding gradient problems.
  10. [10]
    Cybernetics or Control and Communication in the Animal and the ...
    With the influential book Cybernetics, first published in 1948, Norbert Wiener laid the theoretical foundations for the multidisciplinary field of cybernetics ...
  11. [11]
    Cybernetics - an overview | ScienceDirect Topics
    Cybernetics formalized feedback mechanisms as the source of intelligent behaviors, with negative feedback control loops serving as basic models for autonomy and ...
  12. [12]
    A logical calculus of the ideas immanent in nervous activity
    A logical calculus of the ideas immanent in nervous activity. Published: December 1943. Volume 5, pages 115–133, (1943); Cite this ...
  13. [13]
    Backpropagation through time: what it does and how to do it
    Oct 31, 1990 · Basic backpropagation, which is a simple method now being widely used in areas like pattern recognition and fault diagnosis, is reviewed.
  14. [14]
    Learning representations by back-propagating errors - Nature
    Oct 9, 1986 · We describe a new learning procedure, back-propagation, for networks of neurone-like units. The procedure repeatedly adjusts the weights of the connections in ...
  15. [15]
    [PDF] Connectionist Temporal Classification: Labelling Unsegmented ...
    This paper presents a novel method for training RNNs to label un- segmented sequences directly, thereby solv- ing both problems. An experiment on the. TIMIT ...
  16. [16]
    Sequence to Sequence Learning with Neural Networks - arXiv
    Sep 10, 2014 · In this paper, we present a general end-to-end approach to sequence learning that makes minimal assumptions on the sequence structure.
  17. [17]
    [1601.06759] Pixel Recurrent Neural Networks - arXiv
    We present a deep neural network that sequentially predicts the pixels in an image along the two spatial dimensions.
  18. [18]
    [PDF] The Mamba in the Llama: Distilling and Accelerating Hybrid Models
    May 3, 2025 · We show that by reusing weights from attention layers, it is possible to distill a large transformer into a large hybrid-linear RNN with minimal ...
  19. [19]
    None
    ### Summary of Mathematical Formulation for Vanilla RNN
  20. [20]
    Experimental Analysis of the Real-time Recurrent Learning Algorithm
    Apr 5, 2007 · The real-time recurrent learning algorithm is a gradient-following learning algorithm for completely recurrent networks running in continually sampled time.
  21. [21]
    [PDF] Finding Structure in Time
    Finding Structure in Time. JEFFREY L. ELMAN. University of Calcfornia, San Riego. Time underlies many interesting human behaviors. Thus, the question of how to.Missing: paper | Show results with:paper
  22. [22]
    Finding structure in time - ScienceDirect.com
    Time is represented implicitly by its effects on processing using recurrent links, where hidden unit patterns are fed back to themselves.
  23. [23]
    [PDF] arXiv:1601.06581v2 [cs.CL] 28 Jan 2016
    Jan 28, 2016 · The algorithm employs a speech-to-character unidirectional recurrent neural network (RNN), which is end-to-end trained with connectionist ...
  24. [24]
    Recurrent neural network based language model - ISCA Archive
    A new recurrent neural network based language model (RNN LM) with applications to speech recognition is presented.
  25. [25]
    [PDF] A Neural Transducer - NIPS papers
    A Neural Transducer makes incremental predictions as input arrives, unlike sequence-to-sequence models, and can produce outputs as data comes in.
  26. [26]
    Bidirectional recurrent neural networks | IEEE Journals & Magazine
    Abstract: In the first part of this paper, a regular recurrent neural network (RNN) is extended to a bidirectional recurrent neural network (BRNN).
  27. [27]
    Offline Handwriting Recognition with Multidimensional Recurrent ...
    This paper introduces a globally trained offline handwriting recogniser that takes raw pixel data as input.Missing: 2005 | Show results with:2005
  28. [28]
    [1211.5063] On the difficulty of training Recurrent Neural Networks
    Nov 21, 2012 · There are two widely known issues with properly training Recurrent Neural Networks, the vanishing and the exploding gradient problems.
  29. [29]
    [PDF] Understanding Feature Selection and Feature Memorization ... - arXiv
    Mar 3, 2019 · In the modern literature, it is referred to as Vanilla RNN. Its state- update equation is given by (3). st. = tanh(Wxt + Ust−1 + b).
  30. [30]
    [PDF] Learning long-term dependencies with gradient descent is difficult
    Bengio, P. Frasconi, P. Simard, "The problem of learning long- term dependencies in recurrent networks," invited paper at the IEEE. International ...
  31. [31]
    [PDF] LONG SHORT-TERM MEMORY 1 INTRODUCTION
    Hochreiter, S. and Schmidhuber, J. (1997). LSTM can solve hard long time lag problems. In. Advances in Neural Information Processing Systems 9. MIT ...
  32. [32]
    [PDF] Recurrent Nets that Time and Count Felix A. Gers Jiurgen ... - IDSIA
    Peephole connections from within the cell (or recurrent connections from gates) require a refinement of L STM 's update scheme. Updates for peephole LSTM. E ach ...
  33. [33]
    [PDF] An Empirical Exploration of Recurrent Network Architectures
    We conducted a thor- ough architecture search where we evaluated over ten thousand different RNN architectures, and identified an architecture that outperforms.
  34. [34]
    Echo state network - Scholarpedia
    Sep 6, 2007 · Echo state networks (ESN) provide an architecture and supervised learning principle for recurrent neural networks (RNNs).
  35. [35]
    [PDF] The “echo state” approach to analysing and training recurrent neural ...
    Jan 25, 2010 · Jaeger(2001): The ”echo state” approach to analysing and training recurrent neural networks. GMD Report. 148, German National Research Center ...
  36. [36]
    [PDF] Real-Time Computing Without Stable States - IGI, TU-Graz
    Our approach provides new perspectives for the interpretation of neural coding, the design of experiments and data analysis in neurophysiology, and the so-.
  37. [37]
    Neural Machine Translation by Jointly Learning to Align and ... - arXiv
    Sep 1, 2014 · Abstract:Neural machine translation is a recently proposed approach to machine translation. Unlike the traditional statistical machine ...
  38. [38]
    [1602.06023] Abstractive Text Summarization Using Sequence-to ...
    Feb 19, 2016 · In this work, we model abstractive text summarization using Attentional Encoder-Decoder Recurrent Neural Networks, and show that they achieve state-of-the-art ...Missing: rnn seminal
  39. [39]
    [PDF] Grounded Compositional Semantics for Finding and Describing ...
    The DT-RNN has several important differences to previous RNN models of Socher et al. (2011a) and. (Socher et al., 2011b; Socher et al., 2011c). These.
  40. [40]
    [PDF] Parsing Natural Scenes and Natural Language with Recursive ...
    Abstract. Recursive structure is commonly found in the inputs of different modalities such as natural scene images or natural language sentences.
  41. [41]
    [PDF] Recursive Deep Models for Semantic Compositionality Over a ...
    In particular we will de- scribe and experimentally compare our new RNTN model to recursive neural networks (RNN) (Socher et al., 2011b) and matrix-vector RNNs ...
  42. [42]
    [1410.5401] Neural Turing Machines - arXiv
    Oct 20, 2014 · Neural Turing Machines can infer simple algorithms such as copying, sorting, and associative recall from input and output examples.
  43. [43]
    Hybrid computing using a neural network with dynamic external ...
    Oct 12, 2016 · Here we introduce a machine learning model called a differentiable neural computer (DNC), which consists of a neural network that can read from and write to an ...
  44. [44]
    [PDF] Neural Networks for Machine Learning Lecture 6a Overview of mini
    rmsprop: Divide the learning rate for a weight by a running average of the magnitudes of recent gradients for that weight. – This is the mini-‐batch version of ...
  45. [45]
    [1412.6980] Adam: A Method for Stochastic Optimization - arXiv
    Dec 22, 2014 · We introduce Adam, an algorithm for first-order gradient-based optimization of stochastic objective functions, based on adaptive estimates of lower-order ...
  46. [46]
    A Theoretically Grounded Application of Dropout in Recurrent ...
    Dec 16, 2015 · A Theoretically Grounded Application of Dropout in Recurrent Neural Networks. Authors:Yarin Gal, Zoubin Ghahramani.
  47. [47]
    Scheduled Sampling for Sequence Prediction with Recurrent Neural ...
    Jun 9, 2015 · We propose a curriculum learning strategy to gently change the training process from a fully guided scheme using the true previous token, towards a less guided ...
  48. [48]
    Professor Forcing: A New Algorithm for Training Recurrent Networks
    Oct 27, 2016 · The Teacher Forcing algorithm trains recurrent networks by supplying observed sequence values as inputs during training and using the network's own one-step- ...
  49. [49]
    Deep Speech: Scaling up end-to-end speech recognition - arXiv
    Dec 17, 2014 · We present a state-of-the-art speech recognition system developed using end-to-end deep learning. Our architecture is significantly simpler than traditional ...
  50. [50]
    Hybrid CTC/Attention Architecture for End-to-End Speech Recognition
    Oct 16, 2017 · This paper proposes hybrid CTC/attention end-to-end ASR, which effectively utilizes the advantages of both architectures in training and decoding.
  51. [51]
    Learning long-term dependencies with gradient descent is difficult
    We show why gradient based learning algorithms face an increasingly difficult problem as the duration of the dependencies to be captured increases.
  52. [52]
    Framewise phoneme classification with bidirectional LSTM and ...
    In this paper, we present bidirectional Long Short Term Memory (LSTM) networks, and a modified, full gradient version of the LSTM learning algorithm.
  53. [53]
    [PDF] DataStories at SemEval-2017 Task 4: Bidirectional LSTM with ...
    In this paper, we present two deep-learning sys- tems for short text sentiment analysis developed for SemEval-2017 Task 4 “Sentiment Analysis in. Twitter”.
  54. [54]
    [PDF] Generating Text with Recurrent Neural Networks
    RNNs, with a high-dimensional hidden state, are used to predict the next character in text, using a new MRNN architecture with multiplicative connections.Missing: seminal | Show results with:seminal
  55. [55]
    Sequence Transduction with Recurrent Neural Networks - arXiv
    Nov 14, 2012 · This paper introduces an end-to-end, probabilistic sequence transduction system, based entirely on RNNs, that is in principle able to transform any input ...
  56. [56]
    [PDF] Recurrent Neural Networks for Time Series Forecasting - arXiv
    Jan 1, 2019 · This article presents a recurrent neural network based time series forecasting frame- work covering feature engineering, feature importances, ...Missing: seminal | Show results with:seminal
  57. [57]
    (PDF) RNN-Autoencoder Approach for Anomaly Detection in Power ...
    This research proposes to use a combined recurrent neural network (RNN)-autoencoder approach as a "normal" behavior model (NBM) with the Mahalanobis Distance ( ...
  58. [58]
    Efficient Neural Audio Synthesis
    We first describe a single-layer recurrent neural network, the WaveRNN, with a dual softmax layer that matches the quality of the state-of-the-art WaveNet model ...
  59. [59]
    In-Depth Insights into the Application of Recurrent Neural Networks ...
    By capturing long-term dependencies in time series, RNNs can accurately forecast future changes in traffic conditions. Mainly, variants of RNNs, like LSTMs and ...
  60. [60]
    Deep Recurrent Q-Learning for Partially Observable MDPs - arXiv
    Jul 23, 2015 · This article investigates the effects of adding recurrency to a Deep Q-Network (DQN) by replacing the first post-convolutional fully-connected layer with a ...
  61. [61]
    Long-term Recurrent Convolutional Networks for Visual Recognition ...
    Nov 17, 2014 · We develop a novel recurrent convolutional architecture suitable for large-scale visual learning which is end-to-end trainable.
  62. [62]
    [PDF] Solving Deep Memory POMDPs with Recurrent Policy Gradients
    This paper presents Recurrent Policy Gradients, a model- free reinforcement learning (RL) method creating limited-memory sto- chastic policies for partially ...
  63. [63]
    Implementation of edge AI for early fault detection in IoT networks
    Oct 16, 2025 · The architecture leverages recurrent neural networks and autoencoders optimized for time-series anomaly detection, enabling local inference ...