Fact-checked by Grok 2 weeks ago
References
-
[1]
Recurrent Neural Networks (RNNs): Architectures, Training Tricks ...Jul 23, 2023 · Recurrent neural network (RNN) is a specialized neural network with feedback connection for processing sequential data or time-series data in ...
-
[2]
Neural networks and physical systems with emergent collective ...Neural networks and physical systems with emergent collective computational abilities. J J Hopfield ... ArticleApril 15, 1982. Sequence-specific ...
-
[3]
Finding Structure in Time - Elman - 1990 - Cognitive ScienceThe current report develops a proposal along these lines first described by Jordan (1986) which involves the use of recurrent links in order to provide networks ...
-
[4]
A Comprehensive Review of Architectures, Variants, and ApplicationsRecurrent neural networks (RNNs) are a class of deep learning models that are fundamentally designed to handle sequential data [10,11]. Unlike feedforward ...
-
[5]
Long Short-Term Memory | Neural Computation - MIT Press DirectNov 15, 1997 · We briefly review Hochreiter's (1991) analysis of this problem, then address it by introducing a novel, efficient, gradient based method called ...Missing: original | Show results with:original
-
[6]
Learning Phrase Representations using RNN Encoder-Decoder for ...Jun 3, 2014 · In this paper, we propose a novel neural network model called RNN Encoder-Decoder that consists of two recurrent neural networks (RNN).
-
[7]
[PDF] A PARALLEL DISTRmUTED PROCESSING APPROACH - UCSD CSEThe approach would therefore seem to have some potential for reconciling problems of serial order with problems relating to the continuous nature of behavior.
-
[8]
Recurrent Neural Networks (RNNs): A gentle Introduction and ...Nov 23, 2019 · In this work we give a short overview over some of the most important concepts in the realm of Recurrent Neural Networks.
-
[9]
On the difficulty of training recurrent neural networksThere are two widely known issues with properly training recurrent neural networks, the vanishing and the exploding gradient problems.
-
[10]
Cybernetics or Control and Communication in the Animal and the ...With the influential book Cybernetics, first published in 1948, Norbert Wiener laid the theoretical foundations for the multidisciplinary field of cybernetics ...
-
[11]
Cybernetics - an overview | ScienceDirect TopicsCybernetics formalized feedback mechanisms as the source of intelligent behaviors, with negative feedback control loops serving as basic models for autonomy and ...
-
[12]
A logical calculus of the ideas immanent in nervous activityA logical calculus of the ideas immanent in nervous activity. Published: December 1943. Volume 5, pages 115–133, (1943); Cite this ...
-
[13]
Backpropagation through time: what it does and how to do itOct 31, 1990 · Basic backpropagation, which is a simple method now being widely used in areas like pattern recognition and fault diagnosis, is reviewed.
-
[14]
Learning representations by back-propagating errors - NatureOct 9, 1986 · We describe a new learning procedure, back-propagation, for networks of neurone-like units. The procedure repeatedly adjusts the weights of the connections in ...
-
[15]
[PDF] Connectionist Temporal Classification: Labelling Unsegmented ...This paper presents a novel method for training RNNs to label un- segmented sequences directly, thereby solv- ing both problems. An experiment on the. TIMIT ...
-
[16]
Sequence to Sequence Learning with Neural Networks - arXivSep 10, 2014 · In this paper, we present a general end-to-end approach to sequence learning that makes minimal assumptions on the sequence structure.
-
[17]
[1601.06759] Pixel Recurrent Neural Networks - arXivWe present a deep neural network that sequentially predicts the pixels in an image along the two spatial dimensions.
-
[18]
[PDF] The Mamba in the Llama: Distilling and Accelerating Hybrid ModelsMay 3, 2025 · We show that by reusing weights from attention layers, it is possible to distill a large transformer into a large hybrid-linear RNN with minimal ...
-
[19]
None### Summary of Mathematical Formulation for Vanilla RNN
-
[20]
Experimental Analysis of the Real-time Recurrent Learning AlgorithmApr 5, 2007 · The real-time recurrent learning algorithm is a gradient-following learning algorithm for completely recurrent networks running in continually sampled time.
-
[21]
[PDF] Finding Structure in TimeFinding Structure in Time. JEFFREY L. ELMAN. University of Calcfornia, San Riego. Time underlies many interesting human behaviors. Thus, the question of how to.Missing: paper | Show results with:paper
-
[22]
Finding structure in time - ScienceDirect.comTime is represented implicitly by its effects on processing using recurrent links, where hidden unit patterns are fed back to themselves.
-
[23]
[PDF] arXiv:1601.06581v2 [cs.CL] 28 Jan 2016Jan 28, 2016 · The algorithm employs a speech-to-character unidirectional recurrent neural network (RNN), which is end-to-end trained with connectionist ...
-
[24]
Recurrent neural network based language model - ISCA ArchiveA new recurrent neural network based language model (RNN LM) with applications to speech recognition is presented.
-
[25]
[PDF] A Neural Transducer - NIPS papersA Neural Transducer makes incremental predictions as input arrives, unlike sequence-to-sequence models, and can produce outputs as data comes in.
-
[26]
Bidirectional recurrent neural networks | IEEE Journals & MagazineAbstract: In the first part of this paper, a regular recurrent neural network (RNN) is extended to a bidirectional recurrent neural network (BRNN).
-
[27]
Offline Handwriting Recognition with Multidimensional Recurrent ...This paper introduces a globally trained offline handwriting recogniser that takes raw pixel data as input.Missing: 2005 | Show results with:2005
-
[28]
[1211.5063] On the difficulty of training Recurrent Neural NetworksNov 21, 2012 · There are two widely known issues with properly training Recurrent Neural Networks, the vanishing and the exploding gradient problems.
-
[29]
[PDF] Understanding Feature Selection and Feature Memorization ... - arXivMar 3, 2019 · In the modern literature, it is referred to as Vanilla RNN. Its state- update equation is given by (3). st. = tanh(Wxt + Ust−1 + b).
-
[30]
[PDF] Learning long-term dependencies with gradient descent is difficultBengio, P. Frasconi, P. Simard, "The problem of learning long- term dependencies in recurrent networks," invited paper at the IEEE. International ...
-
[31]
[PDF] LONG SHORT-TERM MEMORY 1 INTRODUCTIONHochreiter, S. and Schmidhuber, J. (1997). LSTM can solve hard long time lag problems. In. Advances in Neural Information Processing Systems 9. MIT ...
-
[32]
[PDF] Recurrent Nets that Time and Count Felix A. Gers Jiurgen ... - IDSIAPeephole connections from within the cell (or recurrent connections from gates) require a refinement of L STM 's update scheme. Updates for peephole LSTM. E ach ...
-
[33]
[PDF] An Empirical Exploration of Recurrent Network ArchitecturesWe conducted a thor- ough architecture search where we evaluated over ten thousand different RNN architectures, and identified an architecture that outperforms.
-
[34]
Echo state network - ScholarpediaSep 6, 2007 · Echo state networks (ESN) provide an architecture and supervised learning principle for recurrent neural networks (RNNs).
-
[35]
[PDF] The “echo state” approach to analysing and training recurrent neural ...Jan 25, 2010 · Jaeger(2001): The ”echo state” approach to analysing and training recurrent neural networks. GMD Report. 148, German National Research Center ...
-
[36]
[PDF] Real-Time Computing Without Stable States - IGI, TU-GrazOur approach provides new perspectives for the interpretation of neural coding, the design of experiments and data analysis in neurophysiology, and the so-.
-
[37]
Neural Machine Translation by Jointly Learning to Align and ... - arXivSep 1, 2014 · Abstract:Neural machine translation is a recently proposed approach to machine translation. Unlike the traditional statistical machine ...
-
[38]
[1602.06023] Abstractive Text Summarization Using Sequence-to ...Feb 19, 2016 · In this work, we model abstractive text summarization using Attentional Encoder-Decoder Recurrent Neural Networks, and show that they achieve state-of-the-art ...Missing: rnn seminal
-
[39]
[PDF] Grounded Compositional Semantics for Finding and Describing ...The DT-RNN has several important differences to previous RNN models of Socher et al. (2011a) and. (Socher et al., 2011b; Socher et al., 2011c). These.
-
[40]
[PDF] Parsing Natural Scenes and Natural Language with Recursive ...Abstract. Recursive structure is commonly found in the inputs of different modalities such as natural scene images or natural language sentences.
-
[41]
[PDF] Recursive Deep Models for Semantic Compositionality Over a ...In particular we will de- scribe and experimentally compare our new RNTN model to recursive neural networks (RNN) (Socher et al., 2011b) and matrix-vector RNNs ...
-
[42]
[1410.5401] Neural Turing Machines - arXivOct 20, 2014 · Neural Turing Machines can infer simple algorithms such as copying, sorting, and associative recall from input and output examples.
-
[43]
Hybrid computing using a neural network with dynamic external ...Oct 12, 2016 · Here we introduce a machine learning model called a differentiable neural computer (DNC), which consists of a neural network that can read from and write to an ...
-
[44]
[PDF] Neural Networks for Machine Learning Lecture 6a Overview of minirmsprop: Divide the learning rate for a weight by a running average of the magnitudes of recent gradients for that weight. – This is the mini-‐batch version of ...
-
[45]
[1412.6980] Adam: A Method for Stochastic Optimization - arXivDec 22, 2014 · We introduce Adam, an algorithm for first-order gradient-based optimization of stochastic objective functions, based on adaptive estimates of lower-order ...
-
[46]
A Theoretically Grounded Application of Dropout in Recurrent ...Dec 16, 2015 · A Theoretically Grounded Application of Dropout in Recurrent Neural Networks. Authors:Yarin Gal, Zoubin Ghahramani.
-
[47]
Scheduled Sampling for Sequence Prediction with Recurrent Neural ...Jun 9, 2015 · We propose a curriculum learning strategy to gently change the training process from a fully guided scheme using the true previous token, towards a less guided ...
-
[48]
Professor Forcing: A New Algorithm for Training Recurrent NetworksOct 27, 2016 · The Teacher Forcing algorithm trains recurrent networks by supplying observed sequence values as inputs during training and using the network's own one-step- ...
-
[49]
Deep Speech: Scaling up end-to-end speech recognition - arXivDec 17, 2014 · We present a state-of-the-art speech recognition system developed using end-to-end deep learning. Our architecture is significantly simpler than traditional ...
-
[50]
Hybrid CTC/Attention Architecture for End-to-End Speech RecognitionOct 16, 2017 · This paper proposes hybrid CTC/attention end-to-end ASR, which effectively utilizes the advantages of both architectures in training and decoding.
-
[51]
Learning long-term dependencies with gradient descent is difficultWe show why gradient based learning algorithms face an increasingly difficult problem as the duration of the dependencies to be captured increases.
-
[52]
Framewise phoneme classification with bidirectional LSTM and ...In this paper, we present bidirectional Long Short Term Memory (LSTM) networks, and a modified, full gradient version of the LSTM learning algorithm.
-
[53]
[PDF] DataStories at SemEval-2017 Task 4: Bidirectional LSTM with ...In this paper, we present two deep-learning sys- tems for short text sentiment analysis developed for SemEval-2017 Task 4 “Sentiment Analysis in. Twitter”.
-
[54]
[PDF] Generating Text with Recurrent Neural NetworksRNNs, with a high-dimensional hidden state, are used to predict the next character in text, using a new MRNN architecture with multiplicative connections.Missing: seminal | Show results with:seminal
-
[55]
Sequence Transduction with Recurrent Neural Networks - arXivNov 14, 2012 · This paper introduces an end-to-end, probabilistic sequence transduction system, based entirely on RNNs, that is in principle able to transform any input ...
-
[56]
[PDF] Recurrent Neural Networks for Time Series Forecasting - arXivJan 1, 2019 · This article presents a recurrent neural network based time series forecasting frame- work covering feature engineering, feature importances, ...Missing: seminal | Show results with:seminal
-
[57]
(PDF) RNN-Autoencoder Approach for Anomaly Detection in Power ...This research proposes to use a combined recurrent neural network (RNN)-autoencoder approach as a "normal" behavior model (NBM) with the Mahalanobis Distance ( ...
-
[58]
Efficient Neural Audio SynthesisWe first describe a single-layer recurrent neural network, the WaveRNN, with a dual softmax layer that matches the quality of the state-of-the-art WaveNet model ...
-
[59]
In-Depth Insights into the Application of Recurrent Neural Networks ...By capturing long-term dependencies in time series, RNNs can accurately forecast future changes in traffic conditions. Mainly, variants of RNNs, like LSTMs and ...
-
[60]
Deep Recurrent Q-Learning for Partially Observable MDPs - arXivJul 23, 2015 · This article investigates the effects of adding recurrency to a Deep Q-Network (DQN) by replacing the first post-convolutional fully-connected layer with a ...
-
[61]
Long-term Recurrent Convolutional Networks for Visual Recognition ...Nov 17, 2014 · We develop a novel recurrent convolutional architecture suitable for large-scale visual learning which is end-to-end trainable.
-
[62]
[PDF] Solving Deep Memory POMDPs with Recurrent Policy GradientsThis paper presents Recurrent Policy Gradients, a model- free reinforcement learning (RL) method creating limited-memory sto- chastic policies for partially ...
-
[63]
Implementation of edge AI for early fault detection in IoT networksOct 16, 2025 · The architecture leverages recurrent neural networks and autoencoders optimized for time-series anomaly detection, enabling local inference ...