Fact-checked by Grok 2 weeks ago
References
-
[1]
[PDF] Gradient-Based Learning Applied to Document RecognitionThis paper reviews var- ious methods applied to handwritten character recognition and compares them on a standard handwritten digit recog- nition task.
-
[2]
[PDF] Convolutional Networks and Applications in Vision - Yann LeCunNowlan and J. Platt, “A convolutional neural network hand tracker.” San Mateo, CA: Morgan Kaufmann, 1995, pp. 901–908.Missing: definition | Show results with:definition
-
[3]
[PDF] Convolutional Neural Networks - SciTePressIn this paper, the Convolution Neural Network's history and structure are summarized. And then several areas of Convolutional Neural Network applications are ...Missing: definition | Show results with:definition
-
[4]
[PDF] Handwritten Digit Recognition with a Back-Propagation NetworkThe main point of this paper is to show that large back-propagation (BP) net- works can be applied to real image-recognition problems without a large, complex.
-
[5]
[PDF] Evolution of Convolutional Neural Network (CNN) - arXivubiquitous in various applications running in cloud to edge applications ... Convolutional-Neural-Network-CNN-Compute- vs-Memory-bandwidth-for-Edge-AI.Missing: definition | Show results with:definition
-
[6]
NoneBelow is a merged summary of Convolutional Neural Networks (CNNs) based on the provided segments from http://yann.lecun.com/exdb/publis/pdf/lecun-98.pdf. To retain all information in a dense and organized manner, I’ll use a combination of text and a table to capture the details efficiently. The summary consolidates overlapping information while preserving unique details from each segment.
-
[7]
[1511.08458] An Introduction to Convolutional Neural Networks - arXivNov 26, 2015 · This document provides a brief introduction to CNNs, discussing recently published papers and newly formed techniques in developing these brilliantly fantastic ...Missing: principles | Show results with:principles
-
[8]
Convolutional neural networks: an overview and application in ...Jun 22, 2018 · Convolutional neural network is composed of multiple building blocks, such as convolution layers, pooling layers, and fully connected layers, ...Missing: core seminal
-
[9]
Visual Feature Extraction by a Multilayered Network of Analog ...Abstract: A new type of visual feature extracting network has been synthesized, and the response of the network has been simulated on a digital computer.
-
[10]
Cognitron: A self-organizing multilayered neural networkFukushima, K.: Visual feature extraction by a multilayered network of analog threshold elements. IEEE Trans. Syst. Sci. Cybernetics SSC-5, 322–333 (1969).
-
[11]
[PDF] Gradient-based learning applied to document recognitionThis paper reviews var- ious methods applied to handwritten character recognition and compares them on a standard handwritten digit recog- nition task.
-
[12]
Deep Learning in a Nutshell: Core Concepts | NVIDIA Technical BlogNov 3, 2015 · In the early 2010s, it was shown that combining GPUs with activation functions that offered better gradient flow was sufficient to train deep ...
-
[13]
Very Deep Convolutional Networks for Large-Scale Image ... - arXivSep 4, 2014 · In this work we investigate the effect of the convolutional network depth on its accuracy in the large-scale image recognition setting.
-
[14]
[1409.4842] Going Deeper with Convolutions - arXivSep 17, 2014 · Access Paper: View a PDF of the paper titled Going Deeper with Convolutions, by Christian Szegedy and 7 other authors. View PDF · TeX Source.
-
[15]
[1512.03385] Deep Residual Learning for Image Recognition - arXivDec 10, 2015 · We present a residual learning framework to ease the training of networks that are substantially deeper than those used previously.
-
[16]
EfficientNet: Rethinking Model Scaling for Convolutional Neural ...May 28, 2019 · In this paper, we systematically study model scaling and identify that carefully balancing network depth, width, and resolution can lead to ...
-
[17]
[2201.03545] A ConvNet for the 2020s - arXivJan 10, 2022 · The outcome of this exploration is a family of pure ConvNet models dubbed ConvNeXt. Constructed entirely from standard ConvNet modules, ...
-
[18]
Convolutional Neural Networks (CNNs / ConvNets)Convolutional Neural Networks take advantage of the fact that the input consists of images and they constrain the architecture in a more sensible way. In ...
-
[19]
[PDF] Pooling Methods in Deep Neural Networks, a Review - arXivIn this article, we studied some of the pooling methods used in CNNs. We divided pooling methods into two categories: popular methods and novel methods. In ...
-
[20]
[PDF] arXiv:1312.4400v3 [cs.NE] 4 Mar 2014Mar 4, 2014 · We propose a novel deep network structure called “Network In Network”(NIN) to enhance model discriminability for local patches within the ...
-
[21]
[PDF] Quantifying Translation-Invariance in Convolutional Neural NetworksUsing translation-sensitivity maps, we quantify the degree of translation-invariance of several CNN architectures trained on an expanded version of the MNIST ...
-
[22]
How Convolutional Neural Networks Deal with Aliasing - arXivFeb 15, 2021 · Aliasing, a downsampling side-effect, occurs in CNNs. This paper explores how CNNs implicitly deal with it, showing they don't prevent it in ...
-
[23]
Gradient-based learning applied to document recognitionThis paper reviews various methods applied to handwritten character recognition and compares them on a standard handwritten digit recognition task.
-
[24]
Efficient Convolutional Neural Networks for Mobile Vision ApplicationsApr 17, 2017 · MobileNets are based on a streamlined architecture that uses depth-wise separable convolutions to build light weight deep neural networks.
-
[25]
Learning Spatiotemporal Features with 3D Convolutional NetworksWe propose a simple, yet effective approach for spatiotemporal feature learning using deep 3-dimensional convolutional networks (3D ConvNets)
-
[26]
[PDF] Rectified Linear Units Improve Restricted Boltzmann MachinesIn ICASSP, Dallas, TX,. USA, 2010. Nair, V. and Hinton, G. E. Implicit mixtures of restricted boltzmann machines. In Neural information processing systems, 2008 ...
-
[27]
[PDF] Rectifier Nonlinearities Improve Neural Network Acoustic ModelsThe leaky rectified linear function (LReL) has a non-zero gradient over its entire domain, unlike the standard ReL function. final system performance. We ...
-
[28]
Batch Normalization: Accelerating Deep Network Training by ... - arXivFeb 11, 2015 · Batch Normalization allows us to use much higher learning rates and be less careful about initialization. It also acts as a regularizer, in some cases ...
-
[29]
[1607.06450] Layer Normalization - arXivJul 21, 2016 · In this paper, we transpose batch normalization into layer normalization by computing the mean and variance used for normalization from all of the summed ...
-
[30]
[1803.08494] Group Normalization - arXivMar 22, 2018 · In this paper, we present Group Normalization (GN) as a simple alternative to BN. GN divides the channels into groups and computes within each group the mean ...
-
[31]
[PDF] ImageNet Classification with Deep Convolutional Neural NetworksKrizhevsky, I. Sutskever, and R.R. Salakhutdinov. Improving neural net- works by preventing co-adaptation of feature detectors. arXiv preprint arXiv:1207.0580, ...
-
[32]
Deep learning for pedestrians: backpropagation in CNNs - arXivNov 29, 2018 · The goal of this document is to provide a pedagogical introduction to the main concepts underpinning the training of deep neural networks using gradient ...
-
[33]
[PDF] Backpropagation Applied to Handwritten Zip Code RecognitionPrevious work performed on recognizing simple digit images (LeCun. 1989) showed that good generalization on complex tasks can be obtained by designing a network ...
-
[34]
On the momentum term in gradient descent learning algorithmsThe momentum term improves the speed of convergence by bringing some eigen components of the system closer to critical damping.
-
[35]
[1412.6980] Adam: A Method for Stochastic Optimization - arXivDec 22, 2014 · We introduce Adam, an algorithm for first-order gradient-based optimization of stochastic objective functions, based on adaptive estimates of lower-order ...
-
[36]
[PDF] Convolutional Neural Networks Are Not Invariant to Translation, but ...In fact, multiple studies have reported highly limited translation invariance in CNNs (Kauderer-. Abrams, 2017; Gong et al., 2014; Azulay and Weiss, 2019; Chen ...
-
[37]
[1602.07576] Group Equivariant Convolutional Networks - arXivFeb 24, 2016 · We introduce Group equivariant Convolutional Neural Networks (G-CNNs), a natural generalization of convolutional neural networks that reduces sample complexity.
-
[38]
How transferable are features in deep neural networks? - arXivNov 6, 2014 · Title:How transferable are features in deep neural networks? Authors:Jason Yosinski, Jeff Clune, Yoshua Bengio, Hod Lipson. View a PDF of the ...
-
[39]
Visualizing and Understanding Convolutional Networks - arXivNov 12, 2013 · View a PDF of the paper titled Visualizing and Understanding Convolutional Networks, by Matthew D Zeiler and 1 other authors. View PDF.
-
[40]
U-Net: Convolutional Networks for Biomedical Image SegmentationMay 18, 2015 · U-Net: Convolutional Networks for Biomedical Image Segmentation. Authors:Olaf Ronneberger, Philipp Fischer, Thomas Brox.
-
[41]
Unsupervised Domain Adaptation by Backpropagation - arXivSep 26, 2014 · View a PDF of the paper titled Unsupervised Domain Adaptation by Backpropagation, by Yaroslav Ganin and 1 other authors. View PDF. Abstract ...
-
[42]
[PDF] Rich Feature Hierarchies for Accurate Object Detection and ...1A tech report describing R-CNN first appeared at http://arxiv. org/abs/1311.2524v1 in Nov. 2013. Our system is also quite efficient. The only class ...
-
[43]
You Only Look Once: Unified, Real-Time Object Detection - arXivJun 8, 2015 · YOLO is a real-time object detection approach using a single neural network to predict bounding boxes and class probabilities from full images. ...
-
[44]
[PDF] Fully Convolutional Networks for Semantic SegmentationFully convolutional networks can efficiently learn to make dense predictions for per-pixel tasks like semantic segmen- tation. We show that a fully ...
-
[45]
[PDF] Learning Spatiotemporal Features With 3D Convolutional NetworksIn this paper we propose to learn spatio-temporal features using deep 3D ConvNet. We empirically show that these learned features with a sim- ple linear ...
-
[46]
Optimized deep learning for brain tumor detection: a hybrid ... - NatureAug 26, 2025 · Their model demonstrated effective tumor detection with high performance when validated on different datasets. Choudhury et al. proposed a CNN ...
-
[47]
Convolutional Neural Networks for Sentence Classification - arXivAug 25, 2014 · We report on a series of experiments with convolutional neural networks (CNN) trained on top of pre-trained word vectors for sentence-level classification ...
-
[48]
Character-level Convolutional Networks for Text Classification - arXivSep 4, 2015 · This article offers an empirical exploration on the use of character-level convolutional networks (ConvNets) for text classification.
-
[49]
An Empirical Evaluation of Generic Convolutional and Recurrent ...Mar 4, 2018 · Our results indicate that a simple convolutional architecture outperforms canonical recurrent networks such as LSTMs across a diverse range of tasks and ...
-
[50]
Chemi-Net: A Molecular Graph Convolutional Network for Accurate ...We developed Chemi-Net, a completely data-driven, domain knowledge-free, deep learning method for ADME property prediction.2. Results · 2.1. Mt-Dnn Method Of... · 4. Materials And Methods
-
[51]
PyTorch: An Imperative Style, High-Performance Deep Learning ...Dec 3, 2019 · In this paper, we detail the principles that drove the implementation of PyTorch and how they are reflected in its architecture. We emphasize ...
-
[52]
PyTorchPyTorch Foundation is the deep learning community home for the open source PyTorch framework and ecosystem.Get Started · Tutorials · Previous PyTorch Versions · Torch.nnMissing: 2016 dynamic intuitive pre-
-
[53]
TensorFlow - Google's latest machine learning system, open ...We're proud to announce the open source release of TensorFlow -- our second-generation machine learning system, specifically designed to correct these ...
-
[54]
Keras: The high-level API for TensorFlowJun 8, 2023 · Keras is the high-level API of the TensorFlow platform. It provides an approachable, highly-productive interface for solving machine learning (ML) problems.The Sequential model · Working with RNNs · Serialization and saving
-
[55]
TensorFlowTensorFlow makes it easy to create ML models that can run in any environment. Learn how to use the intuitive APIs through interactive code samples.Tutorials · TensorFlow API Versions · About · TensorFlow.js
- [56]
-
[57]
An in-depth look at Google's first Tensor Processing Unit (TPU)May 12, 2017 · Because a TPU runs at 700MHz, a TPU can compute 65,536 × 700,000,000 = 46 × 1012 multiply-and-add operations or 92 Teraops per second (92 × 1012) ...
-
[58]
TPU architecture | Google Cloud DocumentationTo perform the matrix operations, the TPU loads the parameters from HBM memory into the Matrix Multiplication Unit (MXU). An illustration of how a TPU loads ...Missing: CNN | Show results with:CNN
-
[59]
[PDF] compressing deep neural networks with pruning, trained quantizationFor pruned AlexNet, we are able to quantize to 8-bits (256 shared weights) for each. CONV layers, and 5-bits (32 shared weights) for each FC layer without any ...
-
[60]
[PDF] A Survey of Model Compression and Acceleration for Deep Neural ...Jun 14, 2020 · [8] showed that 8-bit quantization of the parameters can result in significant speed-up with minimal loss of accuracy. The work in [9] used 16- ...<|control11|><|separator|>
-
[61]
[PDF] Pruning and Quantization for Deep Neural Network AccelerationQuantization reduces computations by reducing the precision of the datatype. Weights, biases, and activations may be quantized typically to 8-bit integers ...
-
[62]
[PDF] Systems for Parallel and Distributed Large-Model Deep Learning ...Jan 6, 2023 · Standard DL training typically holds a model's parameters on the memory of an accelerator (e.g. a GPU) and uses sampled data to compute gradient ...
-
[63]
Perception Model Training for Autonomous Vehicles with Tensor ...Apr 26, 2024 · We explore how tensor parallel convolutional neural network (CNN) training can help reduce the GPU memory footprint.
-
[64]
ZeRO: Memory Optimizations Toward Training Trillion Parameter ...Oct 4, 2019 · We develop a novel solution, Zero Redundancy Optimizer (ZeRO), to optimize memory, vastly improving training speed while increasing the model size that can be ...
-
[65]
[PDF] Scaling Distributed Deep Learning Workloads beyond the Memory ...Aug 26, 2020 · Abstract—The dedicated memory of hardware accelerators can be insufficient to store all weights and/or intermediate states of large deep ...<|control11|><|separator|>
- [66]
-
[67]
A Hybrid Edge Classifier: Combining TinyML-Optimised CNN ... - arXivIn this work, we are proposing a hybrid software-hardware edge classifier aimed at the extreme edge near-sensor systems.