Fact-checked by Grok 2 weeks ago

Quantum neural network

A quantum neural network (QNN) is a architecture that integrates principles with classical structures, typically implemented via parameterized quantum circuits to encode, process, and learn from data—often quantum data—exploiting phenomena like superposition and entanglement for potentially enhanced expressivity and efficiency compared to classical counterparts. These models function as a subclass of variational quantum algorithms, where quantum gates with tunable parameters mimic the weights and activations of traditional neural networks, enabling tasks such as classification and optimization on noisy intermediate-scale quantum (NISQ) devices. The conceptual origins of QNNs trace back to the early 2000s, when initial proposals vaguely combined with neurocomputing ideas, inspired by models like the McCulloch-Pitts neuron from 1943 but adapted to quantum constraints such as polylogarithmic usage. A more systematic exploration emerged around 2014, addressing the lack of unified frameworks amid scattered ideas, while modern definitions solidified post-2020 as hybrid quantum-classical systems for , driven by advances in variational quantum algorithms and NISQ . Key milestones include demonstrations of universal quantum computation via feedforward QNNs in 2020 and dissipative variants for low-memory training in 2022. Structurally, QNNs often employ quantum circuit models such as quantum Boltzmann machines (QBMs) or convolutional variants (QCNNs), where data encoding occurs through or embedding, followed by layers of parameterized unitaries and measurements to output predictions. Advantages include higher representational capacity due to —potentially achieving exponential scaling in certain expressive dimensions—and faster training on quantum hardware for specific tasks, as evidenced by numerical benchmarks showing superior effective dimensions over classical networks. However, challenges persist, including sensitivity to noise in current quantum devices and the "barren plateau" problem, where gradients vanish during optimization, limiting scalability. In applications, QNNs have been realized for supervised learning like image classification using repeat-until-success circuits for nonlinear activations, generative modeling via adversarial setups, and even Gaussian process approximations in physics simulations, with experimental validations on platforms like IBM Quantum. Ongoing research explores hybrid quantum-classical convolutional networks for hierarchical feature extraction and noise-resilient variants, positioning QNNs as a promising frontier in quantum machine learning despite debates over their precise advantages in the NISQ era.

Overview

Definition and motivation

Quantum neural networks (QNNs) are hybrid computational models that integrate with , typically implemented as parameterized quantum circuits designed to process either classical or quantum for tasks such as , , and optimization. These circuits encode input into quantum states, apply a series of unitary operations via quantum gates with tunable parameters, and extract outputs through measurements, enabling the network to learn patterns by adjusting parameters to minimize a . Unlike purely classical neural networks, which operate on bits, QNNs leverage qubits to perform computations that exploit quantum phenomena, potentially offering advantages in handling complex, high-dimensional datasets. Central to QNNs are data encoding strategies that map classical inputs to , with common methods including amplitude encoding, where data vectors are directly represented in the amplitudes of a for compact storage in logarithmic qubits, and angle encoding, which embeds features into rotation angles of quantum gates for simpler implementation on near-term . Outputs are typically obtained by expectation values of operators, such as Pauli strings, which provide probabilistic estimates that can be post-processed to yield predictions. These elements allow QNNs to perform learning tasks in a variational , where classical optimizers update quantum parameters iteratively. The primary motivation for developing QNNs stems from the potential for exponential speedups in processing high-dimensional feature spaces, enabled by —which allows simultaneous evaluation of multiple states—and entanglement, which correlates qubits to capture intricate data dependencies that challenge classical neural networks in areas like optimization and . Classical models often struggle with of dimensionality in such tasks, requiring vast computational resources, whereas QNNs could theoretically navigate these spaces more efficiently by representing exponentially many configurations compactly. This promise has driven interest, particularly for applications in , , and image analysis, where quantum advantages might address limitations in and expressivity. Early conceptual foundations for QNNs emerged in the 1990s from quantum computing proposals, with independent works by exploring quantum analogs to neural processing for enhanced information representation and by Ron Chrisley investigating quantum effects in learning systems. However, practical motivation intensified during the Noisy Intermediate-Scale Quantum (NISQ) era around 2018–2020, as accessible quantum hardware enabled initial implementations of variational QNNs, shifting focus from theoretical speculation to empirical exploration of hybrid quantum-classical learning.

Historical development

The concept of quantum neural networks (QNNs) originated in the mid-1990s, drawing from early explorations in theory and classical neural models. In 1995, proposed the foundational idea of quantum neural computing, focusing on quantizing probabilistic distributions in associative memories like the to leverage for enhanced storage capacity. This was followed in 1998 by Dan Ventura and Tony R. Martinez, who introduced a quantum associative memory model based on the , utilizing to achieve exponential scaling in pattern storage compared to classical counterparts. During the and early , theoretical developments expanded to quantum analogs of probabilistic neural architectures. A key advancement came in with the proposal of the quantum (QBM) by Mohammad H. Amin et al., which generalized classical using a transverse-field Ising to model quantum distributions for generative learning tasks. This period emphasized quantum-inspired enhancements to classical models, laying groundwork for hybrid approaches amid limited quantum hardware. The 2010s marked a shift toward variational and gate-based QNNs, enabled by advances in frameworks. Maria Schuld and colleagues provided a seminal overview in 2015, systematically classifying QNN proposals and highlighting their potential for feature mapping in Hilbert spaces. In 2019, Iris Cong, Soonwon Choi, and Mikhail D. Lukin introduced quantum convolutional neural networks (QCNNs), a parameterized architecture inspired by classical CNNs, demonstrating efficient phase recognition and error correction with logarithmic depth scaling on near-term devices. Entering the 2020s, the field transitioned to noisy intermediate-scale quantum (NISQ) implementations, focusing on practical demonstrations. In 2021, a quantum Hopfield associative was experimentally realized on IBM's quantum , showcasing retrieval with up to four qubits and fidelity around 60% for three despite . By 2022, variational QCNNs were implemented on superconducting processors, achieving an average deviation of 0.23 from ideal values in recognizing symmetry-protected topological phases using seven qubits. Recent advancements in have emphasized modular and secure QNN designs for . For instance, Nouhaila Innan et al. proposed next-generation QNNs incorporating optimization strategies and quantum to mitigate NISQ noise while enhancing privacy in distributed settings. This evolution reflects a progression from abstract quantum-inspired classical models to fully gate-based QNNs deployable on NISQ hardware, driven by interdisciplinary contributions from researchers like Maria Schuld in establishing rigorous paradigms.

Theoretical Foundations

Relation to classical neural networks

Quantum neural networks (QNNs) draw direct analogies to classical neural networks by mapping foundational components to quantum equivalents. In classical architectures, neurons inputs through weighted sums followed by functions, whereas in QNNs, quantum gates serve as the analogue to neurons, applying unitary transformations to states representing inputs. Layers in QNNs correspond to sequential applications of these unitary operators, similar to how classical layers stack linear transformations and nonlinearities. functions in classical networks introduce nonlinearity; in QNNs, this role is fulfilled by quantum measurements, which collapse superposed states into probabilistic outcomes, effectively providing a nonlinear readout. A key extension lies in training mechanisms. Classical backpropagation computes gradients via the chain rule on deterministic functions, enabling efficient updates like the perceptron rule: \mathbf{w}_{\text{new}} = \mathbf{w}_{\text{old}} + \eta (y - \hat{y}) \mathbf{x}, where \mathbf{w} are weights, \eta is the , y is the true , \hat{y} is the , and \mathbf{x} is the input. In contrast, QNNs employ the parameter-shift rule to evaluate gradients of values from parameterized quantum , shifting circuit parameters by specific offsets (e.g., \pm \pi/2 for Pauli rotations) and measuring the difference in outputs, which avoids explicit differentiation of non-differentiable quantum operations. Nonlinearity in QNNs arises inherently from quantum interference, where superposed paths in the circuit amplify or suppress amplitudes constructively or destructively, enabling complex pattern separation beyond classical linear models. Fundamental differences highlight QNNs' quantum advantages. Outputs in QNNs are inherently probabilistic due to projective measurements on quantum states, yielding expectation values over repeated runs rather than deterministic results, which introduces stochasticity akin to but distinct from classical softmax probabilities. Moreover, provides inherent parallelism, allowing QNNs to process exponentially large feature spaces simultaneously; for instance, quantum kernel methods map classical data to high-dimensional Hilbert spaces via feature maps like angle embedding, capturing correlations intractable for classical kernels. This enables QNNs to handle datasets with exponential effective dimensionality, contrasting with classical networks' sequential processing.

Quantum computing prerequisites

Quantum computing relies on fundamental , which differ markedly from classical computing. At its core is the , the basic unit of , which can exist in a superposition of states. Unlike a classical bit that is either or 1, a qubit's state is represented by a |\psi\rangle = \alpha |0\rangle + \beta |1\rangle, where \alpha and \beta are complex numbers satisfying |\alpha|^2 + |\beta|^2 = 1, allowing it to encode more information through probabilistic amplitudes. This superposition enables a single qubit to represent multiple possibilities simultaneously, providing an exponential scaling in computational power for systems with many qubits. Quantum operations are performed using unitary that manipulate states while preserving the normalization condition. The Hadamard gate, for instance, creates superposition by transforming |0\rangle to \frac{1}{\sqrt{2}} (|0\rangle + |1\rangle) and |1\rangle to \frac{1}{\sqrt{2}} (|0\rangle - |1\rangle), enabling parallel exploration of computational paths. Entanglement, another key feature, arises from gates like the controlled-NOT (CNOT), which links the state of two or more qubits such that the measurement of one instantly determines the others, regardless of distance; this correlation cannot be replicated classically and allows for non-local information processing. Quantum circuits compose sequences of these to evolve the system unitarily, often parameterized as U(\theta) where \theta represents trainable angles, facilitating adaptive computations. Measurement in collapses the superposition into a classical outcome, with probabilities given by |\alpha|^2 for |0\rangle and |\beta|^2 for |1\rangle, yielding probabilistic results that require repeated executions for . For systems involving mixed states—due to or partial knowledge—density matrices \rho = |\psi\rangle\langle\psi| for pure states or more general forms for ensembles provide a complete description, enabling the tracking of decoherence effects. These quantum primitives are essential for quantum neural networks (QNNs), where superposition allows parallel processing of multiple data representations, potentially accelerating feature extraction in high-dimensional spaces. Entanglement, meanwhile, captures complex correlations between features that classical networks struggle with, enhancing the modeling of interdependent variables in tasks like pattern recognition. Parameterized unitary circuits in QNNs leverage these properties to mimic neural transformations, bridging quantum mechanics with machine learning paradigms.

Models and Architectures

Quantum perceptrons

The quantum perceptron represents a basic building block in quantum neural networks, functioning as a single-layer that processes classical input vectors by encoding them into quantum states on . This model draws a direct analogy to the classical , but exploits and to potentially enhance expressivity for certain tasks. In its simplest form, the quantum perceptron operates on a single ancilla qubit, with inputs and weights integrated through parameterized quantum gates to produce a measurable output for . Input encoding typically involves mapping the components of an input \mathbf{x} = (x_1, \dots, x_n) into rotations on qubits, such as using angle encoding where each x_i (scaled to [0, \pi]) controls an R_y(x_i) gate applied to an initial |[0](/page/0)\rangle state, preparing a that embeds the classical data. The weights \mathbf{w} = (w_1, \dots, w_n) and bias b are then incorporated via additional parameterized unitaries, often implemented as R_y(2w_i) gates controlled by the input qubits on the ancilla, effectively computing a quantum analog of the \sum w_i x_i + b. This results in a final |\psi\rangle on the ancilla that captures the weighted input through phase and amplitude adjustments. The core operation concludes with a projective measurement on the ancilla qubit in the Z-basis, yielding an expectation value \langle Z \rangle = \langle \psi | Z | \psi \rangle that determines the classification. The output probability for the positive class is P(y=1) = \frac{1 - \langle Z \rangle}{2}, with \langle Z \rangle producing a signed value in [-1, 1] that can be thresholded at 0 to assign binary labels, mimicking the sign function in classical perceptrons but with inherent quantum nonlinearity. Early implementations of quantum perceptrons have been applied to XOR-like tasks, where the model processes entangled or superposed inputs to achieve accuracies unattainable by classical single-layer due to the problem's non-linearity. For example, a quantum perceptron network with two input qubits and controlled rotations demonstrated near-perfect performance on the quantum XOR problem, highlighting an advantage in handling entangled inputs that classical counterparts cannot efficiently learn without multiple layers.

Variational quantum circuits

Variational quantum circuits (VQCs) serve as the foundational trainable architecture in modern quantum neural networks (QNNs), enabling the approximation of complex functions through parameterized quantum operations optimized via classical feedback. These circuits consist of alternating layers of entangling gates, such as gates, and single-qubit rotations, forming a unitary operator U(\theta) = \prod R(\theta_i), where R(\theta_i) represents parameterized rotation gates like , RY, or RZ. This layered structure allows VQCs to generate entangled quantum states that capture non-local correlations, distinguishing them from classical neural networks and providing expressive power for tasks. Quantum perceptrons can be viewed as basic building blocks within these multi-layer variational forms. The output of a VQC is typically a quantum state |\psi(\theta)\rangle = U(\theta) |\phi_{\text{input}}\rangle, where |\phi_{\text{input}}\rangle encodes classical input data into the quantum system via or encoding. For measurement-based QNNs, the circuit's functionality is extracted through values of observables O, yielding f(\theta) = \langle \psi(\theta) | O | \psi(\theta) \rangle, which serves as the model's and is minimized or maximized during . This variational approach leverages the quantum-classical paradigm, where quantum hardware evaluates the and classical optimizers adjust the parameters \theta to fit data. Several variants of VQCs enhance their suitability for near-term quantum devices. Data re-uploading techniques repeatedly embed classical input data into the circuit by interleaving data-encoding layers with trainable unitary blocks, effectively increasing the circuit's depth and expressivity without requiring deeper native quantum operations, which is particularly useful for NISQ-era hardware with limited coherence times. Hardware-efficient ansatze prioritize shallow circuits tailored to specific quantum architectures, using native gate sets to minimize compilation overhead and error accumulation, as demonstrated in applications to molecular simulations and tasks. Recent advancements integrate VQCs with probabilistic models, showing that Haar-random QNNs—where unitaries are drawn from the —converge to in the large-dimensional limit, enabling kernel-based interpretations and improved in predictions. This connection, established through theoretical analysis of deep QNN architectures, opens avenues for hybrid models combining quantum expressivity with classical regression.

Quantum recurrent and convolutional networks

Quantum recurrent neural networks (QRNNs) extend classical recurrent architectures to quantum settings by incorporating parameterized quantum circuits that process sequential data through feedback mechanisms. These models typically employ time-step unitaries or partial measurements to simulate recurrence, allowing the quantum state to evolve while preserving quantum coherence for tasks involving temporal dependencies. A key variant is the quantum gated recurrent unit (QGRU), which adapts the classical GRU by using controlled rotations and variational quantum circuits to manage information flow via update and reset gates. In QGRUs, the hidden state update at time step t is given by |h_t\rangle = U_g \left( |h_{t-1}\rangle \otimes |x_t\rangle \right), where U_g represents parameterized unitaries, such as rotation operators controlled by the input |x_t\rangle, enabling selective retention of prior . This formulation leverages to potentially capture complex correlations more efficiently than classical counterparts, though it requires careful design to mitigate decoherence in near-term devices. Quantum convolutional neural networks (QCNNs) draw inspiration from classical CNNs by applying local entangling gates to mimic convolutional filters on quantum data encoded in qubit registers. These architectures use translationally invariant layers of parameterized two-qubit gates to extract spatial features, followed by pooling operations implemented via partial traces over subsets of qubits, which reduce dimensionality while preserving essential quantum information. The resulting structure scales logarithmically with the number of qubits, offering advantages in expressivity for pattern recognition in quantum states. Recent advancements include modularized quantum neural networks incorporating recurrent and convolutional elements for , achieving competitive accuracy on datasets like MNIST through designs that enhance trainability. Additionally, explorations of quantum in recurrent models have demonstrated improved sequential learning by distributing computations across multiple quantum processors, as shown in encoder-decoder frameworks integrating QGRUs.

Training and Optimization

Cost functions

In quantum neural networks (QNNs), cost functions quantify the discrepancy between predicted outputs—derived from quantum measurements—and target values, guiding the training of parameterized quantum circuits such as variational quantum circuits. These functions are typically formulated to account for the probabilistic nature of quantum measurements, where outputs are expectation values of observables or probabilities obtained via the Born rule. A common approach for regression tasks involves the mean squared error (MSE) on expectation values, defined as L(\theta) = \frac{1}{N} \sum_{i=1}^N \left( y_i - \langle O \rangle_i \right)^2, where y_i are the target values, \langle O \rangle_i is the measured expectation value of the observable O for the i-th input under parameters \theta, and N is the number of data points. This loss measures how closely the quantum model's predictions align with classical labels, as demonstrated in applications like materials science decoding where the QNN processes quantum states to approximate classical outputs. For classification tasks, the loss is adapted to the probabilities emerging from quantum measurements under the , which governs the collapse of the upon measurement. The loss takes the form of a categorical between the predicted p_k = |\langle k | \psi(\theta) \rangle|^2 for class k and the true label distribution, penalizing deviations in the quantum-encoded probability space. This adaptation enables QNNs to handle multi-class problems by leveraging the inherent superposition and in quantum states, as explored in text frameworks that treat documents as quantum superpositions. Quantum-specific cost functions often incorporate fidelity measures to directly compare quantum states or density matrices, particularly when training on quantum data. The fidelity loss, such as $1 - F(\rho, \sigma) where F is the quantum state fidelity between the predicted state \rho(\theta) and target \sigma, captures overlaps in the and is suitable for tasks like quantum state preparation or generative modeling. These losses are especially useful in quantum generative adversarial networks, where fidelity-based objectives ensure high-fidelity state generation while mitigating classical approximation errors. Additionally, estimating these costs involves handling from finite measurements, which introduces statistical variance; techniques like error mitigation or increased sampling shots are employed to refine estimates, reducing the impact on convergence in noisy intermediate-scale quantum devices. A foundational formulation for many QNN costs is the expectation value C(\theta) = \langle \psi(\theta) | H | \psi(\theta) \rangle, where |\psi(\theta)\rangle is the parameterized and H is a encoding the classical , such as an for optimization problems. This variational form allows the quantum expectation to proxy the desired , minimizing C(\theta) to find optimal parameters. Recent variations, particularly in 2025, have integrated privacy-preserving mechanisms into QNN costs, such as those leveraging to secure gradient sharing without exposing raw quantum data, enhancing applications in distributed .

Parameter optimization techniques

Parameter optimization in quantum neural networks (QNNs) relies on techniques that compute gradients or approximations thereof to minimize functions derived from quantum measurements. These methods must account for the unique challenges of quantum hardware, such as limited access to intermediate states and the need for circuit evaluations that can be executed on near-term devices. A primary approach is the parameter-shift rule, which provides an exact method for obtaining analytic gradients of expectation values in parameterized quantum circuits composed of Pauli rotation gates. The parameter-shift rule expresses the partial derivative of a quantum function f(\theta), typically an expectation value \langle \psi(\theta) | O | \psi(\theta) \rangle where O is a Pauli observable, as \frac{\partial f}{\partial \theta} = \frac{f(\theta + s/2) - f(\theta - s/2)}{s}, with shift s = \pi/2 for standard Pauli rotations e^{-i \theta P / 2} where P is a Pauli operator. This rule arises from the eigenvalue spectrum of Pauli generators, enabling gradient estimation through two additional circuit evaluations without auxiliary qubits or parameter decompositions. For general single-parameter gates, extensions of the rule use multiple shifts to handle arbitrary spectra, maintaining compatibility with variational QNN architectures. Another advanced technique is the quantum natural gradient (QNG), which incorporates the of the quantum parameter manifold to precondition standard . Unlike classical natural gradients that use the matrix, QNG employs the real part of the Fubini-Study , g_{ij} = \Re \langle \partial_i \psi | (I - |\psi\rangle\langle\psi|) | \partial_j \psi \rangle, where |\psi\rangle is the parameterized and \partial_i = \partial / \partial \theta_i. This metric quantifies infinitesimal distances in , leading to updates \Delta \theta = -\eta g^{-1} \nabla f that converge faster than vanilla gradients by aligning steps with the circuit's natural geometry. Computing g_{ij} requires additional circuit executions, often approximated block-diagonally for efficiency in QNN training. Hybrid quantum-classical optimization integrates classical algorithms with quantum oracles for gradient or function evaluation. First-order methods like leverage quantum-computed gradients via the parameter-shift rule, adapting learning rates based on moment estimates to handle noisy quantum outputs in QNNs. For derivative-free alternatives, the simultaneous perturbation stochastic approximation (SPSA) estimates gradients using only two circuit evaluations per parameter by perturbing parameters with random directions, making it robust to noise and suitable for high-dimensional QNN parameter spaces. These classical optimizers bridge the gap, enabling scalable training of QNNs on hardware. Recent advances in efficiency exploit for quantum computations, allowing multiple data samples to be processed simultaneously via and entanglement. This approach distributes evaluations across quantum resources, reducing wall-clock time for QNN on large datasets without increasing per-circuit overhead.

Barren plateaus

Barren plateaus represent a significant challenge in quantum neural networks (QNNs), characterized by the exponential vanishing of variances as the number of qubits increases, resulting in flat regions in the optimization landscape that hinder effective learning. This phenomenon leads to trainability cliffs, where the probability of finding non-zero gradients diminishes , making it difficult for optimization algorithms to navigate the parameter meaningfully. In deep QNNs, particularly those employing variational quantum circuits, the gradients become vanishingly small, often approaching zero across the vast majority of the parameter , which severely limits the of these models. The primary causes of barren plateaus stem from the concentration of measure in the high-dimensional of , where states and observables tend to cluster around their global averages due to the in dimensionality (scaling as $2^n for n qubits). When QNNs utilize Haar-random unitaries or random parameterized s that approximate 2-designs, this concentration induces flat landscapes, as the partial derivatives of the with respect to parameters exhibit zero and exponentially decaying variance. Specifically, the variance of the is given by \mathrm{Var}[\partial_\theta f] \approx 2^{-n}, illustrating the with the number of qubits n. This effect arises because the output of the quantum circuit, being a normalized state, concentrates sharply, amplifying the flatness in deeper or more complex architectures. To mitigate barren plateaus, several strategies have been developed, including the use of shallow circuits to limit depth and reduce the onset of variance , as deeper layers exacerbate the issue. Initialization techniques, such as those employing structured distributions that preserve , can prevent initial entrapment in flat regions by ensuring higher variance at the start of training. Additionally, layered ansätze, like those in quantum convolutional neural networks, have been shown to avoid barren plateaus entirely by restricting entanglement growth and maintaining polynomial scaling in variances. More recent advancements, including measurement-induced methods to control entanglement and optimization strategies in next-generation QNN frameworks, further address this challenge by enhancing trainability in larger systems.

Applications

Quantum machine learning tasks

Quantum neural networks (QNNs) have been applied to a variety of tasks, including , regression, and generative modeling, where they leverage and entanglement to process data in high-dimensional Hilbert spaces. In tasks, QNNs demonstrate particular efficacy for and multiclass problems by encoding classical features into quantum states and using variational circuits to learn decision boundaries. For instance, modular QNNs (mQNNs) have achieved high accuracy on quantum-adapted versions of the MNIST dataset for digit , reaching up to 98% test accuracy with shallow circuits. Single-qudit QNNs have shown competitive performance while using minimal quantum resources. Regression tasks with QNNs focus on predicting continuous outputs from quantum datasets, such as those generated from evolutions under noisy channels. Single-qubit QNNs (SQQNNs) have been employed for on synthetic quantum data, exhibiting mean squared errors comparable to classical neural networks but with advantages in capturing quantum correlations. The QDataSet repository provides quantum datasets for such tasks. Quantum convolutional neural networks (QCNNs) serve as an example architecture for applications. Generative modeling via quantum Boltzmann machines (QBMs) allows QNNs to sample from quantum probability distributions, outperforming classical Boltzmann machines in modeling complex quantum states. QNNs offer advantages in kernel estimation for (SVM)-like tasks, where quantum feature maps compute high-dimensional kernels exponentially faster than classical methods. Trained QNNs as neural quantum kernels have enhanced SVM classification on small datasets, reducing from quadratic to linear in feature dimensions. Empirical studies from 2025 highlight lower generalization errors in small-data regimes, attributing this to QNNs' equivalence to Gaussian processes, which provide and achieve test errors 10-15% below classical models on datasets with fewer than 100 samples.

Hybrid quantum-classical approaches

Hybrid quantum-classical approaches integrate quantum neural networks (QNNs) with classical neural networks to leverage the strengths of both paradigms, particularly in the noisy intermediate-scale quantum (NISQ) era where fully quantum implementations face hardware limitations. These methods typically involve embedding variational quantum circuits as differentiable layers within classical architectures, allowing quantum components to handle high-dimensional feature mappings while classical layers manage scalable computations like and post-processing. For instance, classical neural networks can preprocess input data—such as via —before encoding it into quantum states for processing by QNN layers, enabling efficient handling of complex datasets that exceed pure quantum capacities. Key frameworks facilitate this integration, such as PennyLane, which supports of hybrid computations through quantum circuits using techniques like the parameter-shift rule, and TensorFlow Quantum (TFQ), which enables rapid prototyping of hybrid models by interleaving quantum algorithms with 's classical ML pipelines. These tools allow for end-to-end differentiable pipelines, where gradients flow seamlessly from classical loss functions back through quantum layers via compatible optimization methods, such as adapted for quantum parameters. A primary benefit is the mitigation of through classical error correction mechanisms, where classical networks can refine noisy quantum outputs or implement redundancy in measurements to improve overall model robustness without requiring fault-tolerant quantum hardware. In practice, hybrid QNNs often employ a combined to balance classical and quantum contributions during training: L_{\text{total}} = L_{\text{classical}} + \lambda L_{\text{quantum}} Here, L_{\text{classical}} captures errors from classical components (e.g., on processed features), L_{\text{quantum}} measures quantum circuit fidelity or expectation values, and \lambda is a hyperparameter the trade-off; proceeds through quantum oracles by approximating gradients of parameterized quantum gates, enabling joint optimization. This setup has been applied in , where quantum feature maps extract molecular representations from classical classifiers to predict drug responses, achieving improved accuracy over purely classical models on benchmarks like the Genomics of Drug Sensitivity in Cancer (GDSC) dataset. Recent 2025 advancements extend QNNs to secure , incorporating quantum components for enhanced privacy in distributed training across devices; for example, quantum-secure aggregation protocols combined with classical federated averaging mitigate data leakage while preserving model utility in sensitive applications like healthcare. These developments highlight approaches' role in bridging quantum advantages, such as exponential feature spaces, with classical efficiency for real-world tasks.

Challenges and Limitations

Hardware and noise issues

The implementation of quantum neural networks (QNNs) on current hardware faces significant challenges due to the noisy intermediate-scale quantum (NISQ) era limitations, where devices typically feature 50-1,000 qubits with gate error rates around 0.1-1%. These constraints arise primarily from imperfect quantum operations and environmental interactions, hindering the reliable execution of the variational quantum circuits that underpin most QNN architectures. Key noise sources in QNNs include gate errors from imprecise control pulses, decoherence due to interactions with the —manifesting as relaxation and with times typically ranging from 50 μs to several milliseconds in advanced superconducting qubits, with leading examples exceeding 1 ms as of 2025—and readout noise during state . Gate errors distort unitary operations essential for entanglement generation, while decoherence limits circuit depth to shallow configurations viable only for small-scale QNNs. Readout noise, stemming from inaccuracies, further corrupts output probabilities, with error rates typically 1-5% in NISQ systems. These noise sources profoundly impact QNN performance by degrading entanglement , which is crucial for quantum in learning tasks, and introducing high variance in estimates. To achieve reliable or value estimates, QNN training often requires thousands to millions of shots per evaluation, escalating computational overhead and limiting applicability to modest datasets. also exacerbates trainability issues like barren plateaus by amplifying fluctuations in landscapes. Mitigation strategies for these hardware issues include error mitigation techniques such as zero-noise extrapolation (ZNE), which artificially amplifies noise in simulations and extrapolates to an ideal noiseless result, improving accuracy in QNN inference without full error correction. For future scalable QNNs, transitioning to fault-tolerant regimes relies on quantum error correction codes like the surface code, which can maintain logical fidelity below physical error thresholds of approximately 1%. In 2025, NISQ platforms enable proof-of-concept QNNs for simple classification tasks but remain insufficient for full-scale deployment due to cumulative error accumulation over multi-layer circuits. As of 2025, advances like Quantinuum's accelerated roadmap target universal fault-tolerant by 2030, with demonstrations of logical qubits achieving error rates as low as 10^{-5} to 10^{-6}, offering pathways to overcome current limitations.

Scalability and theoretical hurdles

One of the primary scalability challenges in quantum neural networks (QNNs) arises from the exponential growth in the dimension of the Hilbert space, which scales as d = 2^n for n qubits, leading to a curse of dimensionality that complicates optimization in high-dimensional parameter spaces. This exponential resource requirement manifests in the need for increasingly complex circuits and measurements as the number of qubits grows, rendering simulations and training infeasible on classical hardware beyond modest scales. Furthermore, trainability is severely limited beyond approximately 50 qubits due to the barren plateau phenomenon, where gradients of the loss landscape concentrate exponentially to near-zero values, with variance decaying as O(2^{-2n}), making parameter updates ineffective during optimization. Theoretical hurdles further constrain QNN performance, including expressivity bounds imposed by the unitary nature of quantum operations. Specifically, the ability of QNNs to approximate target functions, such as physical observables or entropies, is limited unless input states span a where the effective dimension satisfies D_{\text{in}} < D_{\text{total}}/2, often requiring ancillary qubits to restore full expressivity. The exacerbates these issues by prohibiting the duplication of unknown quantum states, which prevents straightforward data reuse across layers or in backpropagation-like training protocols, necessitating alternative strategies like measurement-based approaches. Additionally, open questions persist regarding proofs of quantum advantage for QNNs over classical neural networks, particularly whether QNNs can consistently outperform in tasks or if advantages are task-specific and bounded by the "No Free Lunch" theorem. Looking ahead, fault-tolerant QNNs are anticipated post-2030, enabled by roadmaps targeting universal quantum computers with millions of logical gates by 2029–2033, which could support scalable hybrid quantum-classical learning without current limitations serving as practical manifestations of these theoretical barriers. Emerging theoretical models, such as those showing QNNs with Haar-random unitaries converging to Gaussian processes in the large limit, offer promising avenues for understanding and mitigating expressivity constraints through kernel methods and .

References

  1. [1]
    The power of quantum neural networks - Nature
    Jun 24, 2021 · Quantum neural networks are a subclass of variational quantum algorithms that comprise quantum circuits containing parameterized gate ...
  2. [2]
    [2205.08154] Quantum neural networks - arXiv
    May 17, 2022 · We introduce dissipative quantum neural networks (DQNNs), which are designed for fully quantum learning tasks, are capable of universal quantum computation.
  3. [3]
    A review of Quantum Neural Networks: Methods, Models, Dilemma
    ### Summary of Quantum Neural Networks from https://arxiv.org/abs/2109.01840
  4. [4]
    [2104.07106] On quantum neural networks - arXiv
    Apr 12, 2021 · Abstract:The early definition of a quantum neural network as a new field that combines the classical neurocomputing with quantum computing ...
  5. [5]
    [quant-ph/0201144] Quantum Neural Networks - arXiv
    Jan 30, 2002 · This paper initiates the study of quantum computing within the constraints of using a polylogarithmic (O(\log^kn), k\geq 1) number of qubits and a ...
  6. [6]
    [1408.7005] The quest for a Quantum Neural Network - arXiv
    Aug 29, 2014 · This article presents a systematic approach to QNN research, which so far consists of a conglomeration of ideas and proposals.
  7. [7]
    Training deep quantum neural networks | Nature Communications
    Feb 10, 2020 · Here we propose a truly quantum analogue of classical neurons, which form quantum feedforward neural networks capable of universal quantum computation.
  8. [8]
    Hybrid quantum-classical-quantum convolutional neural networks
    Aug 28, 2025 · QCNNs implement convolutional and pooling operations directly on quantum states, inspired by the hierarchical structures of classical CNNs and ...
  9. [9]
    [2011.00027] The power of quantum neural networks - arXiv
    Oct 30, 2020 · We show that quantum neural networks are able to achieve a significantly better effective dimension than comparable classical neural networks.
  10. [10]
    [2106.04975] The dilemma of quantum neural networks - arXiv
    Jun 9, 2021 · Through systematic numerical experiments, we empirically observe that current QNNs fail to provide any benefit over classical learning models.
  11. [11]
    Realization of a quantum neural network using repeat-until-success ...
    Nov 21, 2023 · In this paper, we use repeat-until-success circuits enabled by real-time control-flow feedback to realize quantum neurons with non-linear activation functions.Results · Constructing A Qnn Using Rus... · Training A Qnn From...
  12. [12]
    Quantum neural networks form Gaussian processes | Nature Physics
    May 21, 2025 · One of the most interesting results regarding NNs is that fully connected models with a single hidden layer converge to Gaussian processes (GPs) ...
  13. [13]
    Parameterized quantum circuits as machine learning models
    Parameterized quantum circuits can be regarded as machine learning models with remarkable expressive power. This Review presents the components of these models.
  14. [14]
    Quantum data encoding: a comparative analysis of classical-to ...
    Oct 25, 2024 · We explored various classical-to-quantum mapping methods; ranging from basis encoding and angle encoding to amplitude encoding; for encoding ...
  15. [15]
    Design and analysis of quantum machine learning: a survey
    Mar 29, 2024 · In the paper, we survey the basic concepts, algorithms, applications and challenges of quantum machine learning.
  16. [16]
    On quantum neural computing - ScienceDirect.com
    This paper examines the notion of quantum neural computing in the context of several new directions in neural network research.
  17. [17]
    (PDF) Quantum Learning - ResearchGate
    Quantum Learning. July 1995. In book: New directions in cognitive science: Proceedings of the International Symposium, 4-9 ...
  18. [18]
    (PDF) The quest for a Quantum Neural Network - ResearchGate
    Aug 7, 2025 · This article presents a systematic approach to QNN research, which so far consists of a conglomeration of ideas and proposals.
  19. [19]
    [quant-ph/9807053] Quantum Associative Memory - arXiv
    Jul 18, 1998 · View a PDF of the paper titled Quantum Associative Memory, by Dan Ventura and Tony Martinez ... Hopfield network. The paper covers ...
  20. [20]
    [1601.02036] Quantum Boltzmann Machine - arXiv
    Jan 8, 2016 · We propose a new machine learning approach based on quantum Boltzmann distribution of a transverse-field Ising Hamiltonian.Missing: first | Show results with:first
  21. [21]
    [1409.3097] An introduction to quantum machine learning - arXiv
    Sep 10, 2014 · This contribution gives a systematic overview of the emerging field of quantum machine learning. It presents the approaches as well as technical details in an ...
  22. [22]
    [1810.03787] Quantum Convolutional Neural Networks - arXiv
    Oct 9, 2018 · We introduce and analyze a novel quantum machine learning model motivated by convolutional neural networks.
  23. [23]
    A quantum Hopfield associative memory implemented on an actual ...
    Dec 3, 2021 · In this work, we present a Quantum Hopfield Associative Memory (QHAM) and demonstrate its capabilities in simulation and hardware using IBM ...
  24. [24]
    Realizing quantum convolutional neural networks on a ... - NIH
    Jul 16, 2022 · Here, we realize a quantum convolutional neural network (QCNN) on a 7-qubit superconducting quantum processor to identify symmetry-protected topological (SPT) ...Variational State... · Fig. 2. Variational Ground... · Quantum Phase Recognition
  25. [25]
    Quantum working groups push for near-term use cases - IBM
    May 22, 2024 · In the past two years, IBM and collaborators launched five working groups to spur quantum algorithmic development in domains with promising ...
  26. [26]
    [2507.20537] Next-Generation Quantum Neural Networks - arXiv
    Jul 28, 2025 · This paper presents a framework to enhance QNN efficiency, security, and privacy using optimization strategies, defensive mechanisms, and  ...
  27. [27]
    Machine Learning with Quantum Computers - SpringerLink
    In stockMaria Schuld works as a researcher for the Toronto-based quantum computing ... Besides her numerous contributions to the field, she is a co-developer ...
  28. [28]
    Speeding up quantum perceptron via shortcuts to adiabaticity - Nature
    Mar 11, 2021 · In analogy with classical neurons, a quantum perceptron can be constructed as a qubit that encodes the nonlinear response to an input ...
  29. [29]
    Quantum activation functions for quantum neural networks - arXiv
    Jan 10, 2022 · Here we fill this gap with a quantum algorithm which is capable to approximate any analytic activation functions to any given order of its power series.
  30. [30]
    [1811.11184] Evaluating analytic gradients on quantum hardware
    Nov 27, 2018 · Evaluating analytic gradients on quantum hardware. Authors:Maria Schuld, Ville Bergholm, Christian Gogolin, Josh Izaac, Nathan Killoran.
  31. [31]
    Non-linear classification capability of quantum neural networks due ...
    Aug 20, 2024 · Here we show that effective non-linearities can be implemented in these platforms by exploiting the relationship between information processing and many-body ...Missing: interference | Show results with:interference<|separator|>
  32. [32]
    Quantum Neuron: an elementary building block for machine ... - arXiv
    Nov 30, 2017 · View a PDF of the paper titled Quantum Neuron: an elementary building block for machine learning on quantum computers, by Yudong Cao and ...
  33. [33]
    [2006.14619] Recurrent Quantum Neural Networks - arXiv
    Jun 25, 2020 · In this work we construct a quantum recurrent neural network (QRNN) with demonstrable performance on non-trivial tasks such as sequence learning and integer ...
  34. [34]
    A variational approach to quantum gated recurrent units - IOPscience
    Aug 21, 2024 · The Quantum Gated Recurrent Units outperformed the Quantum Long Short-Term Memory network despite having a simpler internal configuration.Introduction · Related works · Basic concepts on variational... · Quantum GRU
  35. [35]
    Quantum Recurrent Neural Networks with Encoder-Decoder for ...
    Feb 19, 2025 · This study explores Quantum Recurrent Neural Networks within an encoder-decoder framework, integrating Variational Quantum Circuits into Gated ...
  36. [36]
    Quantum Recurrent Neural Networks: Predicting the Dynamics of ...
    Apr 19, 2024 · Quantum Gated Recurrent Unit (QGRU) represents an evolution of traditional GRU networks, integrating with VQCs. The structure of a single QGRU ...<|control11|><|separator|>
  37. [37]
    Quantum convolutional neural networks | Nature Physics
    Aug 26, 2019 · We introduce and analyse a quantum circuit-based algorithm inspired by convolutional neural networks, a highly effective model in machine learning.
  38. [38]
    Exploring quantum neural networks for binary classification on ...
    In this study, we propose a novel modularized Quantum Neural Network (mQNN) model tailored to address the binary classification problem on the MNIST dataset.
  39. [39]
    Practical application of quantum neural network to materials ... - Nature
    Apr 13, 2024 · The mean squared error (MSE) between the labeled data and model ... The QNN decoder takes the expectation value of an observable quantum ...
  40. [40]
    [PDF] Text Classification with Born's Rule
    The Born rule provides a link between the mathematical formalism of quantum theory and experiment, and as such is almost single-handedly responsible for ...
  41. [41]
    Quantum inspired qubit qutrit neural networks for real time financial ...
    Aug 6, 2025 · This research investigates the performance and efficacy of machine learning models in stock prediction, comparing Artificial Neural Networks ...
  42. [42]
    [PDF] A Quantum State Fidelity based Generative Adversarial Network
    The key contributions are summarized below. • Based on quantum fidelity measurements, we propose quantum-state based loss functions with quantum gradi- ents ...
  43. [43]
    [PDF] QuClassi: A Hybrid Deep Neural Network Architecture based on ...
    We propose. QuClassi that employs a quantum state fidelity based loss function and a quantum-classic hybrid architecture to address the current limitations. 3 ...
  44. [44]
    [PDF] Reduction of finite sampling noise in quantum neural networks
    Jun 25, 2024 · In this work, we address the challenge of handling finite sampling noise in QNNs within the constraints of the NISQ era, where a massive number ...
  45. [45]
    [PDF] Implementation of a Quantum Approximate Optimization Algorithm ...
    Oct 3, 2025 · Here, the cost function is defined as in [16]. C(θ) = ⟨ψ(θ)|H|ψ(θ)⟩. (3). The goal is to minimize the expectation value of H over the next ...
  46. [46]
    [PDF] Next-Generation Quantum Neural Networks: Enhancing Efficiency ...
    Next-generation QNNs enhance efficiency, security, and privacy using optimization, defensive mechanisms, and Quantum Federated Learning, addressing challenges ...
  47. [47]
    Barren plateaus in quantum neural network training landscapes
    Nov 16, 2018 · However, multiple techniques have been proposed to mitigate this problem24,35,48,49, and the amount of training data and computational power ...
  48. [48]
    Barren plateaus in quantum neural network training landscapes - arXiv
    Mar 29, 2018 · The paper discusses how the exponential dimension of Hilbert space and gradient estimation complexity make random circuits unsuitable for  ...
  49. [49]
    Absence of Barren Plateaus in Quantum Convolutional Neural ...
    Oct 15, 2021 · Despite their tremendous potential, QNNs have been shown to exhibit a “barren plateau,” where the gradient of a cost function vanishes ...
  50. [50]
    An initialization strategy for addressing barren plateaus in ...
    Dec 9, 2019 · In this technical note we theoretically motivate and empirically validate an initialization strategy which can resolve the barren plateau problem for practical ...
  51. [51]
    A repetitive amplitude encoding method for enhancing the mapping ...
    Sep 1, 2025 · The most commonly used encoding methods currently include amplitude encoding23,24,25,26,27,28, angle encoding29,30,31,32,33, and hybrid ...
  52. [52]
    Single-Qudit Quantum Neural Networks for Multiclass Classification
    Mar 12, 2025 · This paper proposes a single-qudit quantum neural network for multiclass classification, by using the enhanced representational capacity of high-dimensional ...
  53. [53]
    [2412.02059] Lean classical-quantum hybrid neural network model ...
    Dec 3, 2024 · Our experiments demonstrate that LCQHNN achieves 100\%, 99.02\%, and 85.55\% classification accuracy on MNIST, FashionMNIST, and CIFAR-10 ...
  54. [54]
    Regression and Classification with Single-Qubit Quantum Neural ...
    Dec 12, 2024 · We use a resource-efficient and scalable Single-Qubit Quantum Neural Network (SQQNN) for both regression and classification tasks.
  55. [55]
    QDataSet, quantum datasets for machine learning | Scientific Data
    Sep 23, 2022 · The QDataSet comprises 52 high-quality publicly available datasets derived from simulations of one- and two-qubit systems evolving in the presence and/or ...
  56. [56]
    Regressions on quantum neural networks at maximal expressivity
    Dec 30, 2024 · The maximal expressive power increases with the depth of the network and the number of qubits, it is fundamentally bounded by the data encoding mechanism.
  57. [57]
    Implementing Large Quantum Boltzmann Machines as Generative ...
    Feb 5, 2025 · This study explores the implementation of large Quantum Restricted Boltzmann Machines (QRBMs), a key advancement in Quantum Machine Learning ( ...
  58. [58]
    On the sample complexity of quantum Boltzmann machine learning
    Aug 14, 2024 · Quantum Boltzmann machines (QBMs) are machine-learning models for both classical and quantum data. We give an operational definition of QBM ...
  59. [59]
    Experimental quantum-enhanced kernel-based machine learning ...
    Jun 2, 2025 · Here we demonstrate a kernel method on a photonic integrated processor to perform a binary classification task.Classification Task · Experiment · Results
  60. [60]
    TensorFlow Quantum
    TensorFlow Quantum (TFQ) is a quantum machine learning library for rapid prototyping of hybrid quantum-classical ML models.Quantum machine learning · TensorFlow Quantum design · Quantum data · Install
  61. [61]
    Hybrid computation - PennyLane
    Hybrid refers to the strategy of mixing classical and quantum computations. This lies at the heart of optimizing variational circuits.
  62. [62]
    PennyLane: Automatic differentiation of hybrid quantum-classical ...
    PennyLane's core feature is the ability to compute gradients of variational quantum circuits in a way that is compatible with classical techniques such as ...
  63. [63]
    TensorFlow Quantum
    Apr 26, 2024 · TensorFlow Quantum focuses on quantum data and building hybrid quantum-classical models. It provides tools to interleave quantum algorithms ...
  64. [64]
    [PDF] On quantum backpropagation, information reuse, and ... - arXiv
    May 22, 2023 · We present our proposed quan- tum backpropagation algorithm in Figure 1 which highlights the reduction in quantum resources due to the ability ...
  65. [65]
    Hello, many worlds | TensorFlow Quantum
    Jan 10, 2025 · Now that you've seen the basics, let's use TensorFlow Quantum to construct a hybrid quantum-classical neural net. You will train a classical ...
  66. [66]
    Hybrid Quantum Neural Network for Drug Response Prediction - NIH
    This work successfully employs a novel approach in processing patient and drug data to predict the drug response for cancer patients.
  67. [67]
    Hybrid Quantum Neural Networks for Efficient Protein-Ligand ... - arXiv
    Sep 14, 2025 · This study highlights the potential of hybrid QML in computational drug discovery, offering insights into its applicability and advantages ...
  68. [68]
    [2508.15998] Quantum Federated Learning: A Comprehensive Survey
    This paper presents a comprehensive survey on QFL, exploring its key concepts, fundamentals, applications, and emerging challenges in this ...
  69. [69]
    Quantum federated learning: a comprehensive literature review of ...
    Jul 21, 2025 · Federated learning (FL) is a recent technique that emerged to handle the vast amount of training data needed in machine learning algorithms ...
  70. [70]
    Advances in Quantum Computation in NISQ Era - PMC
    Oct 15, 2025 · This Special Issue includes fourteen recent studies advancing quantum computing across algorithms, applications, and hardware. In physics ...
  71. [71]
    A comparative analysis and noise robustness evaluation in quantum ...
    Sep 29, 2025 · In current noisy intermediate-scale quantum (NISQ) devices, hybrid quantum neural networks (HQNNs) offer a promising solution, combining the ...
  72. [72]
  73. [73]
    A Comprehensive Analysis of Noise Robustness in Hybrid Quantum ...
    May 6, 2025 · ... noise interference, such as decoherence, gate errors, and readout errors. This paper presents an extensive comparative analysis of two HQNN ...
  74. [74]
    Quantum Error Corrections for Fault-Tolerant Quantum Computers
    Aug 23, 2025 · The error rates for current quantum computers, typically ranging from 0.1% to 1%, are orders of magnitude higher than those in classical systems ...
  75. [75]
    Reduction of finite sampling noise in quantum neural networks
    Jun 25, 2024 · We reduce this noise by introducing the variance regularization, a technique for reducing the variance of the expectation value during the quantum model ...Missing: issues | Show results with:issues
  76. [76]
    [PDF] Best practices for quantum error mitigation with digital zero-noise ...
    Jul 20, 2023 · Fig. 1. Zero-noise extrapolation is a quantum error mitigation technique, where the zero-noise result E∗ of a quantum computation is estimated ...
  77. [77]
    [PDF] The Decade of Fault-Tolerant Quantum Computing - Preprints.org
    Sep 26, 2025 · The surface code has emerged as the gold standard for fault-tolerant quantum computing, combining high thresholds (~1%) with a simple 2D ...
  78. [78]
    Expressivity of quantum neural networks
    ### Summary of Findings on Expressivity of Quantum Neural Networks
  79. [79]
    Quantum Data Breach: Reusing Training Dataset by Untrusted Quantum Clouds
    ### Summary: Impact of No-Cloning Theorem on Quantum Neural Networks (QNNs)
  80. [80]
  81. [81]
    Quantinuum Unveils Accelerated Roadmap to Achieve Universal ...
    Quantinuum's roadmap unveils its fifth-generation quantum computer, Apollo, which will be a fully fault-tolerant and universal quantum computer.Missing: neural | Show results with:neural