Fact-checked by Grok 2 weeks ago

Halton sequence

The Halton sequence is a deterministic designed to provide more uniform coverage of the unit hypercube than traditional pseudorandom sequences, making it a key tool in quasi-Monte Carlo methods for approximating multidimensional integrals and simulations. Introduced by John H. Halton in 1960 as an extension of the one-dimensional van der Corput sequence, it constructs points by representing successive integers in different prime bases and reversing their digits to form fractional parts, ensuring low discrepancy for efficient sampling. In one dimension, the sequence is generated using a single prime base b: for the index n, express n in base b as digits a_k a_{k-1} \dots a_1 a_0, then compute x_n = \sum_{i=0}^k a_i b^{-(i+1)}, which reverses the digits after the radix point. For example, in base 10, the sequence begins with 0.0, 0.1, 0.2, ..., 0.9, 0.01, 0.11, and so on, cycling through permutations to fill the unit interval evenly. Multidimensional versions employ distinct prime bases (e.g., 2, 3, 5 for three dimensions) for each coordinate to avoid correlations and achieve low discrepancy across the unit , typically up to several thousand points before correlations may increase. Halton sequences are particularly valued in applications requiring high accuracy with fewer samples, such as maximum simulated likelihood estimation in econometric models like mixed logit and financial risk assessment via Monte Carlo integration. Their deterministic nature allows for reproducible results and error bounds based on discrepancy measures, outperforming random sampling in convergence rates for smooth integrands, though they can suffer from higher discrepancy in higher dimensions without modifications like scrambling or randomization. Despite these limitations, the sequence remains a foundational method in numerical analysis due to its simplicity and effectiveness in moderate dimensions.

Mathematical Definition

One-Dimensional Halton Sequence

The one-dimensional Halton sequence is fundamentally a van der Corput sequence constructed in a prime base b. The van der Corput sequence, introduced by J. G. van der Corput in 1935, generates points in the unit interval [0,1) using the radical inverse function \phi_b(n). For a nonnegative integer n with base-b expansion n = \sum_{k=0}^\infty a_k b^k, where $0 \leq a_k < b are the digits, the radical inverse is defined as \phi_b(n) = \sum_{k=0}^\infty \frac{a_k}{b^{k+1}}. $&#36; [](https://arxiv.org/pdf/1506.03764) This function effectively reverses the digits of $n$ in base $b$ and interprets them as a fractional value in $[0,1)&#36;.[](https://web.maths.unsw.edu.au/~josefdick/preprints/KuipersNied_book.pdf) A concrete example is the binary van der Corput sequence with $b=2$, where the radical inverse $\phi_2(n)$ is obtained by reflecting the binary representation of $n$ after the binary point. For $n=0$, the binary expansion is &#36;0$, so $\phi_2(0) = 0$. For $n=1$ (binary &#36;1$), $\phi_2(1) = 0.1_2 = 0.5$. For $n=2$ (binary &#36;10$), $\phi_2(2) = 0.01_2 = 0.25$. For $n=3$ (binary &#36;11$), $\phi_2(3) = 0.11_2 = 0.75$. Subsequent terms continue this pattern, filling the interval with successively finer binary fractions.[](https://web.maths.unsw.edu.au/~josefdick/preprints/KuipersNied_book.pdf) In the context of the Halton sequence, the one-dimensional case selects a prime base $b$ (such as &#36;2$, &#36;3$, or &#36;5$) and defines the $n$th term as $X_n = \phi_b(n)$ for $n = 0, 1, 2, \dots$. This choice of prime base ensures compatibility with multidimensional extensions while maintaining low-discrepancy properties in one dimension.[](https://link.springer.com/article/10.1007/BF01386213) ### Multi-Dimensional Extension The multi-dimensional Halton sequence extends the one-dimensional version by constructing points in the unit hypercube $[0,1)^d$ for $d$ dimensions, where each coordinate is generated independently using the radical inverse function with a distinct prime base for that dimension.[](https://link.springer.com/article/10.1007/BF01386213) Specifically, the first $d$ prime numbers—such as $p_1 = 2$, $p_2 = 3$, $p_3 = 5$, and so on—are selected as bases $p_i$ for $i = 1$ to $d$, ensuring mutual coprimality among the bases to promote low correlation and good distribution properties across dimensions.[](https://link.springer.com/article/10.1007/BF01386213) The $n$th point in the $d$-dimensional sequence is then given by $\mathbf{x}_n = (\phi_{p_1}(n), \phi_{p_2}(n), \dots, \phi_{p_d}(n))$, where $\phi_{p_i}(n)$ denotes the radical inverse of $n$ in base $p_i$.[](https://link.springer.com/article/10.1007/BF01386213) This construction leverages the radical inverse from the one-dimensional case but applies it component-wise with different bases to fill the higher-dimensional space uniformly.[](https://pbr-book.org/4ed/Sampling_and_Reconstruction/Halton_Sampler) The choice of the first $d$ primes minimizes potential correlations that could arise from shared factors in the bases, as distinct primes are inherently coprime.[](https://link.springer.com/article/10.1007/BF01386213) For a concrete illustration in two dimensions ($d=2$), using bases 2 and 3, the sequence generates points in $[0,1)^2$ as follows: the 0th point is $(0, 0)$; the 1st is $(0.5, 1/3)$; the 2nd is $(0.25, 2/3)$; and the 3rd is $(0.75, 1/3)$.[](https://pbr-book.org/4ed/Sampling_and_Reconstruction/Halton_Sampler) These points are computed by expressing $n$ in the respective base, reversing the digits, and interpreting the result as a fractional value in $[0,1)$.[](https://link.springer.com/article/10.1007/BF01386213) ## Properties ### Discrepancy Measures The discrepancy of a finite set of points $ P_N = \{ \mathbf{x}_1, \dots, \mathbf{x}_N \} $ in the unit cube $[0,1]^d$ quantifies the deviation from uniform distribution and is defined as the extreme discrepancy D_N(P_N) = \sup_{J \subset [0,1]^d} \left| \frac{A(J; P_N)}{N} - \lambda(J) \right|, where the supremum is over all subintervals $ J $ (axis-aligned boxes), $ A(J; P_N) $ is the number of points in $ J $, and $ \lambda(J) $ is the [Lebesgue measure](/page/Lebesgue_measure) (volume) of $ J $. The star discrepancy $ D_N^*(P_N) $ restricts the supremum to anchored subintervals $ J^* = \prod_{j=1}^d [0, z_j) $ with $ 0 \leq z_j \leq 1 $, providing a computationally simpler measure that bounds the extreme discrepancy via $ D_N^*(P_N) \leq D_N(P_N) \leq 2^d D_N^*(P_N) $. These measures are central to analyzing the low-discrepancy properties of Halton sequences, as lower discrepancy implies tighter error bounds in quasi-Monte Carlo integration via the Koksma–Hlawka inequality.[](https://www.math.ucla.edu/~caflisch/Pubs/Pubs1990-1994/Morokoffsisc1994.pdf)[](https://www.ricam.oeaw.ac.at/files/people/siambook_nied.pdf) For the one-dimensional case, the Halton sequence reduces to the van der Corput sequence in base $ b \geq 2 $. Halton's theorem establishes that the star discrepancy satisfies $ D_N^* = O\left( \frac{\log_b N}{N} \right) $. This bound arises from estimating the number of points falling into subintervals using the base-$ b $ digit expansion of the indices; specifically, the proof involves a direct count of how the radical-inverse function distributes points by considering the contributions over up to $ m \approx \log_b N $ digit levels, where discrepancies accumulate via base-$ b $ harmonic-like sums, leading to the logarithmic term. More refined asymptotics show $ N D_N^* \sim \frac{\log N}{\log b} $, confirming the order $ O\left( \frac{\log N}{N} \right) $.[](https://www.ricam.oeaw.ac.at/files/people/siambook_nied.pdf) In multiple dimensions, the Halton sequence extends the van der Corput construction using distinct prime bases $ b_1, \dots, b_d $. Halton's theorem generalizes to show that both the extreme and star discrepancies satisfy $ D_N = O( (\log N)^d / N ) $ and $ D_N^* = O( (\log N)^d / N ) $, with the leading constant depending on the product over bases $ \prod_{j=1}^d (b_j - 1)/(2 \log b_j) $. The proof relies on the tensor product structure: the discrepancy factors across dimensions via the independence of digit expansions in coprime bases, bounding the local discrepancy in each coordinate and multiplying the logarithmic factors, yielding the $ d $-th power. This asymptotic rate grows slower than the expected $ O(\sqrt{\log N / N}) $ for pseudorandom sequences in fixed $ d $, but excels in integration error bounds of order $ O( (\log N)^d / N ) $ for functions of bounded variation. Improved variants, such as generalized or scrambled Halton sequences, achieve similar or slightly better constants through digit permutations that reduce the discrete discrepancy in individual coordinates.[](https://www.math.uwaterloo.ca/~clemieux/papers/hfclMCM09_Sep10.pdf)[](https://www.math.ucla.edu/~caflisch/Pubs/Pubs1990-1994/Morokoffsisc1994.pdf) To illustrate, consider the first $ N=4 $ points of the one-dimensional base-2 van der Corput sequence: $ 0, 0.5, 0.25, 0.75 $. Sorted as $ x_{(1)}=0, x_{(2)}=0.25, x_{(3)}=0.5, x_{(4)}=0.75 $, the star discrepancy is computed as D_4^* = \max\left( \max_{1 \leq k \leq 4} \left| \frac{k}{4} - x_{(k)} \right|, \max_{1 \leq k \leq 4} \left| x_{(k)} - \frac{k-1}{4} \right|, \frac{1}{4} \right) = 0.25, achieved at multiple points (e.g., $ 1/4 - 0 = 0.25 $). This value is consistent with the order of the bound $ O((\log N)/N) $.[](https://www.ricam.oeaw.ac.at/files/people/siambook_nied.pdf) ### Correlation Issues and Remedies One notable limitation of the Halton sequence arises in low dimensions, particularly the first two dimensions using bases 2 and 3, where correlations between coordinates produce visible ray-like patterns along diagonals in two-dimensional projections.[](https://www.caee.utexas.edu/prof/bhat/abstracts/scrambled_halton_sequences.pdf) These artifacts stem from the synchronized periodic structures in the radical inverse functions for small primes, which cause points to cluster along linear trajectories rather than distributing uniformly.[](https://www.sciencedirect.com/science/article/abs/pii/S037847540500087X) Small prime bases exacerbate these correlations because their short cycle lengths in the base-$b$ expansion lead to repetitive digit patterns that align across dimensions, making discrepancies apparent even in simple projections.[](https://www.sciencedirect.com/science/article/pii/S0377042705003717) In contrast, larger primes introduce longer cycles that delay such visible alignments to higher dimensions, but the initial dimensions remain susceptible due to the fundamental arithmetic similarities in low-radix representations.[](https://www.caee.utexas.edu/prof/bhat/abstracts/scrambled_halton_sequences.pdf) To mitigate these issues, generalized Halton sequences employ digit permutations in the radical inverse, known as scrambling, to disrupt correlations while preserving low discrepancy. Seminal approaches include the deterministic permutations of Braaten and Weller, which optimize for reduced discrepancy by selecting digit rearrangements that minimize clustering.[](https://www.sciencedirect.com/science/article/pii/S0377042705003717) Another influential method is Owen scrambling, which uses random-like permutations to achieve probabilistic uniformity bounds, enhancing performance in quasi-Monte Carlo applications. Recent advances, such as improved scrambling techniques, further reduce correlations in projections while maintaining low discrepancy.[](https://artowen.su.domains/reports/rhalton.pdf)[](https://arxiv.org/abs/2405.15799) The permuted radical inverse is formally defined as \phi_b^\pi(n) = \sum_{k=0}^\infty \frac{\pi(d_k)}{b^{k+1}}, where $d_k$ are the digits of $n$ in base $b$, and $\pi$ is a permutation of $\{0, 1, \dots, b-1\}$ that fixes 0 to maintain the low-discrepancy property.[](https://www.sciencedirect.com/science/article/pii/S0377042705003717) For instance, applying the simple permutation $\pi = (0, 2, 1)$ to the base-3 component of a two-dimensional [Halton sequence](/page/Halton_sequence) rearranges the digits to break the diagonal alignments, yielding plots with more isotropic distribution and fewer ray artifacts compared to the unscrambled version.[](https://www.caee.utexas.edu/prof/bhat/abstracts/scrambled_halton_sequences.pdf) Such permutations demonstrably improve two-dimensional uniformity, as visualized in comparisons of the first 100 points before and after scrambling.[](https://www.sciencedirect.com/science/article/pii/S0377042705003717) ## History and Development ### Origins in Van der Corput Sequences The van der Corput sequences were introduced in 1935 by the Dutch mathematician [J.G. van der Corput](/page/J.G._van_der_Corput) as the pioneering construction of low-discrepancy sequences in a given integer base $b \geq 2$.[](https://arxiv.org/pdf/1506.03764) These sequences generate points in the unit interval [0,1) by applying the radical-inverse function to the natural numbers, reversing their base-$b$ digits after the decimal point. Van der Corput's innovation built upon earlier notions of uniform distribution, aiming to create deterministic point sets that mimic the even spread of random samples but with provably superior uniformity properties.[](https://arxiv.org/pdf/1506.03764) Van der Corput's motivation stemmed from the study of distribution functions and the desire to construct sequences that exhibit strong uniform distribution modulo 1, addressing limitations in earlier equidistribution results. Uniform distribution modulo 1 requires that the proportion of sequence points falling into any subinterval of [0,1) approaches the interval's length as the number of points increases, but van der Corput sought constructions with faster convergence rates for practical applications in analysis. This work was part of a broader effort in the 1930s to refine tools for approximating integrals and studying Diophantine approximations through explicitly defined sequences.[](https://www.researchgate.net/publication/278048144_From_van_der_Corput_to_modern_constructions_of_sequences_for_quasi-Monte_Carlo_rules) A key prerequisite for understanding the discrepancy in such sequences is Hermann Weyl's equidistribution theorem from 1916, which established that the sequence $\{n \alpha\}$ (where $\{\cdot\}$ denotes the fractional part and $\alpha$ is irrational) is uniformly distributed modulo 1. Weyl's result, proved using exponential sums, provided the theoretical foundation for quantifying how well a sequence fills the unit interval, highlighting the need for measures beyond mere equidistribution to capture clustering or gaps. Van der Corput drew on this framework to design his sequences, ensuring they not only satisfy equidistribution but also minimize deviations from uniformity.[](https://link.springer.com/article/10.1007/s00605-025-02057-2) Van der Corput sequences achieve logarithmic discrepancy growth by leveraging the digit-reversal mechanism, which decorrelates successive points and prevents large gaps in the unit interval. Specifically, the extreme discrepancy $D_N$ satisfies $D_N = O((\log N)/N)$, where $N$ is the number of points, indicating that the normalized error $N D_N$ grows only logarithmically with $N$. This bound, sharper than the $O(1/\sqrt{N})$ rate for pseudorandom sequences, underscores their low-discrepancy nature and was later refined in subsequent analyses. These sequences laid the groundwork for multidimensional extensions, such as those developed by [John Halton](/page/John_Halton) in the 1960s.[](https://www.fq.math.ca/Papers1/50-3/Pillichshammer.pdf) ### Contributions of John Halton John H. Halton introduced the multidimensional [Halton sequence](/page/Halton_sequence) in his seminal 1960 paper published in *Numerische Mathematik*, where he analyzed the efficiency of quasi-random point sequences for evaluating multi-dimensional integrals.[](https://link.springer.com/article/10.1007/BF01386213) Building on the one-dimensional [van der Corput sequence](/page/van_der_Corput_sequence) developed in the 1930s, Halton extended the radical-inverse construction to higher dimensions by assigning a distinct prime number as the base for each dimension, ensuring the bases are pairwise coprime.[](https://link.springer.com/article/10.1007/s00605-018-1225-4) This key insight allowed the sequence to maintain low-discrepancy properties across dimensions, avoiding correlations that could degrade uniformity in multidimensional spaces.[](https://link.springer.com/article/10.1007/s00605-018-1225-4) In the same paper, Halton demonstrated the practical advantages of these sequences through their application to quasi-Monte Carlo numerical integration, showing empirically that they outperform traditional pseudorandom Monte Carlo methods in convergence speed for certain integral evaluations.[](https://link.springer.com/article/10.1007/BF01386213) He emphasized their deterministic nature and superior point distribution, which leads to bounded error estimates superior to the probabilistic error in standard Monte Carlo approaches.[](https://link.springer.com/article/10.1007/BF01386213) Halton's innovation established the Halton sequence as a foundational tool in quasi-Monte Carlo methods, earning it his namesake due to the originality and impact of his multidimensional generalization.[](https://link.springer.com/article/10.1007/s00605-018-1225-4) The use of coprime prime bases, in particular, became a standard technique for preserving low discrepancy, influencing subsequent developments in sampling and integration algorithms.[](https://link.springer.com/article/10.1007/s00605-018-1225-4) ## Applications ### Quasi-Monte Carlo Integration The quasi-Monte Carlo (QMC) method employs low-discrepancy sequences, such as the Halton sequence, to deterministically approximate the value of a multidimensional integral over the unit hypercube $[0,1]^d$. Specifically, the integral $\int_{[0,1]^d} f(\mathbf{x}) \, d\mathbf{x}$ is estimated as $\frac{1}{N} \sum_{i=1}^N f(\mathbf{x}_i)$, where $\{\mathbf{x}_i\}_{i=1}^N$ are points generated from the Halton sequence, providing a more uniform coverage of the integration domain compared to pseudorandom points.[](https://link.springer.com/article/10.1007/BF01386213) This approach was introduced by Halton to enhance the efficiency of numerical integration in high dimensions by minimizing clustering and gaps in point distribution.[](https://link.springer.com/article/10.1007/BF01386213) A fundamental theoretical foundation for QMC methods is the Koksma-Hlawka inequality, which bounds the integration error as $ \left| \frac{1}{N} \sum_{i=1}^N f(\mathbf{x}_i) - \int_{[0,1]^d} f(\mathbf{x}) \, d\mathbf{x} \right| \leq V(f) \, D_N^*(\{\mathbf{x}_i\}) $, where $V(f)$ denotes the total variation of $f$ in the sense of Hardy and Krause, and $D_N^*$ is the star discrepancy of the point set. For Halton sequences, the star discrepancy satisfies $D_N^* = O((\log N)^d / N)$, leading to an error bound that reflects this rate for functions of bounded variation.[](https://www.sciencedirect.com/science/article/pii/S0021999185712090) Compared to classical Monte Carlo integration, which achieves a probabilistic convergence rate of $O(1/\sqrt{N})$ independent of dimension, QMC with Halton sequences offers a deterministic convergence rate of $O((\log N)^d / N)$ for sufficiently smooth integrands, providing faster error reduction in low to moderate dimensions where the logarithmic factor remains manageable.[](https://www.sciencedirect.com/science/article/pii/S0021999185712090) This advantage is particularly pronounced for integrands with low effective dimensionality or smoothness, as demonstrated in computational studies comparing Halton sequences to pseudorandom methods.[](https://www.sciencedirect.com/science/article/pii/S0021999185712090) A representative example of applying Halton sequences in QMC integration is the approximation of $\pi$ via the integral over the unit square $[0,1]^2$, where $\pi/4 = \int_0^1 \int_0^1 \mathbf{1}_{x^2 + y^2 \leq 1}(x,y) \, dx \, dy$, and the estimate is $ \frac{4}{N} \sum_{i=1}^N \mathbf{1}_{x_i^2 + y_i^2 \leq 1} $ using 2D Halton points $(x_i, y_i)$. This simple test case illustrates the sequence's uniform filling of the domain, yielding more accurate approximations than equivalent [Monte Carlo](/page/Monte_Carlo) samples for moderate $N$. ### Sampling in Computer Graphics In computer graphics, Halton sequences are employed to generate low-discrepancy point distributions for sampling tasks in rendering pipelines, particularly in global illumination algorithms where uniform coverage of the sample space is crucial to avoid clustering artifacts. This application leverages the sequence's ability to produce points that fill multidimensional domains more evenly than pseudorandom methods, enabling efficient generation of primary rays from the image plane without introducing unwanted patterns or gaps in coverage. For instance, in global illumination techniques such as [photon mapping](/page/Photon_Mapping) or [radiosity](/page/Radiosity), Halton-based sampling ensures that rays probe the scene's lighting environment comprehensively, supporting accurate estimation of indirect light contributions across surfaces.[](https://pbr-book.org/4ed/Sampling_and_Reconstruction/Halton_Sampler)[](https://cg.ivd.kit.edu/publications/2017/svgf/svgf_preprint.pdf) A key benefit of Halton sequences in path tracing—the core method for simulating [global illumination](/page/Global_Illumination)—is their capacity to reduce variance in radiance estimates compared to purely random sampling, resulting in images with noticeably less noise at equivalent sample counts. By distributing path origins and directions more uniformly, these sequences minimize clustering in high-importance regions, leading to faster convergence and smoother renders, especially in scenes with complex light interactions. Quantitative evaluations in rendering benchmarks demonstrate that Halton sampling achieves up to 2000 times lower mean squared error ([MSE](/page/Mean_Squared_Error)) than independent random sampling for certain integrands, while in practical path-traced scenes, it yields approximately 1.09 times lower MSE per pixel than independent sampling at 4096 samples. This variance reduction is particularly valuable for real-time or progressive rendering workflows, where computational budgets limit the number of paths traced per pixel.[](https://pbr-book.org/4ed/Sampling_and_Reconstruction/Halton_Sampler)[](https://ttwong12.github.io/papers/udpoint/udpoint.pdf)[](https://cg.ivd.kit.edu/publications/2017/svgf/svgf_preprint.pdf) For specific tasks like texture sampling and anti-aliasing, two-dimensional Halton sequences provide deterministic yet stochastic-appearing point sets that enhance image quality by mitigating aliasing artifacts in rendered outputs. In a classic example, applying a 2D Halton sequence (using bases 2 and 3) to sample a checkerboard texture reduces visible banding and improves sharpness compared to uniform grids, as the points avoid periodic repetitions while maintaining low discrepancy. This approach is integrated into established rendering libraries, such as the Physically Based Rendering (PBRT) system, where the HaltonSampler class generates scrambled sequences for pixel and lens sampling, further randomized via Owen scrambling to preserve low-discrepancy properties across dimensions and support adaptive refinement in global illumination computations.[](https://pbr-book.org/4ed/Sampling_and_Reconstruction/Halton_Sampler)[](https://ttwong12.github.io/papers/udpoint/udpoint.pdf) ## Comparisons with Other Sequences ### Versus Sobol Sequences Sobol sequences, introduced by I. M. Sobol in 1967, are constructed using direction numbers in base-2 representations to generate (t, s)-sequences, which ensure improved uniformity and independence in higher dimensions through linear recurrences and permutations of the van der Corput sequence.[](https://www.sciencedirect.com/science/article/abs/pii/0041555367901449) In contrast, Halton sequences, developed by John H. Halton in 1960, employ radical inverse functions based on distinct prime bases, providing a straightforward extension of one-dimensional van der Corput sequences to multiple dimensions. A fundamental difference in their construction is that Halton's approach uses simple radical inverses with consecutive primes, which can lead to correlations between dimensions due to the multiplicative nature of the bases, whereas Sobol sequences leverage direction numbers—precomputed integers that define linear feedback shift registers—to achieve better pairwise independence and reduced correlations.[](https://www.math.ucla.edu/~caflisch/Pubs/Pubs1990-1994/Morokoffsisc1994.pdf) This makes Sobol sequences more robust for applications requiring high-dimensional uniformity, as the direction numbers can be optimized for specific properties like low partial correlations. Performance-wise, Sobol sequences generally exhibit lower star discrepancy in dimensions greater than 2, with theoretical bounds showing a coefficient of order $ 1 / [s! (\log 2)^s] $, which is smaller than Halton's $ \prod_{k=1}^s \log p_k $ for large s, leading to potentially tighter error estimates in higher dimensions.[](https://www.math.ucla.edu/~caflisch/Pubs/Pubs1990-1994/Morokoffsisc1994.pdf) Halton sequences are simpler to implement without needing precomputed tables, but they suffer from higher initial correlations and rapid discrepancy growth beyond 30 dimensions, often forming visible patterns like lines in projections. Empirical studies confirm Sobol's superiority in stability; for instance, in quasi-Monte Carlo integration up to 60 dimensions with sample sizes of 10,000 to 100,000, Sobol yields absolute errors 20-50% lower than Halton, maintaining uniform distributions while Halton shows increasing gaps.[](https://core.ac.uk/download/pdf/201144413.pdf) | Dimension (d) | Sequence | Star Discrepancy Example (for N=1024) | Relative Performance Note | |---------------|----------|---------------------------------------|---------------------------| | 2 | Halton | Higher than Sobol | Comparable to Sobol | | 2 | Sobol | Lower than Halton | Slightly better | | 4 | Halton | Noticeably higher | Higher error growth | | 4 | Sobol | Lower than Halton | Better uniformity | These values, derived from numerical tests on test integrals, illustrate Sobol's advantage in moderate dimensions, where Halton's prime-based correlations begin to degrade coverage.[](https://www.math.ucla.edu/~caflisch/Pubs/Pubs1990-1994/Morokoffsisc1994.pdf)[](https://core.ac.uk/download/pdf/201144413.pdf) ### Versus Pseudorandom Sequences Halton sequences differ fundamentally from pseudorandom sequences used in standard [Monte Carlo methods](/page/Monte_Carlo_method), as the former are deterministic low-discrepancy point sets designed to fill the unit hypercube more uniformly, while the latter rely on stochastic generation that approximates uniformity in the limit but often shows clustering in finite samples.[](https://www.math.ucla.edu/~caflisch/Pubs/Pubs1990-1994/Morokoffsisc1994.pdf) Pseudorandom sequences produce points that are statistically independent and uniformly distributed in expectation, yet their finite realizations can lead to local clustering, resulting in a root-mean-square error convergence rate of $O(N^{-1/2})$ for integration problems, where $N$ is the number of points.[](https://www.math.pku.edu.cn/teachers/litj/notes/numer_anal/MCQMC_Caflisch.pdf) In contrast, Halton sequences exhibit low discrepancy, ensuring even distribution without such clustering, which supports a theoretically superior error bound of $O((\log N)^d / N)$ under the [Koksma-Hlawka inequality](/page/Koksma-Hlawka_inequality) for functions of bounded variation in $d$ dimensions.[](https://www.math.ucla.edu/~caflisch/Pubs/Pubs1990-1994/Morokoffsisc1994.pdf) The primary advantages of Halton sequences over pseudorandom ones lie in their deterministic nature and enhanced uniformity, leading to faster convergence for smooth integrands in quasi-Monte Carlo applications.[](https://www.math.pku.edu.cn/teachers/litj/notes/numer_anal/MCQMC_Caflisch.pdf) For instance, in estimating the integral $\int_{[0,1]^2} x y \, dx \, dy = 1/4$, standard Monte Carlo with pseudorandom points yields errors on the order of $N^{-1/2}$, whereas quasi-Monte Carlo using a Halton sequence achieves errors closer to $(\log N)^2 / N$, often reducing the required $N$ by orders of magnitude for a given accuracy in low dimensions.[](https://www.math.ucla.edu/~caflisch/Pubs/Pubs1990-1994/Morokoffsisc1994.pdf) However, Halton sequences can underperform relative to pseudorandom methods for discontinuous or highly oscillatory functions, where the low-discrepancy structure fails to adapt to sharp variations, potentially leading to larger errors without additional smoothing techniques.[](https://www.math.pku.edu.cn/teachers/litj/notes/numer_anal/MCQMC_Caflisch.pdf) Halton sequences are preferable to pseudorandom ones in low- to moderate-dimensional problems (typically $d \leq 10$) involving smooth integrands, such as certain financial pricing models or global illumination in graphics, due to their efficiency in achieving high precision with fewer points.[](https://www.math.ucla.edu/~caflisch/Pubs/Pubs1990-1994/Morokoffsisc1994.pdf) In higher dimensions or with nonsmooth functions, pseudorandom Monte Carlo remains more robust, as the discrepancy growth in Halton sequences can degrade performance, though variance reduction strategies can mitigate this in practice.[](https://www.math.pku.edu.cn/teachers/litj/notes/numer_anal/MCQMC_Caflisch.pdf) ## Implementation ### Radical Inverse Function The radical inverse function $\phi_b(n)$ serves as the essential mechanism for generating the [one-dimensional components](/page/one-dimensional_components) of the Halton sequence, where $b$ denotes a chosen base (typically prime) and $n$ is a non-negative integer.[](https://link.springer.com/article/10.1007/BF01386213) Introduced by van der Corput in 1935 for base 2 and generalized by Halton in 1960 to support multiple prime bases, this function transforms the integer $n$ into a point in the unit interval $[0, 1)$ with favorable uniformity properties.[](https://link.springer.com/article/10.1007/BF01386213) To define $\phi_b(n)$, first expand $n$ in its base-$b$ representation: $n = \sum_{k=0}^{\infty} d_k b^k$, where each digit satisfies $0 \leq d_k < b$ and $d_0$ is the least significant digit. The radical inverse is then given by interchanging the role of the digits to form a fractional part: \phi_b(n) = \sum_{k=0}^{\infty} d_k b^{-(k+1)}. By convention, $\phi_b(0) = 0$.[](https://link.springer.com/article/10.1007/BF01386213) This function admits a straightforward iterative computation that avoids explicit digit storage, making it suitable for numerical implementation. The standard algorithm proceeds by successively extracting the base-$b$ digits of $n$ from least to most significant and accumulating their contributions to the fractional value. A pseudocode implementation is as follows: ``` function radical_inverse(b, n): if n == 0: return 0.0 result = 0.0 power = 1.0 / b while n > 0: result += (n % b) * power power /= b n = n // b # integer division return result ``` This approach ensures $O(\log_b n)$ time complexity, proportional to the number of digits in $n$'s base-$b$ expansion. For illustration, consider $b = 3$ and $n = 10$. The base-3 expansion of 10 is $101_3$, yielding digits $d_0 = 1$, $d_1 = 0$, $d_2 = 1$. Thus, \phi_3(10) = \frac{1}{3} + \frac{0}{9} + \frac{1}{27} = \frac{9}{27} + \frac{1}{27} = \frac{10}{27} \approx 0.37037. Applying [the algorithm](/page/The_Algorithm) confirms this: first iteration ($n=10$): digit 1, result $1 \times \frac{1}{3} = \frac{1}{3}$, $n=3$; second ($n=3$): digit 0, result $\frac{1}{3} + 0 \times \frac{1}{9} = \frac{1}{3}$, $n=1$; third ($n=1$): digit 1, result $\frac{1}{3} + 1 \times \frac{1}{27} = \frac{10}{27}$. ### Practical Algorithms and Permutations The Halton sequence in $d$ [dimensions](/page/Dimension) is generated by computing the radical inverse function $\phi_{p_i}(n)$ independently for each [dimension](/page/Dimension) $i = 1$ to $d$, where $p_i$ are the first $d$ prime numbers serving as [bases](/page/Base), and $n$ ranges from 0 to $N-1$ to produce $N$ points.[](https://web.maths.unsw.edu.au/~josefdick/MCQMC_Proceedings/MCQMC_Proceedings_2010_Preprints/OktenShahGoncharov-final.pdf) For each $n$, the value in [dimension](/page/Dimension) $i$ is obtained by expressing $n$ in [base](/page/Base) $p_i$ as digits $a_k a_{k-1} \dots a_1 a_0$, then forming $\phi_{p_i}(n) = \sum_{j=0}^k a_j / p_i^{j+1}$.[](https://artowen.su.domains/reports/rhalton.pdf) To reduce correlations between dimensions, especially in higher dimensions where bases are large, permutations are applied to the digits of the radical inverse. These permutations $\pi$ scramble the digit values $\{0, 1, \dots, p_i - 1\}$ for each base $p_i$, yielding a generalized Halton point $\phi_{p_i}^\pi(n) = \sum_{j=0}^k \pi(a_j) / p_i^{j+1}$.[](https://web.maths.unsw.edu.au/~josefdick/MCQMC_Proceedings/MCQMC_Proceedings_2010_Preprints/OktenShahGoncharov-final.pdf) Deterministic permutations, such as those based on linear congruential generators, decorrelate dimensions effectively; for example, $\pi(k) = (a k + c) \mod p_i$, where $a$ and $c$ are chosen to ensure $\pi$ is a full-period permutation (e.g., $a$ coprime to $p_i$, $c \neq 0$).[](https://www.cirm-math.fr/RepOrga/2255/Slides/slides-Okten1and2.pdf) Random permutations can also be used, independently for each digit position and dimension, to introduce variability while preserving low discrepancy.[](https://artowen.su.domains/reports/rhalton.pdf) For improved distribution in finite samples, the first $2^d$ points are sometimes omitted, as the initial segment may exhibit clustering near the origin or axes.[](https://people.sc.fsu.edu/~jburkardt/datasets/halton/halton.html) This skipping helps achieve better uniformity, particularly when $N$ is small relative to the dimension $d$.[](https://people.sc.fsu.edu/~jburkardt/datasets/halton/halton.html) A practical consideration for large $N$ is the base leap method, where instead of generating all points sequentially, every $l$-th index is taken ($n = m \cdot l$ for $m = 0, 1, \dots$), with $l$ chosen coprime to all bases $p_i$ to maintain low discrepancy in subsequences.[](https://www.mathworks.com/help/stats/haltonset.html) The following Python-like pseudocode illustrates generating a 2D Halton sequence with $N=100$ points using bases 2 and 3, applying a linear congruential permutation $\pi(k) = (k + 1) \mod p_i$ (with $a=1$, $c=1$) to each dimension's digits, and skipping the first $2^2 = 4$ points: ``` def radical_inverse_permuted(n, base, a=1, c=1): if n == 0: return 0.0 perm = lambda k: (a * k + c) % base result = 0.0 f = 1.0 / base while n > 0: digit = n % base result += perm(digit) * f f /= base n //= base return result def halton_2d(N, skip=4): points = [] base1, base2 = 2, 3 for i in range(skip, N + skip): x1 = radical_inverse_permuted(i, base1) x2 = radical_inverse_permuted(i, base2) points.append((x1, x2)) return points # Example usage sequence = halton_2d(100) ```[](https://artowen.su.domains/reports/rhalton.pdf)[](https://www.cirm-math.fr/RepOrga/2255/Slides/slides-Okten1and2.pdf)[](https://people.sc.fsu.edu/~jburkardt/datasets/halton/halton.html)

References

  1. [1]
    On the efficiency of certain quasi-random sequences of points in ...
    Halton, J.H. On the efficiency of certain quasi-random sequences of points in evaluating multi-dimensional integrals. Numer. Math. 2, 84–90 (1960). https ...
  2. [2]
    Quasi-Random Sequences - Prof. Michael T. Heath
    Halton sequences, one of the earliest methods for generating quasi-random sequences, depend on choosing a prime number b as the base or radix to generate the ...
  3. [3]
    [PDF] Quasi-random simulation of discrete choice models
    A Halton sequence is defined in terms of a base. The sequence in base 10 is most easily ex- plained; sequences in other bases are created the same as in base 10 ...
  4. [4]
    [PDF] From van der Corput to modern constructions of sequences for quasi ...
    Jun 11, 2015 · Abstract. In 1935 J.G. van der Corput introduced a sequence which has excellent uniform distribution properties modulo 1.
  5. [5]
    [PDF] Uniform Distribution of Sequences
    Uniform distribution of sequences. (Pure and applied mathematics) ... KUIPERS. H. NIEDERREITER. Page 10. Page 11. CONTENTS. Chapter 1 Uniform Distribution Mod 1.
  6. [6]
    8.6 Halton Sampler
    The -dimensional Halton sequence is defined using the radical inverse base , with a different base for each dimension. The bases used must all be relatively ...
  7. [7]
    [PDF] Quasi-Random Sequences and Their Discrepancies
    Examples of such sequences that will be considered below include the Halton sequence, Sobol' sequence, and Faure sequence. Bounds on the discrepancy of ...
  8. [8]
    [PDF] Random Number Generation and QuasimMonte Carlo Methods
    the star discrepancy, the current "record holder" is a generalized van der Corput ... Recall that the Euler-Mascheroni constant is given by. -y = lim (F 1 ...
  9. [9]
    [PDF] Improved Halton sequences and discrepancy bounds
    Using such scrambled one-dimensional sequences as coordinates of generalized Halton or generalized Faure sequences gives very good results in experi- ments with ...
  10. [10]
    [PDF] Simulation Estimation of Mixed Discrete Choice Models Using ...
    2.1 The standard Halton sequence​​ The standard Halton sequence is designed to span the domain of the S-dimensional unit. cube uniformly and efficiently (the ...
  11. [11]
    On the optimal Halton sequence - ScienceDirect.com
    First a brief review of the Halton sequence is presented in Section 2. This is followed by the analysis of poor two-dimensional projections and correlations ...
  12. [12]
    Good permutations for deterministic scrambled Halton sequences in ...
    Apparently, the results are best for the L 2 -star discrepancy, but also in the case of L 2 -extreme discrepancy the curve for the reverse Halton permutations ...
  13. [13]
    [PDF] A randomized Halton algorithm in R - Art Owen
    Some QMC methods work best with specially chosen sample sizes such as large prime numbers or powers of small prime numbers. Halton se- quences can be used with ...
  14. [14]
    From van der Corput to modern constructions of sequences for quasi ...
    Aug 6, 2025 · In 1935 J.G. van der Corput introduced a sequence which has excellent uniform distribution properties modulo 1. This sequence is based on a very ...
  15. [15]
    A note on Weyl's equidistribution theorem
    Feb 14, 2025 · Weyl proved in Weyl (Eins Math Ann 77(3):313–352, 1916) that integer evaluations of polynomials are equidistributed mod 1 whenever at least ...
  16. [16]
    [PDF] on the discrepancy of the van der corput sequence indexed by ...
    It is well-known that for any base b ≥ 2 the van der Corput sequence is uniformly distributed modulo one and that NDN (xn) = O(log N), see, for example, [1]. In ...
  17. [17]
    The generalized and modified Halton sequences in Cantor bases
    Oct 22, 2018 · Perhaps the most famous example of a low-discrepancy sequence is the van der Corput sequence. In 1935, van der Corput [4] introduced a ...<|control11|><|separator|>
  18. [18]
    Quasi-Monte Carlo Integration - ScienceDirect.com
    In this paper quasi-random (Halton, Sobol', and Faure) and pseudo-random sequences are compared in computational experiments designed to determine the effects ...Missing: seminal | Show results with:seminal
  19. [19]
    [PDF] Real-Time Reconstruction for Path-Traced Global Illumination
    A low-discrepancy Halton sequence [Halton and Smith 1964] is used to sample light sources and scattering directions. We loop through a small set of Halton ...
  20. [20]
    [PDF] Sampling with Hammersley and Halton Points - Tien-Tsin Wong
    We discuss the implementation issues and experience in choosing suitable bases of. Hammersley and Halton points on 2D plane and spherical surface. The ...Missing: diagonal | Show results with:diagonal<|control11|><|separator|>
  21. [21]
    On the distribution of points in a cube and the approximate ...
    Volume 7, Issue 4, 1967, Pages 86-112 On the distribution of points in a cube and the approximate evaluation of integrals.Missing: original | Show results with:original
  22. [22]
    [PDF] ZANCO Journal of Pure and Applied Sciences Comparing Halton ...
    The results show that the performance of Sobol sequence is better and more stable than Halton sequence. The results also show that Sobol sequence maintains the ...
  23. [23]
    [PDF] Monte Carlo and quasi-Monte Carlo methods
    The resulting convergence rate is O((logN)kN~l). Because of the correlations, quasi-random sequences are less versatile than random or pseudo-random sequences.
  24. [24]
    [PDF] Random and Deterministic Digit Permutations of the Halton ...
    The star discrepancy of the Halton sequence, or any low-discrepancy sequence, is O(N−1(logs N)). This fact, together with the Koksma-Hlawka inequality, lay the.
  25. [25]
    [PDF] Number sequences for simulation - CIRM
    Example: Linear Congruential Generators. Introduced by Lehmer in 1949: xn ... The transformation can be generalized so that the orbit gives the permuted Halton ...
  26. [26]
    HALTON - Halton Datasets
    In some cases, it is recommended that the initial portion of the sequence be skipped over. A general suggestion is to let STEP be the first power of 2 that is ...
  27. [27]
    haltonset - Halton quasirandom point set - MATLAB - MathWorks
    Generate a three-dimensional Halton point set, skip the first 1000 values, and then retain every 101st point. p = haltonset(3,Skip=1e3,Leap=1e2). p = Halton ...