Fact-checked by Grok 2 weeks ago

Urn problem

In , the urn problem, also known as the urn model, is a foundational for representing random sampling and selection processes, where an serves as a for a finite containing distinguishable objects—typically visualized as balls of different colors or types—from which items are drawn at random, either with (returning the drawn item) or without, to compute probabilities of specific outcomes. This model simplifies complex probabilistic scenarios by focusing on the composition of the urn and the drawing rules, enabling calculations for events like the sequence of draws or the proportion of types selected. The model has historical roots in early probability work, with basic forms appearing in discussions of combinatorial sampling, but it gained prominence through George Pólya's 1923 introduction of the reinforced variant known as Pólya's , which begins with an initial set of balls (e.g., one red and one black) and adds additional balls of the drawn color upon each draw, modeling self-reinforcing processes that lead to exchangeable sequences and limiting proportions following a . Generalized urn models extend this by incorporating multiple colors, variable reinforcement matrices, and removal rules, as in Friedman's urn (where both drawn and undrawn types are reinforced), allowing analysis of Markov chains, martingales, and convergence to stable attractors. A classical variant, the urn problem, shifts focus from sampling to distribution: balls are sequentially placed into multiple urns at random, with interest in metrics like the probability of empty urns after m placements into n urns, which underpins occupancy theory and approximations like the coupon collector problem. Urn models are versatile tools in modern probability and statistics, illustrating core concepts such as , , the (for without-replacement sampling), and the beta-binomial (for reinforced draws), while finding applications in diverse fields including (modeling allele frequencies), economics (market share dynamics), clinical trials (adaptive designs), and (preferential attachment in networks). Their analytical tractability—often via generating functions, , or diffusion limits—makes them enduring for both theoretical advancements and pedagogical purposes, with ongoing research exploring multi-urn interactions and non-constant reinforcements.

Historical Development

Origins in Classical Probability

Urn problems originated as metaphors for random selection processes in lotteries and elections during the late medieval and early modern periods, providing a tangible for probabilistic events. A prominent historical example is the election of the , where, starting in 1268, urns filled with balls representing eligible oligarchs were used in multi-stage lot-drawing protocols to select electoral colleges randomly. This method, designed to prevent factional dominance and ensure fairness, involved drawing balls from urns to narrow down candidates from an initial pool of around 480 to a final shortlist, thereby introducing elements of chance into political decision-making. These urn-based analogies drew early influences from games of chance prevalent in gambling culture, such as and card games, which posed questions about of stakes in interrupted play—known as the . Addressed by mathematicians like and in their 1654 correspondence, such problems highlighted the need for systematic probability calculations in uncertain outcomes, paving the way for urn models to represent equiprobable selections akin to draws. Jacob Bernoulli formalized the urn problem in his seminal work Ars Conjectandi, published posthumously in 1713, where he employed urns containing pebbles of different colors to explore —the inference of an urn's underlying proportions from observed draws. Bernoulli's approach emphasized estimating population characteristics from samples, transforming urns into a cornerstone for in probability. In this context, he applied the to urn draws, demonstrating that repeated extractions from an urn with unknown black and white pebble ratios would, with sufficient trials, approximate the true proportions reliably. This foundational use of urns in Bernoulli's framework influenced subsequent developments, such as 's extensions in the early .

Key Contributors and Milestones

advanced through his seminal work The Doctrine of Chances (1718), where he extended Bernoulli's ideas to approximate probabilities for large populations using combinatorial methods and early , facilitating developments in the normal approximation to the . This contribution marked a shift toward applying these analogies to complex combinatorial problems beyond simple games of chance. Thomas Bayes further developed inverse inference in his posthumously published essay "An Essay towards Solving a Problem in the Doctrine of Chances" (1763), using a analogy with balls to represent updating through conditional probabilities, which laid the foundation for Bayesian methods in inferring causes from observed effects. In the , refined these ideas in Bayesian estimation, using analogies to model the inversion of probabilities and estimate unknown proportions from samples, as detailed in his Théorie analytique des probabilités (1812). Laplace's work emphasized the urn model's utility in practical inference, influencing applications in astronomy and demographics by formalizing beliefs in probabilistic reasoning. In 1923, Felix Eggenberger and introduced the reinforcement-based in their paper on probability logic in failures, which generalized traditional urn schemes by adding balls of the drawn color, modeling and exchangeable sequences in probability. Pólya further elaborated on these concepts in his 1930 paper "Sur quelques points de la théorie des probabilités." Key milestones in urn problem theory thus progressed chronologically: de Moivre's 1718 approximations established descriptive tools for large-scale events; Bayes's 1763 pioneered inferential applications; Laplace's 1812 refinements integrated priors for ; and the 1923 Pólya-Eggenberger model shifted focus to dynamic reinforcement processes. This timeline reflects the evolution from static descriptive uses in classical probability to dynamic inferential frameworks in modern statistics.

Core Concepts

The Basic Urn Model

The basic urn model in consists of an containing a finite number of balls of distinct types, typically represented by colors or categories, where balls of the same type are indistinguishable and symbolize units from a . This setup provides a simple, abstract framework for modeling random selection processes. In the standard configuration, the urn holds x balls of type A and y balls of type B, yielding a total of N = x + y balls. The model assumes that balls are drawn uniformly at random, meaning each ball has an equal probability of selection, and that the types are mutually exclusive and exhaustive, encompassing all relevant categories without overlap. The urn acts as a neutral container to simulate sampling from a finite , isolating the probabilistic structure from real-world complexities or biases that might affect direct observation. utilized this model in (1713) for inference, employing white and black tokens to represent outcomes in estimating unknown ratios. Variations in urn size and composition, such as adjusting the values of x and y or introducing additional types, allow the model to illustrate diverse sampling scenarios while preserving its foundational simplicity.

Sampling Procedures

In urn problems, sampling procedures define the mechanisms by which balls are drawn from the urn, directly influencing the dependence structure of the outcomes. The basic urn model consists of an urn containing balls of distinct types, such as colors, from which draws are made according to specified rules. These procedures typically involve selecting a ball at random, observing its type, and deciding whether to return it to the urn, with the goal of modeling random selection processes. Sampling with entails a , noting its type, and immediately returning it to the , thereby maintaining the original composition for all subsequent draws. This approach ensures that each draw is independent of the previous ones, as the probability of selecting any particular type remains constant across trials. It is commonly employed to simulate scenarios requiring repeated, unaffected selections, such as independent trials in experimental designs. In contrast, sampling without involves removing the drawn from the after observation, which alters the urn's composition and makes successive draws dependent, as the probabilities shift based on prior outcomes. This method is particularly relevant for modeling finite populations where depletion occurs, like selecting items from a limited set without reuse. For multiple draws, the sequence can be treated as ordered, where the order of selection matters (e.g., distinguishing the first draw from the second), or unordered, focusing solely on the of types obtained regardless of sequence. Decision rules govern the observation process: typically, only the type (e.g., color) is recorded, and in without-replacement sampling, this observation informs the updated probabilities for future draws by reducing the count of that type. Single draws represent the simplest case, equivalent to one-step selection, while multiple draws extend to chains of selections, amplifying dependence in without-replacement scenarios. Practical considerations emphasize achieving true in draws; the is often shaken or mixed thoroughly before each selection to ensure uniform probability across all balls. In without- sampling, this dependence implies that early draws can significantly impact later ones, necessitating careful accounting of the evolving urn state to avoid in applications like population surveys. With replacement, the preserved simplifies analysis but may not reflect real-world depletion effects.

Mathematical Framework

Probability Distributions in Urn Models

In urn models, the emerges from sampling with replacement, where each draw is independent and the composition of the urn remains unchanged. Consider an containing N balls, of which x are of type A and y = N - x are of type B. The probability of drawing a type A ball on any single draw is p = x/N. For n independent draws, the number K of type A balls follows a : P(K = k) = \binom{n}{k} p^k (1-p)^{n-k}, \quad k = 0, 1, \dots, n. This formula arises from combinatorial counting: there are \binom{n}{k} ways to select the positions of the k type A draws among n trials, each sequence having probability p^k (1-p)^{n-k}. When sampling without replacement from the same urn setup, the draws are dependent, leading to the for the number K of type A balls in n draws. The is P(K = k) = \frac{\binom{x}{k} \binom{y}{n-k}}{\binom{N}{n}}, \quad \max(0, n - y) \leq k \leq \min(n, x). The derivation relies on combinatorial enumeration: the numerator counts the ways to choose k type A balls from x and n-k type B from y, while the denominator is the total ways to choose any n balls from N. This distribution was first derived in the context of urn problems by in 1711, solving a problem posed by . For urns with more than two types, the generalizes the case under sampling with replacement. Suppose the urn has m types of balls with counts x_i for i = 1, \dots, m, so N = \sum_{i=1}^m x_i and probabilities p_i = x_i / N. In n draws, let K_i be the count for type i, with \sum_{i=1}^m K_i = n. Then, P(K_1 = k_1, \dots, K_m = k_m) = \frac{n!}{k_1! \cdots k_m!} p_1^{k_1} \cdots p_m^{k_m}. The derivation follows from multinomial coefficients counting the ways to the n draws into groups of sizes k_1, \dots, k_m, multiplied by the product of probabilities for each type. This extends the two-type case naturally to multiple categories. Variants for waiting times in repeated draws with replacement also yield familiar distributions. The describes the number of draws X until the first type A ball, with P(X = k) = (1-p)^{k-1} p for k = 1, 2, \dots, derived by noting that the first k-1 draws must be type B (each with probability $1-p) followed by a type A (probability p). The extends this to the number of draws X until the r-th type A ball, with P(X = k) = \binom{k-1}{r-1} p^r (1-p)^{k-r}, \quad k = r, r+1, \dots, where the counts the ways to place the r-1 successes among the first k-1 draws, followed by the r-th success. These arise as limits or special cases of repeated Bernoulli trials modeled via the urn.

Generalizations and Extensions

One prominent generalization of the basic urn model is Pólya's urn, introduced by Eggenberger and Pólya to model contagious processes such as disease spread. In this scheme, the urn begins with x balls of type A and y balls of type B. A ball is drawn at random, replaced, and supplemented by a additional balls of the same type (with symmetric treatment for both types when a > 0). This reinforcement creates positive dependence among draws, unlike the independence in sampling with replacement. The number of type A balls drawn in n successive draws follows a , which arises as a compounded with a beta prior on the success probability, reflecting the model's Bayesian interpretation. Another extension involves mixed replacement schemes, where the replacement rule differs by type to capture asymmetric dynamics. For an with x balls of type A (replaced upon drawing) and y balls of type B (removed without ), the process simulates scenarios like selective retention. The probability of observing exactly k type A draws in m trials is P(m,k) = \binom{m}{k} \frac{x^{(k)} y^{(m-k)}}{N^{(m)}} where N = [x + y](/page/X+Y) and z^{(r)} = z(z-1)\cdots(z-r+1) denotes the falling factorial (with z^{(0)} = 1). This formula accounts for the decreasing total balls due to non-replaced type B draws, yielding a distribution distinct from standard hypergeometric. (Feller, 1950, p. 50) For urns with more than two types, the multivariate generalizes sampling without to c \geq 2 categories. Suppose the urn contains N balls distributed as N_1, N_2, \dots, N_c of each type, with \sum_{i=1}^c N_i = N. a sample of size n without replacement, the probability of drawing k_1, k_2, \dots, k_c balls of each type (where \sum k_i = n) is P(K_1 = k_1, \dots, K_c = k_c) = \frac{\prod_{i=1}^c \binom{N_i}{k_i}}{\binom{N}{n}}. This model applies to multi-category populations, such as ecological sampling or , preserving the without-replacement structure while extending to higher dimensions. Marginal distributions for individual types reduce to the univariate hypergeometric./12%3A_Finite_Sampling_Models/12.03%3A_The_Multivariate_Hypergeometric_Distribution) Considering limits as the urn size grows infinite provides connections to broader probabilistic structures, particularly exchangeability. In Pólya's urn, the sequence of draw colors forms an exchangeable process, meaning any of outcomes has identical distribution. De Finetti's theorem asserts that any infinite exchangeable sequence of binary random variables is a mixture of i.i.d. Bernoulli sequences, with the mixing measure given by a in the Pólya case; this links the finite urn dynamics to infinite exchangeable limits. Key derivations for Pólya's urn involve recursive probability updates and martingale properties. The probability of drawing type A at step t+1 given previous outcomes simplifies due to the model's exchangeability, yielding P(A_{t+1} = A \mid \mathcal{F}_t) = \frac{x + a S_t}{N + a t}, where S_t is the number of prior type A draws and \mathcal{F}_t the history up to t. The proportion of type A balls, Y_t = \frac{x + a S_t}{N + a t}, forms a bounded martingale, converging almost surely to a beta(x/a, y/a) random variable by the martingale convergence theorem, ensuring long-run stability despite reinforcement. These properties underpin asymptotic analyses and extensions to multi-type urns.

Notable Examples

Simple Sampling Problems

Simple sampling problems in urn models provide foundational illustrations of probability principles, using basic setups with balls of distinct types drawn from a finite collection. These scenarios assume uniform randomness in draws and focus on direct calculations without complicating factors like . In the simplest case, consider an containing N balls, of which x are of type A (e.g., ) and N - x are of type B (e.g., blue). The probability of drawing a single type A ball is the direct proportion P(\text{A}) = \frac{x}{N}. For multiple draws with replacement, where each drawn ball is returned to the urn before the next draw, the outcomes are independent. The expected number of type A balls in n draws is thus n \cdot \frac{x}{N}. Sampling without replacement introduces dependence between draws, as the urn's composition changes after each selection. The probability of drawing n type A balls in succession is the product of decreasing ratios: P(\text{all A}) = \frac{x}{N} \cdot \frac{x-1}{N-1} \cdot \ldots \cdot \frac{x-n+1}{N-n+1}. This follows the hypergeometric distribution for the number of type A balls in a fixed sample size. A notable variant is Bertrand's box paradox, which extends the urn concept to multiple urns (or boxes) with fixed compositions to explore conditional probability. Three boxes contain: one with two gold coins (GG), one with two silver coins (SS), and one with one gold and one silver (GS). A box is selected at random, and a gold coin is drawn from it. The probability that the other coin in the selected box is also gold is \frac{2}{3}, as there are two ways to draw a gold from the GG box but only one from the GS box, out of three equally likely gold draws overall. To compute these probabilities manually, consider a small with 3 red and 2 blue balls (N=5, x=3). A probability tree for two draws without replacement branches from the first draw: probability of red first is \frac{3}{5}, then red second is \frac{2}{4} = \frac{1}{2} (joint \frac{3}{5} \cdot \frac{1}{2} = \frac{3}{10}); blue first is \frac{2}{5}, then red second is \frac{3}{4} (joint \frac{2}{5} \cdot \frac{3}{4} = \frac{3}{10}). Summing paths yields the probability of exactly one red in two draws as \frac{6}{10} = \frac{3}{5}.

Advanced Urn Variants

In Pólya's urn model, an urn initially contains one red ball and one white ball. At each step, a ball is drawn at random, observed, replaced, and an additional ball of the same color is added. After n such draws, the number of red balls drawn, denoted Y_n, follows a over \{0, 1, \dots, n\}, with P(Y_n = k) = \frac{1}{n+1} for each k. This counterintuitive uniformity arises because the model induces exchangeability in the sequence of draws, where every specific sequence with k reds and n-k whites has equal probability \prod_{i=0}^{k-1} \frac{1+i}{2+i} \prod_{j=0}^{n-k-1} \frac{1+j}{2+k+j}, which simplifies identically across orders. As n \to \infty, the proportion of red balls in the urn converges to a random distributed as Uniform[0,1], reflecting the model's reinforcement mechanism that preserves initial symmetry. An urn adaptation of the illustrates with labeled balls. Suppose an urn contains two balls, each independently painted gold (representing "boy") or black (representing "girl") with probability $1/2. The possible configurations are GG, GB, BG, BB, each with probability $1/4, but GB and BG are indistinguishable as mixed. A ball is drawn at random and observed to be gold; the conditional probability that both are gold is then $1/3, as the gold observation equally likely comes from GG (two golds) or the mixed cases (one gold each), weighting the mixed twice as heavily. This labeled-ball setup mirrors the paradox's reliance on how information is revealed, emphasizing that the probability hinges on the sampling mechanism rather than naive symmetry. In the multivariate urn case, models extend to three or more types, such as outcomes with s. Starting with one ball of each of d colors and adding one of the drawn color per step, the counts after n draws follow a Dirichlet-multinomial distribution, generalizing the beta-binomial. For d=3, the probability of specific counts (k_1, k_2, k_3) with k_1 + k_2 + k_3 = n is \frac{n!}{k_1! k_2! k_3!} \frac{\Gamma(3) \Gamma(1+k_1) \Gamma(1+k_2) \Gamma(1+k_3)}{\Gamma(3+n) \Gamma(1)^3}, which integrates to marginals uniform over partitions for equal initials. The limiting proportions converge to a Dirichlet(1,1,1) distribution, uniform on the 2-simplex \{ (p_1,p_2,p_3) : p_i \geq 0, \sum p_i =1 \}, capturing where draws favor observed types. Urn models also reveal paradoxes like the inspection paradox in sequential draws. When modeling renewal processes via an urn where balls represent interarrival intervals drawn sequentially without , a randomly inspected draw (e.g., at a fixed time) tends to select a longer interval due to length-biased sampling, with expected length E[L^2]/E[L] exceeding the typical E[L]. In sequential urn sampling, this manifests as oversampling extended runs of the same color, where the inspected run's length is size-biased, leading to counterintuitive overestimation of average run .

Applications

In Statistics and Inference

Urn models serve as foundational tools in , particularly for estimating unknown population parameters from finite samples without replacement. In the context of , pioneered by , these models enable the estimation of the proportion \theta = x/N, where x is the unknown number of favorable items (e.g., white balls) in a total population of N items, based on observed draws. Laplace applied Bayes' rule with a uniform prior over possible values of x (from 0 to N), yielding a posterior proportional to the hypergeometric likelihood of observing k favorable outcomes in n draws. This approach, detailed in his 1774 memoir, provides a principled way to invert observed data to infer underlying causes or proportions, such as estimating the likelihood of future events like sunrises from historical records. For , urn models underpin confidence intervals for population proportions via the , which exactly describes sampling without replacement. When estimating \theta = x/N from k observed successes in n draws, exact confidence intervals can be constructed by inverting the , ensuring coverage probabilities that account for the finite population correction. Methods like the Clopper-Pearson interval, adapted for hypergeometric settings, provide conservative bounds that guarantee at least the nominal coverage level (e.g., 95%), avoiding overestimation of in small or finite populations. These intervals are particularly valuable in or ecological sampling, where the total population size N is known but the proportion of interest is not. Bayesian updating in urn models is vividly illustrated by Pólya's urn scheme, where sequential draws reinforce beliefs about underlying proportions, mirroring the revision of posterior distributions. Starting with an urn containing a balls of one type and b of another (reflecting a prior on the success probability), each draw of a ball is followed by replacing it along with an additional ball of the same type, leading to a process. This setup induces a beta-binomial for the number of successes, with the posterior after observations updating as in standard conjugate for binomial data, but adapted for the urn's self-reinforcing dynamic. The process, formalized by Eggenberger and Pólya in and interpreted Bayesianly by Blackwell and MacQueen in , elegantly demonstrates how beliefs evolve with data in finite sampling scenarios. Urn models also facilitate hypothesis testing for homogeneity, or equality of proportions, by leveraging statistics derived from hypergeometric sampling. To test whether two subpopulations have the same proportion of favorable items, the observed frequencies are modeled as a hypergeometric draw from a combined , with deviations assessed via exact p-values or approximations like the chi-squared statistic under the null of homogeneity. , introduced in 1922, computes the probability of the observed or more extreme tables under this null, providing a distribution-free method for small samples where asymptotic assumptions fail. This urn-based framework ensures precise control of Type I error in finite populations, as seen in applications to genetic association studies or randomized experiments. In modern , urn models extend to metaphorically represent complex prior processes, notably the introduced by Ferguson in 1973. Here, an urn scheme—often a Pólya-like with a base measure—generates exchangeable sequences that underlie infinite mixture models, allowing inference on unknown numbers of latent categories without assumptions. This connection bridges classical urn sampling to contemporary methods like mixtures for clustering or , where sequential "draws" update beliefs about the data-generating distribution in an adaptive, infinite-dimensional manner.

In Other Fields

In , urn models illustrate through scenarios like the , where individuals prefer bets on urns with known compositions over those with unknown but identical expected values, highlighting deviations from expected utility theory due to about probabilities. This setup uses two urns—one with a fixed of red and black balls (known risk) and another with an unknown (ambiguity)—to demonstrate how people avoid the ambiguous option, even when objective probabilities match, as evidenced in experimental preferences for known-risk gambles. In statistical physics, Pólya's urn model serves as an for contagion processes and branching phenomena, such as the spread of or particle interactions in networks, where drawing a "colored" ball reinforces similar outcomes, mimicking self-reinforcing growth. For instance, interacting Pólya urns with finite memory model epidemic dynamics on graphs, capturing spatial and temporal where each node's urn evolves based on neighboring draws, leading to power-law distributions in infection clusters. Similarly, continuous-time variants extend the model to branching processes, representing gamma-distributed in physical systems like neutron chains or molecular clustering. Recent extensions as of 2025 use urn models to simulate discrete-time , providing a classical for quantum probability processes in . In computer science, urn schemes underpin random graph generation, particularly through preferential attachment mechanisms that produce scale-free networks by assigning "balls" to vertices proportional to their degree, favoring connections to high-degree nodes. This Pólya-inspired process starts with an initial graph and iteratively adds vertices linked to existing ones via urn draws, yielding degree distributions with heavy tails akin to real-world networks like the internet or citation graphs. Generalizations allow weighted attachments, enabling models of directed or multilayer graphs while preserving the "rich-get-richer" dynamics. In , reinforced urn processes model evolution, where consumer choices draw from an representing product alternatives, adding balls for chosen options to simulate increasing returns and . For example, in a basic Pólya urn setup for duopoly s, initial market shares determine long-run dominance via a Dirichlet-distributed limit, explaining phenomena like first-mover advantages where early leads amplify over time through . Extensions incorporate quality differences or entry barriers, showing how reinforcement can inefficient equilibria despite superior alternatives. As of 2025, generalized Pólya urns have been applied to model growth through wages and returns, analyzing dynamics. Modern applications in leverage Pólya urns for topic modeling, enhancing (LDA) variants by introducing sparsity and parallelization in sampling word-topic assignments. The Pólya urn LDA sampler replaces traditional Gibbs methods with urn-based draws that enforce double sparsity—over topics and words—reducing computational costs for large corpora while maintaining the model's generative interpretation of documents as mixtures of topics. Further adaptations, such as weighted urn schemes, optimize topic coherence by self-reinforcing semantically related terms, improving interpretability in tasks like document clustering. Generalized urns also address limitations in standard LDA by incorporating hierarchical structures for better handling of correlated topics. Emerging uses as of 2025 include urn models in , such as analyzing pass sequences with Pólya urns to model player interactions and .

References

  1. [1]
    [PDF] A survey of random processes with reinforcement - arXiv
    Section 2 provides an overview of the basic models, primarily urn models, and corresponding known methods of analysis. Section 3 is devoted to urn models, ...
  2. [2]
    [PDF] Introduction to Probability, Statistics and Random Processes ...
    Find the probability that both these balls will be white. 4 / 11. Page 5. More examples. The urn problem (1). Let the urn contain 3 white balls and 4 black ...
  3. [3]
    [PDF] This chapter discusses Pólya's urn process and some ... - Penn Math
    An urn contains R red balls and B black balls. A ball is drawn from the urn and replaced in the urn along with A balls of the same color.
  4. [4]
    [PDF] ON A CLASSICAL PROBLEM OF PROBABILITY THEORY b We ...
    We consider the following classical “urn-problem”. Suppose that there are 71 urns given, and that balls are placed at random in these urns one after.
  5. [5]
    [PDF] Urn Models, Replicator Processes, and Random Genetic Drift
    Sep 6, 2022 · Under the assumption that the expected number of balls being added always exceeds the expected number of balls being removed whenever balls are ...
  6. [6]
    [PDF] Electing the Doge of Venice: analysis of a 13th Century protocol
    Jul 12, 2007 · voting theory, leader election,. Venice. This paper discusses the protocol used for electing the Doge of Venice between 1268 and the end of the ...
  7. [7]
    The Origin of Probability and The Problem of Points
    Indeed, many mathematicians chose problems that arose specifically with relation to gambling as examples to study in regard to the computation of probabilities.
  8. [8]
    [PDF] The Significance of Jacob Bernoulli's Ars Conjectandi - Glenn Shafer
    If there are no relevant limiting frequencies in a problem, it is not, according to the frequentists, a problem for probability. This would have astounded Jacob ...Missing: urn inverse
  9. [9]
    Turning points: Bernoulli's golden theorem - Raper - 2018
    Jul 26, 2018 · Simon Raper begins a new series on key moments in the history of statistics, starting in seventeenth-century Switzerland with theologian-turned-mathematician ...
  10. [10]
    [PDF] Probability Theory and Interpretations
    ❑ Jacob Bernoulli worked on the subject from 1685 to his death in. 1705. His Ars Conjectandi , the first systematic and analytically ... the equation of inverse ...
  11. [11]
    [PDF] Laplace-a pioneer of statistical inference - WordPress.com
    Laplace's inverse probability theory was based on the urn model. It was the model that mathematicians and statisticians subsequently applied throughout the.Missing: analogies | Show results with:analogies
  12. [12]
    [PDF] The Urn Model, Simula6on, and Resampling methods
    Jul 21, 2012 · Why use the Urn Model? • Connects sta6s6cal techniques,. – Hypothesis Tests. – Confidence Interval to underlying chance process. • If clear on ...Missing: definition | Show results with:definition
  13. [13]
    [PDF] Law of Large Numbers The Bernoulli family is the most
    Bernoulli's version of the law focused on proportions of success and failure, especially on the ratio of success to failures. For example, the theoretical ...
  14. [14]
    [PDF] Probability and Statistics: The Science of Uncertainty
    It is important that students acquire a sound mathematical foundation in the basic techniques of probability and statistics, which we believe this book will ...
  15. [15]
    [PDF] 1957-feller-anintroductiontoprobabilitytheoryanditsapplications-1.pdf
    ... FELLER - An Introduction to Probability Theory and Its Applications,. VolumeI, Third Edition. FELLER - An Introduction to Probability Theory and Its Applica ...
  16. [16]
    [PDF] Introduction to Probability - Yale Statistics and Data Science
    ... Feller's classic text, An Introduction to Probability Theory and Its. Applications. This book had its start with a course given jointly at Dartmouth College ...
  17. [17]
    The Hypergeometric Distribution
    ### Procedural Description of Sampling Without Replacement in Urn Model
  18. [18]
    [PDF] Introduction to Probability Theory and Its Applications
    FELLER · An Introduction to Probability Theory and Its Applications,. Volume ... Urn Models. 118. ·. 3. Stochastic Independence. 125. 4. Product Spaces ...
  19. [19]
    [PDF] Uses of the Hypergeometric Distribution for Determining Survival or ...
    In 1711, De Moivre solved “Huygens' fourth proposed problem” and his solution gave the probability of selecting x white and c-x black balls from a population ...
  20. [20]
    Über die Statistik verketteter Vorgänge - Eggenberger - 1923 - ZAMM
    Über die Statistik verketteter Vorgänge. F. Eggenberger,. F. Eggenberger. Zürich. Search for more papers by this author · G. Pólya, ... Download PDF. back ...
  21. [21]
    8. Pólya's Urn Process - Random Services
    Pólya's urn process involves drawing a ball from an urn, returning it with new balls of the same color. It generalizes hypergeometric and Bernoulli models.Missing: source | Show results with:source
  22. [22]
    [PDF] Exchangeability and de Finetti's Theorem
    Apr 26, 2007 · If X1,...,Xn,... are independent and identically distributed, they are exchangeable, but not conversely.Missing: limits | Show results with:limits
  23. [23]
    A martingale approach for Pólya urn processes - Project Euclid
    This paper is devoted to a direct martingale approach for Pólya urn models asymptotic behaviour. A Pólya process is said to be small when the ratio of its ...
  24. [24]
    [PDF] Polya's Urn and the Beta-Bernoulli Process - UChicago Math
    Aug 13, 2025 · In addition, Polya's Urn is a multivariate distribution whose variables are exchangeable but not independent.
  25. [25]
    [PDF] Polya's urn and martingales - staff.math.su.se
    May 19, 2021 · The above property is known as the martingale property, and a sequence ... usually referred to as the continuity property of probability measures.Missing: recursive | Show results with:recursive
  26. [26]
    [PDF] On Reinforcement-Depletion Compartmental Urn Models
    In this section we treat the special case of one stage of reinforcement and depletion. The analysis is based on the following sampling problem: suppose that we ...
  27. [27]
    Boy or Girl
    Imagine somebody selecting two balls (with replacement) from an urn containing an equal number of white and black balls. If you see the colour of the first ball ...
  28. [28]
    [PDF] Pólya-splitting distributions as stationary solutions of multivariate ...
    Dec 7, 2022 · Abstract Multivariate count distributions are crucial for the inference of eco- logical processes underpinning biodiversity.<|separator|>
  29. [29]
    Sampling Bias and the Inspection Paradox - jstor
    These examples of sampling bias are a special case of the well-known inspection paradox in ... successive yarn lengths are independent and identically distributed ...Missing: sequential | Show results with:sequential
  30. [30]
    [PDF] Memoir on the probability of the causes of events - University of York
    LAPLACE'S 1774 MEMOIR event given the causes, and the probability of the existence of each of these is equal to the probability of the event given that ...
  31. [31]
    Laplace's 1774 Memoir on Inverse Probability - jstor
    Abstract. Laplace's first major article on mathematical statistics was pub- lished in 1774. It is arguably the most influential article in this field to.
  32. [32]
    [PDF] A Finite Memory Interacting P´olya Contagion Network and its ... - arXiv
    Dec 21, 2021 · The objective is to model the spread of infection (commonly referred to as contagion) through this interacting Pólya urn network, where each ...
  33. [33]
    A Finite Memory Interacting Pólya Contagion Network and Its ...
    We introduce a new model for contagion spread using a network of interacting finite memory two-color Pólya urns.
  34. [34]
    [PDF] The continuous-time triangular Pólya process
    Jan 2, 2017 · Keywords Urn · Pólya urn · Pólya process · Branching process · Gamma distribution · ... Ehrenfest (1907), where the authors used urns as models ...
  35. [35]
    On a preferential attachment and generalized Pólya's urn model
    We study a general preferential attachment and Pólya's urn model. At each step a new vertex is introduced, which can be connected to at most one.
  36. [36]
    [PDF] P´olya Urn Latent Dirichlet Allocation: a doubly sparse massively ...
    Abstract—Latent Dirichlet Allocation (LDA) is a topic model widely used in natural language processing and machine learning. Most approaches to training the ...
  37. [37]
    Generating Preferential Attachment Graphs via a Pólya Urn ... - arXiv
    Jan 17, 2024 · Preferential attachment graphs are an important class of randomly generated graphs which are often used to capture the “rich gets richer” ...Missing: schemes | Show results with:schemes
  38. [38]
    First mover or higher quality? Optimal product strategy in markets ...
    Oct 26, 2017 · In this paper, we test that intuition within a generalized Pólya urn model. We find that if we assume constant feedbacks, in the long run, ...
  39. [39]
    Wages and capital returns in a generalized Pólya urn
    Dec 16, 2024 · In this paper we study this phenomenon using a new extension of Pólya's urn, modelling wealth growth through wages and capital returns.
  40. [40]
    [PDF] Review Topic Discovery with Phrases using the Pólya Urn Model
    We start by briefly reviewing the Latent Dirichlet Allocation (LDA) model (Blei et al., 2003). Then we describe the simple Pólya urn (SPU) model, which is ...
  41. [41]
    [PDF] Optimizing Semantic Coherence in Topic Models - David Mimno
    This new topic model retains the document– topic component of standard LDA, but replaces the usual Pólya urn topic–word component with a gen- eralized Pólya urn ...