Fact-checked by Grok 2 weeks ago

Beta regression

Beta regression is a statistical technique for modeling continuous response variables bounded in the open interval (0, 1), such as proportions, rates, or percentages, using the as the underlying probability model. The is parameterized by a mean μ (where 0 < μ < 1) and a precision parameter φ (> 0), which together control the , , and of the distribution, enabling the model to capture heteroscedasticity and asymmetry inherent in bounded data. This approach addresses limitations of , which can yield predictions outside (0, 1) and assumes constant variance, by linking the mean μ to predictors via a suitable function (e.g., or ) and allowing the precision φ to depend on covariates as well. The method was developed in the early 2000s, with foundational work by Kieschnick and McCullough (2003), who recommended beta regression alongside approaches for proportions on (0, 1), and by Ferrari and Cribari-Neto (2004), who provided a comprehensive framework including , asymptotic inference, and diagnostic tools. Their parameterization, where the density is expressed as f(y | μ, φ) = [Γ(φ + 1) / {Γ(μ φ) Γ((1 - μ) φ)}] y^{μ φ - 1} (1 - y)^{(1 - μ) φ - 1} for 0 < y < 1, facilitates direct interpretation of effects on the response scale and closed-form expressions for score functions and Fisher's information matrix. Subsequent extensions include zero- or one-inflated variants for data with boundary values and time-series adaptations for autocorrelated bounded outcomes. Beta regression offers advantages over alternatives like arcsine or logit transformations, which distort variance and complicate interpretation, by providing unbiased estimates and valid inference under the beta assumption. It has been widely applied in fields such as ecology for modeling species abundance proportions, economics for firm profitability rates, and medicine for diagnostic test accuracies, with software implementations available in R via the betareg package for estimation and diagnostics. Despite its utility, adoption remains limited in some disciplines, including the natural sciences, where it constitutes less than 1% of regression analyses in recent ecological literature.

Introduction

Definition and Motivation

Beta regression is a statistical modeling technique framed within the generalized linear model (GLM) framework, where the response variable follows a beta distribution reparameterized in terms of its mean \mu and precision \phi. This approach is specifically designed for analyzing continuous response variables that are bounded between 0 and 1, such as rates, proportions, or percentages, providing a flexible way to relate these outcomes to explanatory covariates. The primary motivation for beta regression arises from the limitations of traditional models when dealing with proportion-like data, which often exhibit heteroscedasticity with variance that depends on the mean, reaching a maximum at \mu = 0.5 and approaching zero as the mean nears 0 or 1. Unlike ordinary linear regression, which can produce predictions outside the [0,1] interval and assumes constant variance, beta regression inherently constrains predictions to (0,1) and accommodates mean-dependent variance through the beta distribution's structure. Similarly, it offers advantages over logistic or binomial regression, which assume a variance proportional to \mu(1-\mu) as in Bernoulli trials and are less suitable for non-integer or continuously observed proportions without aggregation. Key assumptions of the model include that the response variable takes values strictly within (0,1), excluding exact zeros or ones, and that covariates influence the mean \mu through a specified link function, such as the logit. In this setup, \mu represents the expected value of the proportion, interpretable as the predicted rate under the model, while \phi controls the dispersion, with larger values indicating reduced variability around \mu for a given mean.

Historical Development

The beta distribution, a flexible family of continuous probability distributions defined on the interval (0,1), has been utilized in statistics since the 18th century for modeling proportions and probabilities, with early mathematical foundations laid by and its role as a conjugate prior in Bayesian analysis formalized by the early 20th century. Although the beta distribution was applied in various statistical contexts, including uncertainty modeling, its adaptation to regression frameworks for bounded response variables remained limited until the early 2000s. Kieschnick and McCullough (2003) recommended beta regression and quasi-likelihood approaches for analyzing proportions observed on (0,1). Beta regression was formally introduced in 2004 by and , who proposed a model where the response variable follows a beta distribution parameterized by the mean and a precision parameter linked to covariates, enabling direct regression on rates and proportions without transformations. This innovation addressed limitations of prior approaches, such as arcsine transformations in linear regression, which often violated normality assumptions and hindered interpretability. Subsequent advancements expanded the model's flexibility. In 2010, Alexandre B. Simas, Wagner Barreto-Souza, and Andréa V. Rocha developed improved estimators for a general class of beta regression models, incorporating variable dispersion to allow the precision parameter to depend on covariates, thus accommodating heteroscedasticity in bounded responses. This was followed in 2012 by Ricardo Ospina and Silvia L. P. Ferrari, who introduced a class of zero-or-one inflated beta regression models to handle data with point masses at the boundaries, common in empirical proportions. Post-2015 developments focused on robustness, with methods like L_q-likelihood maximization proposed to mitigate the influence of outliers and model misspecification in beta regression estimation. The model's adoption surged in the 2010s across disciplines, driven by computational implementations such as the in R, facilitating its use in ecology for analyzing species abundance proportions, economics for modeling market shares, and health sciences for health-related quality-of-life scores.

Model Specification

The Beta Distribution

The beta distribution is a continuous probability distribution defined on the interval (0, 1), making it suitable for modeling random variables that represent proportions, rates, or fractions bounded away from 0 and 1. It is parameterized by two positive shape parameters, often denoted as p and q, and arises as the distribution of the ratio of two independent gamma-distributed variables with shape parameters p and q, respectively. The probability density function (PDF) of the beta distribution is given by f(y \mid p, q) = \frac{\Gamma(p + q)}{\Gamma(p) \Gamma(q)} y^{p-1} (1 - y)^{q-1}, \quad 0 < y < 1, where \Gamma(\cdot) denotes the gamma function. In the context of beta regression, a reparameterization in terms of the mean \mu \in (0,1) and a precision parameter \phi > 0 is commonly used, where the shape parameters are expressed as \alpha = \mu \phi and \beta = (1 - \mu) \phi. This yields the PDF f(y \mid \mu, \phi) = \frac{\Gamma(\phi)}{\Gamma(\mu \phi) \Gamma((1 - \mu) \phi)} y^{\mu \phi - 1} (1 - y)^{(1 - \mu) \phi - 1}, \quad 0 < y < 1. The mean of the distribution is E(Y) = \mu, and the variance is \operatorname{Var}(Y) = \frac{\mu (1 - \mu)}{1 + \phi}, which decreases as \phi increases, reflecting higher precision around the mean. The beta distribution exhibits flexible shapes depending on the values of \mu and \phi: it can be U-shaped (when both shape parameters are less than 1), J-shaped (when one shape parameter is less than 1 and the other exceeds 1), unimodal (when both exceed 1), or uniform (specifically when \mu = 0.5 and \phi = 2). Additionally, the beta distribution serves as the conjugate prior for the binomial likelihood in Bayesian inference, allowing the posterior to remain in the beta family after updating with binomial data. This property facilitates analytical tractability in probabilistic modeling of bounded responses.

Regression Formulation

Beta regression adapts the beta distribution to model response variables Y_i that lie strictly between 0 and 1, such as proportions or rates, by incorporating covariates through a structured parameterization of the mean while assuming a constant precision parameter. In the standard formulation, the observations Y_1, \dots, Y_n are assumed independent, with each Y_i \sim \text{Beta}(\mu_i \phi, (1 - \mu_i) \phi), where \mu_i \in (0,1) is the mean for the i-th observation and \phi > 0 is a global precision parameter that controls the variance for a fixed \mu_i. This parameterization allows the variance to be expressed as \text{Var}(Y_i) = \frac{\mu_i (1 - \mu_i)}{1 + \phi}, highlighting how higher \phi reduces . The core of the structure lies in the mean submodel, which links the \mu_i to a linear predictor via a suitable g(\cdot): g(\mu_i) = \mathbf{x}_i^T \boldsymbol{\beta}, where \mathbf{x}_i is the of covariates for the i-th (including an intercept), and \boldsymbol{\beta} is the of coefficients. The g(\cdot) must be strictly increasing, twice differentiable, and map the interval (0,1) onto the real line to ensure \mu_i \in (0,1) and maintain model by avoiding boundary values that could lead to non-convergence or redundancy in estimates. For the , commonly used as the default due to its symmetry around 0.5 and interpretability in terms of ratios, the form is g(\mu) = \log\left(\frac{\mu}{1 - \mu}\right), yielding \mu_i = \frac{\exp(\mathbf{x}_i^T \boldsymbol{\beta})}{1 + \exp(\mathbf{x}_i^T \boldsymbol{\beta})}. Other common link functions include the probit (g(\mu) = \Phi^{-1}(\mu), where \Phi^{-1} is the inverse cumulative distribution function of the standard normal, suitable for symmetric responses near 0.5) and the complementary log-log (g(\mu) = \log(-\log(1 - \mu)), often selected for responses skewed toward 1). Link function choice depends on the expected symmetry of the response distribution, with the logit preferred for balanced proportions to facilitate straightforward interpretation of coefficients as changes in the log-odds. In this setup, the precision \phi remains constant across all observations, capturing global heteroscedasticity without covariate dependence.

Parameter Estimation

Maximum Likelihood Estimation

Maximum likelihood estimation (MLE) is the primary method for estimating the parameters of the beta regression model, including the regression coefficients \beta and the precision parameter \phi. The response variable y_i follows a beta distribution with mean \mu_i = g^{-1}(x_i^T \beta), where g is a link function such as the , and precision \phi > 0, ensuring $0 < y_i < 1. This approach maximizes the likelihood function derived from the beta density, providing asymptotically efficient estimators under standard regularity conditions. The log-likelihood function for a sample of n independent observations is given by \ell(\beta, \phi) = \sum_{t=1}^n \left[ \log \Gamma(\phi) - \log \Gamma(\mu_t \phi) - \log \Gamma((1 - \mu_t) \phi) + (\mu_t \phi - 1) \log y_t + ((1 - \mu_t) \phi - 1) \log(1 - y_t) \right], where \mu_t = g^{-1}(\eta_t) and \eta_t = x_t^T \beta. This expression leverages the 's parameterization by mean and precision, facilitating numerical maximization. The score vector, or gradient of the log-likelihood, consists of components for \beta and \phi. For the regression coefficients, U_\beta(\beta, \phi) = \phi X^T T (y^* - \mu^*), where y^*_t = \log(y_t / (1 - y_t)), \mu^*_t = \psi(\mu_t \phi) - \psi((1 - \mu_t) \phi), T = \operatorname{diag}(1 / g'(\mu_t)), \psi(\cdot) is the , and X is the design matrix. The score for \phi is U_\phi(\beta, \phi) = \sum_{t=1}^n \left[ \mu_t (y^*_t - \mu^*_t) + \log(1 - y_t) - \psi((1 - \mu_t) \phi) + \psi(\phi) \right]. Setting these to zero yields the MLEs (\hat{\beta}, \hat{\phi}). For the precision parameter, the score involves to account for the distribution's shape. Inference relies on the expected Fisher information matrix K(\beta, \phi), which is block-partitioned as K = \begin{pmatrix} K_{\beta\beta} & K_{\beta\phi} \\ K_{\phi\beta} & K_{\phi\phi} \end{pmatrix}, with K_{\beta\beta} = \phi X^T W X, where W_t = \phi \left[ \psi'(\mu_t \phi) + \psi'((1 - \mu_t) \phi) \right] / [g'(\mu_t)]^2 and \psi'(\cdot) is the trigamma function; the off-diagonal blocks are K_{\beta\phi} = X^T T c with c_t = \phi \left[ \psi'(\mu_t \phi) \mu_t - \psi'((1 - \mu_t) \phi) (1 - \mu_t) \right]; and K_{\phi\phi} = \operatorname{tr}(D) for a diagonal matrix D involving trigamma terms. A common block-diagonal approximation simplifies K_{\beta\beta} to \phi X^T W X with W = \operatorname{diag}\left( \frac{1}{\mu_t (1 - \mu_t) [g'(\mu_t)]^2} \right), treating \beta and \phi separately for efficiency in large samples. The inverse K^{-1} provides the asymptotic covariance matrix. Optimization of the log-likelihood typically employs Fisher scoring or , which iteratively update parameters using the score and information matrix until convergence. Initial values for \beta can be obtained via ordinary least squares on the transformed response g(y_t), while \phi starts from a method-of-moments estimator based on residuals. These methods ensure reliable convergence for moderate sample sizes. Under standard assumptions, the MLEs are asymptotically normal: \sqrt{n} (\hat{\beta} - \beta, \hat{\phi} - \phi)^T \sim N_{k+1}(0, K^{-1}), where k is the number of regressors. Standard errors are derived from the inverse observed or expected information matrix, enabling Wald tests, confidence intervals, and likelihood ratio tests for hypotheses about \beta or \phi. This framework supports reliable inference in applications involving proportions or rates.

Alternative Methods

Bayesian estimation provides an alternative to maximum likelihood for by incorporating prior distributions and enabling full posterior inference. In this approach, weakly informative priors, such as independent normal distributions for the regression coefficients of the mean and precision link functions, are often employed. Posterior inference is typically conducted using methods, including , which updates parameters conditionally through normal transition kernels or , often relying on for intractable posteriors. These methods, as detailed in Cepeda-Cuervo and Gamerman (2005), offer advantages in small sample sizes by leveraging prior information to stabilize estimates and provide credible intervals that quantify uncertainty more comprehensively than asymptotic approximations. Robust methods address the sensitivity of standard estimation to outliers and heavy-tailed data in beta regression. M-estimators, such as those based on maximum L_q-likelihood with a tuning constant q (0 < q ≤ 1), reparameterize the likelihood to bound the influence of contaminants, achieving Fisher-consistency and asymptotic normality while minimizing efficiency loss in clean data. Developments since 2015, including minimum density power divergence estimators (MDPDE), further enhance robustness against outliers by incorporating bias corrections and data-driven tuning, as shown in simulation studies with 5% contamination where these estimators maintain stable performance for sample sizes from 40 to 320. For small samples, bias-corrected robust estimators reduce finite-sample bias in parameter recovery, particularly under high leverage points. Other approaches include penalized likelihood for variable selection and quasi-likelihood approximations suited to large datasets. Penalized methods, such as Lasso regularization on the negative log-likelihood, enable simultaneous parameter estimation and selection in high-dimensional settings by shrinking irrelevant coefficients to zero, with non-asymptotic error bounds ensuring consistency under sparsity conditions like s log(p)/n → 0, where s is the true support size. Quasi-likelihood approximations model the mean and a flexible variance structure, such as Var(Y) = φ μ (1-μ)^p with estimated power p, using estimating equations for efficient computation via Newton scoring, which is particularly advantageous for large data due to its speed and adaptability to bounded responses without full distributional assumptions. Bayesian methods are preferable when uncertainty quantification beyond asymptotic standard errors is needed, such as in small samples or when incorporating domain-specific priors, while robust approaches excel in contaminated data, and penalized or quasi-likelihood techniques suit high-dimensional or large-scale applications requiring efficiency and selection.

Model Extensions

Variable Dispersion

In beta regression models, the precision parameter \phi is typically assumed constant across observations to capture symmetric dispersion around the mean. However, this assumption may not hold in many applications where variability changes systematically with covariates, necessitating an extension to variable dispersion. This extension models the precision as varying with a set of covariates z_i, allowing the model to account for heteroscedasticity in proportions or rates data. The variable dispersion formulation parameterizes the precision parameter as \log(\phi_i) = z_i^T \gamma, where z_i is a vector of dispersion covariates (which may overlap with the mean covariates x_i), and \gamma is the vector of corresponding coefficients. This log-link function ensures \phi_i > 0 for all i, maintaining the validity of the . The approach was introduced by Simas, Barreto-Souza, and Rocha in 2010 to extend the beta regression , motivated by the need to capture extra variability in scenarios such as econometric where may vary with factors like levels. For estimation, the model jointly optimizes the mean parameters \beta and dispersion parameters \gamma via maximum likelihood, with the score function for \gamma given by U_\gamma(\beta, \gamma) = \tilde{Z}^T T_2 v, where \tilde{Z} is the derivative, T_2 is a diagonal matrix of derivatives, and v incorporates response deviations and functions. The matrix for \gamma forms part of the block-diagonal structure K(\zeta) = P^T W P, ensuring asymptotic under regularity conditions. is achieved through the strictly monotonic function and full column of the matrices for both submodels, with k + p < n where k and p are the dimensions of \beta and \gamma. Bias-corrected estimators can further improve finite-sample performance by adjusting the maximum likelihood estimates. The coefficients in \gamma provide interpretable insights into dispersion drivers: a positive \gamma_j indicates that an increase in the j-th covariate z_{ij} is associated with higher \phi_i and thus lower variability around the \mu_i, as the beta variance is \mu_i(1 - \mu_i)/(1 + \phi_i). Conversely, a negative \gamma_j suggests widening with higher z_{ij}. This allows researchers to identify factors that stabilize or amplify variability beyond mean effects. To assess the necessity of variable dispersion, likelihood ratio tests compare the full model against the constant \phi case by testing H_0: \gamma_j = 0 for dispersion-specific covariates (i.e., excluding the intercept). Under H_0, the test statistic follows a chi-squared distribution with degrees of freedom equal to the number of tested coefficients, enabling formal evaluation of whether covariates significantly influence precision.

Variable Precision in Time Series

Beta regression has been extended to time-series data to model autocorrelated bounded outcomes, such as serially dependent proportions. The beta autoregressive (BAR) model, introduced by Jordan and Smithson in 2014, incorporates lagged responses into the mean equation via a link function, allowing for first-order autoregression while preserving the for conditional responses. This addresses in longitudinal data, with estimation via maximum likelihood and diagnostics for residual . Applications include modeling time-varying rates in and .

Zero-One Inflated Models

Zero-one inflated models extend the standard beta regression framework to accommodate response data that include point masses at the boundaries 0 and/or 1, which occur frequently in proportions such as rates of , coverage, or success where exact zeros or ones are overrepresented beyond what the continuous predicts. These models treat the data as arising from a : a component capturing the excess probability at the boundaries and a continuous beta component for interior values in (0,1). The zero-inflated beta (ZIB) model specifies the distribution as P(Y=0) = \pi_0, P(Y=1) = 0, and P(0 < Y < 1) = (1 - \pi_0) f(y \mid \mu, \phi), where f(y \mid \mu, \phi) is the beta density with mean \mu and precision \phi, and \pi_0 represents the probability of excess zeros modeled via a logit link, such as \text{logit}(\pi_0) = \mathbf{w}^T \boldsymbol{\delta}, with covariates \mathbf{w} and coefficients \boldsymbol{\delta} separate from those for \mu and \phi. The one-inflated beta (OIB) model analogously handles excess ones with P(Y=1) = \pi_1, P(Y=0) = 0, and P(0 < Y < 1) = (1 - \pi_1) f(y \mid \mu, \phi), using a similar link for \pi_1. The more general zero-one inflated beta (ZOIB) model incorporates both, with P(Y=0) = \pi_0, P(Y=1) = \pi_1, and P(0 < Y < 1) = (1 - \pi_0 - \pi_1) f(y \mid \mu, \phi), ensuring \pi_0 + \pi_1 \leq 1, and allowing independent regression structures for the inflation parameters \pi_0 and \pi_1. These models were introduced by Ospina and Ferrari in 2012 as a unified class for on proportions with boundary values. estimation relies on maximizing the extended likelihood, typically via direct (MLE) using numerical optimization methods like Newton-Raphson. In applications, such as analyzing proportions of traffic accident mortality across municipalities where some districts reported 0% or 100% rates, the ZOIB model revealed that demographic factors like the proportion of young adults influenced both inflation probabilities and the conditional mean, outperforming standard beta . This approach ensures the model adequately captures structural zeros or ones, such as 0% adoption in conservative regions or 100% compliance in controlled settings, while maintaining separate estimation for inflation and parameters to avoid in interior predictions.

Applications and Implementation

Practical Examples

One prominent practical example of beta regression is Prater's yield , comprising 32 observations on the proportion of crude oil converted to (bounded between 0 and 1) as a function of covariates including dummy variables for the type of crude oil batch and the vaporization temperature. Fitted using a logit link function, the model estimates a precision parameter φ ≈ 440, reflecting relatively low in the response. The regression coefficients indicate positive effects from higher process conditions: for instance, the coefficient for vaporization temperature is 0.011 (p < 0.001), implying higher temperatures enhance output proportions. The model's pseudo-R² of 0.962 suggests strong explanatory power. Another illustrative application involves food expenditure shares from a sample of 38 U.S. households, where the response variable is the proportion of allocated to , regressed on total and using a logit link. The estimated precision parameter is φ ≈ 36, and the pseudo-R² ≈ 0.39 captures moderate fit while accommodating heteroscedasticity observed in simpler linear models of the data. Key coefficients show a negative with (β = -0.012, p < 0.001), indicating wealthier households devote smaller shares to , contrasted by a positive effect of (β = 0.118, p < 0.001), where larger groups spend higher proportions. With the logit link, these can be interpreted as changes in the log-odds of the food share. Beta regression extends to diverse fields, such as , where it models proportions like relative plant coverage in vegetation surveys to assess environmental influences on abundance without assuming . In , applications include analyzing bounded rates, such as regional shares, to quantify covariate impacts like variables on employment proportions via interpretable odds ratios from the link. In , a 2024 application used beta regression to model proportions of adherence to a mobile app for chronic disease management, identifying key predictors such as patient age and level. Model diagnostics in these examples typically involve residual plots to detect non-random patterns suggesting misspecification, and quantile-quantile (Q-Q) plots of beta-transformed residuals to evaluate adherence to the assumed . For the gasoline yield data, such plots identify one influential observation (e.g., via high ) but confirm overall adequacy, with half-normal envelopes showing no significant deviations. In the food expenditure case, s highlight the model's success in addressing heteroscedasticity through the precision parameter.

Software Packages

Beta regression models can be fitted and analyzed using several software packages in popular statistical environments, with the betareg being one of the most comprehensive and widely adopted tools for this purpose. Introduced in 2010, the betareg package provides the betareg() function for of standard beta regression models, where both the and parameters can be modeled as functions of explanatory variables via link functions such as or . It supports variable by allowing separate regression formulas for the (μ) and (φ), enabling flexible modeling of heteroscedasticity in proportions or rates. Additional features include diagnostic tools like plots and measures, prediction methods for fitted values and intervals, and capabilities for assessing model power or generating response data from fitted models. The package includes vignettes with practical examples demonstrating its use on datasets such as gasoline proportions. In Stata, the betareg command, available since version 14 in 2015, facilitates beta regression for fractional outcomes in the open unit interval (0,1), using maximum likelihood estimation with support for various link functions and robust standard errors. Postestimation commands enable hypothesis tests, marginal effects calculations, and graphical diagnostics such as predicted versus observed plots. It also accommodates extensions like zero- or one-inflated beta models through finite mixture specifications via fmm: betareg. Other implementations include approximate fitting in using PROC GENMOD with a for the under a beta-like assumption, though more precise modeling requires PROC GLIMMIX or nonlinear procedures like NLMIXED for full beta parameterization. In , the statsmodels library offers the BetaModel class for estimating beta regressions, parameterizing both and with explanatory variables and supporting functions, while custom implementations can leverage SciPy's optimization routines for maximum likelihood. For Bayesian approaches, beta regression can be specified in JAGS or , with the rstanarm package providing stan_betareg() for MCMC-based inference incorporating priors on coefficients and handling variable dispersion. Common features across these tools include variable selection methods like (often via add-on functions in ) and for , though limitations persist, such as the absence of built-in robust estimation in base 's betareg without extensions. Open-source options, particularly the R betareg package, dominate accessibility for research due to their flexibility, extensive documentation, and integration with broader ecosystems like for data manipulation.

References

  1. [1]
    A case for beta regression in the natural sciences - ESA Journals
    Feb 23, 2022 · This paper provides an overview of the current state of beta regression in natural sciences to date, as well as an evaluation of the two ...INTRODUCTION · METHODS · RESULTS · DISCUSSION
  2. [2]
    [PDF] BETA REGRESSION FOR MODELLING RATES AND PROPORTIONS
    This paper proposes a regression model where the response is beta distributed using a parameterization of the beta law that is indexed by mean and dispersion pa ...
  3. [3]
    Beta Regression for Modelling Rates and Proportions
    Aug 2, 2010 · This paper proposes a regression model where the response is beta distributed using a parameterization of the beta law that is indexed by mean and dispersion ...
  4. [4]
    Regression analysis of variates observed on (0, 1) - Sage Journals
    This paper focuses on regression models for proportions observed on (0, 1), recommending beta or quasi-likelihood models, and comparing prior research ...
  5. [5]
    A general class of zero-or-one inflated beta regression models
    This paper proposes a general class of regression models for continuous proportions when the data contain zeros or ones.Missing: original | Show results with:original
  6. [6]
    Beta regression for time series analysis of bounded data, with ...
    This manuscript proposes a practical approach to analyze bounded time series, through a beta regression model. The method allows the direct interpretation of ...
  7. [7]
    Beta Regression in R - CRAN
    A variant of the beta regression model that allows for nonlinearities and variable dispersion was proposed by Simas, Barreto-Souza, and Rocha (2010). In ...
  8. [8]
    Longitudinal beta regression models for analyzing health-related ...
    Sep 17, 2012 · A regression technique that is gaining increasing attention in the analysis of doubly bounded outcome measures is the beta regression as ...Missing: ecology sciences
  9. [9]
    [PDF] Bayesian Data Analysis Third edition (with errors fixed as of 20 ...
    ... beta distribution (see Appendix A). It is left as an exercise to reproduce ... Conjugate prior distributions. Conjugacy is formally defined as follows ...
  10. [10]
  11. [11]
    [PDF] Robust estimation in beta regression via maximum Lg-likelihood
    May 24, 2022 · In this work, we develop a robust estimation procedure for beta regression models based on the maximization of a reparameterized Lq-likelihood.
  12. [12]
    [PDF] Lasso Penalization for High-Dimensional Beta Regression Models
    Jul 26, 2025 · This paper analyzes Lasso penalization for beta regression in high-dimensional settings, addressing non-convexity and proposing a debiasing ...
  13. [13]
    Variable selection for varying dispersion beta regression model
    Based on a penalized likelihood method, the consistency and the oracle property of the penalized estimators are established.
  14. [14]
    Flexible quasi-beta regression models for continuous bounded data
    Sep 2, 2018 · We propose a flexible class of regression models for continuous bounded data based on second-moment assumptions.
  15. [15]
    Improved estimators for a general class of beta regression models
    Our objective in this paper is to reduce the bias of the maximum likelihood estimators (MLEs) for this extended class of beta regression models. We extend the ...
  16. [16]
  17. [17]
    Analysing continuous proportions in ecology and evolution: A ...
    Jun 6, 2019 · Some recent examples of beta regression in ecological contexts include analyses of: the contribution of food derived from different energy ...INTRODUCTION · EXTENSIONS TO DIRICHLET... · ISSUES WITH BETA AND...
  18. [18]
    Beta Regression in R | Journal of Statistical Software
    Apr 5, 2010 · This paper describes the betareg package which provides the class of beta regressions in the R system for statistical computing. The ...
  19. [19]
    [PDF] betareg — Beta regression - Description Quick start Menu
    Beta regression models were proposed by Ferrari and Cribari-Neto (2004) and extended by Smithson and Verkuilen (2006) to allow the scale parameter to depend on ...
  20. [20]
    [PDF] betareg postestimation - Stata
    Postestimation commands. The following postestimation commands are available after betareg: Command. Description contrast contrasts and ANOVA-style joint ...
  21. [21]
    [PDF] fmm: betareg — Finite mixtures of beta regression models - Stata
    fmm: betareg fits mixtures of beta regression models to a fractional outcome whose values are greater than 0 and less than 1; see [FMM] fmm and [R] betareg for ...
  22. [22]
    Modeling continuous proportions: Normal and Beta Regression ...
    Jan 20, 2016 · The beta model can be used to fit more complex models to continuous proportion data. For example, the following statements fit a random effects ...Missing: economic | Show results with:economic
  23. [23]
    statsmodels.othermod.betareg.BetaModel - statsmodels 0.15.0 (+841)
    Beta Regression. The Model is parameterized by mean and precision. Both can depend on explanatory variables through link functions.
  24. [24]
    Bayesian beta regression models via Stan — stan_betareg • rstanarm
    Bayesian beta regression models use optional prior distributions and perform full Bayesian estimation via MCMC, adding priors on the coefficients.Missing: JAGS | Show results with:JAGS