Fact-checked by Grok 2 weeks ago

Kendall tau distance

The Kendall tau distance, also known as the Kendall rank distance, is a that quantifies the dissimilarity between two rankings of the same set of n items by counting the number of pairwise disagreements, where a disagreement occurs for a pair of items that appear in opposite relative orders in the two rankings. Introduced by statistician Maurice G. Kendall in his 1938 paper on methods, it serves as the foundational measure underlying Kendall's tau rank correlation coefficient, providing a non-parametric way to assess ordinal associations without assuming underlying distributions. Mathematically, for two permutations σ and π of {1, 2, ..., n}, the dK(σ, π) equals the number of pairs (i, j) with i < j such that the relative ordering of i and j differs between σ and π, formally dK(σ, π) = |{ (i,j) | i < j, sign(σ(i) - σ(j)) ≠ sign(π(i) - π(j)) }|. This count is equivalent to the minimum number of adjacent transpositions (swaps of neighboring elements) required to transform one ranking into the other, making it the graph in the Cayley graph of the symmetric group generated by adjacent swaps. It satisfies the properties of a metric: non-negativity (dK(σ, π) ≥ 0, with equality σ = π), (dK(σ, π) = dK(π, σ)), and the (dK(σ, ρ) ≤ dK(σ, π) + dK(π, ρ) for any permutations σ, π, ρ). The maximum possible between any two rankings is n choose 2 = n(n-1)/2, achieved when one is the reverse of the other. The Kendall tau distance is closely related to Kendall's tau correlation coefficient τ, defined as \tau = 1 - \frac{4 d_K(\sigma, \pi)}{n(n-1)}, which normalizes the distance to the range [-1, 1] and measures the strength and direction of the rank association (positive for agreement, negative for disagreement). Extensions handle ties and partial rankings, such as the generalized Kendall tau with parameter p ∈ [0,1] that penalizes inconsistencies involving tied pairs, formulated as dKp(σ₁, σ₂) = |D(σ₁, σ₂)| + p · (|R1(σ₁, σ₂)| + |R2(σ₁, σ₂)|), where D counts order disagreements and R1, R2 count tie inconsistencies. In applications, the Kendall tau distance is widely used in statistics for non-parametric hypothesis testing of rank correlations, in machine learning for evaluating ranking models (e.g., in recommendation systems and search engines via metrics like normalized discounted cumulative gain proxies), and in information retrieval and bioinformatics to compare ordered lists such as gene rankings or document relevance scores. It also appears in social choice theory for aggregating voter preferences under the Kemeny-Young method, where the optimal ranking minimizes the total distance to all input rankings, though this is NP-hard to compute exactly. Efficient algorithms exist for its computation, including O(n log n) time via sorting or dynamic programming for special cases, enabling scalability to large datasets.

Introduction

Historical Background

The Kendall tau distance traces its origins to the work of statistician Kendall, who introduced it in 1938 as a measure for assessing agreement between two rankings in the context of statistical analysis of . This development addressed the need for a robust to quantify concordance and discordance in paired rankings, emerging amid early 20th-century challenges in and where from preference orders and social surveys required systematic comparison. Kendall's initial formulation built on prior ideas, such as Gustav Fechner's 1897 proposal for a similar concordance measure in time series analysis, but tailored it specifically for in statistical applications. The measure evolved through Kendall's subsequent research, with a key formalization appearing in his paper on handling ties in ranking problems, where he extended the framework to account for equal ranks commonly encountered in real-world data. This work solidified the distance's role in non-parametric statistics, influencing its adoption beyond initial statistical contexts. The , "," derives from letter τ (τ), which Kendall employed in his notation to denote the rank correlation coefficient associated with the distance. Post-1970s, the Kendall tau distance saw significant uptake in bioinformatics, particularly for ranking tasks, where it provided a way to compare ordered structures. By the 2000s, it had integrated into as a standard metric for assessing models, aiding evaluations in areas like recommendation systems and predictive ordering. Today, it continues to inform modern applications, such as optimizing rankings in search engines.

Basic Concepts

In and ranking theory, a of a set of n distinct items is defined as a on that set, where every pair of items is comparable, the relation is transitive, antisymmetric, and total (i.e., for any two distinct items i and j, either i precedes j or j precedes i). This structure ensures a complete linear arrangement without ambiguities in relative positions, often represented as a \pi of the labels \{1, 2, \dots, n\}, where \pi(i) denotes the assigned to item i. Such rankings are fundamental in contexts like preference modeling, results, and statistical analysis of , providing a way to impose a strict on the items. To compare two rankings, the focus shifts to pairwise comparisons between items. For any two distinct items i and j, a pair is considered concordant if the relative ordering is the same in both rankings—meaning both i before j or both j before i—and discordant if the ordering is reversed between the two rankings. These comparisons capture the agreement or disagreement in the implied preferences, forming the basis for measures of similarity or distance between rankings. The total number of such unique pairs across n items is given by the \binom{n}{2} = \frac{n(n-1)}{2}, representing all possible two-item subsets that can be evaluated. Rankings can sometimes incorporate ties, where two or more items are deemed equally ranked, resulting in a weak order rather than a strict ; in such cases, tied pairs are neither concordant nor discordant. However, the foundational concepts of Kendall tau distance initially apply to tie-free rankings, assuming a strict with no equalities to simplify the analysis of pairwise agreements. This distinction highlights the need for extensions when ties are present in real-world data. The Kendall tau distance itself arises as a measure that counts these discordant pairs to quantify the divergence between two rankings.

Definition and Formulation

Formal Definition

The Kendall tau distance between two rankings \sigma and \tau of n items is defined as the number of discordant pairs, that is, the number of pairs (i, j) with i < j such that the relative order of i and j differs between the two rankings, i.e., one ranking places i before j while the other places j before i. This measure quantifies the extent of disagreement in the pairwise comparisons induced by the rankings. Mathematically, the distance is given by K(\sigma, \tau) = \sum_{1 \leq i < j \leq n} \mathbf{1}_{(\sigma(i) - \sigma(j))(\tau(i) - \tau(j)) < 0}, where \mathbf{1} is the indicator function that equals 1 if the condition holds (indicating a discordant pair) and 0 otherwise, and \sigma(k) denotes the rank assigned to item k in ranking \sigma. This distance is equivalent to the minimum number of adjacent transpositions (swaps of neighboring elements) required in a bubble sort to convert one ranking into the other. The value of K(\sigma, \tau) ranges from 0, when the rankings are identical (no discordant pairs), to \binom{n}{2} = \frac{n(n-1)}{2}, when one ranking is the complete reverse of the other (all pairs discordant). It is often normalized by dividing by \frac{n(n-1)}{2} to scale the distance to the interval [0, 1].

Normalized Distance

The normalized Kendall tau distance provides a scaled version of the raw count of discordant pairs, bounding the metric between 0 and 1 to facilitate comparisons and probabilistic analyses across rankings of varying sizes. It is formally defined as d(\sigma, \tau) = \frac{2K(\sigma, \tau)}{n(n-1)}, where K(\sigma, \tau) is the number of discordant pairs between two permutations \sigma and \tau of n elements. This scaling divides by the total number of possible ordered pairs, twice the number of unordered pairs \binom{n}{2} = \frac{n(n-1)}{2}, ensuring d(\sigma, \sigma) = 0 for identical rankings and d(\sigma, \tau) = 1 for fully reversed rankings. This normalized form interprets the distance as the probability that a pair of items selected uniformly at random from all possible pairs is discordant in the two rankings. In models assuming uniform random permutations, such as random permutation models in statistical ranking analysis, the expected value of d(\sigma, \tau) is 0.5, as each pair is equally likely to be concordant or discordant. When rankings include ties, resulting in partial orders, the normalization must adjust to account for incomparable pairs. One approach excludes pairs tied in either ranking, computing the distance as the proportion of discordant pairs among only the comparable (untied in both) pairs, effectively replacing n(n-1) with the count of such pairs in the denominator. Alternatively, a generalized form inspired by adjusts for ties by normalizing the raw discordant count against the effective variability, using \tau_b = \frac{C - D}{\sqrt{(T - T_x)(T - T_y)}}, where C and D are concordant and discordant pairs, T = \frac{n(n-1)}{2} is the total unordered pairs, and T_x, T_y are the numbers of tied pairs in each ranking; the associated distance can then be derived as d = \frac{1 - \tau_b}{2}. This tau-b variant, which reduces to the standard tau without ties, is particularly useful in contexts with moderate ties to maintain interpretability.

Properties

Metric Properties

The Kendall tau distance d(\sigma, \tau) between two permutations \sigma and \tau of n elements satisfies non-negativity, as the count of disagreeing pairs is always greater than or equal to zero. Furthermore, d(\sigma, \tau) = 0 if and only if \sigma = \tau, establishing the identity of indiscernibles, since no disagreements occur precisely when the rankings are identical. The distance is symmetric, meaning d(\sigma, \tau) = d(\tau, \sigma) for any permutations \sigma and \tau, because the number of pairwise disagreements remains unchanged when swapping the roles of the two rankings. It also satisfies the triangle inequality: for any three permutations \sigma, \tau, and \upsilon, d(\sigma, \upsilon) \leq d(\sigma, \tau) + d(\tau, \upsilon). This follows from the interpretation of the Kendall tau distance as the minimum number of adjacent transpositions required to transform one permutation into another; such path lengths inherently obey the triangle inequality. As a strict metric, the Kendall tau distance exhibits positivity in the space of rankings, where d(\sigma, \tau) > 0 whenever \sigma \neq \tau.

Invariance and Symmetry

The Kendall tau distance possesses label invariance, meaning that relabeling the items in both rankings consistently does not alter the computed . This property stems from the metric's reliance on relative ordering: it counts the number of pairwise discordant pairs—where the order of two items differs between rankings—independent of their specific identifiers or numerical labels. For instance, renaming items from numbers to letters or vice versa yields the same value, as the focus remains on ordinal relationships rather than absolute identifiers. Closely related is the permutation invariance, or more precisely right-invariance, of the Kendall tau distance: for any permutations \pi, \sigma, \rho \in S_n, d_K(\pi, \sigma) = d_K(\pi \circ \rho, \sigma \circ \rho). This ensures that applying the same permutation to the labels of both rankings preserves the distance, reflecting that equivalent rankings under group actions (such as relabeling via \rho) maintain identical relative order structures. A proof follows by observing that discordant pairs (i, j) in (\pi, \sigma) map bijectively to pairs (\rho^{-1}(i), \rho^{-1}(j)) in (\pi \circ \rho, \sigma \circ \rho), preserving the sign differences in their orderings and thus the total count. The distance is also symmetric, satisfying d_K(\pi, \sigma) = d_K(\sigma, \pi) for all \pi, \sigma \in S_n. In contrast to position-dependent metrics, such as the on rank position vectors (which underpins variants like Spearman's footrule), the depends exclusively on order relations and ignores absolute positional values. This order-centric nature reinforces its invariances, as transformations affecting positions but not pairwise orders leave the distance unaffected. As a result, in practical applications like rank aggregation or on , preprocessing to standardize labels is unnecessary, allowing direct use of raw ranking information without loss of metric fidelity.

Relations to Other Measures

Kendall Tau Rank Correlation

The Kendall tau rank correlation , denoted as \tau, measures the ordinal association between two rankings by assessing the proportion of concordant and discordant pairs among all possible pairs of elements. A concordant pair occurs when two elements are ordered in the same way in both rankings (i.e., both rankings agree on their relative order), while a discordant pair occurs when the order is reversed between the rankings. For two rankings \sigma and \pi of n distinct elements, let C be the number of concordant pairs and D the number of discordant pairs; the total number of pairs is \binom{n}{2} = n(n-1)/2. The is defined as \tau(\sigma, \pi) = \frac{C - D}{n(n-1)/2} = \frac{2(C - D)}{n(n-1)}. This formulation arises from the probability interpretation: \tau equals the probability of a random pair being concordant minus the probability of it being discordant, assuming no ties. The Kendall tau distance K(\sigma, \pi), which counts the number of discordant pairs D, directly relates to the through normalization. Since C + D = n(n-1)/2, substituting yields C - D = n(n-1)/2 - 2D, so \tau(\sigma, \pi) = \frac{n(n-1)/2 - 2D}{n(n-1)/2} = 1 - \frac{4D}{n(n-1)} = 1 - \frac{4K(\sigma, \pi)}{n(n-1)}. This inverse relationship highlights that the is a normalized, signed of the : a distance of zero corresponds to \tau = 1 (perfect agreement), while the maximum distance K = n(n-1)/2 yields \tau = -1 (perfect reversal). The coefficient thus ranges from -1 to $1, with values near zero indicating weak or no monotonic association. This connection was originally developed in Kendall's work on , later extended to metrics in ranking analysis. When ties are present in the rankings, the basic \tau (often called \tau_a) overestimates the association by not adjusting for unresolved pairs. The tie-adjusted variant, Kendall's \tau_b, corrects for this by scaling the numerator while accounting for tied pairs in each ranking separately. Let T_x be the number of pairs tied in the first ranking and T_y in the second; then the total pairs decompose as n(n-1)/2 = C + D + T_x + T_y - T_{xy} (where T_{xy} adjusts for double-counted ties), but \tau_b is computed as \tau_b(\sigma, \pi) = \frac{C - D}{\sqrt{(C + D + T_x)(C + D + T_y)}}. This version maintains the range [-1, 1] and reduces to \tau in the absence of ties, providing a more robust measure for real-world data with equal ranks. The adjustment was formalized in subsequent developments of Kendall's methods to handle partial orders.

Comparison with Spearman Distance

The Spearman rank distance between two rankings \sigma and \pi of n items is defined as d_S(\sigma, \pi) = \sum_{i=1}^n (\sigma(i) - \pi(i))^2, which quantifies the total squared deviations in the assigned positions of each item across the rankings. In contrast, the Kendall tau distance d_K(\sigma, \pi) counts the number of pairwise inversions, equivalent to the minimum number of adjacent transpositions required to transform one ranking into the other, focusing on discordant pairs where the relative order of two items differs. A fundamental difference lies in their measurement approach: Kendall tau distance is inversion-based, penalizing only the pairwise order disagreements without regard to the magnitude of positional shifts, while Spearman distance aggregates all positional deviations, emphasizing the extent of each item's displacement through squaring. This makes Kendall tau distance inherently tied to bubble sort-like operations on adjacent elements, whereas Spearman treats rankings as points in \mathbb{R}^n under the metric on rank vectors. Kendall tau distance exhibits greater robustness to outliers compared to Spearman distance, as the former's pairwise counting avoids amplifying extreme rank deviations through squaring, leading to lower in contaminated distributions. Specifically, Kendall penalizes local swaps (e.g., adjacent inversions) minimally with a cost of 1 per pair, while global shifts incur costs proportional to the number of affected pairs; Spearman, however, applies penalties that escalate sharply for large displacements, making it more sensitive to such outliers. The distances coincide up to for the identity ranking (measuring disorder identically in scale for monotonic cases), but diverge notably for non-monotonic rankings with clustered inversions, where Kendall yields smaller values for nearby discordant pairs. Note that Kendall tau distance relates directly to the via d_K = \frac{n(n-1)}{4} (1 - \tau), distinct from the analogous relation for Spearman distance and Spearman's \rho. To illustrate, consider n=4 items with the reference ranking \sigma = (1,2,3,4) and three alternatives:
Alternative \piDescriptionKendall \tau DistanceSpearman Distance
(1,3,2,4)Single adjacent swap12
(2,1,3,4)Single non-adjacent swap12
(4,3,2,1)Full reversal620
These values highlight Kendall's uniform penalty for single inversions regardless of proximity, versus Spearman's consistent small cost for minor shifts but explosive growth for extremes.

Computation Methods

Naive Computation

The naive computation of the Kendall tau distance between two rankings involves systematically examining every possible pair of distinct items to determine if their relative ordering disagrees between the rankings. Given two permutations σ and τ, which represent the rankings by assigning ranks to items labeled from 1 to n (where σ(i) denotes the rank of item i in the first ranking and τ(i) the rank in the second), the distance is the count of discordant pairs. A pair (i, j) with i < j is discordant if the product (σ(i) - σ(j)) × (τ(i) - τ(j)) is negative, indicating opposite relative orders in the two rankings. This approach can be implemented using nested loops to iterate over all pairs, as shown in the following pseudocode (assuming 1-based indexing for ranks and items):
function kendall_tau_naive(sigma, tau, n):
    count = 0
    for i = 1 to n-1:
        for j = i+1 to n:
            if (sigma[i] - sigma[j]) * (tau[i] - tau[j]) < 0:
                count += 1
    return count
The algorithm performs exactly \binom{n}{2} = n(n-1)/2 pairwise checks, each involving constant-time arithmetic operations, resulting in O(n^2) time complexity. Beyond storing the input rankings (which requires O(n) space), the method uses only a constant amount of additional space for the counter and loop variables, achieving O(1) auxiliary space complexity. This naive method is straightforward to implement in any programming language and is suitable for small to moderate dataset sizes, such as n < 1000, where the quadratic time remains computationally feasible on standard hardware. For larger n, efficient O(n log n) algorithms based on sorting or merge-sort variants for inversion counting are preferable to avoid excessive runtime.

Efficient Algorithms

The Kendall tau distance can be computed efficiently in O(n log n) time for rankings without ties by reducing the problem to counting inversions in the permutation obtained by composing the two rankings, which is achieved using a modified merge sort algorithm that accumulates discordant pairs during the merge steps. This approach adapts the standard inversion-counting technique from sorting algorithms, where splits and merges track relative orders across subarrays without enumerating all pairs. Advanced variants maintain O(n log n) complexity using data structures like (binary indexed trees) for range queries during insertion of sorted elements, allowing efficient determination of how many prior elements violate the order for each new item. These methods, which employ dynamic updates and prefix sum queries, are particularly suited for implementations requiring repeated computations or extensions to weighted distances. When rankings contain ties, forming partial orders, the distance computation excludes tied pairs from discordance counts or applies penalties based on variant definitions, such as treating tied pairs as neither concordant nor discordant in the basic Kendall tau-a. Specialized algorithms for partial rankings preprocess the data into buckets, then use divide-and-conquer merging or subset rank structures to count disagreements across tied groups in O(n log n / log log n) time, improving upon quadratic baselines for sparse ties. Practical implementations, such as the kendalltau function in SciPy's statistics module, adopt Fenwick tree-based counting of concordant and discordant pairs for O(n log n) performance on large datasets, supporting both correlation and distance variants with tie handling via normalization. These achieve the information-theoretic lower bound of Ω(n log n) in the comparison-based model, as the problem inherently requires distinguishing orderings akin to sorting. For parallelization, the recursive divide-and-conquer structure of merge sort enables multi-core distribution by processing independent subproblems concurrently, while GPU-accelerated frameworks using Fenwick trees compute full distance matrices for thousands of rankings in seconds, scaling to sparse or high-dimensional data in genomics and recommendation systems.

Examples and Applications

Illustrative Example

Consider four items labeled A, B, C, and D, ranked by height and by weight. The rank vector for height is σ = [1, 2, 3, 4], meaning A has the highest rank (tallest, rank 1), followed by B (rank 2), C (rank 3), and D (rank 4, shortest). The rank vector for weight is τ = [1, 3, 4, 2], meaning A has the highest rank (heaviest, rank 1), followed by D (rank 2), B (rank 3), and C (rank 4, lightest). Lower rank numbers indicate higher positions in both cases. The Kendall tau distance measures the disagreement between these rankings by counting the number of discordant pairs, where the relative order of two items differs across the criteria. With four items, there are \binom{4}{2} = 6 possible pairs to evaluate. The pairs are analyzed as follows:
PairHeight relative orderWeight relative orderStatus
A-BA > BA > BConcordant
A-CA > CA > CConcordant
A-DA > DA > DConcordant
B-CB > CB > CConcordant
B-DB > DD > BDiscordant
C-DC > DD > CDiscordant
Two pairs (B-D and C-D) are discordant. The raw Kendall tau distance is K = 2, the count of discordant pairs. The normalized distance, which scales the raw value to the range [0, 1], is d = \frac{K}{\binom{4}{2}} = \frac{2}{6} = \frac{1}{3} \approx 0.333. This value indicates moderate disagreement between the height and weight rankings, stemming from D's elevated position in the weight ranking relative to B and C.

Applications in Statistics and Machine Learning

In statistics, the Kendall tau distance serves as a key tool in non-parametric tests for assessing associations in rank data, particularly in contexts like testing where distributional assumptions on covariates are avoided. For instance, generalized versions of Kendall's tau enable robust evaluation of between variables given factors, using U-statistic-based approaches to handle pairwise comparisons without restrictions. This application is evident in econometric analyses, such as validating variables by testing conditions like Y ⊥ S | X, where rank-based distances provide a non-parametric alternative to traditional measures. Additionally, extensions to employ conditional Kendall's tau for estimating concordance in censored or interval-censored data, facilitating inference on hazard functions or time-to-event rankings under non-proportional hazards. In , the Kendall tau distance is widely applied in evaluation, especially within systems, where it quantifies the number of pairwise inversions between predicted and ground-truth rankings to assess model performance. It correlates with metrics like normalized (NDCG) by emphasizing ordinal disagreements, making it suitable for optimizing search engines and recommendation systems. In preference learning and ensemble methods, the distance supports robust by minimizing misrankings in noisy training data, as seen in methods that aggregate rankings while accounting for inconsistencies in pairwise preferences. Efficient algorithms for its computation ensure in large-scale ranking tasks, allowing integration into training pipelines for tasks like top-k retrieval. In bioinformatics, the Kendall tau distance measures similarity in ranked gene expression profiles or protein homology rankings derived from BLAST searches, enabling stratification of co-evolving genomic groups across species. For example, it compares rank-based phylogenetic profiles to cluster proteins into evolutionary categories, achieving high accuracy (e.g., 92% for certain microbial genomes) in identifying vertical versus horizontal gene transfer patterns. In phylogenetic tree analysis, adaptations like the Kendall-Colijn metric extend the distance to ranked tree shapes and genealogies, quantifying evolutionary divergences by encoding branch orders and lengths, which aids in summarizing tree distributions from genomic data such as influenza phylogenies. Beyond these domains, the Kendall tau distance finds use in for voting systems, where it models aggregation as error-correcting codes to recover ground-truth rankings from noisy voter inputs, bounding recovery errors within twice the average pairwise disagreement distance. In , weighted variants analyze data from attitude scaling and structures, such as month preferences or semantic recall tasks, by incorporating item importance or similarity weights to reveal clusters in ordinal responses like seasonal s or . A primary advantage of the Kendall tau distance lies in its suitability for , as its rank-based formulation captures monotonic associations without requiring interval scales or normality assumptions, outperforming parametric metrics in non-continuous settings. It also exhibits robustness to and outliers, since discrepancies are evaluated via concordant/discordant pairs rather than raw values, maintaining in sparse or contaminated datasets like those in or surveys.

References

  1. [1]
    [PDF] Computing Distances Between Partial Rankings
    The Kendall tau distance between two full rankings (over the same set of n elements) counts the number of pairwise disagreements between the two rank- ings, and ...
  2. [2]
    A New Measure of Rank Correlation - jstor
    In psychological work the problem of comparing two different rankings of the same set of individuals may be divided into two types. In the first type the.
  3. [3]
    Kendall tau distance in nLab
    Dec 25, 2023 · The Kendall tau distance between two permutations is the minimum number of transpositions (swaps) of adjacent elements needed to turn one into the other.
  4. [4]
    [PDF] The Kendall Rank Correlation Coefficient
    The Kendall (1955) rank correlation coefficient evaluates the de- gree of similarity between two sets of ranks given to a same set of objects.
  5. [5]
    GPU-accelerated Kendall distance computation for large or sparse ...
    Dec 9, 2024 · Kendall- τ correlation, being a robust nonparametric similarity measure, is widely employed for multidimensional data analysis in cases when the ...
  6. [6]
    [PDF] COMPARING TOP k LISTS 1. Introduction. The ... - biz.uiowa.edu
    Kendall's tau turns out to be equal to the number of exchanges needed in a bubble sort to convert one permutation to the other. Spearman's footrule metric is ...
  7. [7]
    RankAggreg, an R package for weighted rank aggregation
    Feb 19, 2009 · Kendall's tau distance. The Kendall's tau distance takes a different approach at measuring the distance between two ordered lists. It ...
  8. [8]
    Generalized distances between rankings - ACM Digital Library
    Abstract. Spearman's footrule and Kendall's tau are two well established distances between rankings. They, however, fail to take into account concepts crucial ...Missing: original paper
  9. [9]
    18.3 - Kendall Tau-b Correlation Coefficient | STAT 509
    The Kendall tau-b is a nonparametric measure of association based on concordances and discordances, ranging from -1 to +1, and is preferred over Spearman.
  10. [10]
    [PDF] Rank Aggregation Methods for the Web - Brown Computer Science
    Rank Aggregation Methods for the Web. Cynthia Dwork. Ravi Kumar y. Moni Naor z. D. Sivakumar x. ABSTRACT. We consider the problem of combining ranking results ...
  11. [11]
    [PDF] Extending Kendall Tau from Ranks to Sequences - arXiv
    In this paper, we begin with a metric on permutations, and adapt it to measure the distance between sequences (i.e., strings, arrays, or any other sequential ...
  12. [12]
    [PDF] Generalized Distances between Rankings - Stanford CS Theory
    Apr 30, 2010 · [18, 19] give a version of Kendall's tau distance, where every inversion is weighted in proportion to the product of the two element weights.
  13. [13]
    [PDF] arXiv:2401.10719v1 [math.CO] 19 Jan 2024
    Jan 19, 2024 · The Kendall-Tau metric is right invariant, that is, for any σ,τ,α ∈ Sn, dK(σ, ρ) = dK(σα, ρα). Proof. Let (i, j) be any pair such that 1 ...Missing: invariance | Show results with:invariance
  14. [14]
    [PDF] EXPLAINABLE SEQUENTIAL OPTIMIZATION - OpenReview
    The Spearman rank distance aggregates these squared differences across all elements to provide a cumulative measure of the rank divergence between the two ...<|control11|><|separator|>
  15. [15]
    [PDF] RankAggreg, an R package for weighted rank aggregation
    2.2 Kendall's tau distance. The Kendall's tau distance takes a different approach at measuring the dis- tance between two ordered lists. It utilizes pairs of ...
  16. [16]
    [PDF] A comparative analysis of Spearman's rho and Kendall's tau in ...
    Comparison of BIASz, z 2 fP,S,K,Mg for (a) ... In this paper we have investigated systematically the properties of Spearman's rho and Kendall's tau for sam-.
  17. [17]
    Computing distances between partial rankings - ScienceDirect.com
    We give two efficient algorithms for computing distances between partial rankings (i.e. rankings with ties). Given two partial rankings over n elements, and ...
  18. [18]
    kendalltau — SciPy v1.16.2 Manual
    Kendall's tau is a measure of the correspondence between two rankings. Values close to 1 indicate strong agreement, and values close to -1 indicate strong ...
  19. [19]
    NEW MEASURE OF RANK CORRELATION | Biometrika
    M. G. KENDALL; A NEW MEASURE OF RANK CORRELATION, Biometrika, Volume 30, Issue 1-2, 1 June 1938, Pages 81–93, https://doi.org/10.1093/biomet/30.1-2.81.