Analysis of ordinal outcome#

Analysis methods#

1. Proportional odds (PO) model#

  • Assumptions [1]

    • Parallelism: the logit of the CDFs of different groups should be parallel (equal distance) to each other

      • Check this assumption: calculated the empirical CDF(ECDF) for the response variable by group; take logit transformation of the ECDF; check for parallelism

    • Linearity: required ONLY if using a parametric logistic distribution instead of semi-parametric PO model

  • It’s semi-parametric model in the sense that it assumes additivity and linearity (by default) in formulating the probabilities but it doesn’t assume the distribution of the response variable \(Y\)

    • PO model assumes nothing about the actual \(Y\) distribution; it only assumes how the distribution of \(Y\) for one group is connected to the distribution of another group, i.e. the logit of the cumulative distribution functions are parallel

Connection with Wilcoxon/Kruskal-Wallis test#

  • PO is also what the Wilcoxon/Kruskal-Wallis test assumes to have optimal power

    • By treating all the distinct values in the response variable as a level, then the outcome can be seen as a ordinal variable with many categories

    • The hypothesis that the Wilcoxon test possess, i.e. the probability a random sample from group A is greater than a random sample from group B, is technically identical to that the general PO model tries to test (OR = 1 for any cutoff)

  • Numerator of the score test for the PO model, when there is only the grouping variable in the model (i.e. not adjusted), is exactly the Wilcoxon statistic

  • Kruskal-Wallis test can be formed using a PO model with more than one indicator variable

    • This can solve the transivity problem when conducting the pairwise Wilxoson tests [2]

    • Model (with 4 groups): \(logit[P(Y\ge y|\text{group})] = \alpha_y + \beta_1B + \beta_2C + \beta_3D\)

Advantages of PO model vs purely non-parametric counterparts#

  • can adjust for (continuous) covariates

  • more accurate p-values even with extreme number of tied values

  • provides a framework for consistent pairwise comparisons [2]

  • provides estimates of means, quantiles, and exceedance probabilities [3]

  • sets the sage for Bayesian PO model, so can get a Bayesian Wilcoxon test

Software#

  • rms package in R

    • rms::orm(y ~ group)

  • Hmisc package in R

    • Power calculation with popower and posamsize

    • simRegOrd function can also simulate power for an adjusted two-sample comparison

      if there is one adjustment covariate

2. Check PO assumption#

  • Score test in PROC LOGISTIC: extreme anti-conservatism in many cases

  • Compare means of \(X|Y\) with and without assuming PO

  • Stratifying one each predictor and computing the logits of all proportions of the form \(Y\ge j, j = \text{1, 2, }\cdots, k\). When proportional odds holds, the difference in logits between different values of \(j\) should be the same at all levels of \(X\), because the model dictates that \(logit(Y\ge j| X) - logit(Y\ge i|X) = \alpha_j - \alpha_i\), for any constant \(X\).

    • We check both the consistency of \(logit(Y\ge j| X) - logit(Y\ge i|X)\) at different \(X\) levels

    • Or we can check the parallelism of \(logit(Y\ge j| X) - logit(Y\ge j|X+k)\) for \(Y = 1, \cdots, J\)

      image-20200828135732415

3. Method for PO is not met (?)#

  • When analyzing ordinal data where PO is not met, the Wilcoxon test may still work OK but it will not be optimal.

  • One way to generate a non-parametric version of OR (called Wilcoxon Mann Whitney Generalized OR [WMW GenOR]) is introduced in [Agresti, 1980] and later discussed in [Churilov et al., 2014]. [O’Brien and Castelloe, 2006] offered a way to eliminate ties and apply Agresti’s formula directly.

    • Note: this statistic isn’t really an “odds ratio”, rather it’s an odds per its mathematic definition

  • WMW GenOR test statistic is defined as

    输入图片描述

    \[\hat{\alpha} = \sum_{j>i} p_{1i}p_{2j}/\sum_{j<i}p_{1i}p_{2j}\]

    where \(p_{1i} = N_{1i}/\sum_i N_{1i}\) and \(p_{2i} = N_{2i}/\sum_i N_{2i}\)

    which follows normal distribution with mean \(\alpha = Pr(Y_2 > Y_1)/Pr(Y_1 > Y_2)\) and variance

    \[\hat{\sigma}^2 = \left\{\frac{1}{N_1}\sum_{j}p_{ij}\left(\hat{\alpha}\sum_{i<j}p_{2i} - \sum_{i>j}p_{2i}\right)^2 + \frac{1}{N_2}\left(\hat{\alpha}\sum_{i>j}p_{1i} - \sum_{i<j}p_{2i}\right)^2\right\} \Bigg/ \left(\sum_{i>j}p_{1i}p_{2j}\right)^2\]

    Furthermore, with patients being stratified into total \(M\) strata, the logrithm of test statistic can be written as

    \[\log(\hat{\alpha}) = \sum_m^M \hat{\alpha}_m^2 \log\hat{\alpha}_m/\hat{\sigma}_m^2\Bigg/ \sum_{m}^M\hat{\alpha}_m^2/\hat{\sigma}_m^2\]

    with variance

    \[\hat{\sigma}^2 = \left(\sum_m^M\hat{\alpha}_m^2/\hat{\sigma^2_m}\right)^{-1}\]
  • However, I think as long as it uses Wilcoxon type of method, it actually assumes PO in the data

Sample size calculation with ordinal outcome#

Method 1: with proportional odds assumption [Whitehead, 1993]#

  • Let \(Q_{ie\text{ or }c} = \sum_{u=1}^i p_{ue\text{ or }c}, \text{for } i = 1,\cdots, k-1\)

  • Log odds ratio at cutoff \(i\) is defined as \(\theta_{iR} = \log\left\{\frac{Q_{ie}(1-Q_{ic})}{(1-Q_{ie})Q_{ic}}\right\}\)

    • Proportional odds assumption says: \(\theta_{iR} = \theta_R = \frac{Q_{e}(1-Q_{c})}{(1-Q_{e})Q_{c}}\) for all \(i = 1, \cdots, k-1\)

Basic steps#

Set up the assumption for outcome distribution#

  • Assume the percentages at the most commonly used cutoff (e.g. mRS \(\le\) 2 vs > 2)

  • Based on this assumption, calculate the common odds ratio

  • With the assumption of proportional odds, calculate the percentages for each ordinal category and the full distribution of the ordinal outcome can be obtained accordingly

Sample size calculation#

using normal approximation

\[n = \frac{3(A+1)^2}{A(1-\sum_{i=1}^k \bar{p}_i^3)}*\frac{(z_{\alpha/2}+z_{\beta})^2}{\theta_R^2},\]

where

  • \(A = n_c/n_e\), is the randomization ratio between control and treated arms

  • \(\theta_R\) is the log of the common odds ratio for treated vs control arm

  • \(z_{\alpha/2} = \Phi^{-1}(1-\alpha/2), z_{\beta} = \Phi^{-1}(1-\beta)\)

  • \(\bar{p}_i = \frac{p_{ie} + p_{ic}}{2}\) for \(i= 1, \cdots, k\)

Method 2: without PO assumtion [Zhao et al., 2008]#

Link to notes.

Method 3: without PO assumption [Tang, 2011]#

  • Exact variance method for the size and power calculation for theWilcoxon–Mann–Whitney test for ordered categorical data.

  • Notes to be done

Resources for Ordinal Regression Models#

General references#

  • https://hbiostat.org/bib/ordinal.html which includes tutorials such as the one by Susan Scott et al.

[Agr80]

Alan Agresti. Generalized odds ratios for ordinal data. Biometrics, pages 59–67, 1980.

[CAJ+14]

Leonid Churilov, Sarah Arnup, Hayden Johns, Tiffany Leung, Stuart Roberts, Bruce CV Campbell, Stephen M Davis, and Geoffrey A Donnan. An improved method for simple, assumption-free ordinal analysis of the modified rankin scale using generalized odds ratios. International Journal of Stroke, 9(8):999–1005, 2014.

[OBrienC06]

Ralph G O’Brien and John Castelloe. Exploiting the link between the wilcoxon-mann-whitney test and a simple odds statistic. In Proceedings of the Thirty-first Annual SAS Users Group International Conference, 209–31. Citeseer, 2006.

[Tan11]

Yongqiang Tang. Size and power estimation for the wilcoxon–mann–whitney test for ordered categorical data. Statistics in Medicine, 30(29):3461–3470, 2011.

[Whi93]

John Whitehead. Sample size calculations for ordered categorical data. Statistics in medicine, 12(24):2257–2271, 1993.

[ZRQ08]

Yan D Zhao, Dewi Rahardja, and Yongming Qu. Sample size calculation for the wilcoxon–mann–whitney test adjusting for ties. Statistics in medicine, 27(3):462–468, 2008.