skip to main content
US FlagAn official website of the United States government
dot gov icon
Official websites use .gov
A .gov website belongs to an official government organization in the United States.
https lock icon
Secure .gov websites use HTTPS
A lock ( lock ) or https:// means you've safely connected to the .gov website. Share sensitive information only on official, secure websites.


Title: Moments and Cumulants of The Two-Stage Mann-Whitney Statistic
This paper illustrates how to calculate the moments and cumulants of the two-stage Mann-Whitney statistic. These results may be used to calculate the asymptotic critical values of the two-stage Mann-Whitney test. In this paper, a large amount of deductions will be showed.  more » « less
Award ID(s):
1712839
PAR ID:
10094404
Author(s) / Creator(s):
;
Date Published:
Journal Name:
ArXiv.org
ISSN:
2331-8422
Format(s):
Medium: X
Sponsoring Org:
National Science Foundation
More Like this
  1. Summary A general framework is set up to study the asymptotic properties of the intent-to-treat Wilcoxon–Mann–Whitney test in randomized experiments with nonignorable noncompliance. Under location-shift alternatives, the Pitman efficiencies of the intent-to-treat Wilcoxon–Mann–Whitney and $$t$$ tests are derived. It is shown that the former is superior if the compliers are more likely to be found in high-density regions of the outcome distribution or, equivalently, if the noncompliers tend to reside in the tails. By logical extension, the relative efficiency of the two tests is sharply bounded by least and most favourable scenarios in which the compliers are segregated into regions of lowest and highest density, respectively. Such bounds can be derived analytically as a function of the compliance rate for common location families such as Gaussian, Laplace, logistic and $$t$$ distributions. These results can help empirical researchers choose the more efficient test for existing data, and calculate sample size for future trials in anticipation of noncompliance. Results for nonadditive alternatives and other tests follow along similar lines. 
    more » « less
  2. The relationship between the robustness of HRV derived by linear and nonlinear methods to the required minimum data lengths has yet to be well understood. The normal electrocardiography (ECG) data of 14 healthy volunteers were applied to 34 HRV measures using various data lengths, and compared with the most prolonged (2000 R peaks or 750 s) by using the Mann–Whitney U test, to determine the 0.05 level of significance. We found that SDNN, RMSSD, pNN50, normalized LF, the ratio of LF and HF, and SD1 of the Poincaré plot could be adequately computed by small data size (60–100 R peaks). In addition, parameters of RQA did not show any significant differences among 60 and 750 s. However, longer data length (1000 R peaks) is recommended to calculate most other measures. The DFA and Lyapunov exponent might require an even longer data length to show robust results. Conclusions: Our work suggests the optimal minimum data sizes for different HRV measures which can potentially improve the efficiency and save the time and effort for both patients and medical care providers. 
    more » « less
  3. Lack of racial diversity has been an ongoing issue in higher education. Recently, the Theory of Racialized Organizations has been used to help explain why, despite many calls for diversity, the demographics of higher education have not changed. Considering this framework, we seek to understand what aspects of the graduate school application process are viewed as barriers by minoritized students for applying. As part of a larger study of undergraduate student knowledge of the graduate school application process, we analyze 515 responses from undergraduate math majors using Mann-Whitney U tests to identify differences in what participants view as a barrier to apply to graduate school by race/ethnicity. We discuss two main results and recommend changes to graduate programs wishing to recruit more minoritized students. 
    more » « less
  4. Leitner, Stephan (Ed.)
    ObjectivePeer review frequently follows a process where reviewers first provide initial reviews, authors respond to these reviews, then reviewers update their reviews based on the authors’ response. There is mixed evidence regarding whether this process is useful, including frequent anecdotal complaints that reviewers insufficiently update their scores. In this study, we aim to investigate whether reviewersanchorto their original scores when updating their reviews, which serves as a potential explanation for the lack of updates in reviewer scores. DesignWe design a novel randomized controlled trial to test if reviewers exhibit anchoring. In the experimental condition, participants initially see a flawed version of a paper that is corrected after they submit their initial review, while in the control condition, participants only see the correct version. We take various measures to ensure that in the absence of anchoring, reviewers in the experimental group should revise their scores to be identically distributed to the scores from the control group. Furthermore, we construct the reviewed paper to maximize the difference between the flawed and corrected versions, and employ deception to hide the true experiment purpose. ResultsOur randomized controlled trial consists of 108 researchers as participants. First, we find that our intervention was successful at creating a difference in perceived paper quality between the flawed and corrected versions: Using a permutation test with the Mann-WhitneyUstatistic, we find that the experimental group’s initial scores are lower than the control group’s scores in both the Evaluation category (Vargha-DelaneyA= 0.64,p= 0.0096) and Overall score (A= 0.59,p= 0.058). Next, we test for anchoring by comparing the experimental group’s revised scores with the control group’s scores. We find no significant evidence of anchoring in either the Overall (A= 0.50,p= 0.61) or Evaluation category (A= 0.49,p= 0.61). The Mann-WhitneyUrepresents the number of individual pairwise comparisons across groups in which the value from the specified group is stochastically greater, while the Vargha-DelaneyAis the normalized version in [0, 1]. 
    more » « less
  5. Water quality is affected by multiple spatial and temporal factors, including the surrounding land characteristics, human activities, and antecedent precipitation amounts. However, identifying the relationships between water quality and spatially and temporally varying environmental variables with a machine learning technique in a heterogeneous urban landscape has been understudied. We explore how seasonal and variable precipitation amounts and other small-scale landscape variables affect E. coli, total suspended solids (TSS), nitrogen-nitrate, orthophosphate, lead, and zinc concentrations in Portland, Oregon, USA. Mann–Whitney tests were used to detect differences in water quality between seasons and COVID-19 periods. Spearman’s rank correlation analysis was used to identify the relationship between water quality and explanatory variables. A Random Forest (RF) model was used to predict water quality using antecedent precipitation amounts and landscape variables as inputs. The performance of RF was compared with that of ordinary least squares (OLS). Mann–Whitney tests identified statistically significant differences in all pollutant concentrations (except TSS) between the wet and dry seasons. Nitrate was the only pollutant to display statistically significant reductions in median concentrations (from 1.5 mg/L to 1.04 mg/L) during the COVID-19 lockdown period, likely associated with reduced traffic volumes. Spearman’s correlation analysis identified the highest correlation coefficients between one-day precipitation amounts and E. coli, lead, zinc, and TSS concentrations. Road length is positively associated with E. coli and zinc. The Random Forest (RF) model best predicts orthophosphate concentrations (R2 = 0.58), followed by TSS (R2 = 0.54) and nitrate (R2 = 0.46). E. coli was the most difficult to model and had the highest RMSE, MAE, and MAPE values. Overall, the Random Forest model outperformed OLS, as evaluated by RMSE, MAE, MAPE, and R2. The Random Forest was an effective approach to modeling pollutant concentrations using both categorical seasonal and COVID data along with continuous rain and landscape variables to predict water quality in urban streams. Implementing optimization techniques can further improve the model’s performance and allow researchers to use a machine learning approach for water quality modeling. 
    more » « less