scholarly journals Reaction times and other skewed distributions: problems with the mean and the median

2018 ◽  
Author(s):  
Guillaume A. Rousselet ◽  
Rand R. Wilcox

ABSTRACTTo summarise skewed (asymmetric) distributions, such as reaction times, typically the mean or the median are used as measures of central tendency. Using the mean might seem surprising, given that it provides a poor measure of central tendency for skewed distributions, whereas the median provides a better indication of the location of the bulk of the observations. However, the sample median is biased: with small sample sizes, it tends to overestimate the population median. This is not the case for the mean. Based on this observation, Miller (1988) concluded that “sample medians must not be used to compare reaction times across experimental conditions when there are unequal numbers of trials in the conditions.” Here we replicate and extend Miller (1988), and demonstrate that his conclusion was ill-advised for several reasons. First, the median’s bias can be corrected using a percentile bootstrap bias correction. Second, a careful examination of the sampling distributions reveals that the sample median is median unbiased, whereas the mean is median biased when dealing with skewed distributions. That is, on average the sample mean estimates the population mean, but typically this is not the case. In addition, simulations of false and true positives in various situations show that no method dominates. Crucially, neither the mean nor the median are sufficient or even necessary to compare skewed distributions. Different questions require different methods and it would be unwise to use the mean or the median in all situations. Better tools are available to get a deeper understanding of how distributions differ: we illustrate a powerful alternative that relies on quantile estimation. All the code and data to reproduce the figures and analyses in the article are available online.

2019 ◽  
Author(s):  
Guillaume A Rousselet ◽  
Rand R. Wilcox

To summarise skewed (asymmetric) distributions, such as reaction times, typically the mean or the median are used as measures of central tendency. Using the mean might seem surprising, given that it provides a poor measure of central tendency for skewed distributions, whereas the median provides a better indication of the location of the bulk of the observations. However, the sample median is biased: with small sample sizes, it tends to overestimate the population median. This is not the case for the mean. Based on this observation, Miller (1988) concluded that ”sample medians must not be used to compare reaction times across experimental conditions when there are unequal numbers of trials in the conditions.” Here we replicate and extend Miller (1988), and demonstrate that his conclusion was ill-advised for several reasons. First, the median’s bias can be corrected using a percentile bootstrap bias correction. Second, a careful examination of the sampling distributions reveals that the sample median is median unbiased, whereas the mean is median biased when dealing with skewed distributions. That is, on average the sample mean estimates the population mean, but typically this is not the case. In addition, simulations of false and true positives in various situations show that no method dominates. Crucially, neither the mean nor the median are sufficient or even necessary to compare skewed distributions. Different questions require different methods and it would be unwise to use the mean or the median in all situations. Better tools are available to get a deeper understanding of how distributions differ: we illustrate the hierarchical shift function, a powerful alternative that relies on quantile estimation. All the code and data to reproduce the figures and analyses in the article are available online.


2020 ◽  
Vol 4 ◽  
Author(s):  
Guillaume A Rousselet ◽  
Rand R Wilcox

To summarise skewed (asymmetric) distributions, such as reaction times, typically the mean or the median are used as measures of central tendency. Using the mean might seem surprising, given that it provides a poor measure of central tendency for skewed distributions, whereas the median provides a better indication of the location of the bulk of the observations. However, the sample median is biased: with small sample sizes, it tends to overestimate the population median. This is not the case for the mean. Based on this observation, Miller (1988) concluded that "sample medians must not be used to compare reaction times across experimental conditions when there are unequal numbers of trials in the conditions". Here we replicate and extend Miller (1988), and demonstrate that his conclusion was ill-advised for several reasons. First, the median's bias can be corrected using a percentile bootstrap bias correction. Second, a careful examination of the sampling distributions reveals that the sample median is median unbiased, whereas the mean is median biased when dealing with skewed distributions. That is, on average the sample mean estimates the population mean, but typically this is not the case. In addition, simulations of false and true positives in various situations show that no method dominates. Crucially, neither the mean nor the median are sufficient or even necessary to compare skewed distributions. Different questions require different methods and it would be unwise to use the mean or the median in all situations. Better tools are available to get a deeper understanding of how distributions differ: we illustrate the hierarchical shift function, a powerful alternative that relies on quantile estimation. All the code and data to reproduce the figures and analyses in the article are available online.


2020 ◽  
Author(s):  
Jeff Miller

Contrary to the warning of Miller (1988), Rousselet and Wilcox (2020) argued that it is better to summarize each participant’s single-trial reaction times (RTs) in a given condition with the median than with the mean when comparing the central tendencies of RT distributions across experimental conditions. They acknowledged that median RTs can produce inflated Type I error rates when conditions differ in the number of trials tested, consistent with Miller’s warning, but they showed that the bias responsible for this error rate inflation could be eliminated with a bootstrap bias correction technique. The present simulations extend their analysis by examining the power of bias-corrected medians to detect true experimental effects and by comparing this power with the power of analyses using means and regular medians. Unfortunately, although bias-corrected medians solve the problem of inflated Type I error rates, their power is lower than that of means or regular medians in many realistic situations. In addition, even when conditions do not differ in the number of trials tested, the power of tests (e.g., t-tests) is generally lower using medians rather than means as the summary measures. Thus, the present simulations demonstrate that summary means will often provide the most powerful test for differences between conditions, and they show what aspects of the RT distributions determine the size of the power advantage for means.


1990 ◽  
Vol 24 (1) ◽  
pp. 1-4 ◽  
Author(s):  
R. Sanz Sampelayo ◽  
J. Fonolla ◽  
F. Gil Extremera

A study was carried out to examine the distribution of individual weights in Helix aspersa snails, the aims being to establish the best estimate of the ponderal growth and also to obtain a model growth curve. Four groups of 20 snails from the same clutch were analysed and kept under experimental conditions from birth up to 6 months. The variability of their individual weights within groups was studied by calculating the coefficients of variation every 15 days. At the same time, the assumed normal distribution of those weights was being tested. The coefficients of variation increased with age and the assumed normal distribution of individual weights had to be rejected. By means of a log transformation of the original data, a model growth curve was constructed, and was used to assess the possibility of estimating age from weight. We finally reached the conclusion that median weight, rather than the mean, would be a better measure of central tendency to use until it is possible to obtain selected populations. The difficulty of estimating age from weight is emphasized.


2020 ◽  
Vol 18 (2) ◽  
pp. 2-16
Author(s):  
Christina Chatzipantsiou ◽  
Marios Dimitriadis ◽  
Manos Papadakis ◽  
Michail Tsagris

Re-sampling based statistical tests are known to be computationally heavy, but reliable when small sample sizes are available. Despite their nice theoretical properties not much effort has been put to make them efficient. Computationally efficient method for calculating permutation-based p-values for the Pearson correlation coefficient and two independent samples t-test are proposed. The method is general and can be applied to other similar two sample mean or two mean vectors cases.


Author(s):  
Michael J. Kalsher ◽  
William G. Obenauer ◽  
Christopher F. Weiss

Debate continues regarding the relative effectiveness of the ANSI Z535 guidelines for the design and placement of warnings. Research shows consistent positive effects of these guidelines on precursors to warning compliance (e.g., noticing, reading, intended compliance), but less consistency on compliance behavior. Challenges in interpreting these findings stem from factors such as small sample sizes, varying research designs and experimental conditions, and treating the ANSI Z535 guidelines as a singular entity rather than as an integrative system of separable features. Here, we address these issues by testing perceptions of warning label effectiveness using a large sample ( n=533) and systematically manipulating variables cited in the Z535 guidelines. Collectively, we tested eight label designs for a (2-drawer and 4-drawer) file cabinet and found statistically significant relationships between design recommendations from the ANSI Z535 guidelines and perceptions of effectiveness. The presence of a warning header and pictogram exerted the largest effects. Bulleted text and (larger) font size were also related to increased perceptions of effectiveness.


2020 ◽  
Vol 29 (9) ◽  
pp. 2520-2537 ◽  
Author(s):  
Sean McGrath ◽  
XiaoFei Zhao ◽  
Russell Steele ◽  
Brett D. Thombs ◽  
Andrea Benedetti ◽  
...  

Researchers increasingly use meta-analysis to synthesize the results of several studies in order to estimate a common effect. When the outcome variable is continuous, standard meta-analytic approaches assume that the primary studies report the sample mean and standard deviation of the outcome. However, when the outcome is skewed, authors sometimes summarize the data by reporting the sample median and one or both of (i) the minimum and maximum values and (ii) the first and third quartiles, but do not report the mean or standard deviation. To include these studies in meta-analysis, several methods have been developed to estimate the sample mean and standard deviation from the reported summary data. A major limitation of these widely used methods is that they assume that the outcome distribution is normal, which is unlikely to be tenable for studies reporting medians. We propose two novel approaches to estimate the sample mean and standard deviation when data are suspected to be non-normal. Our simulation results and empirical assessments show that the proposed methods often perform better than the existing methods when applied to non-normal data.


1998 ◽  
Vol 25 (5) ◽  
pp. 475 ◽  
Author(s):  
D. G. Barratt

Information on the amount of vertebrate prey caught by house cats in Canberra was collected by recording prey deposited at cat owners’ residences over 12 months. The amount of prey taken was not significantly influenced by cat gender, age when neutered, or cat breed. Nor did belling or the number of meals provided per day have a significant influence on the amount of prey caught. The age of the cat and the proportion of nights spent outside explained approximately 11% of the variation in the amount of prey caught by individual cats. In all, 43% of variation in predation on introduced species (predominantly rodents) was explained by distance from potential prey source areas (i.e. rural/grassland habitat) and cat density. The mean number of prey reported per cat over 12 months (10.2) was significantly lower than mean predation per cat per year based on estimates made by cat owners before the prey survey began (23.3). Counts of the amount of prey caught by house cats were highly positively skewed. In all, 70% of cats were observed to catch less than 10 prey over 12 months, but for 6% of cats, more than 50 prey were recorded. Estimates of predation by house cats, particularly extrapolated estimates, should be treated with caution. The total number of prey caught by house cats in Canberra estimated using the sample median was approximately half the estimate based on the sample mean. Predation estimates alone do not prove that prey populations are detrimentally affected, especially in highly disturbed and modified environments such as suburbs. Impacts on native fauna are likely to be most significant in undisturbed habitat adjacent to new residential developments.


Author(s):  
Petr Zvyagin ◽  
Jaakko Heinonen

Sets of measurements of underwater ridge parts usually contain a limited amount of data. Outcomes need to be made while relying on small sample sizes. In this event, the chance of making inaccurate estimations increases. This paper proposes to use stochastic confidence regions in the estimation of the unknown parameters of keel depths. A model for a random variable with a lognormal distribution for keel depths is assumed. Regions for the mean and standard deviation of keel depths are obtained from Mood’s and minimum-area confidence regions for parameters of the normally distributed random variable. Conservative safety probability of non-exceeding the critical keel depth in one random interaction of the ridge with structure is estimated. An algorithm for statistically assessment of ice ridge keel data by means of confidence region building is here offered. The assessment of a set of ridge keel depths for the Gulf of Bothnia (Baltic Sea) is performed.


2016 ◽  
Author(s):  
Brian Keith Lohman ◽  
Jesse N Weber ◽  
Daniel I Bolnick

RNAseq is a relatively new tool for ecological genetics that offers researchers insight into changes in gene expression in response to a myriad of natural or experimental conditions. However, standard RNAseq methods (e.g., Illumina TruSeq® or NEBNext®) can be cost prohibitive, especially when study designs require large sample sizes. Consequently, RNAseq is often underused as a method, or is applied to small sample sizes that confer poor statistical power. Low cost RNAseq methods could therefore enable far greater and more powerful applications of transcriptomics in ecological genetics and beyond. Standard mRNAseq is costly partly because one sequences portions of the full length of all transcripts. Such whole-mRNA data is redundant for estimates of relative gene expression. TagSeq is an alternative method that focuses sequencing effort on mRNAs 3-prime end, thereby reducing the necessary sequencing depth per sample, and thus cost. Here we present a revised TagSeq protocol, and compare its performance against NEBNext®, the gold-standard whole mRNAseq method. We built both TagSeq and NEBNext® libraries from the same biological samples, each spiked with control RNAs. We found that TagSeq measured the control RNA distribution more accurately than NEBNext®, for a fraction of the cost per sample (~10%). The higher accuracy of TagSeq was particularly apparent for transcripts of moderate to low abundance. Technical replicates of TagSeq libraries are highly correlated, and were correlated with NEBNext® results. Overall, we show that our modified TagSeq protocol is an efficient alternative to traditional whole mRNAseq, offering researchers comparable data at greatly reduced cost.


Sign in / Sign up

Export Citation Format

Share Document