scholarly journals Using Split Samples to Improve Inference on Causal Effects

2017 ◽  
Vol 25 (4) ◽  
pp. 465-482 ◽  
Author(s):  
Marcel Fafchamps ◽  
Julien Labonne

We discuss a statistical procedure to carry out empirical research that combines recent insights about preanalysis plans (PAPs) and replication. Researchers send their datasets to an independent third party who randomly generates training and testing samples. Researchers perform their analysis on the training sample and are able to incorporate feedback from both colleagues, editors, and referees. Once the paper is accepted for publication the method is applied to the testing sample and it is those results that are published. Simulations indicate that, under empirically relevant settings, the proposed method delivers more power than a PAP. The effect mostly operates through a lower likelihood that relevant hypotheses are left untested. The method appears better suited for exploratory analyses where there is significant uncertainty about the outcomes of interest. We do not recommend using the method in situations where the treatment are very costly and thus the available sample size is limited. An interpretation of the method is that it allows researchers to perform direct replication of their work. We also discuss a number of practical issues about the method’s feasibility and implementation.

2021 ◽  
Vol 66 (18) ◽  
pp. 185012
Author(s):  
Yingtao Fang ◽  
Jiazhou Wang ◽  
Xiaomin Ou ◽  
Hongmei Ying ◽  
Chaosu Hu ◽  
...  

2022 ◽  
Vol 13 ◽  
Author(s):  
Niklas Wulms ◽  
Lea Redmann ◽  
Christine Herpertz ◽  
Nadine Bonberg ◽  
Klaus Berger ◽  
...  

Introduction: White matter hyperintensities of presumed vascular origin (WMH) are an important magnetic resonance imaging marker of cerebral small vessel disease and are associated with cognitive decline, stroke, and mortality. Their relevance in healthy individuals, however, is less clear. This is partly due to the methodological challenge of accurately measuring rare and small WMH with automated segmentation programs. In this study, we tested whether WMH volumetry with FMRIB software library v6.0 (FSL; https://fsl.fmrib.ox.ac.uk/fsl/fslwiki) Brain Intensity AbNormality Classification Algorithm (BIANCA), a customizable and trainable algorithm that quantifies WMH volume based on individual data training sets, can be optimized for a normal aging population.Methods: We evaluated the effect of varying training sample sizes on the accuracy and the robustness of the predicted white matter hyperintensity volume in a population (n = 201) with a low prevalence of confluent WMH and a substantial proportion of participants without WMH. BIANCA was trained with seven different sample sizes between 10 and 40 with increments of 5. For each sample size, 100 random samples of T1w and FLAIR images were drawn and trained with manually delineated masks. For validation, we defined an internal and external validation set and compared the mean absolute error, resulting from the difference between manually delineated and predicted WMH volumes for each set. For spatial overlap, we calculated the Dice similarity index (SI) for the external validation cohort.Results: The study population had a median WMH volume of 0.34 ml (IQR of 1.6 ml) and included n = 28 (18%) participants without any WMH. The mean absolute error of the difference between BIANCA prediction and manually delineated masks was minimized and became more robust with an increasing number of training participants. The lowest mean absolute error of 0.05 ml (SD of 0.24 ml) was identified in the external validation set with a training sample size of 35. Compared to the volumetric overlap, the spatial overlap was poor with an average Dice similarity index of 0.14 (SD 0.16) in the external cohort, driven by subjects with very low lesion volumes.Discussion: We found that the performance of BIANCA, particularly the robustness of predictions, could be optimized for use in populations with a low WMH load by enlargement of the training sample size. Further work is needed to evaluate and potentially improve the prediction accuracy for low lesion volumes. These findings are important for current and future population-based studies with the majority of participants being normal aging people.


2020 ◽  
pp. 28-63
Author(s):  
A. G. Vinogradov

The article belongs to a special modern genre of scholar publications, so-called tutorials – articles devoted to the application of the latest methods of design, modeling or analysis in an accessible format in order to disseminate best practices. The article acquaints Ukrainian psychologists with the basics of using the R programming language to the analysis of empirical research data. The article discusses the current state of world psychology in connection with the Crisis of Confidence, which arose due to the low reproducibility of empirical research. This problem is caused by poor quality of psychological measurement tools, insufficient attention to adequate sample planning, typical statistical hypothesis testing practices, and so-called “questionable research practices.” The tutorial demonstrates methods for determining the sample size depending on the expected magnitude of the effect size and desired statistical power, performing basic variable transformations and statistical analysis of psychological research data using language and environment R. The tutorial presents minimal system of R functions required to carry out: modern analysis of reliability of measurement scales, sample size calculation, point and interval estimation of effect size for four the most widespread in psychology designs for the analysis of two variables’ interdependence. These typical problems include finding the differences between the means and variances in two or more samples, correlations between continuous and categorical variables. Practical information on data preparation, import, basic transformations, and application of basic statistical methods in the cloud version of RStudio is provided.


2019 ◽  
Vol 67 (1) ◽  
pp. 67-79 ◽  
Author(s):  
Katarzyna Haverkamp

Zusammenfassung In der empirischen Wirtschaftsforschung zeigt sich ein zunehmendes Interesse an der Untersuchung der Fragen der Gründungsdynamik und des Gründungserfolgs im Kontext der deutschen Handwerkswirtschaft. Eine besondere Herausforderung für diese Analysen besteht jedoch darin, dass eine statistische Abgrenzung des juristisch definierten Handwerkssektors in den vorliegenden Sekundärdatensätzen meist nur mit Einschränkungen möglich ist. Vor diesem Hintergrund analysiert dieser Beitrag Möglichkeiten und Grenzen einer statistischen Abgrenzung des Handwerks in Mikrodatensätzen und untersucht unterschiedliche, bislang verwendete Identifikationsverfahren im Hinblick auf die Repräsentativität der jeweils gewonnen Stichproben. Im Ergebnis zeigt der Beitrag die Stärken und Schwächen unterschiedlicher Identifikationsverfahren und formuliert Empfehlungen hinsichtlich ihrer Verwendung in der Entrepreneurship-Forschung. Abstract Recently, several empirical studies investigate the causal effects of regulation on market entry and exit using the example of the German crafts sector. However, since the definition of the sector is made on legal- and not statistical basis, the identification of crafts companies and employees in microdata records is an intricate process. This paper examines different identification strategies that have been used so far in empirical research and investigates whether the resulting samples are consistent with the overall population in question. The paper contributes to existing economic research by providing an understanding for the potential pitfalls when analyzing sub-groups in larger datasets and by formulating an explicit recommendation for the case of the research on regulation and entry in the German crafts sector.


Sign in / Sign up

Export Citation Format

Share Document