Research topic
Report
Detailed summary
Empirical studies quantifying the prevalence of the error where significance is inferred from a nonsignificant result alongside a significant one are scarce, but some studies highlight similar statistical misinterpretations in specific disciplines such as animal cognition [1], urology [2], and psychology [3].
Detailed Summary:
- Empirical Quantification of the Error:
- The specific error of interpreting the difference between significant and nonsignificant results as itself significant is not directly quantified across multiple disciplines in the studies retrieved. However, related statistical misinterpretations are analyzed within specific fields.
Field-Specific Studies:
- Animal Cognition: Research [1] quantifies how often non-significant results are reported as no effect, highlighting a lack of proper interpretation of statistical results.
- Urology: Study [2] examines how often abstracts misuse the term “trend toward significance” as a misinterpretation near p-value thresholds.
- Psychology: Literature [3, 6] reports frequent misinterpretations of nonsignificant results as accepting the null hypothesis and inadequate evidential support for such claims through Bayes factors.
General Discussions and Concerns:
- Articles such as [4] and [10] discuss the broader implications of statistical misinterpretations concerning significance, stressing the need for statistical literacy and correct method application.
- General calls for re-evaluating the use of p-values and significance thresholds due to their contribution to misinterpretations and replication issues in research [7, 9].
Overall, while studies directly quantifying the prevalence of the specific error across disciplines are lacking, similar issues with statistical significance are documented in specialized analyses within individual fields.
Categories of papers
- Papers that explicitly empirically quantify the prevalence of the statistical error or closely related errors, such as misinterpretation of nonsignificant results as significant.
- References: [1], [3], [5], [6], [9]
- Details:
- [1] examines misinterpretation of nonsignificant p-values in animal cognition studies.
- [3] investigates misinterpretation of nonsignificant results in Chinese psychological research.
- [5] looks at terms like "marginally significant" in interpretation of results in personality and social psychology studies.
- [6] analyzes nonsignificant results' evidential impact in psychology.
- [9] investigates statistical significance misuses in public health research.
- Papers that discuss the theoretical aspects of the statistical error without necessarily providing empirical quantification data.
- References: [4], [10]
- Details:
- [4] discusses the conceptual problem of inferring significance from nonsignificant results.
- [10] critiques the overreliance on 'statistical significance' as a decision criterion.
- Papers focusing broadly on common statistical misinterpretations related to p-values, including descriptions of “trends toward significance.”
- References: [2], [7]
- Details:
- [2] analyzes use of "trend toward significance" in urology.
- [7] reviews the impact of dichotomizing p-values into significance categories on research reproducibility.
- Papers proposing methodological or policy changes to address broader issues in statistical interpretation.
- References: [8], [10]
- Details:
- [8] compares correct vs. incorrect methods for interpreting statistical effects.
- [10] proposes retirement of the statistical significance threshold.
This categorization highlights papers of highest relevance to the researcher's interest in empirical quantification of a specific statistical error, with additional context on theoretical discussions and broader contexts or solutions offered.