Next Article in Journal
FedTKD: A Trustworthy Heterogeneous Federated Learning Based on Adaptive Knowledge Distillation
Next Article in Special Issue
Likelihood Ratio Test and the Evidential Approach for 2 × 2 Tables
Previous Article in Journal
Probabilistic Hesitant Fuzzy Evidence Theory and Its Application in Capability Evaluation of a Satellite Communication System
Previous Article in Special Issue
Profile Likelihood for Hierarchical Models Using Data Doubling
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

How the Post-Data Severity Converts Testing Results into Evidence for or against Pertinent Inferential Claims

Department of Economics, Virginia Tech, Blacksburg, VA 24061, USA
Entropy 2024, 26(1), 95; https://doi.org/10.3390/e26010095
Submission received: 29 October 2023 / Revised: 29 November 2023 / Accepted: 29 December 2023 / Published: 22 January 2024

Abstract

The paper makes a case that the current discussions on replicability and the abuse of significance testing have overlooked a more general contributor to the untrustworthiness of published empirical evidence, which is the uninformed and recipe-like implementation of statistical modeling and inference. It is argued that this contributes to the untrustworthiness problem in several different ways, including [a] statistical misspecification, [b] unwarranted evidential interpretations of frequentist inference results, and [c] questionable modeling strategies that rely on curve-fitting. What is more, the alternative proposals to replace or modify frequentist testing, including [i] replacing p-values with observed confidence intervals and effects sizes, and [ii] redefining statistical significance, will not address the untrustworthiness of evidence problem since they are equally vulnerable to [a]–[c]. The paper calls for distinguishing between unduly data-dependant ‘statistical results’, such as a point estimate, a p-value, and accept/reject H0, from ‘evidence for or against inferential claims’. The post-data severity (SEV) evaluation of the accept/reject H0 results, converts them into evidence for or against germane inferential claims. These claims can be used to address/elucidate several foundational issues, including (i) statistical vs. substantive significance, (ii) the large n problem, and (iii) the replicability of evidence. Also, the SEV perspective sheds light on the impertinence of the proposed alternatives [i]–[iii], and oppugns [iii] the alleged arbitrariness of framing H0 and H1 which is often exploited to undermine the credibility of frequentist testing.
Keywords: replication; untrustworthy evidence; statistical misspecification; statistical vs. substantive significance; pre-data vs. post-data error probabilities; p-hacking; post-data severity evaluation; observed confidence intervals; effect sizes replication; untrustworthy evidence; statistical misspecification; statistical vs. substantive significance; pre-data vs. post-data error probabilities; p-hacking; post-data severity evaluation; observed confidence intervals; effect sizes

Share and Cite

MDPI and ACS Style

Spanos, A. How the Post-Data Severity Converts Testing Results into Evidence for or against Pertinent Inferential Claims. Entropy 2024, 26, 95. https://doi.org/10.3390/e26010095

AMA Style

Spanos A. How the Post-Data Severity Converts Testing Results into Evidence for or against Pertinent Inferential Claims. Entropy. 2024; 26(1):95. https://doi.org/10.3390/e26010095

Chicago/Turabian Style

Spanos, Aris. 2024. "How the Post-Data Severity Converts Testing Results into Evidence for or against Pertinent Inferential Claims" Entropy 26, no. 1: 95. https://doi.org/10.3390/e26010095

APA Style

Spanos, A. (2024). How the Post-Data Severity Converts Testing Results into Evidence for or against Pertinent Inferential Claims. Entropy, 26(1), 95. https://doi.org/10.3390/e26010095

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop