Switch to: References

Add citations

You must login to add citations.
  1. Classical versus Bayesian Statistics.Eric Johannesson - 2020 - Philosophy of Science 87 (2):302-318.
    In statistics, there are two main paradigms: classical and Bayesian statistics. The purpose of this article is to investigate the extent to which classicists and Bayesians can agree. My conclusion is that, in certain situations, they cannot. The upshot is that, if we assume that the classicist is not allowed to have a higher degree of belief in a null hypothesis after he has rejected it than before, then he has to either have trivial or incoherent credences to begin with (...)
    Download  
     
    Export citation  
     
    Bookmark   2 citations  
  • Why is Bayesian confirmation theory rarely practiced.Robert W. P. Luk - 2019 - Science and Philosophy 7 (1):3-20.
    Bayesian confirmation theory is a leading theory to decide the confirmation/refutation of a hypothesis based on probability calculus. While it may be much discussed in philosophy of science, is it actually practiced in terms of hypothesis testing by scientists? Since the assignment of some of the probabilities in the theory is open to debate and the risk of making the wrong decision is unknown, many scientists do not use the theory in hypothesis testing. Instead, they use alternative statistical tests that (...)
    Download  
     
    Export citation  
     
    Bookmark   3 citations  
  • Significance testing, p-values and the principle of total evidence.Bengt Autzen - 2016 - European Journal for Philosophy of Science 6 (2):281-295.
    The paper examines the claim that significance testing violates the Principle of Total Evidence. I argue that p-values violate PTE for two-sided tests but satisfy PTE for one-sided tests invoking a sufficient test statistic independent of the preferred theory of evidence. While the focus of the paper is to evaluate a particular claim about the relationship of significance testing and PTE, I clarify the reading of this methodological principle along the way.
    Download  
     
    Export citation  
     
    Bookmark   2 citations  
  • Cosmic Bayes. Datasets and priors in the hunt for dark energy.Michela Massimi - 2021 - European Journal for Philosophy of Science 11 (1):1-21.
    Bayesian methods are ubiquitous in contemporary observational cosmology. They enter into three main tasks: cross-checking datasets for consistency; fixing constraints on cosmological parameters; and model selection. This article explores some epistemic limits of using Bayesian methods. The first limit concerns the degree of informativeness of the Bayesian priors and an ensuing methodological tension between task and task. The second limit concerns the choice of wide flat priors and related tension between parameter estimation and model selection. The Dark Energy Survey and (...)
    Download  
     
    Export citation  
     
    Bookmark   1 citation  
  • Severity and Trustworthy Evidence: Foundational Problems versus Misuses of Frequentist Testing.Aris Spanos - 2022 - Philosophy of Science 89 (2):378-397.
    For model-based frequentist statistics, based on a parametric statistical model ${{\cal M}_\theta }$, the trustworthiness of the ensuing evidence depends crucially on the validity of the probabilistic assumptions comprising ${{\cal M}_\theta }$, the optimality of the inference procedures employed, and the adequateness of the sample size to learn from data by securing –. It is argued that the criticism of the postdata severity evaluation of testing results based on a small n by Rochefort-Maranda is meritless because it conflates [a] misuses (...)
    Download  
     
    Export citation  
     
    Bookmark  
  • Bernoulli’s golden theorem in retrospect: error probabilities and trustworthy evidence.Aris Spanos - 2021 - Synthese 199 (5-6):13949-13976.
    Bernoulli’s 1713 golden theorem is viewed retrospectively in the context of modern model-based frequentist inference that revolves around the concept of a prespecified statistical model Mθx\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$${\mathcal{M}}_{{{\varvec{\uptheta}}}} \left( {\mathbf{x}} \right)$$\end{document}, defining the inductive premises of inference. It is argued that several widely-accepted claims relating to the golden theorem and frequentist inference are either misleading or erroneous: (a) Bernoulli solved the problem of inference ‘from probability to frequency’, and thus (b) the golden theorem (...)
    Download  
     
    Export citation  
     
    Bookmark  
  • Inflated effect sizes and underpowered tests: how the severity measure of evidence is affected by the winner’s curse.Guillaume Rochefort-Maranda - 2021 - Philosophical Studies 178 (1):133-145.
    My aim in this paper is to show how the problem of inflated effect sizes corrupts the severity measure of evidence. This has never been done. In fact, the Winner’s Curse is barely mentioned in the philosophical literature. Since the severity score is the predominant measure of evidence for frequentist tests in the philosophical literature, it is important to underscore its flaws. It is also crucial to bring the philosophical literature up to speed with the limits of classical testing. The (...)
    Download  
     
    Export citation  
     
    Bookmark   1 citation  
  • The Jeffreys–Lindley paradox and discovery criteria in high energy physics.Robert D. Cousins - 2017 - Synthese 194 (2):395-432.
    The Jeffreys–Lindley paradox displays how the use of a \ value ) in a frequentist hypothesis test can lead to an inference that is radically different from that of a Bayesian hypothesis test in the form advocated by Harold Jeffreys in the 1930s and common today. The setting is the test of a well-specified null hypothesis versus a composite alternative. The \ value, as well as the ratio of the likelihood under the null hypothesis to the maximized likelihood under the (...)
    Download  
     
    Export citation  
     
    Bookmark   6 citations  
  • On the Jeffreys-Lindley Paradox.Christian P. Robert - 2014 - Philosophy of Science 81 (2):216-232,.
    This article discusses the dual interpretation of the Jeffreys-Lindley paradox associated with Bayesian posterior probabilities and Bayes factors, both as a differentiation between frequentist and Bayesian statistics and as a pointer to the difficulty of using improper priors while testing. I stress the considerable impact of this paradox on the foundations of both classical and Bayesian statistics. While assessing existing resolutions of the paradox, I focus on a critical viewpoint of the paradox discussed by Spanos in Philosophy of Science.
    Download  
     
    Export citation  
     
    Bookmark   7 citations  
  • History and nature of the Jeffreys–Lindley paradox.Eric-Jan Wagenmakers & Alexander Ly - 2023 - Archive for History of Exact Sciences 77 (1):25-72.
    The Jeffreys–Lindley paradox exposes a rift between Bayesian and frequentist hypothesis testing that strikes at the heart of statistical inference. Contrary to what most current literature suggests, the paradox was central to the Bayesian testing methodology developed by Sir Harold Jeffreys in the late 1930s. Jeffreys showed that the evidence for a point-null hypothesis $${\mathcal {H}}_0$$ H 0 scales with $$\sqrt{n}$$ n and repeatedly argued that it would, therefore, be mistaken to set a threshold for rejecting $${\mathcal {H}}_0$$ H 0 (...)
    Download  
     
    Export citation  
     
    Bookmark  
  • Indices of Effect Existence and Significance in the Bayesian Framework.Dominique Makowski, Mattan S. Ben-Shachar, S. H. Annabel Chen & Daniel Lüdecke - 2019 - Frontiers in Psychology 10.
    Download  
     
    Export citation  
     
    Bookmark   10 citations  
  • A Paradoxical Feature of the Severity Measure of Evidence.Guillaume Rochefort-Maranda - unknown
    The main point of this paper is to underscore that tests with very low power will be significant only if the observations are deviant under both H0 and H1. Therefore, the results of those significant tests will generate misleadingly high severity scores for differences between H0 and H1 that are excessively overestimated. In other words, that measure of evidence is bound to fail in those cases. It will inevitably fail to adequately measure the strength of the evidence provided by tests (...)
    Download  
     
    Export citation  
     
    Bookmark