Switch to: References

Add citations

You must login to add citations.
  1. Algorithms Advise, Humans Decide: the Evidential Role of the Patient Preference Predictor.Nicholas Makins - forthcoming - Journal of Medical Ethics.
    An AI-based “patient preference predictor” (PPP) is a proposed method for guiding healthcare decisions for patients who lack decision-making capacity. The proposal is to use correlations between sociodemographic data and known healthcare preferences to construct a model that predicts the unknown preferences of a particular patient. In this paper, I highlight a distinction that has been largely overlooked so far in debates about the PPP–that between algorithmic prediction and decision-making–and argue that much of the recent philosophical disagreement stems from this (...)
    Download  
     
    Export citation  
     
    Bookmark  
  • The Patient preference predictor and the objection from higher-order preferences.Jakob Thrane Mainz - 2023 - Journal of Medical Ethics 49 (3):221-222.
    Recently, Jardas _et al_ have convincingly defended the patient preference predictor (PPP) against a range of autonomy-based objections. In this response, I propose a new autonomy-based objection to the PPP that is not explicitly discussed by Jardas _et al_. I call it the ‘objection from higher-order preferences’. Even if this objection is not sufficient reason to reject the PPP, the objection constitutes a pro tanto reason that is at least as powerful as the ones discussed by Jardas _et al._.
    Download  
     
    Export citation  
     
    Bookmark   5 citations  
  • Commentary on ‘Autonomy-based criticisms of the patient preference predictor’.Collin O'Neil - 2022 - Journal of Medical Ethics 48 (5):315-316.
    When a patient lacks sufficient capacity to make a certain treatment decision, whether because of deficits in their ability to make a judgement that reflects their values or to make a decision that reflects their judgement or both, the decision must be made by a surrogate. Often the best way to respect the patient’s autonomy, in such cases, is for the surrogate to make a ‘substituted’ judgement on behalf of the patient, which is the decision that best reflects the patient’s (...)
    Download  
     
    Export citation  
     
    Bookmark   2 citations  
  • Health Care in Contexts of Risk, Uncertainty, and Hybridity.Daniel Messelken & David Winkler (eds.) - 2021 - Springer.
    This book sheds light on various ethical challenges military and humanitarian health care personnel face while working in adverse conditions. Contexts of armed conflict, hybrid wars or other forms of violence short of war, as well as natural disasters, all have in common that ordinary circumstances can no longer be taken for granted. Hence, the provision of health care has to adapt, for example, to a different level of risk, to scarce resources, or uncommon approaches due to external incentives or (...)
    Download  
     
    Export citation  
     
    Bookmark  
  • Personalized Patient Preference Predictors Are Neither Technically Feasible nor Ethically Desirable.Nathaniel Sharadin - 2024 - American Journal of Bioethics 24 (7):62-65.
    Except in extraordinary circumstances, patients' clinical care should reflect their preferences. Incapacitated patients cannot report their preferences. This is a problem. Extant solutions to the problem are inadequate: surrogates are unreliable, and advance directives are uncommon. In response, some authors have suggested developing algorithmic "patient preference predictors" (PPPs) to inform care for incapacitated patients. In a recent paper, Earp et al. propose a new twist on PPPs. Earp et al. suggest we personalize PPPs using modern machine learning (ML) techniques. In (...)
    Download  
     
    Export citation  
     
    Bookmark  
  • Predicting and Preferring.Nathaniel Sharadin - forthcoming - Inquiry: An Interdisciplinary Journal of Philosophy.
    The use of machine learning, or “artificial intelligence” (AI) in medicine is widespread and growing. In this paper, I focus on a specific proposed clinical application of AI: using models to predict incapacitated patients’ treatment preferences. Drawing on results from machine learning, I argue this proposal faces a special moral problem. Machine learning researchers owe us assurance on this front before experimental research can proceed. In my conclusion I connect this concern to broader issues in AI safety.
    Download  
     
    Export citation  
     
    Bookmark   1 citation  
  • The ethics of biomedical military research: Therapy, prevention, enhancement, and risk.Alexandre Erler & Vincent C. Müller - 2021 - In Daniel Messelken & David Winkler (eds.), Health Care in Contexts of Risk, Uncertainty, and Hybridity. Springer. pp. 235-252.
    What proper role should considerations of risk, particularly to research subjects, play when it comes to conducting research on human enhancement in the military context? We introduce the currently visible military enhancement techniques (1) and the standard discussion of risk for these (2), in particular what we refer to as the ‘Assumption’, which states that the demands for risk-avoidance are higher for enhancement than for therapy. We challenge the Assumption through the introduction of three categories of enhancements (3): therapeutic, preventive, (...)
    Download  
     
    Export citation  
     
    Bookmark   1 citation  
  • A Personalized Patient Preference Predictor for Substituted Judgments in Healthcare: Technically Feasible and Ethically Desirable.Brian D. Earp, Sebastian Porsdam Mann, Jemima Allen, Sabine Salloch, Vynn Suren, Karin Jongsma, Matthias Braun, Dominic Wilkinson, Walter Sinnott-Armstrong, Annette Rid, David Wendler & Julian Savulescu - 2024 - American Journal of Bioethics 24 (7):13-26.
    When making substituted judgments for incapacitated patients, surrogates often struggle to guess what the patient would want if they had capacity. Surrogates may also agonize over having the (sole) responsibility of making such a determination. To address such concerns, a Patient Preference Predictor (PPP) has been proposed that would use an algorithm to infer the treatment preferences of individual patients from population-level data about the known preferences of people with similar demographic characteristics. However, critics have suggested that even if such (...)
    Download  
     
    Export citation  
     
    Bookmark   26 citations  
  • Should Artificial Intelligence be used to support clinical ethical decision-making? A systematic review of reasons.Sabine Salloch, Tim Kacprowski, Wolf-Tilo Balke, Frank Ursin & Lasse Benzinger - 2023 - BMC Medical Ethics 24 (1):1-9.
    BackgroundHealthcare providers have to make ethically complex clinical decisions which may be a source of stress. Researchers have recently introduced Artificial Intelligence (AI)-based applications to assist in clinical ethical decision-making. However, the use of such tools is controversial. This review aims to provide a comprehensive overview of the reasons given in the academic literature for and against their use.MethodsPubMed, Web of Science, Philpapers.org and Google Scholar were searched for all relevant publications. The resulting set of publications was title and abstract (...)
    Download  
     
    Export citation  
     
    Bookmark   5 citations  
  • Will a Patient Preference Predictor Improve Treatment Decision Making for Incapacitated Patients?Annette Rid - 2014 - Journal of Medicine and Philosophy 39 (2):99-103.
    Download  
     
    Export citation  
     
    Bookmark   3 citations  
  • Patient preference predictors and the problem of naked statistical evidence.Nathaniel Paul Sharadin - 2018 - Journal of Medical Ethics 44 (12):857-862.
    Patient preference predictors (PPPs) promise to provide medical professionals with a new solution to the problem of making treatment decisions on behalf of incapacitated patients. I show that the use of PPPs faces a version of a normative problem familiar from legal scholarship: the problem of naked statistical evidence. I sketch two sorts of possible reply, vindicating and debunking, and suggest that our reply to the problem in the one domain ought to mirror our reply in the other. The conclusion (...)
    Download  
     
    Export citation  
     
    Bookmark   13 citations  
  • Autonomy-based criticisms of the patient preference predictor.E. J. Jardas, David Wasserman & David Wendler - 2022 - Journal of Medical Ethics 48 (5):304-310.
    The patient preference predictor is a proposed computer-based algorithm that would predict the treatment preferences of decisionally incapacitated patients. Incorporation of a PPP into the decision-making process has the potential to improve implementation of the substituted judgement standard by providing more accurate predictions of patients’ treatment preferences than reliance on surrogates alone. Yet, critics argue that methods for making treatment decisions for incapacitated patients should be judged on a number of factors beyond simply providing them with the treatments they would (...)
    Download  
     
    Export citation  
     
    Bookmark   12 citations