17 found
Order:
  1. Understanding from Machine Learning Models.Emily Sullivan - 2022 - British Journal for the Philosophy of Science 73 (1):109-133.
    Simple idealized models seem to provide more understanding than opaque, complex, and hyper-realistic models. However, an increasing number of scientists are going in the opposite direction by utilizing opaque machine learning models to make predictions and draw inferences, suggesting that scientists are opting for models that have less potential for understanding. Are scientists trading understanding for some other epistemic or pragmatic good when they choose a machine learning model? Or are the assumptions behind why minimal models provide understanding misguided? In (...)
    Download  
     
    Export citation  
     
    Bookmark   56 citations  
  2. Vulnerability in Social Epistemic Networks.Emily Sullivan, Max Sondag, Ignaz Rutter, Wouter Meulemans, Scott Cunningham, Bettina Speckmann & Mark Alfano - 2020 - International Journal of Philosophical Studies 28 (5):1-23.
    Social epistemologists should be well-equipped to explain and evaluate the growing vulnerabilities associated with filter bubbles, echo chambers, and group polarization in social media. However, almost all social epistemology has been built for social contexts that involve merely a speaker-hearer dyad. Filter bubbles, echo chambers, and group polarization all presuppose much larger and more complex network structures. In this paper, we lay the groundwork for a properly social epistemology that gives the role and structure of networks their due. In particular, (...)
    Download  
     
    Export citation  
     
    Bookmark   16 citations  
  3. Inductive Risk, Understanding, and Opaque Machine Learning Models.Emily Sullivan - 2022 - Philosophy of Science 89 (5):1065-1074.
    Under what conditions does machine learning (ML) model opacity inhibit the possibility of explaining and understanding phenomena? In this article, I argue that nonepistemic values give shape to the ML opacity problem even if we keep researcher interests fixed. Treating ML models as an instance of doing model-based science to explain and understand phenomena reveals that there is (i) an external opacity problem, where the presence of inductive risk imposes higher standards on externally validating models, and (ii) an internal opacity (...)
    Download  
     
    Export citation  
     
    Bookmark   7 citations  
  4. Reliability in Machine Learning.Thomas Grote, Konstantin Genin & Emily Sullivan - 2024 - Philosophy Compass 19 (5):e12974.
    Issues of reliability are claiming center-stage in the epistemology of machine learning. This paper unifies different branches in the literature and points to promising research directions, whilst also providing an accessible introduction to key concepts in statistics and machine learning – as far as they are concerned with reliability.
    Download  
     
    Export citation  
     
    Bookmark   1 citation  
  5. Do ML models represent their targets?Emily Sullivan - forthcoming - Philosophy of Science.
    I argue that ML models used in science function as highly idealized toy models. If we treat ML models as a type of highly idealized toy model, then we can deploy standard representational and epistemic strategies from the toy model literature to explain why ML models can still provide epistemic success despite their lack of similarity to their targets.
    Download  
     
    Export citation  
     
    Bookmark   1 citation  
  6. Universality caused: the case of renormalization group explanation.Emily Sullivan - 2019 - European Journal for Philosophy of Science 9 (3):36.
    Recently, many have argued that there are certain kinds of abstract mathematical explanations that are noncausal. In particular, the irrelevancy approach suggests that abstracting away irrelevant causal details can leave us with a noncausal explanation. In this paper, I argue that the common example of Renormalization Group explanations of universality used to motivate the irrelevancy approach deserves more critical attention. I argue that the reasons given by those who hold up RG as noncausal do not stand up to critical scrutiny. (...)
    Download  
     
    Export citation  
     
    Bookmark   5 citations  
  7. Can Real Social Epistemic Networks Deliver the Wisdom of Crowds?Emily Sullivan, Max Sondag, Ignaz Rutter, Wouter Meulemans, Scott Cunningham, Bettina Speckmann & Mark Alfano - 2014 - In Tania Lombrozo, Joshua Knobe & Shaun Nichols (eds.), Oxford Studies in Experimental Philosophy, Volume 1. Oxford, GB: Oxford University Press UK.
    In this paper, we explain and showcase the promising methodology of testimonial network analysis and visualization for experimental epistemology, arguing that it can be used to gain insights and answer philosophical questions in social epistemology. Our use case is the epistemic community that discusses vaccine safety primarily in English on Twitter. In two studies, we show, using both statistical analysis and exploratory data visualization, that there is almost no neutral or ambivalent discussion of vaccine safety on Twitter. Roughly half the (...)
    Download  
     
    Export citation  
     
    Bookmark   3 citations  
  8. SIDEs: Separating Idealization from Deceptive ‘Explanations’ in xAI.Emily Sullivan - forthcoming - Proceedings of the 2024 Acm Conference on Fairness, Accountability, and Transparency.
    Explainable AI (xAI) methods are important for establishing trust in using black-box models. However, recent criticism has mounted against current xAI methods that they disagree, are necessarily false, and can be manipulated, which has started to undermine the deployment of black-box models. Rudin (2019) goes so far as to say that we should stop using black-box models altogether in high-stakes cases because xAI explanations ‘must be wrong’. However, strict fidelity to the truth is historically not a desideratum in science. Idealizations (...)
    Download  
     
    Export citation  
     
    Bookmark  
  9. Humility in networks.Mark Alfano & Emily Sullivan - 2020 - In Mark Alfano, Michael Patrick Lynch & Alessandra Tanesini (eds.), The Routledge Handbook of the Philosophy of Humility. New York, NY: Routledge.
    What do humility, intellectual humility, and open-mindedness mean in the context of inter-group conflict? We spend most of our time with ingroup members, such as family, friends, and colleagues. Yet our biggest disagreements —— about practical, moral, and epistemic matters —— are likely to be with those who do not belong to our ingroup. An attitude of humility towards the former might be difficult to integrate with a corresponding attitude of humility towards the latter, leading to smug tribalism that masquerades (...)
    Download  
     
    Export citation  
     
    Bookmark   2 citations  
  10. Vectors of epistemic insecurity.Emily Sullivan & Mark Alfano - 2020 - In Ian James Kidd, Quassim Cassam & Heather Battaly (eds.), Vice Epistemology. New York, NY: Routledge.
    Epistemologists have addressed a variety of modal epistemic standings, such as sensitivity, safety, risk, and epistemic virtue. These concepts mark out the ways that beliefs can fail to track the truth, articulate the conditions needed for knowledge, and indicate ways to become a better epistemic agent. However, it is our contention that current ways of carving up epistemic modality ignore the complexities that emerge when individuals are embedded within a community and listening to a variety of sources, some of whom (...)
    Download  
     
    Export citation  
     
    Bookmark   1 citation  
  11. From Explanation to Recommendation: Ethical Standards for Algorithmic Recourse.Emily Sullivan & Philippe Verreault-Julien - forthcoming - Proceedings of the 2022 AAAI/ACM Conference on AI, Ethics, and Society (AIES’22).
    People are increasingly subject to algorithmic decisions, and it is generally agreed that end-users should be provided an explanation or rationale for these decisions. There are different purposes that explanations can have, such as increasing user trust in the system or allowing users to contest the decision. One specific purpose that is gaining more traction is algorithmic recourse. We first pro- pose that recourse should be viewed as a recommendation problem, not an explanation problem. Then, we argue that the capability (...)
    Download  
     
    Export citation  
     
    Bookmark   1 citation  
  12. The wisdom-of-crowds: an efficient, philosophically-validated, social epistemological network profiling toolkit.Colin Klein, Marc Cheong, Marinus Ferreira, Emily Sullivan & Mark Alfano - 2023 - In Hocine Cherifi, Rosario Nunzio Mantegna, Luis M. Rocha, Chantal Cherifi & Salvatore Miccichè (eds.), Complex Networks and Their Applications XI: Proceedings of The Eleventh International Conference on Complex Networks and Their Applications: COMPLEX NETWORKS 2022 — Volume 1. Springer.
    The epistemic position of an agent often depends on their position in a larger network of other agents who provide them with information. In general, agents are better off if they have diverse and independent sources. Sullivan et al. [19] developed a method for quantitatively characterizing the epistemic position of individuals in a network that takes into account both diversity and independence; and presented a proof-of-concept, closed-source implementation on a small graph derived from Twitter data [19]. This paper reports on (...)
    Download  
     
    Export citation  
     
    Bookmark  
  13. A normative framework for sharing information online.Emily Sullivan & Mark Alfano - 2023 - In Carissa Véliz (ed.), The Oxford Handbook of Digital Ethics. Oxford University Press.
    People have always shared information through chains and networks of testimony. It’s arguably part of what makes us human and enables us to live in cooperative communities with populations greater than the Dunbar number. The invention of the Internet and the rise of social media have turbo-charged our ability to share information. In this chapter, we develop a normative framework for sharing information online. This framework takes into account both ethical and epistemic considerations that are intertwined in typical cases of (...)
    Download  
     
    Export citation  
     
    Bookmark  
  14. Motivated numeracy and active reasoning in a Western European sample.Paul Connor, Emily Sullivan, Mark Alfano & Nava Tintarev - 2020 - Behavioral Public Policy 1.
    Recent work by Kahan et al. (2017) on the psychology of motivated numeracy in the context of intracultural disagreement suggests that people are less likely to employ their capabilities when the evidence runs contrary to their political ideology. This research has so far been carried out primarily in the USA regarding the liberal–conservative divide over gun control regulation. In this paper, we present the results of a modified replication that included an active reasoning intervention with Western European participants regarding both (...)
    Download  
     
    Export citation  
     
    Bookmark   1 citation  
  15. How Values Shape the Machine Learning Opacity Problem.Emily Sullivan - 2022 - In Insa Lawler, Kareem Khalifa & Elay Shech (eds.), Scientific Understanding and Representation: Modeling in the Physical Sciences. New York, NY: Routledge. pp. 306-322.
    One of the main worries with machine learning model opacity is that we cannot know enough about how the model works to fully understand the decisions they make. But how much is model opacity really a problem? This chapter argues that the problem of machine learning model opacity is entangled with non-epistemic values. The chapter considers three different stages of the machine learning modeling process that corresponds to understanding phenomena: (i) model acceptance and linking the model to the phenomenon, (ii) (...)
    Download  
     
    Export citation  
     
    Bookmark  
  16. Link Uncertainty, Implementation, and ML Opacity: A Reply to Tamir and Shech.Emily Sullivan - 2022 - In Insa Lawler, Kareem Khalifa & Elay Shech (eds.), Scientific Understanding and Representation: Modeling in the Physical Sciences. New York, NY: Routledge. pp. 341-345.
    This chapter responds to Michael Tamir and Elay Shech’s chapter “Understanding from Deep Learning Models in Context.”.
    Download  
     
    Export citation  
     
    Bookmark  
  17. Ethical pitfalls for natural language processing in psychology.Mark Alfano, Emily Sullivan & Amir Ebrahimi Fard - forthcoming - In Morteza Dehghani & Ryan Boyd (eds.), The Atlas of Language Analysis in Psychology. Guilford Press.
    Knowledge is power. Knowledge about human psychology is increasingly being produced using natural language processing (NLP) and related techniques. The power that accompanies and harnesses this knowledge should be subject to ethical controls and oversight. In this chapter, we address the ethical pitfalls that are likely to be encountered in the context of such research. These pitfalls occur at various stages of the NLP pipeline, including data acquisition, enrichment, analysis, storage, and sharing. We also address secondary uses of the results (...)
    Download  
     
    Export citation  
     
    Bookmark