Switch to: References

Add citations

You must login to add citations.
  1. Models, Parameterization, and Software: Epistemic Opacity in Computational Chemistry.Frédéric Wieber & Alexandre Hocquet - 2020 - Perspectives on Science 28 (5):610-629.
    . Computational chemistry grew in a new era of “desktop modeling,” which coincided with a growing demand for modeling software, especially from the pharmaceutical industry. Parameterization of models in computational chemistry is an arduous enterprise, and we argue that this activity leads, in this specific context, to tensions among scientists regarding the epistemic opacity transparency of parameterized methods and the software implementing them. We relate one flame war from the Computational Chemistry mailing List in order to assess in detail the (...)
    Download  
     
    Export citation  
     
    Bookmark   4 citations  
  • Epistemic opacity, confirmation holism and technical debt: computer simulation in the light of empirical software engineering.Julian Newman - 2016 - In History and Philosophy of Computing (IFIP AICT 487). Cham, Switzerland: Springer. pp. 256-272.
    Epistemic opacity vis a vis human agents has been presented as an essential, ineliminable characteristic of computer simulation models resulting from the characteristics of the human cognitive agent. This paper argues, on the contrary, that such epistemic opacity as does occur in computer simulations is not a consequence of human limitations but of a failure on the part of model developers to adopt good software engineering practice for managing human error and ensuring the software artefact is maintainable. One consequence of (...)
    Download  
     
    Export citation  
     
    Bookmark   10 citations  
  • Transparency in Complex Computational Systems.Kathleen A. Creel - 2020 - Philosophy of Science 87 (4):568-589.
    Scientists depend on complex computational systems that are often ineliminably opaque, to the detriment of our ability to give scientific explanations and detect artifacts. Some philosophers have s...
    Download  
     
    Export citation  
     
    Bookmark   53 citations  
  • Modelos, idealizaciones: una crítica del ficcionalismo.Alejandro Cassini - 2013 - Principia: An International Journal of Epistemology 17 (3):345.
    Download  
     
    Export citation  
     
    Bookmark  
  • Spotting When Algorithms Are Wrong.Stefan Buijsman & Herman Veluwenkamp - 2023 - Minds and Machines 33 (4):541-562.
    Users of sociotechnical systems often have no way to independently verify whether the system output which they use to make decisions is correct; they are epistemically dependent on the system. We argue that this leads to problems when the system is wrong, namely to bad decisions and violations of the norm of practical reasoning. To prevent this from occurring we suggest the implementation of defeaters: information that a system is unreliable in a specific case (undercutting defeat) or independent information that (...)
    Download  
     
    Export citation  
     
    Bookmark   1 citation  
  • Two Dimensions of Opacity and the Deep Learning Predicament.Florian J. Boge - 2021 - Minds and Machines 32 (1):43-75.
    Deep neural networks have become increasingly successful in applications from biology to cosmology to social science. Trained DNNs, moreover, correspond to models that ideally allow the prediction of new phenomena. Building in part on the literature on ‘eXplainable AI’, I here argue that these models are instrumental in a sense that makes them non-explanatory, and that their automated generation is opaque in a unique way. This combination implies the possibility of an unprecedented gap between discovery and explanation: When unsupervised models (...)
    Download  
     
    Export citation  
     
    Bookmark   17 citations  
  • Strengthening Weak Emergence.Nora Berenstain - 2020 - Erkenntnis 87 (5):2457-2474.
    Bedau's influential (1997) account analyzes weak emergence in terms of the non-derivability of a system’s macrostates from its microstates except by simulation. I offer an improved version of Bedau’s account of weak emergence in light of insights from information theory. Non-derivability alone does not guarantee that a system’s macrostates are weakly emergent. Rather, it is non-derivability plus the algorithmic compressibility of the system’s macrostates that makes them weakly emergent. I argue that the resulting information-theoretic picture provides a metaphysical account of (...)
    Download  
     
    Export citation  
     
    Bookmark  
  • Philosophy of science at sea: Clarifying the interpretability of machine learning.Claus Beisbart & Tim Räz - 2022 - Philosophy Compass 17 (6):e12830.
    Philosophy Compass, Volume 17, Issue 6, June 2022.
    Download  
     
    Export citation  
     
    Bookmark   8 citations  
  • Opacity thought through: on the intransparency of computer simulations.Claus Beisbart - 2021 - Synthese 199 (3-4):11643-11666.
    Computer simulations are often claimed to be opaque and thus to lack transparency. But what exactly is the opacity of simulations? This paper aims to answer that question by proposing an explication of opacity. Such an explication is needed, I argue, because the pioneering definition of opacity by P. Humphreys and a recent elaboration by Durán and Formanek are too narrow. While it is true that simulations are opaque in that they include too many computations and thus cannot be checked (...)
    Download  
     
    Export citation  
     
    Bookmark   11 citations  
  • How can computer simulations produce new knowledge?Claus Beisbart - 2012 - European Journal for Philosophy of Science 2 (3):395-434.
    It is often claimed that scientists can obtain new knowledge about nature by running computer simulations. How is this possible? I answer this question by arguing that computer simulations are arguments. This view parallels Norton’s argument view about thought experiments. I show that computer simulations can be reconstructed as arguments that fully capture the epistemic power of the simulations. Assuming the extended mind hypothesis, I furthermore argue that running the computer simulation is to execute the reconstructing argument. I discuss some (...)
    Download  
     
    Export citation  
     
    Bookmark   34 citations  
  • Testing Bottom-Up Models of Complex Citation Networks.Mark A. Bedau - 2014 - Philosophy of Science 81 (5):1131-1143.
    The robust behavior of the patent citation network is a complex target of recent bottom-up models in science. This paper investigates the purpose and testing of three especially simple bottom-up models of the citation count distribution observed in the patent citation network. The complex causal webs in the models generate weakly emergent patterns of behavior, and this explains both the need for empirical observation of computer simulations of the models and the epistemic harmlessness of the resulting epistemic opacity.
    Download  
     
    Export citation  
     
    Bookmark   3 citations  
  • About the warrants of computer-based empirical knowledge.Anouk Barberousse & Marion Vorms - 2014 - Synthese 191 (15):3595-3620.
    Computer simulations are widely used in current scientific practice, as a tool to obtain information about various phenomena. Scientists accordingly rely on the outputs of computer simulations to make statements about the empirical world. In that sense, simulations seem to enable scientists to acquire empirical knowledge. The aim of this paper is to assess whether computer simulations actually allow for the production of empirical knowledge, and how. It provides an epistemological analysis of present-day empirical science, to which the traditional epistemological (...)
    Download  
     
    Export citation  
     
    Bookmark   13 citations  
  • On the presumed superiority of analytical solutions over numerical methods.Vincent Ardourel & Julie Jebeile - 2017 - European Journal for Philosophy of Science 7 (2):201-220.
    An important task in mathematical sciences is to make quantitative predictions, which is often done via the solution of differential equations. In this paper, we investigate why, to perform this task, scientists sometimes choose to use numerical methods instead of analytical solutions. Via several examples, we argue that the choice for numerical methods can be explained by the fact that, while making quantitative predictions seems at first glance to be facilitated by analytical solutions, this is actually often much easier with (...)
    Download  
     
    Export citation  
     
    Bookmark   8 citations  
  • Computer Simulations as Scientific Instruments.Ramón Alvarado - 2022 - Foundations of Science 27 (3):1183-1205.
    Computer simulations have conventionally been understood to be either extensions of formal methods such as mathematical models or as special cases of empirical practices such as experiments. Here, I argue that computer simulations are best understood as instruments. Understanding them as such can better elucidate their actual role as well as their potential epistemic standing in relation to science and other scientific methods, practices and devices.
    Download  
     
    Export citation  
     
    Bookmark   6 citations  
  • AI as an Epistemic Technology.Ramón Alvarado - 2023 - Science and Engineering Ethics 29 (5):1-30.
    In this paper I argue that Artificial Intelligence and the many data science methods associated with it, such as machine learning and large language models, are first and foremost epistemic technologies. In order to establish this claim, I first argue that epistemic technologies can be conceptually and practically distinguished from other technologies in virtue of what they are designed for, what they do and how they do it. I then proceed to show that unlike other kinds of technology (_including_ other (...)
    Download  
     
    Export citation  
     
    Bookmark   4 citations  
  • Interdisciplinarity in the Making: Models and Methods in Frontier Science.Nancy J. Nersessian - 2022 - Cambridge, MA: MIT.
    A cognitive ethnography of how bioengineering scientists create innovative modeling methods. In this first full-scale, long-term cognitive ethnography by a philosopher of science, Nancy J. Nersessian offers an account of how scientists at the interdisciplinary frontiers of bioengineering create novel problem-solving methods. Bioengineering scientists model complex dynamical biological systems using concepts, methods, materials, and other resources drawn primarily from engineering. They aim to understand these systems sufficiently to control or intervene in them. What Nersessian examines here is how cutting-edge bioengineering (...)
    Download  
     
    Export citation  
     
    Bookmark   1 citation  
  • Instruments, agents, and artificial intelligence: novel epistemic categories of reliability.Eamon Duede - 2022 - Synthese 200 (6):1-20.
    Deep learning (DL) has become increasingly central to science, primarily due to its capacity to quickly, efficiently, and accurately predict and classify phenomena of scientific interest. This paper seeks to understand the principles that underwrite scientists’ epistemic entitlement to rely on DL in the first place and argues that these principles are philosophically novel. The question of this paper is not whether scientists can be justified in trusting in the reliability of DL. While today’s artificial intelligence exhibits characteristics common to (...)
    Download  
     
    Export citation  
     
    Bookmark   4 citations  
  • Computational Modeling in Philosophy.Simon Scheller, Merdes Christoph & Stephan Hartmann (eds.) - 2022
    Computational modeling should play a central role in philosophy. In this introduction to our topical collection, we propose a small topology of computational modeling in philosophy in general, and show how the various contributions to our topical collection ft into this overall picture. On this basis, we describe some of the ways in which computational models from other disciplines have found their way into philosophy, and how the principles one found here still underlie current trends in the feld. Moreover, we (...)
    Download  
     
    Export citation  
     
    Bookmark  
  • The Cost of Prediction.Johannes Lenhard, Simon Stephan & Hans Hasse - manuscript
    This paper examines a looming reproducibility crisis in the core of the hard sciences. Namely, it concentrates on molecular modeling and simulation (MMS), a family of methods that predict properties of substances through computing interactions on a molecular level and that is widely popular in physics, chemistry, materials science, and engineering. The paper argues that in order to make quantitative predictions, sophisticated models are needed which have to be evaluated with complex simulation procedures that amalgamate theoretical, technological, and social factors (...)
    Download  
     
    Export citation  
     
    Bookmark  
  • Theorem proving in artificial neural networks: new frontiers in mathematical AI.Markus Pantsar - 2024 - European Journal for Philosophy of Science 14 (1):1-22.
    Computer assisted theorem proving is an increasingly important part of mathematical methodology, as well as a long-standing topic in artificial intelligence (AI) research. However, the current generation of theorem proving software have limited functioning in terms of providing new proofs. Importantly, they are not able to discriminate interesting theorems and proofs from trivial ones. In order for computers to develop further in theorem proving, there would need to be a radical change in how the software functions. Recently, machine learning results (...)
    Download  
     
    Export citation  
     
    Bookmark  
  • On the Opacity of Deep Neural Networks.Anders Søgaard - forthcoming - Canadian Journal of Philosophy:1-16.
    Deep neural networks are said to be opaque, impeding the development of safe and trustworthy artificial intelligence, but where this opacity stems from is less clear. What are the sufficient properties for neural network opacity? Here, I discuss five common properties of deep neural networks and two different kinds of opacity. Which of these properties are sufficient for what type of opacity? I show how each kind of opacity stems from only one of these five properties, and then discuss to (...)
    Download  
     
    Export citation  
     
    Bookmark  
  • AI and the need for justification (to the patient).Anantharaman Muralidharan, Julian Savulescu & G. Owen Schaefer - 2024 - Ethics and Information Technology 26 (1):1-12.
    This paper argues that one problem that besets black-box AI is that it lacks algorithmic justifiability. We argue that the norm of shared decision making in medical care presupposes that treatment decisions ought to be justifiable to the patient. Medical decisions are justifiable to the patient only if they are compatible with the patient’s values and preferences and the patient is able to see that this is so. Patient-directed justifiability is threatened by black-box AIs because the lack of rationale provided (...)
    Download  
     
    Export citation  
     
    Bookmark  
  • La deriva genética como fuerza evolutiva.Ariel Jonathan Roffé - 2015 - In J. Ahumada, N. Venturelli & S. Seno Chibeni (eds.), Selección de Trabajos del IX Encuentro AFHIC y las XXV Jornadas de Epistemología e Historia de la ciencia. pp. 615-626.
    Download  
     
    Export citation  
     
    Bookmark  
  • Models, Fictions and Artifacts.Tarja Knuuttila - 2021 - In Wenceslao J. Gonzalez (ed.), Language and Scientific Research. Springer Verlag. pp. 199-22.
    This paper discusses modeling from the artifactual perspective. The artifactual approach conceives models as erotetic devices. They are purpose-built systems of dependencies that are constrained in view of answering a pending scientific question, motivated by theoretical or empirical considerations. In treating models as artifacts, the artifactual approach is able to address the various languages of sciences that are overlooked by the traditional accounts that concentrate on the relationship of representation in an abstract and general manner. In contrast, the artifactual approach (...)
    Download  
     
    Export citation  
     
    Bookmark  
  • From Models to Simulations.Franck Varenne - 2018 - London, UK: Routledge.
    This book analyses the impact computerization has had on contemporary science and explains the origins, technical nature and epistemological consequences of the current decisive interplay between technology and science: an intertwining of formalism, computation, data acquisition, data and visualization and how these factors have led to the spread of simulation models since the 1950s. -/- Using historical, comparative and interpretative case studies from a range of disciplines, with a particular emphasis on the case of plant studies, the author shows how (...)
    Download  
     
    Export citation  
     
    Bookmark   15 citations  
  • Homepage Eckhart Arnold.Eckhart Arnold (ed.) - 2001 - Munich: Preprint.
    This is my personal homepage. Find my philosophical papers under "Philosophy".
    Download  
     
    Export citation  
     
    Bookmark  
  • Scientific Exploration and Explainable Artificial Intelligence.Carlos Zednik & Hannes Boelsen - 2022 - Minds and Machines 32 (1):219-239.
    Models developed using machine learning are increasingly prevalent in scientific research. At the same time, these models are notoriously opaque. Explainable AI aims to mitigate the impact of opacity by rendering opaque models transparent. More than being just the solution to a problem, however, Explainable AI can also play an invaluable role in scientific exploration. This paper describes how post-hoc analytic techniques from Explainable AI can be used to refine target phenomena in medical science, to identify starting points for future (...)
    Download  
     
    Export citation  
     
    Bookmark   9 citations  
  • Solving the Black Box Problem: A Normative Framework for Explainable Artificial Intelligence.Carlos Zednik - 2019 - Philosophy and Technology 34 (2):265-288.
    Many of the computing systems programmed using Machine Learning are opaque: it is difficult to know why they do what they do or how they work. Explainable Artificial Intelligence aims to develop analytic techniques that render opaque computing systems transparent, but lacks a normative framework with which to evaluate these techniques’ explanatory successes. The aim of the present discussion is to develop such a framework, paying particular attention to different stakeholders’ distinct explanatory requirements. Building on an analysis of “opacity” from (...)
    Download  
     
    Export citation  
     
    Bookmark   53 citations  
  • Two Exploratory Uses for General Circulation Models in Climate Science.Joseph Wilson - 2021 - Perspectives on Science 29 (4):493-509.
    . In this paper I present two ways in which climate modelers use general circulation models for exploratory purposes. The complexity of Earth’s climate system makes it difficult to predict precisely how lower-order climate dynamics will interact over time to drive higher-order dynamics. The same issues arise for complex models built to simulate climate behavior like the Community Earth Systems Model. I argue that as a result of system complexity, climate modelers use general circulation models to perform model dynamic exploration (...)
    Download  
     
    Export citation  
     
    Bookmark   2 citations  
  • Transparency and the Black Box Problem: Why We Do Not Trust AI.Warren J. von Eschenbach - 2021 - Philosophy and Technology 34 (4):1607-1622.
    With automation of routine decisions coupled with more intricate and complex information architecture operating this automation, concerns are increasing about the trustworthiness of these systems. These concerns are exacerbated by a class of artificial intelligence that uses deep learning, an algorithmic system of deep neural networks, which on the whole remain opaque or hidden from human comprehension. This situation is commonly referred to as the black box problem in AI. Without understanding how AI reaches its conclusions, it is an open (...)
    Download  
     
    Export citation  
     
    Bookmark   12 citations  
  • Markets, market algorithms, and algorithmic bias.Philippe van Basshuysen - 2022 - Journal of Economic Methodology 30 (4):310-321.
    Where economists previously viewed the market as arising from a ‘spontaneous order’, antithetical to design, they now design markets to achieve specific purposes. This paper reconstructs how this change in what markets are and can do came about and considers some consequences. Two decisive developments in economic theory are identified: first, Hurwicz’s view of institutions as mechanisms, which should be designed to align incentives with social goals; and second, the notion of marketplaces – consisting of infrastructure and algorithms – which (...)
    Download  
     
    Export citation  
     
    Bookmark  
  • Can we trust Big Data? Applying philosophy of science to software.John Symons & Ramón Alvarado - 2016 - Big Data and Society 3 (2).
    We address some of the epistemological challenges highlighted by the Critical Data Studies literature by reference to some of the key debates in the philosophy of science concerning computational modeling and simulation. We provide a brief overview of these debates focusing particularly on what Paul Humphreys calls epistemic opacity. We argue that debates in Critical Data Studies and philosophy of science have neglected the problem of error management and error detection. This is an especially important feature of the epistemology of (...)
    Download  
     
    Export citation  
     
    Bookmark   15 citations  
  • Epistemic Entitlements and the Practice of Computer Simulation.John Symons & Ramón Alvarado - 2019 - Minds and Machines 29 (1):37-60.
    What does it mean to trust the results of a computer simulation? This paper argues that trust in simulations should be grounded in empirical evidence, good engineering practice, and established theoretical principles. Without these constraints, computer simulation risks becoming little more than speculation. We argue against two prominent positions in the epistemology of computer simulation and defend a conservative view that emphasizes the difference between the norms governing scientific investigation and those governing ordinary epistemic practices.
    Download  
     
    Export citation  
     
    Bookmark   19 citations  
  • Epistemic injustice and data science technologies.John Symons & Ramón Alvarado - 2022 - Synthese 200 (2):1-26.
    Technologies that deploy data science methods are liable to result in epistemic harms involving the diminution of individuals with respect to their standing as knowers or their credibility as sources of testimony. Not all harms of this kind are unjust but when they are we ought to try to prevent or correct them. Epistemically unjust harms will typically intersect with other more familiar and well-studied kinds of harm that result from the design, development, and use of data science technologies. However, (...)
    Download  
     
    Export citation  
     
    Bookmark   12 citations  
  • Cultures of simulations vs. cultures of calculations? The development of simulation practices in meteorology and astrophysics.Mikaela Sundberg - 2010 - Studies in History and Philosophy of Science Part B: Studies in History and Philosophy of Modern Physics 41 (3):273-281.
    Download  
     
    Export citation  
     
    Bookmark   6 citations  
  • Cultures of simulations vs. cultures of calculations? The development of simulation practices in meteorology and astrophysics.Mikaela Sundberg - 2010 - Studies in History and Philosophy of Science Part B: Studies in History and Philosophy of Modern Physics 41 (3):273-281.
    Download  
     
    Export citation  
     
    Bookmark   5 citations  
  • Creating Convincing Simulations in Astrophysics. [REVIEW]Mikaela Sundberg - 2012 - Science, Technology, and Human Values 37 (1):64-87.
    Numerical simulations have come to be widely used in scientific work. Like experiments, simulations generate large quantities of numbers that require analysis and constant concern with uncertainty and error. How do simulationists convince themselves, and others, about the credibility of output? The present analysis reconstructs the perspectives related to performing numerical simulations, in general, and the situations in which simulationists deal with uncertain output, in particular. Starting from a distinction between idealized and realistic simulations, the paper presents the principal methods (...)
    Download  
     
    Export citation  
     
    Bookmark   2 citations  
  • Understanding from Machine Learning Models.Emily Sullivan - 2022 - British Journal for the Philosophy of Science 73 (1):109-133.
    Simple idealized models seem to provide more understanding than opaque, complex, and hyper-realistic models. However, an increasing number of scientists are going in the opposite direction by utilizing opaque machine learning models to make predictions and draw inferences, suggesting that scientists are opting for models that have less potential for understanding. Are scientists trading understanding for some other epistemic or pragmatic good when they choose a machine learning model? Or are the assumptions behind why minimal models provide understanding misguided? In (...)
    Download  
     
    Export citation  
     
    Bookmark   48 citations  
  • Peeking Inside the Black Box: A New Kind of Scientific Visualization.Michael T. Stuart & Nancy J. Nersessian - 2018 - Minds and Machines 29 (1):87-107.
    Computational systems biologists create and manipulate computational models of biological systems, but they do not always have straightforward epistemic access to the content and behavioural profile of such models because of their length, coding idiosyncrasies, and formal complexity. This creates difficulties both for modellers in their research groups and for their bioscience collaborators who rely on these models. In this paper we introduce a new kind of visualization that was developed to address just this sort of epistemic opacity. The visualization (...)
    Download  
     
    Export citation  
     
    Bookmark   7 citations  
  • Agnostic Science. Towards a Philosophy of Data Analysis.D. C. Struppa - 2011 - Foundations of Science 16 (1):1-20.
    In this paper we will offer a few examples to illustrate the orientation of contemporary research in data analysis and we will investigate the corresponding role of mathematics. We argue that the modus operandi of data analysis is implicitly based on the belief that if we have collected enough and sufficiently diverse data, we will be able to answer most relevant questions concerning the phenomenon itself. This is a methodological paradigm strongly related, but not limited to, biology, and we label (...)
    Download  
     
    Export citation  
     
    Bookmark   9 citations  
  • Protecting rainforest realism: James Ladyman, Don Ross: Everything must go: metaphysics naturalized, Oxford: Oxford University Press, 2007, pp. 368 £49.00 HB.P. Kyle Stanford, Paul Humphreys, Katherine Hawley, James Ladyman & Don Ross - 2010 - Metascience 19 (2):161-185.
    Reply in Book Symposium on James Ladyman, Don Ross: 'Everything must go: metaphysics naturalized', Oxford: Oxford University Press, 2007.
    Download  
     
    Export citation  
     
    Bookmark   7 citations  
  • The Automated Laplacean Demon: How ML Challenges Our Views on Prediction and Explanation.Sanja Srećković, Andrea Berber & Nenad Filipović - 2021 - Minds and Machines 32 (1):159-183.
    Certain characteristics make machine learning a powerful tool for processing large amounts of data, and also particularly unsuitable for explanatory purposes. There are worries that its increasing use in science may sideline the explanatory goals of research. We analyze the key characteristics of ML that might have implications for the future directions in scientific research: epistemic opacity and the ‘theory-agnostic’ modeling. These characteristics are further analyzed in a comparison of ML with the traditional statistical methods, in order to demonstrate what (...)
    Download  
     
    Export citation  
     
    Bookmark   4 citations  
  • Validating the Universe in a Box.Chris Smeenk & Sarah C. Gallagher - 2020 - Philosophy of Science 87 (5):1221-1233.
    Computer simulations of the formation and evolution of large-scale structure in the universe are integral to the enterprise of modern cosmology. Establishing the reliability of these simulations ha...
    Download  
     
    Export citation  
     
    Bookmark   2 citations  
  • Metaphysics within Chemical Physics: The Case of Ab Initio Molecular Dynamics. [REVIEW]Carsten Seck - 2012 - Journal for General Philosophy of Science / Zeitschrift für Allgemeine Wissenschaftstheorie 43 (2):361-375.
    This paper combines naturalized metaphysics and a philosophical reflection on a recently evolving interdisciplinary branch of quantum chemistry, ab initio molecular dynamics. Bridging the gaps among chemistry, physics, and computer science, this cutting-edge research field explores the structure and dynamics of complex molecular many-body systems through computer simulations. These simulations are allegedly crafted solely by the laws of fundamental physics, and are explicitly designed to capture nature as closely as possible. The models and algorithms employed, however, involve many approximations and (...)
    Download  
     
    Export citation  
     
    Bookmark   3 citations  
  • Dealing with Molecular Complexity. Atomistic Computer Simulations and Scientific Explanation.Julie Schweer & Marcus Elstner - 2023 - Perspectives on Science 31 (5):594-626.
    Explanation is commonly considered one of the central goals of science. Although computer simulations have become an important tool in many scientific areas, various philosophical concerns indicate that their explanatory power requires further scrutiny. We examine a case study in which atomistic simulations have been used to examine the factors responsible for the transport selectivity of certain channel proteins located at cell membranes. By elucidating how precisely atomistic simulations helped scientists draw inferences about the molecular system under investigation, we respond (...)
    Download  
     
    Export citation  
     
    Bookmark  
  • Computational modeling in philosophy: introduction to a topical collection.Simon Scheller, Christoph Merdes & Stephan Hartmann - 2022 - Synthese 200 (2):1-10.
    Computational modeling should play a central role in philosophy. In this introduction to our topical collection, we propose a small topology of computational modeling in philosophy in general, and show how the various contributions to our topical collection fit into this overall picture. On this basis, we describe some of the ways in which computational models from other disciplines have found their way into philosophy, and how the principles one found here still underlie current trends in the field. Moreover, we (...)
    Download  
     
    Export citation  
     
    Bookmark  
  • A practical philosophy of complex climate modelling.Gavin A. Schmidt & Steven Sherwood - 2015 - European Journal for Philosophy of Science 5 (2):149-169.
    We give an overview of the practice of developing and using complex climate models, as seen from experiences in a major climate modelling center and through participation in the Coupled Model Intercomparison Project. We discuss the construction and calibration of models; their evaluation, especially through use of out-of-sample tests; and their exploitation in multi-model ensembles to identify biases and make predictions. We stress that adequacy or utility of climate models is best assessed via their skill against more naïve predictions. The (...)
    Download  
     
    Export citation  
     
    Bookmark   8 citations  
  • Model Organisms as Scientific Representations.Lorenzo Sartori - forthcoming - British Journal for the Philosophy of Science.
    Download  
     
    Export citation  
     
    Bookmark  
  • What is a Computer Simulation? A Review of a Passionate Debate.Nicole J. Saam - 2017 - Journal for General Philosophy of Science / Zeitschrift für Allgemeine Wissenschaftstheorie 48 (2):293-309.
    Where should computer simulations be located on the ‘usual methodological map’ which distinguishes experiment from theory? Specifically, do simulations ultimately qualify as experiments or as thought experiments? Ever since Galison raised that question, a passionate debate has developed, pushing many issues to the forefront of discussions concerning the epistemology and methodology of computer simulation. This review article illuminates the positions in that debate, evaluates the discourse and gives an outlook on questions that have not yet been addressed.
    Download  
     
    Export citation  
     
    Bookmark   7 citations  
  • Connecting ethics and epistemology of AI.Federica Russo, Eric Schliesser & Jean Wagemans - forthcoming - AI and Society:1-19.
    The need for fair and just AI is often related to the possibility of understanding AI itself, in other words, of turning an opaque box into a glass box, as inspectable as possible. Transparency and explainability, however, pertain to the technical domain and to philosophy of science, thus leaving the ethics and epistemology of AI largely disconnected. To remedy this, we propose an integrated approach premised on the idea that a glass-box epistemology should explicitly consider how to incorporate values and (...)
    Download  
     
    Export citation  
     
    Bookmark   1 citation