Workshop: Machine Learning meets Scientific Understanding

Date: 26.06.25, 27.06.25

Place: TU Dortmund University, Internationales Begegnungszentrum

Poster

Machine Learning meets Scientific Understanding

The Emmy Noether group UDNN: Scientific Understanding and Deep Neural Networks is pleased to announce its upcoming workshop Machine Learning Meets Scientific Understanding, which will take place on June 26th and 27th at TU Dortmund University. This interdisciplinary event seeks to bring together philosophers of science and machine learning (ML), as well as ML practitioners, to explore the intersections between ML and scientific understanding. 

Confirmed speakers 

Cameron Buckner, Philosophy (University of Florida): Predictively-valid “Alien” Features, or Artifacts? Coping with Inscrutable Scientific Progress 

Abstract: Systems like AlphaFold raise the prospect of predictive AI systems that can blow past previous upper bounds on the performance of hand-designed analytical models in many areas of scientific analysis. It is difficult to disagree with the results of these systems, which can achieve predictive accuracy on problems that were thought to be too complex or chaotic for human scientific theory to solve. However, these models may base their predictions on features that are in some sense beyond the cognitive grasp of humans–„alien“ properties that may have predictive utility but which are not natural or cognitively accessible to us. In this talk I will analyze these properties by beginning with a discussion of adversarial attacks, and discuss methods for coping with this epistemic situation in a scientific regime which increasingly relies on complex deep learning models for data analysis. 

Heather Champion, Philosophy (Tübingen University): On scientific discovery with machine learning: what is „strong“ novelty?

Abstract: Recent philosophical accounts of machine learning’s (ML) impact negatively appraise the significance of its contribution to changing scientific theory, understanding, or concepts. While valuable, these analyses tend to focus narrowly on one kind of strong disruption claim owing to one notion of “strong” novelty, which is not always clarified. Meanwhile, some philosophers assess whether the capacities of ML algorithms are sufficient to cause one of these disruptions (e.g. whether they utilize creative processes). But what exactly constitutes strongly novel outcomes of ML-enabled science? Omitting a multifaceted answer to this question risks overemphasizing the non-disruptiveness of ML. Also, while the analysis of novel outcomes and the means that successfully achieve them are inevitably linked, outcome desiderata play an important role in evaluating human-computer interactions. Therefore, I focus on outcomes enabled by ML, such as predictions, ideas, or virtual artifacts. I first raise three difficulties with Ratti (2020) and Boge’s (2022) outcome-based characterizations of strong novelty: (1) Ratti’s domain-specific focus is unnecessary, (2) both underappreciate the scientific impact of token predictions, and (3) Boge is ambiguous about the kind of prior information he takes to be disqualifying for “use novelty,” while I argue that on several interpretations, use novelty does not helpfully discriminate strong novelty. However, their accounts capture useful intuitions: changes to existing theory, scientific knowledge, or research direction are highly impacting, as are many outcomes achieved without a certain kind of informational bias (elaborated below).  

Next, I introduce a new, wide, variety of outcome-based notions of strong novelty from philosophy of creativity, epistemology, and philosophy of science. I illustrate these with cases from various scientific domains, such as economics and astrophysics. First, a creative outcome generates surprise when an idea with low expected utility turns out to be useful. Alternatively, outcomes that reduce uncertainty (“blindness”) regarding an idea’s utility helpfully steer research. These notions of belief revision assume a state of awareness regarding a proposition, but ML might also generate this awareness, eliminating deep ignorance regarding scientifically useful patterns, evidence, or hypotheses. Zooming out from these notions of local epistemic change, ML might make broader impact by prompting conceptual change. Particularly, if deep learning methods directly learn conceptualizations useful for specific tasks, these might diverge from existing human conceptualizations. Finally, using ML to learn from data with some independence of local theory regarding a target phenomenon has generative power for scientific change. I define local theory as theory that demarks or explains a target phenomenon. This “bottom-up” form of learning constitutes strong novelty for science because it signals an aim to find a new research direction, often by relying on a different set of cognitive tools for analyzing multidimensional data. It also clarifies that local theory is the kind of prior information that diminishes the generative impact of an ML prediction. My taxonomy clarifies desiderata for scientific exploration with ML and complements assessments of what algorithmic processes might achieve them. It also invites reflection on how some forms of novelty might co-occur and what problems this raises for scientific understanding.  

Boge, Florian J. “Two Dimensions of Opacity and the Deep Learning Predicament.” Minds and Machines 32 (2022): 43–75. https://doi.org/10.1007/s11023-021-09569-4.  

Ratti, Emanuele. “What Kind of Novelties Can Machine Learning Possibly Generate? The Case of Genomics.” Studies in History and Philosophy of Science Part A 83 (2020): 86–96. https://doi.org/10.1016/j.shpsa.2020.04.001

Edward Chang, Computer Science (Stanford University): From Generative AI to AGI: Multi-LLM Agent Collaboration as a Path Forward 

Abstract: The rise of large language models (LLMs) has transformed AI—shifting it from passive analysis to generative capabilities, from narrow task-specific tools to general-purpose systems, and from monolithic models to collaborative multi-agent frameworks. While some experts anticipate the emergence of Artificial General Intelligence (AGI) by 2040, critics like LeCun (2023) argue that LLMs cannot lead to AGI, citing their lack of world models, persistent memory, structured reasoning, and planning capabilities. Critics also highlight how LLMs require massive training data yet still fail to match the efficient few-shot learning demonstrated by even young children. 

This talk challenges these critiques by positioning LLMs not as complete solutions, but as necessary substrates for AGI emergence, analogous to how unconscious processes enable conscious reasoning in humans. Just as humans aren’t born with blank slates but with neural priors that scaffold learning, LLMs provide foundational capabilities for in-context learning and environmental adaptation. By augmenting LLMs with transactional reliability, self-validation mechanisms, Socratic reasoning, and multi-agent architectures, we can address their current limitations. The proposed Multi-LLM Agent Collaboration framework offers a pragmatic, scalable path toward AGI, where intelligence emerges not from a single model but through structured interaction, persistent memory, and collective reasoning across networked systems.  

 

Finnur Dellsén, Philosophy (University of Iceland): Scientific Progress in the Age of AI

Abstract: What role does artificial intelligence (AI) play – and what role might it play – in scientific progress? Are AI systems best understood as tools for accelerating the scientific progress made by human researchers, or are they capable of making scientific progress in their own right? Could AI systems even become autonomous scientific agents one day, capable of generating, testing, evaluating, and communicating scientific hypotheses in a way that produces scientific progress? And do recent advances in AI research constitute genuine scientific progress, or are these developments better understood as technological advances? This talk explores the relationship between AI and scientific progress through these and related questions, proposing new directions for future research. It explores both how the use of AI in science aligns with or challenges existing accounts of scientific progress, and how the philosophical debate about these accounts sheds light on the value of AI in science. 

Henk W. de Regt, Philosophy (Radboud University) and Eugene Shalugin, Philosophy (Radboud University): Bridging Scientific Understanding and Creativity with an LLM Benchmark for Narrow-Domain Scientific Fields

Abstract: The rapid advancement of large language models (LLMs) raises questions about their potential to understand complex scientific domains and contribute to scientific discovery. It has been argued that LLMs are ‘stochastic parrots’ that make predictions on the basis of large training data sets and are therefore incapable of genuine understanding and creativity [2]. However, such claims presuppose certain philosophical conceptions of understanding and creativity, which remain unexamined. We claim, by contrast, that current philosophical insights into the nature of scientific understanding and creativity allow for the possibility of scientific understanding and creativity with LLMs – at least in a restricted sense. Nonetheless, a benchmark for evaluating such understanding and creativity has not yet been developed. The aim of our paper is to fill this gap by developing a semi-automated LLM benchmark creation pipeline for narrow-domain scientific fields for these scientific capabilities, utilizing both human experts and LLMs. With regard to understanding, we adopt Barman et al.’s [1] behavioural conception, which implies that an agent’s understanding consists in its ability to process, integrate, and apply knowledge, also in unseen scenarios, beyond mere factual retrieval. In particular, we focus on assessing the type of questions an agent can answer. With regard to creativity, we follow Boden [3] and call a scientific product creative if it is valuable, novel, and surprising. We propose a pipeline for the semi-automated creation of questions to evaluate LLMs‘ understanding and creativity in narrow-domain scientific fields. We focus on creating what-, why-, and w-questions (counterfactuals). The questions are generated using human-provided texts with trusted information (e.g. lecture notes, scientific papers, textbooks). Parts of the set of documents are then provided to the LLMs as context with Retrieval Augmented Generation (RAG) and the LLM is tasked with generating questions based on the documents. Domain experts are then invited to review the validity and sensibility of the questions. The resulting question-answer pairs form a high-quality narrow-domain benchmark. We use the field of particle physics as a case study and introduce the platform www.physicsbenchmark.org for generation and curation of high-quality questions. Using this dataset of factual, explanatory, and counterfactual questions, we evaluate how well state-of-the-art LLMs understand particle physics. Contextual grounding is toggled via access to the corpus. We then propose a scientific creativity benchmark (SCB) that challenges LLMs with questions whose answers lie outside their training data. These questions are either unprecedented counterfactual queries or questions about scientific literature published after 2023 (the cutoff date of the tested models). By formulating prompts that fall beyond the statistical patterns the model has learned, we induce a form of novelty: if the model provides a satisfactory answer to, say, a frontier particle physics question, it must have performed non-trivial logical transformations on its outdated information. This deviation serves as an epistemically surprising indicator of creative output, meeting our criteria for responses that are not only novel and unexpected but also valuable. (N.B.: we do not argue that LLMs are creative agents but that they can generate creative products.)  

[1] K. G. Barman, S. Caron, T. Claassen, and H.W. de Regt, “Towards a Benchmark for Scientific Understanding in Humans and Machines,” Minds and Machines, vol. 34, no. 1, 2024, doi: 10.1007/s11023-024-09657-1.  

[2] E. M. Bender, T. Gebru, A. McMillan-Major, and S. Shmitchell, “On the Dangers of Stochastic Parrots: Can Language Models Be Too Big?,” in Proceedings of the 2021 ACM Conference on Fairness, Accountability, and Transparency, Mar. 2021, pp. 610–623. doi: 10.1145/3442188.3445922.  

[3] M. A. Boden, The Creative Mind: Myths and Mechanisms. Psychology Press, 2004 

Timo Freiesleben, Philosophy (Tübingen University): The Benchmarking Epistemology – What Inferences Can Scientists Draw from Competitive Comparisons of Prediction Models?

Abstract: What inferences can scientists draw from competitive comparisons of prediction models? Benchmarking, the practice of evaluating machine learning (ML) models based on their predictive performance on test datasets and ranking them against competitors, is a cornerstone of ML research. Often referred to as the “iron rule” of machine learning, benchmarking comprises four key components: (1) prediction tasks, defining the target variables and predictors; (2) evaluation metrics, determining what constitutes good predictions; (3) datasets, including a training set for model development and a test set for performance assessment; and (4) leaderboards, which rank models based on their test set performance. As ML becomes a widespread tool in the natural and social sciences, benchmarking becomes a common evaluation standard in scientific practice (Kitchin, 2014; Mussgnug, 2022; Pankowska et al., 2023). Despite its enormous practical significance, benchmarking has drawn surprisingly little attention in the philosophy of science community, which has primarily focused on inductive inferences in machine learning  (Sterkenburg and Gr ̈unwald, 2021; Karaca, 2021), opacity (Creel, 2020; Boge, 2022; Sullivan, 2022), and explainable artificial intelligence (Zednik and Boelsen, 2022; Freiesleben et al., 2024). This paper addresses the gap, arguing that benchmarking constitutes a scientific epistemology in its own right, offering a powerful framework for scientific inference. We examine four types of inferences that scientists employ benchmarking for: 1. Identifying the (current) best model on task T. 2. Determining the (current) best learning algorithm for tasks similar to T. 3. Selecting the (current) most suitable model for deployment in a specific application. 4. Estimating the optimal predictability of a target Y given features X. An important insight is that none of these inferences can be drawn from the benchmark result alone, they all require additional assumptions to be valid. Similar to the inferences we draw from psychological tests, we must ensure construct validity (AERA et al., 2014), which means that the additional assumptions must both be specified and justified by additional (sources of) evidence (Messick, 1995). Straightforward in its abstract generality, we analyse three case studies from different research areas to specify what these assumptions amount to for the four inferences above. The case studies are the Imagenet benchmark for image recognition (Deng et al., 2009), the Fragile Families benchmark for predicting life outcomes (Salganik et al., 2019), and the WeatherBench benchmark for global weather forecasting (Rasp et al., 2020, 2024). However, benchmarks play not only an epistemic role but also have an important social role in organizing scientific communities around a shared goal. We point out that these social roles potentially threaten the validity of the scientific inferences drawn by benchmarks. For example, benchmarks are often used iteratively for model validation; a practice that undermines the assumption that the test data is unseen, which is essential for benchmark arguments (Grote et al., 2024; Hardt and Recht, 2022). Moreover, we point out that benchmarks are often used as a standard of scientific significance in peer review processes. This incentivises scientists to fiddle with their results, similar to well-known problems with p-values (Gelman and Loken, 2014; Bzdok et al., 2018).

AERA, APA, and NCME. 2014. Standards for educational and psychological testing. American Educational Research Association (AERA), the American Psychological Association (APA) and the National Council on Measurement in Education (NCME). 

Boge, F.J. 2022. Two dimensions of opacity and the deep learning predicament. Minds and Machines 32(1): 43–75 .  

Bzdok, D., N. Altman, and M. Krzywinski. 2018. Statistics versus machine learning. Nature Methods 15: 233–234 . Creel, K.A. 2020. Transparency in complex computational systems. Philosophy of Science 87(4): 568–589 .  

Deng, J., W. Dong, R. Socher, L.J. Li, K. Li, and L. Fei-Fei 2009. Imagenet: A large-scale hierarchical image database. In 2009 IEEE conference on computer vision and pattern recognition, pp. 248–255. Ieee.  

Freiesleben, T., G. Königg, C. Molnar, and ́A. Tejero-Cantero. 2024. Scientific inference with interpretable machine learning: Analyzing models to learn about real-world phenomena. Minds and Machines 34(3): 32 . 3  

Gelman, A. and E. Loken. 2014. The statistical crisis in science. American scientist 102(6): 460–465 .  

Grote, T., K. Genin, and E. Sullivan. 2024, April. Reliability in Machine Learning. Philosophy Compass 19(5). https://doi.org/10.1111/phc3.12974 .  

Hardt, M. and B. Recht. 2022. Patterns, Predictions, and Actions: Foundations of Machine Learning. Princeton University Press.  

Karaca, K. 2021. Values and inductive risk in machine learning modelling: the case of binary classification models. European Journal for Philosophy of Science 11(4): 102 . 

Kitchin, R. 2014. Big data, new epistemologies and paradigm shifts. Big data & society 1(1): 2053951714528481 . Messick, S. 1995. Standards of validity and the validity of standards in per- formance asessment. Educational measurement: Issues and practice 14(4): 5–8 . 

Mussgnug, A.M. 2022. The predictive reframing of machine learning appli- cations: good predictions and bad measurements. European Journal for Philosophy of Science 12(3): 55. 

Pankowska, P., A. Mendrik, T. Emery, and J. Garcia-Bernardo. 2023, Septem- ber. Accelerating progress in the social sciences: the potential of benchmarks. https://doi.org/10.31235/osf.io/ekfxy .  

Rasp, S., P.D. Dueben, S. Scher, J.A. Weyn, S. Mouatadid, and N. Thuerey. 2020. Weatherbench: a benchmark data set for data-driven weather forecasting. Journal of Advances in Modeling Earth Systems 12(11): e2020MS002203 .  

Rasp, S., S. Hoyer, A. Merose, I. Langmore, P. Battaglia, T. Russell, A. Sanchez-Gonzalez, V. Yang, R. Carver, S. Agrawal, et al. 2024. Weath- erbench 2: A benchmark for the next generation of data-driven global weather models. Journal of Advances in Modeling Earth Systems 16(6): e2023MS004019 .  

Salganik, M.J., I. Lundberg, A.T. Kindel, and S. McLanahan. 2019, January. Introduction to the Special Collection on the Fragile Families Challenge. Socius: Sociological Research for a Dynamic World 5: 237802311987158. https://doi.org/10.1177/2378023119871580 .  

Sterkenburg, T.F. and P.D. Gr ̈unwald. 2021. The no-free-lunch theorems of supervised learning. Synthese 199(3): 9979–10015 Sullivan, E. 2022. Understanding from machine learning models. The British Journal for the Philosophy of Science .  

Zednik, C. and H. Boelsen. 2022. Scientific exploration and explainable artificial intelligence. Minds and Machines 32(1): 219–239.  

Giovanni Galli, Philosophy (University of Teramo): Deep-learning Models and Scientific Understanding through Explanations and Representations

Abstract: In the rapidly evolving landscape of artificial intelligence, the understandability and explainability of AI systems have become crucial concerns. As AI models grow increasingly complex, they often operate as “black boxes”, making decisions without explaining their processes clearly. This opacity can hinder trust, accountability, and ethical compliance, particularly in critical domains such as healthcare, finance, law, and scientific research. Still, deep-learning models (DLMs) are powerful tools in order to understand phenomena, as recognised by Páez (2019), Sullivan (2022), Fleisher (2022), Jumper (2021a) and Abramson et al. (2024). Thus, on the one hand, Explainable Artificial Intelligence (XAI) aims to answer the first issue about the opacity of the DLMs, offering us ways to understand the DLMs; on the other hand, the kind of understanding gained from DLMs leads us to re-define what scientific understanding is. According to Sullivan (2022), the lack of understanding of DLMs does not limit our scientific understanding of phenomena. She argues that when we fail to achieve understanding with DLMs, it is not due to the lack of understanding of the DLMs in question but to the “link uncertainty”, i.e. the lack of evidence, knowledge and understanding of how the model and its target-system are related. On the opposite side, Räz and Beisbart (2022) argue that due to the lack of understanding of DLMs, we may fail to understand a phenomenon scientifically through the use of the models. Along their line, Durán (2021) claims that what we gain from DLMs is not a genuine understanding of the phenomena and that XAI’s explanations are better defined as classifications. In this paper, we first argue that a machine can explain and that some XAI explanations are rule-based. We defend the idea that if specific XAI explanations can capture the rules underlying the scrutinised phenomenon, they are genuinely scientific explanations. Second, we claim that, given understanding as a noetic-mediated state, DLMs play the role of noetic mediators for scientific understanding, even if they present essential differences from other traditionally well-suited mediators, such as explanations, theories, and non-artificial models. Moreover, we highlight a crucial distinction when we speak of scientific understanding with DLMs and with other models and theories. De Regt (2017) and Khalifa (2013, 2017) defend that scientific understanding (SU) is gained via explanatory information about the phenomenon under scrutiny. However, when scientists use DLMs to study and understand a phenomenon they cannot access all the relevant explanatory information. We present the case study of AlphaFold’s DLMs (Jumper et al., 2021a, 2021b) to propose another form of SU, complementary to the explanatory one, namely representational understanding (Galli, 2023). We then present the features of representational and explanatory scientific understanding involved in scientific research with DLMs like AlphaFold’s models. In conclusion, we outline the differences between representational and explanatory understanding in light of the explanations provided by XAI methods. 

Insa Lawler, Philosophy (UNC Greensboro): The Gradability of Explanatory Understanding 

Abstract: It is a common place that explanatory understanding comes in degrees. Some people have more understanding of a subject matter than others or a greater degree of understanding. Its gradability is claimed to be one feature that sets apart understanding from knowledge. But what precisely does it mean that understanding comes in degrees or is gradable? In my talk, I explore how the gradability of understanding can be analyzed, drawing on insights from epistemology, formal semantics, metaphysics, and philosophy of science. 

Holger Lyre, Philosophy (Magdeburg University): Semantic Grounding in Advanced LLMs 


Abstract: ML models serve as scientific tools and help to deliver our scientific understanding in much the same way as models in general. However, ML models are special in a certain respect: namely, when these models themselves acquire a semantic grounding, when they start to become cognitive and thus possess a form of understanding themselves. Obviously, semantically grounded models will be far more powerful than ungrounded models, they could potentially achieve the status of scientific partners rather than tools. In my talk, I will explore the question of whether advanced LLMs already show signs of semantic grounding, and I will argue that this is indeed the case. To assess the question of semantic grounding, five methodological ways will be distinguished. The most promising way, I claim, is to apply core assumptions of theories of meaning in philosophy of mind and language to LLMs. I will demonstrate that grounding proves to be a gradual affair with a three-dimensional distinction between functional, social and causal grounding. Modern LLMs show basic evidence in all three dimensions. A strong argument is that they develop world models. Hence, LLMs are no stochastic parrots, but already understand the language they generate, at least in an elementary sense.

Abstract: ML models serve as scientific tools and help to deliver our scientific understanding in much the same way as models in general. However, ML models are special in a certain respect: namely, when these models themselves acquire a semantic grounding, when they start to become cognitive and thus possess a form of understanding themselves. Obviously, semantically grounded models will be far more powerful than ungrounded models, they could potentially achieve the status of scientific partners rather than tools. In my talk, I will explore the question of whether advanced LLMs already show signs of semantic grounding, and I will argue that this is indeed the case. To assess the question of semantic grounding, five methodological ways will be distinguished. The most promising way, I claim, is to apply core assumptions of theories of meaning in philosophy of mind and language to LLMs. I will demonstrate that grounding proves to be a gradual affair with a three-dimensional distinction between functional, social and causal grounding. Modern LLMs show basic evidence in all three dimensions. A strong argument is that they develop world models. Hence, LLMs are no stochastic parrots, but already understand the language they generate, at least in an elementary sense.  

Daniel Neider, Computer Science (TU Dortmund University): A Gentle Introduction to Neural Network Verification (and How It Might Contribute to Evaluating Scientific Insights)

Abstract: The increasing use of artificial intelligence in safety-critical domains such as autonomous systems and healthcare demands robust methods to ensure the reliability and safety of these technologies. Neural network verification has emerged as a vital research area, providing algorithmic frameworks to rigorously analyze and guarantee critical properties of neural networks across diverse applications. This talk provides an accessible introduction to this critical field, with a specific emphasis on safety-critical properties of neural networks and the algorithmic frameworks designed to automatically verify these properties. Furthermore, we briefly discuss how verification might contribute to evaluating scientific insights obtained through machine learning, inviting exploration of its potential role in advancing scientific understanding.

Sara Pernille Jensen, Philosophy (Oslo University): The Underdetermination of Representational Content in DNNs

Abstract: There is widespread hope of using ML models to make new scientific discoveries. As part of this endeavour, much effort is being put into establishing methods for interpreting the learned basis vectors in the latent spaces of deep neural networks (DNNs) (Räz 2023; Boge 2024), motivated by the belief that the networks implicitly learn scientifically relevant representations or concepts from the data (Buckner 2020; Bau et al. 2020). By studying these learned representations, we may learn about new dependencies and structures in nature. However, there is disagreement regarding how concepts are represented in the hidden layers, specifically whether they are localized or distributed across nodes, and whether they are linear or non-linear. Here, I argue that for distributed representations, linear or not, the conceptual content of the representations will often be underdetermined. For classical scientific representations, one can unambiguously tell whether a concept has been represented or not, since the concepts represented are e.g. those explicitly symbolized in an equation. Such representations, with their explicit conceptual content, stand in contrast with mere informational content. For although the information about some derived property is present in the original representation, the derived property is not thereby itself represented. My worry is whether the different accounts of representations in DNNs lead to any detectable differences between informational and conceptual content in the representations. I assume Harding’s operationalized definition of representations in DNNs, requiring a representation to carry informational content about its target, that the later layers of the network use the representation, and that it comes with a possibility of misrepresentation (Harding 2023). Local representations are unproblematic, since each node is dependent on a single variable, so only represents the concept corresponding to that variable. The problem arises for both linear and non-linear distributed representations, where any compound variable derivable from (non-)linear transformations of a set of activations in a given layer may be represented. Here, the conceptual content will be underdetermined in cases where the concepts include sets of variables which are defined in terms of each other. Examples include the ideal gas law; PV=nRT, the total energy; E_total = E_kin + E_pot, and the Lagrangian; L = E_kin − E_pot. Importantly, these dependencies are often empirical discoveries, not analytic truths, so the concepts themselves are not defined in terms of each other. Yet, when the model represents and conceptualizes two of the three variables, there will be no way for us to tell which of the three that is, due to their interdependence and equivalent model functionality. It will therefore be underdetermined what the conceptual content of the representation is. I consider two implications of this finding. Firstly, it suggests some caution in our use of such anthropomorphic language of representations and concepts, for if the conceptual content of representations is sometimes underdetermined for DNNs, we might need to reconsider what we really mean by ‘representations of concepts’. Secondly, the underdetermination introduces an additional difficulty in learning new concepts and relations from distributed representations in DNNs, which might have implications for their usefulness in scientific discoveries.  

Bau, David, Jun-Yan Zhu, Hendrik Strobelt, Agata Lapedriza, Bolei Zhou, and Antonio Torralba. 2020. ‘Understanding the Role of Individual Units in a Deep Neural Network’. Proceedings of the National Academy of Sciences 117 (48): 30071–78. https://doi.org/10.1073/pnas.1907375117.  

Boge, Florian J. 2024. ‘Functional Concept Proxies and the Actually Smart Hans Problem: What’s Special About Deep Neural Networks in Science’. Synthese 203 (1): 1–39. https://doi.org/10.1007/s11229-023-04440-8.  

Buckner, Cameron. 2020. ‘Understanding Adversarial Examples Requires a Theory of Artefacts for Deep Learning’. Nature Machine Intelligence 2 (12): 731–36. https://doi.org/10.1038/s42256-020-00266-y.  

Harding, Jacqueline. 2023. ‘Operationalising Representation in Natural Language Processing’. The British Journal for the Philosophy of Science, November. https://doi.org/10.1086/728685.  

Räz, Tim. 2023. ‘Methods for Identifying Emergent Concepts in Deep Neural Networks’. Patterns 4 (6): 100761. https://doi.org/10.1016/j.patter.2023.100761

Darrell P. Rowbottom, Philosophy (Lingnan University Hong Kong): tba

tba

Emily Sullivan, Philosophy (Utrecht University): Idealization Failure in ML 

Abstract: Idealizations, deliberate distortions introduced into scientific theories and models, are commonplace in science. This has led to a puzzle in epistemology and philosophy of science: How could a deliberately false claim or representation lead to the epistemic successes of science? In answering this question philosophers have been single-focused on explaining how and why idealizations are successful. But surely some idealizations fail. I propose that if we ask a slightly different question, whether a particular idealization is successful, then that not only gives insight into idealization failure, but will make us realize that our theories of idealization need revision. In this talk I consider idealizations in computation and machine learning.  

Schedule

Day 1 Day 2
09:00 Arrival 09:00 Einführung
10:00 Introduction 09:15 – 10:00 to be announced
Darrell Rowbottom, Lingnan University Hong Kong
10:15-11:00 Semantic Grounding in Advanced LLMs?
Holger Lyre, Magdeburg University
10:15 – 11:00 The Underdetermination of Representational Content in DNNs
Sara Pernille Jensen, Oslo University
11:15–12:00 A Gentle Introduction to Neural Network Verification (and How It Might Contribute to Evaluating Scientific Insights)
Daniel Neider, TU Dortmund
11:15-12:00 The Benchmarking Epistemology – What Inferences Can Scientists Draw from Competitive Comparisons of Prediction Models?
Timo Freiesleben, Tübingen University
12:00-13:00 Lunch 12:00-13:00 Lunch
13:00–13:45 On scientific discovery with machine learning: what is “strong“ novelty?
Heather Champion, Tübingen University
13:00-13:45 Deep-learning Models and Scientific Understanding through Explanations and Representations
Giovanni Galli, University of Teramo
14:00–14:45 From Generative AI to AGI: Multi-LLM Agent Collaboration as a Path Forward 
Edward Chang, Stanford University
14:00–14:45 The Gradability of Explanatory Understanding
Insa Lawler, UNC Greensboro
14:45–15:15 coffee break 14:45 – 15:00 coffee break
15:15–16:00 Idealization Failure in ML 
Emily Sullivan, Utrecht University
15:00 – 15:45 Scientific Progress in the Age of AI
Finnur Dellsén, University of Iceland
16:15–17:00 Bridging Scientific Understanding and Creativity with an LLM Benchmark for Narrow-Domain Scientific Fields
Henk W. de Regt, Radboud University and Eugene Shalugin, Radboud University
16:00 – 16:45 Predictively-valid “Alien” Features, or Artifacts? Coping with Inscrutable Scientific Progress
Cameron Buckner, University of Florida

For inquiries, please contact the organizing committee at udnn.ht@tu-dortmund.de

Main Organizers: Annika Schuster, Frauke Stoll, and Florian J. Boge