Shaping the Digital Future of ErUM Research: Sustainability & Ethics

Europe/Berlin
Erholungs-Gesellschaft Aachen 1837

Erholungs-Gesellschaft Aachen 1837

Reihstraße 13, 52062 Aachen
Description

DIG-UM, the ErUM-Data-Hub and the Program Committee cordially invite you to the second workshop addressing sustainability & ethics in the digital transformation in ErUM-Data. This creative workshop aims to discuss and develop strategic concepts and concrete action for sustainability accompanied by ethical aspects in the digital transformation of basic research on universe & matter (ErUM). 

The workshop program consists of: 

  • A review of the progress and changes since the last in person workshop in 2023.
  • Space for the presentation of new approaches and projects in sustainability. ErUM researchers from all 8 ErUM communities as well as researchers from related disciplines and industry are cordially invited to submit a contribution.
  • We are particularly interested in learning about new approaches that we as a community should consider and implement. 
  • Pleanry discussion of further sustainability measures from the ErUM-Data community.
  • Writing slots for the preparation of the findings in a joint publication.
  • Invited talks on ethical and sustainability topics to inspire and stimulate discussion.
  • Room for contributions on the topic of ethics from the ErUM-Data community. We invite you to submit contributions on topics that consider ethics in the further development of AI (e.g. sustainable use of resources from an ethical perspective, using AI in teaching, does AI replace human creativity in ErUM Research? How will it be possible to realize predominantly human-performed research activities with the support of AI tools in the future? What is our contribution?).

 

Target group:

  • This workshop welcomes contributions as well as participation from ErUM scientists from all academic levels of all 8 ErUM communities, related sciences, policy stakeholders and industry partners. We encourage young scientists and experienced scientists to participate in order to enable broad discussions.

 

Deadlines:

  • Submission deadline is June 6, 2025. The call for abstracts is open. Find all info here. To submmit a contribution, please use the button at the very bottom of the overview page
  • Registration is open until July 11th, 2025. 

 

Registrations are open. Programme is preliminary. 

 

Surveys
Feedback on the Workshop
    • 09:00 09:30
      Coffee 30m
    • 09:30 10:30
      Sustainability: Invited Talks
      • 09:30
        Wolfang Gründinger: Tech for Future (remote) 30m

        Digitale Innovationen sind bereits heute erfolgreich im Einsatz für Energie- und Verkehrswende, Klimaschutz und nachhaltige Entwicklung. Aus dem Management der dezentralen, erneuerbaren Stromversorgung etwa sind digitale Technologien nicht mehr wegzudenken. Warum digitale Transformation und nachhaltige Entwicklung zusammengehören, das zeigt Wolfgang Gründinger in diesem Impulsvortrag.

      • 10:00
        Sustainability aspects in the planning of the Einstein Telescope 30m
        Speaker: Achim Stahl (RWTH Aachen University)
    • 10:30 11:00
      Coffee 30m
    • 11:00 12:30
      Sustainability: Session 1 (Sustainable ErUM research centres of the future)
      • 11:00
        From Data to Duty: A Simulation Framework for Sustainable Distributed Computing 30m

        As scientific experiments in basic research of universe & matter continue to generate vast volumes of data, the need for scalable, efficient, and sustainable computing becomes increasingly critical, not only to match the future requirements for data storage and processing but also to align with the responsibility imposed on the scientific community towards a more sustainable future.

        This contribution presents the simulation framework DCSim — developed and maintained by an international and interdisciplinary collaboration — designed to model the execution of computing workflows on large-scale heterogeneous distributed computing infrastructures. By simulating detailed aspects of workflow execution — including data movement, job scheduling, resource allocation, and workload execution — the model provides reliable insights into system performance and resource utilization.

        A notable feature is its extensibility for energy consumption prediction, enabling users to evaluate and optimize computing workflows not only for efficiency and throughput but also for their environmental impact. This predictive capability supports evidence-based decision-making towards greener computing strategies, aligning with broader ethical imperatives and sustainability goals. The tool is intended to serve both as a planning aid for future infrastructure design and as a guiding basis for developing and promoting responsible digital research strategies and practices.

        Speaker: Maximilian Maria Horzela (Georg-August-Universität Göttingen)
      • 11:30
        Advancing the Environmental Sustainability of Scientific Computing for ErUM 30m

        Scientific computing contributes significantly to the CO2 footprint created by research conducted in ErUM. In order to advance the environmental sustainability in this area the SUSFECIT (Sustainable Federated Compute Infrastructures) research network has been proposed with the goal to contribute to developing a strategy and interlinked software ecosystems to reduce CO2 footprint and to increase the energy efficiency of distributed computing resources. The basic idea is to exploit the dispatchability of compute jobs in space and time and use the partition of a federated computing infrastructure at a place, which at a certain time, is (dominantly) powered by renewable energies such a wind and solar power plants. In order to realize this basic concept, it is foreseen to develop and optimize three interlinked ecosystems: i) for forecasting of the available energy mix, power costs and requested needs for compute power, (ii) for the orchestration of jobs on federated and locally distributed compute clusters taking into account the forecasts and (iii) for the accounting of the used CPU and GPU resources with respect to elapsed time, power consumption and CO2 footprint. A digital twin for the above set of ecosystems shall also be developed in order to optimize e.g. operation parameters. The presentation will discuss the basic concept, the content of the three ecosystems and exploratory work, which has been conducted by partners in the research network (DESY, KIT, Universities in Aachen, Bonn, Göttingen, Freiburg and Öko-Institut).

        Speaker: Markus Schumacher (Albert-Ludwigs-Universität Freiburg)
      • 12:00
        Potential Synergies between the Helmholtz Cluster for a Sustainable and Infrastructure-Compatible Hydrogen Economy and ErUM Research 30m

        In 2020, the German Federal Government and the state government of North Rhine-Westphalia decided that coal firing for electricity generation will be phased out by 2030 to reduce CO2 emission. Coal mining regions like the Rhenish mining area are affected by this decision. Towards this end, the Helmholtz Cluster for a Sustainable and Infrastructure-Compatible Hydrogen Economy (HC-H2) was founded. The HC-H2’s ultimate vision is to research, develop, and demonstrate innovative H2-based technologies for a climate-neutral energy economy of the future, which comprises basic research up to commercialization with partners. Our mission is the transformation of the Rhenish mining area into a H2 model region with Europe-wide appeal, thereby contributing to both the successful “Strukturwandel” and a CO2-neutral energy system of the future. The talk will introduce the overall structure of the HC-H2 and exemplify our approach using recent examples from our demonstration region for future hydrogen technologies, including Multi-SOFC (electricity and heat generation from hydrogen derivatives for a hospital, Erkelenz), HyFRed (H2-based reduction of iron ores, Mönchengladbach), and HyHeat (Hydrogen as a fuel in industrial furnaces for press hardening processes, Simmerath). Finally, a few thoughts on innovative H2-based technologies for shaping the sustainable and digital future of ErUM Research will be shared in the hope to inspire further ideas and discussion.

        Speaker: Hans-Georg Steinrück
    • 12:30 13:30
      Lunch 1h
    • 14:00 15:30
      Sustainability: Plenary Discussion
      Convener: Dwayne Spiteri (IT (IT Scientific Computing))
    • 15:30 16:00
      Coffee Break 30m
    • 16:00 16:15
    • 16:15 17:30
    • 17:30 19:00
      Dinner 1h 30m
    • 09:00 09:30
      Coffee 30m
    • 09:30 10:00
      Ethics: Invited Talk
      • 09:30
        Talk by Prof. Saskia Nagel 30m
    • 10:00 10:30
      Ethics: Submitted Contributions
      • 10:00
        Atmosphere Ownership 30m

        Who owns the atmosphere of the earth? No human can live without the atmosphere. Therefore one can assume that each human owns an indefeasible and equal share of the atmosphere. So I can claim ownership of a fraction of about 1 / 8 billion of the atmosphere. Now if others use my property I have the right charge them for the usage. For a worldwide, annual CO2 emission of 38 billion t this means I can ask for a compensation of the emission of 4.75 t of CO2 into my part of the atmosphere. Taking a commercially available offer for direct air capture of CO2 to estimate the provided value this gives an amount of 4750 Euro. As the duty of my government is to protect my rights I can demand the money from them.

        This chain of thoughts is meant as input to the discussion how to increase motivation to address the climate problem.

        Speaker: Thomas Kuhr (LMU Munich)
    • 10:30 11:00
      Coffee Break 30m
    • 11:00 12:30
      Ethics: Submitted Contributions
      • 11:00
        Responsible use of tools 15m
        Speaker: Jan Burger
      • 11:15
        Stingy Computing (Vijay Kartik) 15m
        Speaker: Vijay Kartik (DESY)
      • 11:30
        AI-Enhanced ErUM: Four New Epistemic and Ethical Considerations 30m

        Scientists are already using generative AI in their research (Furze, 2025; Kwon, 2025; Nazir & Wang, 2023). But while there is extensive and ongoing research into the ethics of AI throughout its lifecycle (Coeckelbergh, 2020), and some research about the ethics of AI-related jobs and how certain job tasks might use AI (Chance & Hammersley, n.d.; Gray & Surrey, 2019; Perrigo, 2023; Williams et al., 2022), relatively little of the research has to do specifically with use of AI to augment knowledge work (Kulkarni et al., 2024; Nah et al., 2023; Resnik & Hosseini, 2024). In scientific research in particular, important critiques charge AI use with risking fabrication of vital data and citations, undermining transparency in data collection and use, and complicating the very nature of authorship (Resnik & Hosseini, 2024).

        However, no one has yet investigated the specific ethical issues surrounding use of AI research in the field of what is known in German as “ErUM” (“Erforschung von Universum und Materie”), i.e., research into the universe and matter. Such research comprises several fields of the natural sciences, including astrophysics, particle physics, nuclear physics, hadron and ion physics, and photon and neutron science. This paper therefore offers a preliminary ethical assessment of the ethics of AI-assisted ErUM. We find that even setting aside traditional concerns about AI use in scientific research in general, AI-enhanced ErUM in particular raises four specific, ethically salient problems, all of which are related to the long-term sustainability of AI-enhanced ErUM research.

        The first problem is that fabrication (sometimes called “hallucination”) may be especially difficult to detect in ErUM compared to other sciences and to non-scientific research fields, given the required investments of money and expertise of running reliable ErUM experiments. Put briefly, while LLMs’ fabrication rates have declined greatly in the last few years (Vectara, n.d.), it is obviously very important that specific scientific claims made in a manuscript be accurate. In other areas of scholarly research, fabrication is either not a problem (because the research is fully creative in nature), easy to detect (because the research’s main evidence is explicitly, publicly presented in the document for anyone to inexpensively evaluate), or at least open to testing by other, interested parties. But because much of ErUM is, by its nature, relatively expensive and abstruse, fabrications may be difficult to detect. Replicating large experiments such as those at CERN, FAIR, or XFEL is practically impossible for almost everyone (Junk & Lyons, 2020), so fabricated interpretations or subtly altered conclusions, especially in internal reports or AI-drafted analyses, may go unchallenged for years. Similarly, raw data in ErUM are typically vast, complex, and not publicly accessible, or only shared within limited collaborations. Even the collaborators rely on multi-layered data-processing pipelines, with specialized software and calibration tools that only a subset of the team understands (cf. Neves et al., 2011; Rumsey, 2025; Werum, n.d.). And while this is more controversial, arguably, some ErUM subfields (e.g., dark-matter searches, early-universe cosmology, string theory, and neutrino physics) involve interpretations under conditions of uncertainty. Small anomalies are often the bases of tentative theoretical frameworks, and new theories are sometimes expected to rest, at first, on minimal or indirect empirical signals.

        Second, and related, what is sometimes called “model collapse” is especially dangerous in ErUM, because so much ErUM is concentrated in a few major institutions, and so the demand for data may outstrip the production of novel research. Model collapse occurs when an LLM is trained, at least partially, on LLM-generated data (cf. Crotty, 2024; Shumailov, 2024; Sun et al., 2024). This is a problem in general, but because cutting-edge ErUM (e.g., collider results, XFEL imaging, and neutrino events) requires highly specialized and expensive equipment, and is generated at a relatively small number of sites (e.g., CERN, DESY, and FAIR), only a few collaborations and labs occupy the pipelines from raw data to public interpretation. So, the number of available human-written interpretations of ErUM data is already very limited compared to fields such as psychology, sociology, and even molecular biology. If even a small fraction of the LLM training corpus on ErUM comes from derivative or LLM-generated summaries of these limited sources, recursive contamination is far more likely, and harder to detect, because replication is nearly impossible. Beyond this, fields such as physics and cosmology have outsized cultural and epistemic influence. Discoveries such as the Higgs boson, gravitational waves, and the Big Bang model are frequently “reprocessed” in popular science, TED talks, and Wikipedia (cf. Levinson, 2013). But LLMs are disproportionately trained on those popularized and digested forms of ErUM. If even a few flawed, AI-generated articles inflect these outlets, they are more likely to get reabsorbed into training corpora and thereby overweight the epistemic authority of incorrect interpretations.

        Third, and also related, given that ErUM is adjacent to the foundations of physics, the use of AI to describe and interpret research results may activate its well-known plausibility bias (Agarwal et al., 2024), exacerbate LLMs’ lack of transparency, and have other misleading effects. So, when asked to summarize or interpret foundational ErUM, LLMs may favor familiar explanatory tropes (e.g., “curved spacetime,” “particle-wave duality,” “the fabric of the universe,” and “uncertainty principle”) and reproduce consensus-sounding narratives, even where the field remains deeply unsettled. Relatedly, because LLMs specialize in semantic mimicry, their statements may miss subtle distinctions in ontological assumptions and theoretical commitments. They may unwittingly conflate rival interpretations, misrepresent theories’ scopes, or invent “bridges” between incompatible frameworks, for the sake of plausibility.

        Fourth, because ErUM commonly requires wide-ranging international collaborations, injecting AI writing and interpretation into internal documents and translations may undermine accountability and introduce further fabrications. ErUM is sometimes conducted by massive international consortia involving hundreds to thousands of researchers (cf. Abbott et al., 2012; ATLAS Collaboration et al., 2012), spread across many countries and time zones (e.g., ATLAS, MCS, IceCube, and FAIR). Internal communication (such as memos, reports, logs, and drafts) often passes through many hands and is written collaboratively. Beyond this, international ErUM teams often operate in multiple working languages and rely on AI translation tools to write or interpret internal communications. When precision is critical, subtle mistranslations can alter apparent meanings. Indeed, LLMs are known to fabricate or “smooth over” unclear concepts, especially in technical domains. And because scientific collaboration relies not only on accuracy but also on understanding who knows what (and to what degree of confidence)—a sort of “epistemic map” of the team—AI-generated text may mask the human judgment behind claims. Finally, large-scale ErUM projects require extensive documentation, which is often delegated to staff or early-career researchers, and increasingly, to LLMs. For all these reasons, there is danger of mistranslation, miscommunication, and semantic drift, and these errors can propagate into the literature or infrastructure plans.

        This paper concludes by issuing some recommendations, based on these ethical problems, for ethically responsible use of AI in ErUM. Put briefly, ErUM researchers who use AI should adopt these five best practices:
        • AI-provenance tagging in publications and preprints: ErUM authors should clearly indicate which portions of a document, and which datasets, were produced, edited, or otherwise modified by AI. Authors should include prompts and source chains as supplementary files.
        • Independence in summaries: ErUM researchers should avoid using LLMs to evaluate or summarize papers that the LLMs themselves may have been trained on, especially in grant or peer review.
        • AI-verification workflows: ErUM researchers should build institutional workflows with checklists or signoff procedures for AI-generated sections, key-claim verification, and audits. They should also create logs or registries of AI-generated content.
        • Epistemic qualifiers: Authors should mark speculative or inferential language with qualifiers such as “tentative,” and discourage AIs’ being used to explain anomalous results, unless a human has first formulated or vetted the anomaly.
        • General caution: ErUM researchers and journalists should treat ErUM as an “epistemic high-risk zone” for fabrication, plausibility bias, and recursive contamination, and should support research into domain-specific collapse detection.
        These practices can help maintain the promise of AI-assisted ErUM research while minimizing the principal ethical and epistemic dangers.

        Given the undeniable promise of AI in academic research, it would be a mistake to completely boycott or forgo the benefits of AI assistance, including in ErUM. By being aware of the domain-specific dangers of AI-enhanced ErUM research, scientists can enhance their understanding of the universe while investing in a reliable and solid foundation for the future of ErUM.

        Speaker: Thomas Metcalf (Sustainable AI Lab, Institute for Science and Ethics, University of Bonn)
      • 12:00
        Ethical Aspects of Using, Writing, and Distributing Software 30m

        I will argue why the classical Four Software Freedoms (Use, inspect, improve, share) are much more than utopian dreams of idealists with long beards. On the contrary, in addition to them providing an inclusive social contract fostering software development in a process resembling the scientific process itself, it is also an important ingredient to FAIR software that is sustainable in both senses of the word.

        Speaker: Markus Demleitner (Universität Heidelberg, Zentrum für Astronomie)
    • 12:30 12:40
      Group Picture 10m
    • 12:40 13:40
      Lunch 1h
    • 14:00 15:30
      Ethics: Plenary discussion
      Convener: Markus Demleitner (Universität Heidelberg)
    • 15:30 16:00
      Coffee 30m
    • 16:00 17:30
    • 17:30 19:00
      Dinner 1h 30m
    • 19:30 20:30
      Public evening lecture: Zukunft gestalten mit Verantwortung: KI, Forschung und Nachhaltigkeit (Johannes Hartl, Eden Palast Aachen)

      Ein öffentlicher Abendvortrag mit Johannes Hartl im Rahmen des ErUM-Workshops zu Ethik & Nachhaltigkeit