2025 European HDF5 Users Group Meeting

Europe/Berlin
FLASH seminar room

FLASH seminar room

FLASH Notkestrasse 85 22607 Hamburg
Description

We are excited to invite you to the 2025 European HDF5 User Group Meeting, taking place from May 26-28, 2025, at DESY in Hamburg.

This meeting presents a great opportunity to engage in in-depth discussions on the evolving needs of data management. While the focus is on HDF5, we encourage participation from the broader community to share experiences, insights, and innovations in working with HDF5 and other data formats and I/O libraries. The goal is to optimize workflows and address data challenges across diverse fields.

The workshop will be held in a hybrid format, with both in-person and remote participation options. If you plan to attend and present in person, we encourage you to register early, as space is limited to 100 participants.

We are seeking presentations and posters on the following topics (feel free to suggest your own!):

  • Real-world applications of HDF5 in various scientific and industrial sectors (e.g., computer simulations, data acquisition, data archiving and sharing in scientific research, bioinformatics and healthcare, manufacturing, automotive, etc.).
  • Best practices in implementing and optimizing HDF5 (or other formats) for complex data storage and analysis.
  • Techniques for overcoming common challenges in HDF5 integration, performance tuning, and scalability.
  • Innovative solutions using HDF5 in HPC, cloud, or hybrid environments.

Submission Guidelines
Proposals should include:

  • A brief abstract (150-250 words) outlining the focus of your presentation.
  • A description of how your work or case study demonstrates using HDF5 to solve real-world data challenges.
  • Keywords that describe the topic.

Please register and submit your abstract if you plan to present.

There is no charge to attend, thanks to our generous sponsors. 

Sponsors

DESY logo
Daphne4NFDI Logo
Lifeboat LLC logo
DESY is assisting by providing the venue and local organizing. Daphne4NFDI is sponsoring the coffee breaks. Llifeboat, LLC is providing additional assistance.
Contact the HUG25 Organizing Committee
Registration
2025 European HDF5 Users Group Meeting Registration
    • 09:00 09:25
      Welcome Session - Anton Barty: Welcome Address
    • 09:25 09:50
      Using HDF5 to Store Scientific Data at the European XFEL 25m

      The European XFEL is an X-ray laser research facility that produces extremely short and intense X-ray flashes, enabling investigations across a wide range of fields—from the structure of matter to the dynamic evolution of molecular systems. A typical experiment can generate petabytes of data within a day, originating from diverse detectors and in multiple formats. Managing this high-volume, heterogeneous data, along with metadata, in real- or near real-time poses unique challenges.

      HDF5 provides a robust solution to these requirements. Its ability to define both data and its structure enables consistent storage across various data sources. HDF5 also supports parallel, real-time writing, by distributing experimental data across multiple files, and linking them coherently. Features such as sparse chunking and compression further optimize storage, which is critical given the facility’s data output.

      In this talk, I will introduce karaboHDF5, a library that enables real-time, space-efficient storage of experimental data, while ensuring the data is well-structured and easy to access for analysis and processing. Developed by the European XFEL on top of the HDF5 core library, it is tailored to the facility’s demanding performance and usability requirements, and fully leverages the capabilities of HDF5.

      Keywords: HDF5, real-time data acquisition, high-volume data, heterogeneous data, X-ray data acquisition.

      Speaker: Jose Vazquez-Garcia (European XFEL)
    • 09:50 10:10
      Coffee Break 20m
    • 10:10 10:35
      The keys to the management of time-evolution signals on HDF5 files in ITER 25m

      ITER (International Thermonuclear Experimental Reactor) is the largest international experiment in the field of generating nuclear fusion energy by magnetic confinement. ITER’s objective is to operate in modes that come as close as possible to the conditions of a commercial fusion reactor, which implies long pulses and systems running continuously.
      From the point of view of the data acquisition and control system, ITER, on the one hand, requires an enormous number of signals (due to its size and experimental nature), many of them with a very high sampling rate (due to the highly chaotic nature of the plasma under extreme conditions), and, on the other hand, it requires a data storage and management system compatible with long pulse and continuous data acquisition. These requirements have driven the design and implementation of ITER’s data storage and management system, where time evolving signals and their real time processing have absorbed the great majority of development effort.
      From the very beginning, ITER chose HDF5. This presentation describes the technical factors that have proved decisive in using HDF5 as the core technology for storing and managing all of ITER’s time evolving signals, as well as for the real time handling of their data. It details the different challenges that have been addressed throughout the project, together with a detailed description of the solutions that have been adopted at the design and implementation level.

      Speaker: Rodrigo Castro (CIEMAT)
    • 10:35 11:00
      Seamless Integration of Blosc2 and HDF5 for High-Performance Data Compression 25m

      HDF5 is the de facto standard for storing large volumes of binary data in files. Blosc2, an award-winning high-performance library, excels at compressing binary data in memory. Both are widely used, making their integration natural. This talk will cover using Blosc2 as an HDF5 filter and HDF5 as a Blosc2 backend.

      We will outline the current state of the Blosc2 plugin for HDF5 (https://github.com/Blosc/HDF5-Blosc2) and provide instructions on its usage. Additionally, we will introduce the b2h5py package (https://github.com/Blosc/b2h5py), which bypasses the slow HDF5 filter pipeline to achieve high performance. Sparse datasets will be used to demonstrate Blosc2's performance on HDF5.

      Finally, we will discuss various codecs available in Blosc2, with a focus on the Grok codec (https://github.com/GrokImageCompression/grok), which efficiently compresses data in the JPEG2000 format. We will also touch on the enhancements made to BTune (https://ironarray.io/btune), a Blosc2 plugin, to support lossy compression and automatically select the best codec/filter.

      This work has been carried out as part of the LEAPS-INNOV program (https://leaps-innov.eu/), which strives to build a European ecosystem for photon sciences. The integration of Blosc2 and HDF5 ensures efficient storage and retrieval of large datasets, a critical factor for the program's success.

      Compression #HighPerformance #HDF5 #Blosc2 #JPEG200 #LEAPS-INNOV

      Speaker: Francesc Alted (Blosc project)
    • 11:00 11:25
      High Performance Storage for HPC: Proven HDF5 Scalability and Consistency on Pure Storage FlashBlade 25m

      The integration of the HDF5 high-level IO library, encompassing Single Writer Multiple Reader (SWMR) and parallel mode (pHDF5), with Pure Storage FlashBlade using NFS, modern Linux Kernels, and Networks, offers a robust and high-performance solution. This combination significantly reduces Total Cost of Ownership (TCO) and technical debt compared to traditional parallel file systems.

      HDF5, particularly in SWMR configuration, ensures data consistency during concurrent read and write operations. Pure Storage FlashBlade's NFS provides a high-performance storage solution, ideal for High-Performance Computing (HPC) environments, facilitating efficient access and modification of large datasets. Enhancements within HDF5, such as superblocks, further bolster data consistency. FlashBlade's POSIX-compliant data access guarantees atomicity during read and write operations to direct flash memory modules, a crucial requirement for SWMR operations to prevent inconsistencies.

      Pure Storage FlashBlade delivers exceptional throughput, low latency, and supports GPU Direct Storage with NFS over RDMA, essential for HPC. Multi-iteration write and read tests have demonstrated robust throughput and consistency, validating FlashBlade's capability to meet HDF5 demands in both SWMR and parallel configurations. Parallel HDF5 leverages MPI-IO, which supports NFS backends, treating it as a unified file system. This integration ensures parallel processes operate efficiently on shared datasets, with MPI layer compatibility enabling effective data handling across multiple nodes.

      With optimized HDF5 settings, SWMR mode operates without exclusive or shared locks, allowing a single writer to handle write operations while multiple readers access data concurrently without conflicts. Comprehensive tests have validated FlashBlade's ability to run HDF5 in both SWMR and parallel modes, ensuring atomic writes and reads. Performance benchmarking using HDF5-specific tools has confirmed high throughput and data consistency across multiple nodes via NFS.

      In conclusion, the synergy of HDF5's robust data handling with the high performance and consistency of Pure Storage FlashBlade using NFS creates a powerful solution for managing large, complex datasets in HPC environments.

      Speaker: Michael Kaspars (Pure Storage)
    • 11:25 11:50
      Revisiting HDF5 SWMR and file versioning features 25m

      In this talk, we will present an overview and demonstration of two HDF5 features implemented via virtual file drivers (VFDs), both currently in a prototype stage: the full single-writer/multiple-reader capability (VFD SWMR) and HDF5 versioning (also known as the Onion VFD).
      The VFD SWMR feature enables file modifications during writing and provides guarantees on the maximum latency before new data becomes available to readers. The Onion VFD allows tracking of modifications to an HDF5 file between open/close calls while preserving—or providing access to—the file's state prior to a given set of changes.
      Although the Onion VFD has been released and is available in current maintenance versions, a more robust implementation is still in development. In this presentation, we aim to gather feedback from the HDF5 user community on both VFDs and to share our roadmap for developing production-ready versions of these features.

      Speaker: Elena Pourmal (Lifeboat, LLC)
    • 11:50 13:20
      Lunch 1h 30m
    • 13:20 13:45
      HDF5: The most versatile container for sharing scientific and engineering data 25m

      When people hear this affirmation, a common reaction is, "When, HDF Group, when?" In her book How to Make Sense of Any Mess, Abby Covert reminds us to be careful and not fall in love with our plans or ideas but with the effects we can have when we communicate clearly. While I would reject the notion that "a mess" aptly describes the current state of HDF5, I want to use this short presentation to clearly communicate our plans and ideas vis-à-vis HDF5 and the community.

      Speaker: Gerd Heber (The HDF Group)
    • 13:45 14:10
      An Open HDF5-Based Format for Industrial Inspection Data with Experimental Integration of the Onion VFD for File Versioning 25m

      The NDE File Format (.nde), developed by Evident, is an open, extensible data format tailored for the non-destructive evaluation (NDE) and testing (NDT) industry. Built upon the HDF5 container and augmented with JSON-based metadata, it offers a platform-independent solution for storing inspection data, primarily for ultrasonic modality. By adopting an open format, .nde files can be accessed and analyzed without proprietary software, promoting integration with third-party tools and advanced analytics platforms.

      To address the challenges of data versioning and traceability inherent in NDT workflows, we explore the application of the Onion Virtual File Driver (VFD), introduced in HDF5 version 1.13.2. The Onion VFD enables in-file revision management by layering modifications atop the original dataset, allowing users to access and revert to previous versions without duplicating entire files. This approach ensures data integrity and auditability, essential for compliance with industry standards and regulations.

      This presentation demonstrates how the integration of the .nde format with the Onion VFD addresses real-world data management challenges in industrial inspection environments. We discuss the technical implementation, benefits observed in data handling and version control, and potential implications for broader adoption in the NDT industry.

      Speaker: Baptiste Gauthier (Evident Scientific)
    • 14:10 14:35
      Using a HDF5 File as a Zarr v3 Shard 25m

      Version 3 of the Zarr specification includes a sharding codec that allows for chunks to contain small inner chunks. The format of the resulting binary file format of shards is reminiscent of a HDF5 file. Both HDF5 files and Zarr v3 shards may contain compressed chunks. Furthermore, the Zarr v3 shard specification is similar to the Fixed Array Data Block structure within a HDF5 file. Additionally, the sharding concept is similar to a subset of the HDF5 virutal dataset feature. I will discuss how a standard HDF5 file could be used a Zarr v3 shard and how a set of such files could be used as a Zarr array or assembled into a HDF5 virtual dataset. Finally, I will discuss potential cooperation between Zarr and HDF5 and alternatives to my approach.

      As a bonus topic, time permitting, I could discuss how a single file could simultaneously act as a valid TIFF file, HDF5 file, and Zarr v3 shard, taking advantage of cloud optimization of each of these formats.

      Speaker: Mark Kittisopikul (Howard Hughes Medical Institute)
    • 14:35 15:00
      Race Ya! An Accelerator-Native I/O Pipeline for HDF5 25m

      GPUs and similar accelerators have become the dominant compute platform for STEM applications, from finance to space flight, and beyond. However, HDF5 continues to execute exclusively on the Host CPU of GPU nodes. This talk will present a design overview of moving the I/O pipeline filters, datatype conversions and other transforms from the CPU to the GPU, including how to perform I/O directly from GPU memory. Implementation details of the revised transform pipeline will be included, along with projected performance benefits.

      Speaker: Quincey Koziol (NVIDIA)
    • 15:00 15:20
      Coffee Break 20m
    • 15:20 15:50
      HDF: Powering the Future of Data - What's Now, What's New, What's Next! 30m

      This talk will provide an overview of the current state of the HDF5 software ecosystem, highlighting recent advancements, key components, and ongoing challenges. We will explore the future directions of the various tools and libraries that empower researchers and developers to manage HDF5-related workflows efficiently. Additionally, the talk will outline potential future initiatives for the HDF5 ecosystem to ensure its continued relevance and evolution in the ever-expanding realm of data-intensive applications.

      Speaker: Scot Breitenfeld (The HDF Group)
    • 15:50 16:50
      HDF5 Roundtable

      The HDF5 Roundtable is an open, interactive forum hosted by The HDF Group during HUG events, designed to bring together users, developers, and stakeholders from the HDF5 community. This session provides a space for sharing insights, discussing current challenges, and exploring the future direction of HDF5. Whether you're a new user or a longtime contributor, the Roundtable is a collaborative opportunity to ask questions, offer feedback, and help shape the evolution of the HDF5 ecosystem.

    • 09:00 09:25
      Innovative Data Acquisition Solutions with HSDS at MAX IV Synchrotron 25m

      MAX IV is an accelerator-based light source located in southern Sweden. It continuously operates 16 experimental stations using X-rays for material and life sciences. User data analysis primarily occurs on a small edge HPC, while automatic scientific data processing, from around 100 data sources, runs in an edge-cloud environment. These pipelines provide rapid feedback on data acquisition and analysis to control workstations and users' laptops at the facility or accessed remotely. The vast majority of raw experimental data is stored in the HDF5 format.

      Data analysis and visualization at synchrotron light sources is inherently a distributed computing problem (HPC, edge, users’ terminals). Traditionally, we have used distributed file systems together with HDF5, which motivated the development of VDS and SWMR. The HDF group introduced HSDS over half a decade ago, and POSIX HSDS storage was implemented shortly after.

      This allowed us at MAX IV to start HSDS microservices for our experimental stations, even without object storage. We routinely use HSDS to store intermediate data for our distributed data reduction pipelines, particularly for streaming tomography. This offers several advantages, including no dependencies on distributed file systems and straightforward data access from users’ terminals. We can also utilize the existing HDF5 tools, such as the silx viewer and other standard software like h5pyd.

      In many cases, the intermediate HDF5 data do not require persistent storage. Therefore, we developed an extension for dranspose – the distributed data analysis pipelines framework, which allows in-memory Python dictionary data to be exposed with the HSDS/h5pyd interface.

      Keywords: data acquisition, distributed applications, hybrid environments, memory resident datasets, HSDS, synchrotron, tomography

      Speaker: Zdenek Matej (MAX IV Laboratory, Lund University)
    • 09:25 09:50
      Down to the bytes: can we simplify alternative access to HDF5? 25m

      HDF5 is an enormously powerful and flexible file format. There are many different ways to use it, and it's difficult to provide one API that works efficiently for all the possible use cases. However, the complexity of the on-disk file format is a high barrier to alternative implementations, so with a few heroic exceptions, most code reading & writing HDF5 does so through the canonical C implementation, including its bindings in various other languages.

      In this open-ended talk, I aim to give an overview of a range of different ways to access HDF5 files, bypassing part or all of libhdf5. This will include practical techniques to speed up access in specific circumstances, projects building additional indexes around HDF5 files, and alternative implementations of the file format. I'll conclude with some ideas on how we might lower the barrier and make it easier to implement optimised HDF5 file access for specific scenarios.

      Speaker: Thomas Kluyver (Eur.XFEL (European XFEL))
    • 09:50 10:10
      Coffee Break 20m
    • 10:10 10:35
      h5pydantic 25m

      h5pydantic is a Pydantic based library aimed at making it easier for scientists to organise their HDF5 files, by writing Python models of their experiments. The library is similar to an Object Relational Mapper (ORM), but instead of targeting a relational database, it targets HDF.

      The library is inspired by the need of the Australian Synchrotron during the commissioning of our new beamlines under the BRIGHT program. The library is currently under development.

      Speaker: Clinton Roy (ANSTO)
    • 10:35 11:00
      HSDS v1.0 – Performance Features 25m

      HSDS (Highly Scalable Data Service) is a REST-based service that provides read/write access to HDF5 data stores – using object storage or posix. By using a combination of multi-processing and asynchronous IO, HSDS can achieve remarkable performance when accessing very large datasets. On the other hand, performance lagged for clients invoking a series of smaller requests (reading or writing a small dataset selection, creating an attribute or a new link). As typical in client-server architectures, the per-request latency is quite high compared to in-process operations (e.g. making a call to the HDF5 library). To address this, the next release of HSDS and h5pyd (the HSDS client library for Python) will look to improve performance by a combination of read-ahead logic and combining write operations into a single request. At the same time, clients can continue using the same h5py-api to achieve better performance without needing to make any code changes.

      Speaker: John Readey (The HDF Group)
    • 11:00 11:25
      Writing HDF5 in Pure Java: Practical Guide 25m

      This contribution presents our experience using a pure Java implementation of the HDF5 file format to support metadata collection at the P05 beamline at Hereon, DESY. Since 2016, we have relied on Java-based solutions to generate HDF5 files for hundreds of experiments with minimal maintenance overhead.

      We will provide a practical overview of how to use the Java HDF5 library effectively in the context of beamline operations, covering:
      – Creating and organizing datasets and groups
      – Handling compound datatypes and attributes
      – Performance considerations and memory management
      – Integration with control systems such as Tango and TINE

      This talk is intended for developers and beamline scientists interested in building robust data workflows using Java, without relying on native code bindings. The approach also supports long-term maintainability and ease of integration into existing Java-based infrastructures.

      Speaker: Igor Khokhriakov (FS-SC (Scientific computing))
    • 11:25 11:50
      Enhancing HDF5 with Multi-Threading, Sparse Data Storage, and Encryption: Preliminary Results and Demonstrations 25m

      Over the past few years, Lifeboat LLC has been focused on advancing the capabilities of HDF5 by incorporating multi-threaded support, enhancing the storage of sparse and variable-length data, and implementing robust encryption mechanisms for data stored within HDF5 files. These improvements are aimed at optimizing performance, increasing flexibility, and strengthening data security.
      In our presentation, we will share the preliminary results of our work on multi-threading and sparse data storage. Specifically, we will showcase the progress made in enabling efficient multi-threaded I/O operations and managing sparse datasets more effectively. Additionally, we will demonstrate the newly introduced encryption feature for securing sensitive data within HDF5, ensuring both confidentiality and integrity.
      We look forward to engaging with the community on these developments and receiving valuable feedback to further refine these features.

      Key words: HDF5 multi-threading, sparse data, variable-size data, data encryption

      Speaker: Elena Pourmal (Lifeboat, LLC)
    • 11:50 13:20
      Lunch 1h 30m
    • 13:20 14:00
      Keynote: From Data Blocks to Code Snippets: Boosting HDF5 with GitHub Copilot 40m

      Working with HDF5 often means navigating large datasets, verbose APIs, and boilerplate-heavy code. In this demo-heavy session, we’ll explore how AI-assisted coding tools—specifically GitHub Copilot—can accelerate common HDF5 workflows across C, C++, and Python. From auto-generating read/write boilerplate, to documenting complex structures, to scaffolding tests and data conversion routines, Copilot is more than autocomplete—it’s your new lab assistant.

      We’ll walk through examples, showing where Copilot shines, where it struggles, and how to make the most of it in a scientific or engineering context.

      Whether you’re wrangling metadata, porting legacy scripts, or just tired of writing H5Fopen() for the hundredth time—this session will show you how AI can actually help.

      Speaker: Julia Kordick (Microsoft)
    • 14:00 14:25
      Shiver Me Timbers: The Design of Sharded Storage for HDF5 25m

      As object storage becomes even more prevalent, HDF5's underlying storage format needs to updated to match the interface that cloud and on-prem object systems provide. This talk will present a design overview of a mapping of the HDF5 data model onto S3-compatible storage systems. An outline of the planned VOL connector implementation and projected performance goals will be part of the talk.

      Speaker: Mr Quincey Koziol (NVIDIA)
    • 14:25 15:15
      HDF5 Viewer Roundtable 50m

      A very common need when presented with an HDF5 file (especially for non-programmers or those new to HDF5), is some way to “see” the contents. Happily, there are a variety of viewers for HDF5 data sources available: HDFView, H5Web (myhdf5), HDF Compass, etc. However, it’s not obvious how these compare or when one or another might be preferable for a particular application. In this session, we’ll have a short review of HDFView, H5Web, and HDF Compass, covering their history, features, intended use, and future roadmap. We will conclude with an open discussion of what future steps would best serve the needs of the HDF5 community.

      Speakers: Axel Bocciarelli (ESRF), Gerd Heber (The HDF Group), John Readey (The HDF Group), Thomas VINCENT (ESRF)
    • 15:15 15:35
      Coffee Break 20m
    • 15:35 16:35
      HDF5 Roundtable

      The HDF5 Roundtable is an open, interactive forum hosted by The HDF Group during HUG events, designed to bring together users, developers, and stakeholders from the HDF5 community. This session provides a space for sharing insights, discussing current challenges, and exploring the future direction of HDF5. Whether you're a new user or a longtime contributor, the Roundtable is a collaborative opportunity to ask questions, offer feedback, and help shape the evolution of the HDF5 ecosystem.

      • 15:35
        Coffee Break 20m
    • 09:00 09:25
      Accelerating Data Compression in HDF5 through Parallel Filter Processing 25m

      Modern science and engineering creates and accumulates huge amounts of data which is
      persisted through tools like HDF5 in order to be available for further analysis, display
      and many other operations. Increasing efficiency in this data processing is critical for
      nowadays growing data quantities, not only for saving time, but also to efficiently use
      available resources.
      This thesis aimed to provide a working prototype and analysis on parallel application of
      data filters within the HDF5 environment with special emphasis on HDF5 registered
      filters, such as LZ4 compression.
      This prototype is embedded into the HDF5 framework, can be freely accessed as any
      other provided function and is available ready to use after building the project, while
      maintaining standard library behavior for all other use cases. Generally implementation
      is based on an all purpose thread pool with variable functionality based on registered
      callback function and can therefore be used in later versions of development. Analysis is
      based on comparison of standard library performance and prototype, based on identical
      example datasets, as well as statical analysis of provided program code. Furthermore
      CPU utilization and I/O performance are evaluated.
      Results suggest a great potential for a fully implemented design including most capabilities
      of the stock HDF5 library. Near full CPU utilisation is shown with little to no wait for
      I/O completion and therefore cutting down runtime by quite an extensive amount.
      This prototype, related work and analysis show the great improvement possible by
      adjusting the existing framework to a multithreaded solution while still maintaining full
      standard behaviour.

      Speaker: Frederick Neu
    • 09:25 09:50
      Zarr access via NetCDF-C 25m

      With the growing adoption of the Zarr data format for scalable and cloud-optimized storage, NetCDF has introduced an interface to support Zarr access. This integration enables a broader range of scientific software, beyond the Python ecosystem, to interact with Zarr datasets through the familiar NetCDF API. In this presentation, we will discuss the current state of the NetCDF-Zarr implementation, explore opportunities for optimizing I/O patterns, and highlight desirable directions for future support of remote data access features.

      Speaker: Manuel Reis (Deutsches Klimarechenzentrum - DKRZ)
    • 09:50 10:10
      Coffee Break 20m
    • 10:10 10:35
      Synergizing Science and Data: IOwarp for Scalable Workflow Management 25m

      Scientific workflows are evolving rapidly, demanding the seamless integration of simulation, experiments, analytics, and AI. This evolution is placing immense pressure on traditional data management systems. To address these challenges, we present IOwarp, a new initiative focused on building a comprehensive data management platform. IOwarp aims to streamline complex scientific workflows by providing scalable and efficient data access, movement, transformation, and sharing capabilities. We will discuss the core architecture of IOwarp and will highlight the complementary roles of IOwarp in modern scientific data management. We welcome feedback and collaboration to further develop and refine these tools for the scientific community.

      Speaker: Scot Breitenfeld (The HDF Group)
    • 10:35 11:00
      Streamlining HDF5's AI Workloads Benchmarking 25m

      Rapid adoption of artificial intelligence (AI) in scientific computing requires new tools to evaluate I/O performance effectively. HDF5 is one of the data formats frequently used not only in HPC but also in modern AI applications. However, existing benchmarks are insufficient to address the current challenges posed by AI workloads. This talk introduces an extension to the existing HDF5 benchmark, called h5bench, by incorporating the workload characteristics from the MLPerf Storage - DLIO benchmark. This extension allows users to test AI workloads together with traditional HPC benchmarks in the same context without the complexities of installing various machine learning libraries. Our experimental analysis demonstrates that the extension can replicate the existing I/O patterns with easily customizable configurations to perform various scaling tests.

      Speaker: Dlyaver Djebarov
    • 11:00 14:00
      Facility Tour 3h

      Come on a tour of the DESY facility. We will leave at 11:00 a.m. in two groups of 20.