Skip to content

Latest commit

 

History

History
39 lines (27 loc) · 2.25 KB

README.md

File metadata and controls

39 lines (27 loc) · 2.25 KB

Particle Mesh Simulation with JAXPM on Multi-GPU and Multi-Host Systems

This collection of notebooks demonstrates how to perform Particle Mesh (PM) simulations using JAXPM, leveraging JAX for efficient computation on multi-GPU and multi-host systems. Each notebook progressively covers different setups, from single-GPU simulations to advanced, distributed, multi-host simulations across multiple nodes.

Table of Contents

  1. Single-GPU Particle Mesh Simulation

    • Introduction to basic PM simulations on a single GPU.
    • Uses JAXPM to run simulations with absolute particle positions and Cloud-in-Cell (CIC) painting.
  2. Advanced Particle Mesh Simulation on a Single GPU

    • Explore using diffrax solvers in the ODE step.
    • Explores second order Lagrangian Perturbation Theory (LPT) simulations.
    • Introduces weighted density field projections
  3. Multi-GPU Particle Mesh Simulation with Halo Exchange

    • Extends PM simulation to multi-GPU setups with halo exchange.
    • Uses sharding and device mesh configurations to manage distributed data across GPUs.
  4. Multi-GPU Particle Mesh Simulation with Advanced Solvers

    • Compares different ODE solvers (Leapfrog and Dopri5) in multi-GPU simulations.
    • Highlights performance, memory considerations, and solver impact on simulation quality.
  5. Multi-Host Particle Mesh Simulation

    • Extends PM simulations to multi-host, multi-GPU setups for large-scale simulations.
    • Guides through job submission, device initialization, and retrieving results across nodes.

Getting Started

Each notebook includes installation instructions and guidelines for configuring JAXPM and required dependencies. Follow the setup instructions in each notebook to ensure an optimal environment.

Requirements

  • JAXPM (included in the installation commands within notebooks)
  • Diffrax for ODE solvers
  • JAX with CUDA support for multi-GPU or TPU setups
  • SLURM for job scheduling on clusters (if running multi-host setups)

Note: These notebooks are tested on the Jean Zay supercomputer and may require configuration changes for different HPC clusters.