The code repository for the NeurIPS 2022 paper PDEBench: An Extensive Benchmark for Scientific Machine Learning
PDEBench provides a diverse and comprehensive set of benchmarks for scientific machine learning, including challenging and realistic physical problems. The repository consists of the code used to generate the datasets, to upload and download the datasets from the data repository, as well as to train and evaluate different machine learning models as baseline. PDEBench features a much wider range of PDEs than existing benchmarks and included realistic and difficult problems (both forward and inverse), larger ready-to-use datasets comprising various initial and boundary conditions, and PDE parameters. Moreover, PDEBench was crated to make the source code extensible and we invite active participation to improve and extent the benchmark.
Created and maintained by Makoto Takamoto <[email protected], [email protected]>
, Timothy Praditia <[email protected]>
, Raphael Leiteritz, Dan MacKinlay, Francesco Alesiani, Dirk Pflüger and Mathias Niepert.
We also provide datasets and pretrained machine learning models.
PDEBench Datasets: https://darus.uni-stuttgart.de/dataset.xhtml?persistentId=doi:10.18419/darus-2986
PDEBench Pre-Trained Models: https://darus.uni-stuttgart.de/dataset.xhtml?persistentId=doi:10.18419/darus-2987
DOIs
pip install .
python3 -m venv ./venv --prompt pde_benchmark --system-site-packages
. ./venv/bin/activate
pip install --upgrade pip wheel
pip install -r requirements.txt
The minimum required packages to train and run the baseline ML models are listed in requirements.txt.
To run the data generation scripts, the complete package requirements are listed in requirements_datagen.txt
For GPU support there are additional platform-specific instructions:
For PyTorch, see here.
For JAX, which is approximately 6 times faster for simulations than PyTorch in our tests, see here
If you like you can also install dependencies using anaconda. We suggest using miniforge (and possibly mamba) as distribution. Otherwise you may have to enable the conda-forge channel for the following commands.
Starting from a fresh environment:
conda create -n myenv python=3.9
conda activate myenv
Install dependencies for model training:
conda install deepxde hydra-core h5py
Install PyTorch according to your hardware requirements:
E.g.
conda install pytorch torchvision torchaudio cudatoolkit=11.3 -c pytorch
Optional dependencies for data generation:
conda install clawpack jax jaxlib python-dotenv
Optional dependencies for data downloading:
pip install pyDarus~=1.0.5
In our tests we used PyTorch as backend for DeepXDE. Please follow the documentation to enable this.
The data generation codes are contained in data_gen:
gen_diff_react.py
to generate the 2D diffusion-reaction data.gen_diff_sorp.py
to generate the 1D diffusion-sorption data.gen_radial_dam_break.py
to generate the 2D shallow-water data.gen_ns_incomp.py
to generate the 2D incompressible inhomogenous Navier-Stokes data.plot.py
to plot the generated data.uploader.py
to upload the generated data to the data repository..env
is the environment data to store Dataverse URL and API token to upload the generated data. Note that the filename should be strictly.env
(i.e. remove theexample
from the filename)configs
directory contains the yaml files storing the configuration for the simulation. Arguments for the simulation are problem-specific and detailed explanation can be found in the simulation scripts.src
directory contains the simulation scripts for different problems:sim_diff_react-py
for 2D diffusion-reaction,sim_diff_sorp.py
for 1D diffusion-sorption, andswe
for the shallow-water equation.
Data Generation for 1D Advection/Burgers/Reaction-Diffusion/2D DarcyFlow/Compressible Navier-Stokes Equations
The data generation codes are contained in data_gen_NLE:
utils.py
util file for data generation, mainly boundary conditions and initial conditions.AdvectionEq
directory with the source codes to generate 1D Advection equation training samplesBurgersEq
directory with the source codes to generate 1D Burgers equation training samplesCompressibleFluid
directory with the source codes to generate compressible Navier-Stokes equations training samplesReactionDiffusionEq
directory with the source codes to generate 1D Reaction-Diffusion equation training samplessave
directory saving the generated training samples
A typical example to generate training samples (1D Advection Equation):
(in data_gen/data_gen_NLE/AdvectionEq/
)
python3 advection_multi_solution_Hydra.py +multi=beta1e0.yaml
which is assumed to be performed in each directory.
Examples for generating other PDEs are provided in run_trainset.sh
in each PDE's directories.
The config files for Hydra are stored in config
directory in each PDE's directory.
1D Advection/Burgers/Reaction-Diffusion/2D DarcyFlow/Compressible Navier-Stokes Equations save data as a numpy array.
So, to read those data via our dataloaders, the data transformation/merge should be performed.
This can be done using data_gen_NLE/Data_Merge.py
whose config file is located at: data_gen/data_gen_NLE/config/config.yaml
.
After properly set parameters in the config file (type: name of PDEs, dim: number of spatial-dimension, bd: boundary condition),
the corresponding HDF5 file could be obtained as:
python3 Data_Merge.py
You can set the default values for data locations for this project by putting config vars like this in the .env
file:
WORKING_DIR=~/Data/Working
ARCHIVE_DATA_DIR=~/Data/Archive
There is an example in example.env
.
The download scripts are provided in data_download. There are two options to download data.
- Using
download_direct.py
(recommended)- Retrieves data shards directly using URLs. Sample command for each PDE is given the README file in the data_download directory.
- Using
download_easydataverse.py
(might encounter errors/issues, hence not recommended)- Use the config files from the
config
directory that contains the yaml files storing the configuration. Any files in the dataset matchingargs.filename
will be downloaded intoargs.data_folder
.
- Use the config files from the
In this work, we provide three different ML models to be trained and evaluated against the benchmark datasets, namely FNO, U-Net, and PINN. The codes for the baseline model implementations are contained in models:
train_models_forward.py
is the main script to train and evaluate the model. It will call on model-specific script based on the input argument.train_models_inverse.py
is the main script to train and evaluate the model for inverse problems. It will call on model-specific script based on the input argument.metrics.py
is the script to evaluate the trained models based on various evaluation metrics described in our paper. Additionally, it also plots the prediction and target data.analyse_result_forward.py
is the script to convert the saved pickle file from the metrics calculation script into pandas dataframe format and save it as a CSV file. Additionally it also plots a bar chart to compare the results between different models.analyse_result_inverse.py
is the script to convert the saved pickle file from the metrics calculation script into pandas dataframe format and save it as a CSV file. This script is used for the inverse problems. Additionally it also plots a bar chart to compare the results between different models.fno
contains the scripts of FNO implementation. These are partly adapted from the FNO repository.unet
contains the scripts of U-Net implementation. These are partly adapted from the U-Net repository.pinn
contains the scripts of PINN implementation. These utilize the DeepXDE library.inverse
contains the model for inverse model based on gradient.config
contains the yaml files for the model training input. The default templates for different equations are provided in the args directory. User just needs to copy and paste them to the args keyword in the config.yaml file.
An example to run the forward model training can be found in run_forward_1D.sh, and an example to run the inverse model training can be found in run_inverse.sh.
- model_name: string, containing the baseline model name, either 'FNO', 'Unet', or 'PINN'.
- if_training: bool, set True for training, or False for evaluation.
- continue_training: bool, set True to continute training from a checkpoint.
- num_workers: int, number of workers for the PyTorch dataloader.
- batch_size: int, training batch size.
- initial_step: int, number of time steps used as input for FNO and U-Net.
- t_train: int, number of the last time step used for training (for extrapolation testing, set this to be < Nt).
- model_update: int, number of epochs to save model.
- filename: str, has to match the dataset filename.
- single_file: bool, set False for 2D diffusion-reaction, 1D diffusion-sorption, and the radial dam break scenarios, and set True otherwise.
- reduced_resolution: int, factor to downsample spatial resolution.
- reduced_resolution_t: int, factor to downsample temporal resolution.
- reduced_batch: int, factor to downsample sample size used for training.
- epochs: int, total epochs used for training.
- learning_rate: float, learning rate of the optimizer.
- scheduler_step: int, number of epochs to update the learning rate scheduler.
- scheduler_gamma: float, decay rate of the learning rate.
- in_channels: int, number of input channels
- out_channels: int, number of output channels
- ar_mode: bool, set True for fully autoregressive or pushforward training.
- pushforward: bool, set True for pushforward training, False otherwise (ar_mode also has to be set True).
- unroll_step: int, number of time steps to backpropagate in the pushforward training.
- num_channels: int, number of channels (variables).
- modes: int, number of Fourier modes to multiply.
- width: int, number of channels for the Fourier layer.
- base_path: string, location of the data directory
- training_type: string, type of training, autoregressive, single
- mcmc_num_samples: int, number of generated samples
- mcmc_warmup_steps: 10
- mcmc_num_chains: 1
- num_samples_max: 1000
- in_channels_hid: 64
- inverse_model_type: string, type of inverse inference model, ProbRasterLatent, InitialConditionInterp
- inverse_epochs: int, number of epochs for the gradient based method
- inverse_learning_rate: float, learning rate for the gradient based method
- inverse_verbose_flag: bool, some printing
- plot: bool, set True to activate plotting.
- channel_plot: int, determines which channel/variable to plot.
- x_min: float, left spatial domain.
- x_max: float, right spatial domain.
- y_min: float, lower spatial domain.
- y_max: float, upper spatial domain.
- t_min: float, start of temporal domain.
- t_max: float, end of temporal domain.
We provide the benchmark datasets we used in the paper through our DaRUS data repository.
The data generation configuration can be found in the paper.
Additionally, the pretrained models are also available to be downloaded from PDEBench Pretrained Models DaRus repository. To use the pretrained models, users can specify the argument continue_training: True
in the config file.
@inproceedings{PDEBench2022,
author = {Takamoto, Makoto and Praditia, Timothy and Leiteritz, Raphael and MacKinlay, Dan and Alesiani, Francesco and Pflüger, Dirk and Niepert, Mathias},
title = {{PDEBench: An Extensive Benchmark for Scientific Machine Learning}},
year = {2022},
booktitle = {36th Conference on Neural Information Processing Systems (NeurIPS 2022) Track on Datasets and Benchmarks},
url = {https://doi.org/10.18419/darus-2986}
}
@data{darus-2986_2022,
author = {Takamoto, Makoto and Praditia, Timothy and Leiteritz, Raphael and MacKinlay, Dan and Alesiani, Francesco and Pflüger, Dirk and Niepert, Mathias},
publisher = {DaRUS},
title = {{PDEBench Datasets}},
year = {2022},
doi = {10.18419/darus-2986},
url = {https://doi.org/10.18419/darus-2986}
}
- Makato Takamoto
- Timothy Praditia
- Raphael Leiteritz
- Francesco Alesiani
- Dan MacKinlay
- John Kim
- Gefei Shan
- Yizhou Yang
- Ran Zhang
- Simon Brown
MIT for solver code and baseline code, and NLE Academic License for selected code