Bug: pppm/dplr + fix dplr hangs or crashes with multi-MPI (except ntasks=1), including inconsistent behavior with restart and read_data #5362
Unanswered
Ziyang-You
asked this question in
Q&A
Replies: 0 comments
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
When using
pair_style deepmd+fix dplr+kspace_style pppm/dplr, the simulation only runs reliably with 1 MPI process (ntasks=1). With multiple MPI processes (ntasks ≥ 2), it frequently hangs indefinitely during PPPM initialization or crashes with MPI collective errors. The behavior is inconsistent — some systems (fromread_data) can run with multi-MPI, while others (especially fromread_restart) fail.This makes DPLR simulations impractical on HPC clusters, as single-MPI runs are extremely slow.
Environment
pip install deepmd-kit[cpu,lmp]Input script, sbatch inputs and out
in.zip
Beta Was this translation helpful? Give feedback.
All reactions