In our testing, only small systems are affected (not your case), but the exact cause and impact are still being investigated. So, for now, the combination is disabled.
We hope to fix the problem by 2025.1 or 2025.2. Also note that this issue issue affects 2024 series as well (when you have GMX_ENABLE_DIRECT_GPU_COMM set), but it was only very recently discovered, so no notices / workaround in GROMACS 2024 yet.
Going off a tangential: the environment variables (“exports”) are used to control experimental (for one reason or the other) features. They should be used with understanding, and not by enabling everything in hopes of things getting faster. It’s not like we as developers want to make your life harder by making you jump through extra hoops to get the best performance. In particular, going through your list:
GMX_ENABLE_DIRECT_GPU_COMM=1: this one is useful for 2024, but is not needed normally for 2025: either because it’s enabled by default (with libMPI), or because it’s disabled altogether due to the aforementioned bug (with threadMPI). This is the reason you get speed up with GROMACS 2024.
GMX_HEFFTE_USE_GPU_AWARE=1: this one only makes sense if you have libMPI (“real MPI”) build with HeFFTe and run with several PME ranks. You do neither of the three, so this setting does nothing for you.
GMX_GPU_PME_DECOMPOSITION=1: same as the previous one.
GMX_USE_GPU_BUFFER_OPS=1: this can have some effect with 2024 when you are not using GPU update; in 2025, it’s on by default. And you are very likely using GPU update, so this is, again, very likely does nothing.
GMX_DISABLE_GPU_TIMING=1: GPU timings are only enabled by default with OpenCL, so this does nothing with CUDA.
GMX_FORCE_UPDATE_DEFAULT_GPU=true: this option is the default behavior since GROMACS 2023, so it does nothing.
GMX_GPU_DD_COMMS=true: was removed as an individual option a few releases ago, now controlled by GMX_ENABLE_DIRECT_GPU_COMM
GMX_GPU_PME_PP_COMMS=true: same as the previous one.
Of course, there is no harm is setting the flags that do nothing (other than them doing something when you don’t expect). But why have eight lines in your script when one is enough?
Yes. There were even plans to include AMBER 19SB in GROMACS 2025, but it did not quite work out, testing took more time than planned. Having new AMBER forcefields in GROMACS 2026 is our current target.
Nope-nope-nope. Forcefield parameterisation is not on the roadmap :)
What we are working on is incorporating neural-network based potentials, so one could use them to improve the accuracy of ligand modeling (as an intermediate step between classical MD and QM/MM).
We plan to make the 2025.1 release, with this issue fixed, later this week.