2 * This file is part of the GROMACS molecular simulation package.
4 * Copyright (c) 1991-2000, University of Groningen, The Netherlands.
5 * Copyright (c) 2001-2004, The GROMACS development team.
6 * Copyright (c) 2013,2014,2015,2016,2017,2018, by the GROMACS development team, led by
7 * Mark Abraham, David van der Spoel, Berk Hess, and Erik Lindahl,
8 * and including many others, as listed in the AUTHORS file in the
9 * top-level source directory and at http://www.gromacs.org.
11 * GROMACS is free software; you can redistribute it and/or
12 * modify it under the terms of the GNU Lesser General Public License
13 * as published by the Free Software Foundation; either version 2.1
14 * of the License, or (at your option) any later version.
16 * GROMACS is distributed in the hope that it will be useful,
17 * but WITHOUT ANY WARRANTY; without even the implied warranty of
18 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
19 * Lesser General Public License for more details.
21 * You should have received a copy of the GNU Lesser General Public
22 * License along with GROMACS; if not, see
23 * http://www.gnu.org/licenses, or write to the Free Software Foundation,
24 * Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA.
26 * If you want to redistribute modifications to GROMACS, please
27 * consider that scientific software is very special. Version
28 * control is crucial - bugs must be traceable. We will be happy to
29 * consider code for inclusion in the official distribution, but
30 * derived work must not be called official GROMACS. Details are found
31 * in the README & COPYING files - if they are missing, get the
32 * official version at http://www.gromacs.org.
34 * To help us fund GROMACS development, we humbly ask that you cite
35 * the research papers on the package. Check out http://www.gromacs.org.
37 /* IMPORTANT FOR DEVELOPERS:
39 * Triclinic pme stuff isn't entirely trivial, and we've experienced
40 * some bugs during development (many of them due to me). To avoid
41 * this in the future, please check the following things if you make
42 * changes in this file:
44 * 1. You should obtain identical (at least to the PME precision)
45 * energies, forces, and virial for
46 * a rectangular box and a triclinic one where the z (or y) axis is
47 * tilted a whole box side. For instance you could use these boxes:
49 * rectangular triclinic
54 * 2. You should check the energy conservation in a triclinic box.
56 * It might seem an overkill, but better safe than sorry.
74 #include "gromacs/compat/make_unique.h"
75 #include "gromacs/domdec/domdec.h"
76 #include "gromacs/ewald/pme.h"
77 #include "gromacs/fft/parallel_3dfft.h"
78 #include "gromacs/fileio/pdbio.h"
79 #include "gromacs/gmxlib/network.h"
80 #include "gromacs/gmxlib/nrnb.h"
81 #include "gromacs/gpu_utils/hostallocator.h"
82 #include "gromacs/math/gmxcomplex.h"
83 #include "gromacs/math/units.h"
84 #include "gromacs/math/vec.h"
85 #include "gromacs/mdtypes/commrec.h"
86 #include "gromacs/mdtypes/inputrec.h"
87 #include "gromacs/timing/cyclecounter.h"
88 #include "gromacs/timing/wallcycle.h"
89 #include "gromacs/utility/fatalerror.h"
90 #include "gromacs/utility/futil.h"
91 #include "gromacs/utility/gmxmpi.h"
92 #include "gromacs/utility/gmxomp.h"
93 #include "gromacs/utility/smalloc.h"
95 #include "pme-internal.h"
96 #include "pme-pp-communication.h"
98 //! Contains information about the PP ranks that partner this PME rank.
101 //! The MPI rank ID of this partner PP rank.
103 //! The number of atoms to communicate with this partner PP rank.
107 /*! \brief Master PP-PME communication data structure */
109 MPI_Comm mpi_comm_mysim
; /**< MPI communicator for this simulation */
110 std::vector
<PpRanks
> ppRanks
; /**< The PP partner ranks */
111 int peerRankId
; /**< The peer PP rank id */
113 /**< Vectors of A- and B-state parameters used to transfer vectors to PME ranks */
114 gmx::HostVector
<real
> chargeA
;
115 std::vector
<real
> chargeB
;
116 std::vector
<real
> sqrt_c6A
;
117 std::vector
<real
> sqrt_c6B
;
118 std::vector
<real
> sigmaA
;
119 std::vector
<real
> sigmaB
;
121 gmx::HostVector
<gmx::RVec
> x
; /**< Vector of atom coordinates to transfer to PME ranks */
122 std::vector
<gmx::RVec
> f
; /**< Vector of atom forces received from PME ranks */
124 /**< Vectors of MPI objects used in non-blocking communication between multiple PP ranks per PME rank */
125 std::vector
<MPI_Request
> req
;
126 std::vector
<MPI_Status
> stat
;
130 /*! \brief Initialize the PME-only side of the PME <-> PP communication */
131 static std::unique_ptr
<gmx_pme_pp
> gmx_pme_pp_init(const t_commrec
*cr
)
133 auto pme_pp
= gmx::compat::make_unique
<gmx_pme_pp
>();
138 pme_pp
->mpi_comm_mysim
= cr
->mpi_comm_mysim
;
139 MPI_Comm_rank(cr
->mpi_comm_mygroup
, &rank
);
140 auto ppRanks
= get_pme_ddranks(cr
, rank
);
141 pme_pp
->ppRanks
.reserve(ppRanks
.size());
142 for (const auto &ppRankId
: ppRanks
)
144 pme_pp
->ppRanks
.push_back({ppRankId
, 0});
146 // The peer PP rank is the last one.
147 pme_pp
->peerRankId
= pme_pp
->ppRanks
.back().rankId
;
148 pme_pp
->req
.resize(eCommType_NR
*pme_pp
->ppRanks
.size());
149 pme_pp
->stat
.resize(eCommType_NR
*pme_pp
->ppRanks
.size());
151 GMX_UNUSED_VALUE(cr
);
157 static void reset_pmeonly_counters(gmx_wallcycle_t wcycle
,
158 gmx_walltime_accounting_t walltime_accounting
,
163 /* Reset all the counters related to performance over the run */
164 wallcycle_stop(wcycle
, ewcRUN
);
165 wallcycle_reset_all(wcycle
);
167 wallcycle_start(wcycle
, ewcRUN
);
168 walltime_accounting_reset_time(walltime_accounting
, step
);
176 static gmx_pme_t
*gmx_pmeonly_switch(std::vector
<gmx_pme_t
*> *pmedata
,
177 const ivec grid_size
,
178 real ewaldcoeff_q
, real ewaldcoeff_lj
,
179 const t_commrec
*cr
, const t_inputrec
*ir
)
181 GMX_ASSERT(pmedata
, "Bad PME tuning list pointer");
182 for (auto &pme
: *pmedata
)
184 GMX_ASSERT(pme
, "Bad PME tuning list element pointer");
185 if (pme
->nkx
== grid_size
[XX
] &&
186 pme
->nky
== grid_size
[YY
] &&
187 pme
->nkz
== grid_size
[ZZ
])
189 /* Here we have found an existing PME data structure that suits us.
190 * However, in the GPU case, we have to reinitialize it - there's only one GPU structure.
191 * This should not cause actual GPU reallocations, at least (the allocated buffers are never shrunk).
192 * So, just some grid size updates in the GPU kernel parameters.
193 * TODO: this should be something like gmx_pme_update_split_params()
195 gmx_pme_reinit(&pme
, cr
, pme
, ir
, grid_size
, ewaldcoeff_q
, ewaldcoeff_lj
);
200 const auto &pme
= pmedata
->back();
201 gmx_pme_t
*newStructure
= nullptr;
202 // Copy last structure with new grid params
203 gmx_pme_reinit(&newStructure
, cr
, pme
, ir
, grid_size
, ewaldcoeff_q
, ewaldcoeff_lj
);
204 pmedata
->push_back(newStructure
);
208 /*! \brief Called by PME-only ranks to receive coefficients and coordinates
210 * \param[in,out] pme_pp PME-PP communication structure.
211 * \param[out] natoms Number of received atoms.
212 * \param[out] box System box, if received.
213 * \param[out] maxshift_x Maximum shift in X direction, if received.
214 * \param[out] maxshift_y Maximum shift in Y direction, if received.
215 * \param[out] lambda_q Free-energy lambda for electrostatics, if received.
216 * \param[out] lambda_lj Free-energy lambda for Lennard-Jones, if received.
217 * \param[out] bEnerVir Set to true if this is an energy/virial calculation step, otherwise set to false.
218 * \param[out] step MD integration step number.
219 * \param[out] grid_size PME grid size, if received.
220 * \param[out] ewaldcoeff_q Ewald cut-off parameter for electrostatics, if received.
221 * \param[out] ewaldcoeff_lj Ewald cut-off parameter for Lennard-Jones, if received.
222 * \param[out] atomSetChanged Set to true only if the local domain atom data (charges/coefficients)
223 * has been received (after DD) and should be reinitialized. Otherwise not changed.
225 * \retval pmerecvqxX All parameters were set, chargeA and chargeB can be NULL.
226 * \retval pmerecvqxFINISH No parameters were set.
227 * \retval pmerecvqxSWITCHGRID Only grid_size and *ewaldcoeff were set.
228 * \retval pmerecvqxRESETCOUNTERS *step was set.
230 static int gmx_pme_recv_coeffs_coords(gmx_pme_pp
*pme_pp
,
242 bool *atomSetChanged
)
248 unsigned int flags
= 0;
253 gmx_pme_comm_n_box_t cnb
;
256 /* Receive the send count, box and time step from the peer PP node */
257 MPI_Recv(&cnb
, sizeof(cnb
), MPI_BYTE
,
258 pme_pp
->peerRankId
, eCommType_CNB
,
259 pme_pp
->mpi_comm_mysim
, MPI_STATUS_IGNORE
);
261 /* We accumulate all received flags */
268 fprintf(debug
, "PME only rank receiving:%s%s%s%s%s\n",
269 (cnb
.flags
& PP_PME_CHARGE
) ? " charges" : "",
270 (cnb
.flags
& PP_PME_COORD
) ? " coordinates" : "",
271 (cnb
.flags
& PP_PME_FINISH
) ? " finish" : "",
272 (cnb
.flags
& PP_PME_SWITCHGRID
) ? " switch grid" : "",
273 (cnb
.flags
& PP_PME_RESETCOUNTERS
) ? " reset counters" : "");
276 if (cnb
.flags
& PP_PME_FINISH
)
278 status
= pmerecvqxFINISH
;
281 if (cnb
.flags
& PP_PME_SWITCHGRID
)
283 /* Special case, receive the new parameters and return */
284 copy_ivec(cnb
.grid_size
, *grid_size
);
285 *ewaldcoeff_q
= cnb
.ewaldcoeff_q
;
286 *ewaldcoeff_lj
= cnb
.ewaldcoeff_lj
;
288 status
= pmerecvqxSWITCHGRID
;
291 if (cnb
.flags
& PP_PME_RESETCOUNTERS
)
293 /* Special case, receive the step (set above) and return */
294 status
= pmerecvqxRESETCOUNTERS
;
297 if (cnb
.flags
& (PP_PME_CHARGE
| PP_PME_SQRTC6
| PP_PME_SIGMA
))
299 *atomSetChanged
= true;
301 /* Receive the send counts from the other PP nodes */
302 for (auto &sender
: pme_pp
->ppRanks
)
304 if (sender
.rankId
== pme_pp
->peerRankId
)
306 sender
.numAtoms
= cnb
.natoms
;
310 MPI_Irecv(&sender
.numAtoms
, sizeof(sender
.numAtoms
),
312 sender
.rankId
, eCommType_CNB
,
313 pme_pp
->mpi_comm_mysim
, &pme_pp
->req
[messages
++]);
316 MPI_Waitall(messages
, pme_pp
->req
.data(), pme_pp
->stat
.data());
320 for (const auto &sender
: pme_pp
->ppRanks
)
322 nat
+= sender
.numAtoms
;
325 if (cnb
.flags
& PP_PME_CHARGE
)
327 pme_pp
->chargeA
.resizeWithPadding(nat
);
329 if (cnb
.flags
& PP_PME_CHARGEB
)
331 pme_pp
->chargeB
.resize(nat
);
333 if (cnb
.flags
& PP_PME_SQRTC6
)
335 pme_pp
->sqrt_c6A
.resize(nat
);
337 if (cnb
.flags
& PP_PME_SQRTC6B
)
339 pme_pp
->sqrt_c6B
.resize(nat
);
341 if (cnb
.flags
& PP_PME_SIGMA
)
343 pme_pp
->sigmaA
.resize(nat
);
345 if (cnb
.flags
& PP_PME_SIGMAB
)
347 pme_pp
->sigmaB
.resize(nat
);
349 pme_pp
->x
.resizeWithPadding(nat
);
350 pme_pp
->f
.resize(nat
);
352 /* maxshift is sent when the charges are sent */
353 *maxshift_x
= cnb
.maxshift_x
;
354 *maxshift_y
= cnb
.maxshift_y
;
356 /* Receive the charges in place */
357 for (int q
= 0; q
< eCommType_NR
; q
++)
361 if (!(cnb
.flags
& (PP_PME_CHARGE
<<q
)))
367 case eCommType_ChargeA
: bufferPtr
= pme_pp
->chargeA
.data(); break;
368 case eCommType_ChargeB
: bufferPtr
= pme_pp
->chargeB
.data(); break;
369 case eCommType_SQRTC6A
: bufferPtr
= pme_pp
->sqrt_c6A
.data(); break;
370 case eCommType_SQRTC6B
: bufferPtr
= pme_pp
->sqrt_c6B
.data(); break;
371 case eCommType_SigmaA
: bufferPtr
= pme_pp
->sigmaA
.data(); break;
372 case eCommType_SigmaB
: bufferPtr
= pme_pp
->sigmaB
.data(); break;
373 default: gmx_incons("Wrong eCommType");
376 for (const auto &sender
: pme_pp
->ppRanks
)
378 if (sender
.numAtoms
> 0)
380 MPI_Irecv(bufferPtr
+nat
,
381 sender
.numAtoms
*sizeof(real
),
384 pme_pp
->mpi_comm_mysim
,
385 &pme_pp
->req
[messages
++]);
386 nat
+= sender
.numAtoms
;
389 fprintf(debug
, "Received from PP rank %d: %d %s\n",
390 sender
.rankId
, sender
.numAtoms
,
391 (q
== eCommType_ChargeA
||
392 q
== eCommType_ChargeB
) ? "charges" : "params");
399 if (cnb
.flags
& PP_PME_COORD
)
401 /* The box, FE flag and lambda are sent along with the coordinates
403 copy_mat(cnb
.box
, box
);
404 *lambda_q
= cnb
.lambda_q
;
405 *lambda_lj
= cnb
.lambda_lj
;
406 *bEnerVir
= ((cnb
.flags
& PP_PME_ENER_VIR
) != 0u);
409 /* Receive the coordinates in place */
411 for (const auto &sender
: pme_pp
->ppRanks
)
413 if (sender
.numAtoms
> 0)
415 MPI_Irecv(pme_pp
->x
[nat
],
416 sender
.numAtoms
*sizeof(rvec
),
418 sender
.rankId
, eCommType_COORD
,
419 pme_pp
->mpi_comm_mysim
, &pme_pp
->req
[messages
++]);
420 nat
+= sender
.numAtoms
;
423 fprintf(debug
, "Received from PP rank %d: %d "
425 sender
.rankId
, sender
.numAtoms
);
433 /* Wait for the coordinates and/or charges to arrive */
434 MPI_Waitall(messages
, pme_pp
->req
.data(), pme_pp
->stat
.data());
437 while (status
== -1);
439 GMX_UNUSED_VALUE(pme_pp
);
440 GMX_UNUSED_VALUE(box
);
441 GMX_UNUSED_VALUE(maxshift_x
);
442 GMX_UNUSED_VALUE(maxshift_y
);
443 GMX_UNUSED_VALUE(lambda_q
);
444 GMX_UNUSED_VALUE(lambda_lj
);
445 GMX_UNUSED_VALUE(bEnerVir
);
446 GMX_UNUSED_VALUE(step
);
447 GMX_UNUSED_VALUE(grid_size
);
448 GMX_UNUSED_VALUE(ewaldcoeff_q
);
449 GMX_UNUSED_VALUE(ewaldcoeff_lj
);
450 GMX_UNUSED_VALUE(atomSetChanged
);
455 if (status
== pmerecvqxX
)
463 /*! \brief Send the PME mesh force, virial and energy to the PP-only ranks. */
464 static void gmx_pme_send_force_vir_ener(gmx_pme_pp
*pme_pp
,
466 matrix vir_q
, real energy_q
,
467 matrix vir_lj
, real energy_lj
,
468 real dvdlambda_q
, real dvdlambda_lj
,
472 gmx_pme_comm_vir_ene_t cve
;
473 int messages
, ind_start
, ind_end
;
476 /* Now the evaluated forces have to be transferred to the PP nodes */
479 for (const auto &receiver
: pme_pp
->ppRanks
)
482 ind_end
= ind_start
+ receiver
.numAtoms
;
483 if (MPI_Isend(const_cast<void *>(static_cast<const void *>(f
[ind_start
])),
484 (ind_end
-ind_start
)*sizeof(rvec
), MPI_BYTE
,
486 pme_pp
->mpi_comm_mysim
, &pme_pp
->req
[messages
++]) != 0)
488 gmx_comm("MPI_Isend failed in do_pmeonly");
492 /* send virial and energy to our last PP node */
493 copy_mat(vir_q
, cve
.vir_q
);
494 copy_mat(vir_lj
, cve
.vir_lj
);
495 cve
.energy_q
= energy_q
;
496 cve
.energy_lj
= energy_lj
;
497 cve
.dvdlambda_q
= dvdlambda_q
;
498 cve
.dvdlambda_lj
= dvdlambda_lj
;
499 /* check for the signals to send back to a PP node */
500 cve
.stop_cond
= gmx_get_stop_condition();
506 fprintf(debug
, "PME rank sending to PP rank %d: virial and energy\n",
509 MPI_Isend(&cve
, sizeof(cve
), MPI_BYTE
,
510 pme_pp
->peerRankId
, 1,
511 pme_pp
->mpi_comm_mysim
, &pme_pp
->req
[messages
++]);
513 /* Wait for the forces to arrive */
514 MPI_Waitall(messages
, pme_pp
->req
.data(), pme_pp
->stat
.data());
516 gmx_call("MPI not enabled");
517 GMX_UNUSED_VALUE(pme_pp
);
519 GMX_UNUSED_VALUE(vir_q
);
520 GMX_UNUSED_VALUE(energy_q
);
521 GMX_UNUSED_VALUE(vir_lj
);
522 GMX_UNUSED_VALUE(energy_lj
);
523 GMX_UNUSED_VALUE(dvdlambda_q
);
524 GMX_UNUSED_VALUE(dvdlambda_lj
);
525 GMX_UNUSED_VALUE(cycles
);
529 int gmx_pmeonly(struct gmx_pme_t
*pme
,
530 const t_commrec
*cr
, t_nrnb
*mynrnb
,
531 gmx_wallcycle
*wcycle
,
532 gmx_walltime_accounting_t walltime_accounting
,
533 t_inputrec
*ir
, PmeRunMode runMode
)
540 int maxshift_x
= 0, maxshift_y
= 0;
541 real energy_q
, energy_lj
, dvdlambda_q
, dvdlambda_lj
;
542 matrix vir_q
, vir_lj
;
545 gmx_bool bEnerVir
= FALSE
;
548 /* This data will only use with PME tuning, i.e. switching PME grids */
549 std::vector
<gmx_pme_t
*> pmedata
;
550 pmedata
.push_back(pme
);
552 auto pme_pp
= gmx_pme_pp_init(cr
);
553 //TODO the variable below should be queried from the task assignment info
554 const bool useGpuForPme
= (runMode
== PmeRunMode::GPU
) || (runMode
== PmeRunMode::Mixed
);
557 changePinningPolicy(&pme_pp
->chargeA
, pme_get_pinning_policy());
558 changePinningPolicy(&pme_pp
->x
, pme_get_pinning_policy());
564 do /****** this is a quasi-loop over time steps! */
566 /* The reason for having a loop here is PME grid tuning/switching */
569 /* Domain decomposition */
571 bool atomSetChanged
= false;
572 real ewaldcoeff_q
= 0, ewaldcoeff_lj
= 0;
573 ret
= gmx_pme_recv_coeffs_coords(pme_pp
.get(),
576 &maxshift_x
, &maxshift_y
,
577 &lambda_q
, &lambda_lj
,
585 if (ret
== pmerecvqxSWITCHGRID
)
587 /* Switch the PME grid to newGridSize */
588 pme
= gmx_pmeonly_switch(&pmedata
, newGridSize
, ewaldcoeff_q
, ewaldcoeff_lj
, cr
, ir
);
593 gmx_pme_reinit_atoms(pme
, natoms
, pme_pp
->chargeA
.data());
596 if (ret
== pmerecvqxRESETCOUNTERS
)
598 /* Reset the cycle and flop counters */
599 reset_pmeonly_counters(wcycle
, walltime_accounting
, mynrnb
, step
, useGpuForPme
);
602 while (ret
== pmerecvqxSWITCHGRID
|| ret
== pmerecvqxRESETCOUNTERS
);
604 if (ret
== pmerecvqxFINISH
)
606 /* We should stop: break out of the loop */
612 wallcycle_start(wcycle
, ewcRUN
);
613 walltime_accounting_start_time(walltime_accounting
);
616 wallcycle_start(wcycle
, ewcPMEMESH
);
625 // TODO Make a struct of array refs onto these per-atom fields
626 // of pme_pp (maybe box, energy and virial, too; and likewise
627 // from mdatoms for the other call to gmx_pme_do), so we have
628 // fewer lines of code and less parameter passing.
629 const int pmeFlags
= GMX_PME_DO_ALL_F
| (bEnerVir
? GMX_PME_CALC_ENER_VIR
: 0);
630 gmx::ArrayRef
<const gmx::RVec
> forces
;
633 const bool boxChanged
= false;
634 //TODO this should be set properly by gmx_pme_recv_coeffs_coords,
635 // or maybe use inputrecDynamicBox(ir), at the very least - change this when this codepath is tested!
636 pme_gpu_prepare_computation(pme
, boxChanged
, box
, wcycle
, pmeFlags
);
637 pme_gpu_launch_spread(pme
, pme_pp
->x
.rvec_array(), wcycle
);
638 pme_gpu_launch_complex_transforms(pme
, wcycle
);
639 pme_gpu_launch_gather(pme
, wcycle
, PmeForceOutputHandling::Set
);
640 pme_gpu_wait_finish_task(pme
, wcycle
, &forces
, vir_q
, &energy_q
);
641 pme_gpu_reinit_computation(pme
, wcycle
);
645 gmx_pme_do(pme
, 0, natoms
, pme_pp
->x
.rvec_array(), as_rvec_array(pme_pp
->f
.data()),
646 pme_pp
->chargeA
.data(), pme_pp
->chargeB
.data(),
647 pme_pp
->sqrt_c6A
.data(), pme_pp
->sqrt_c6B
.data(),
648 pme_pp
->sigmaA
.data(), pme_pp
->sigmaB
.data(), box
,
649 cr
, maxshift_x
, maxshift_y
, mynrnb
, wcycle
,
651 &energy_q
, &energy_lj
, lambda_q
, lambda_lj
, &dvdlambda_q
, &dvdlambda_lj
,
656 cycles
= wallcycle_stop(wcycle
, ewcPMEMESH
);
658 gmx_pme_send_force_vir_ener(pme_pp
.get(), as_rvec_array(forces
.data()),
659 vir_q
, energy_q
, vir_lj
, energy_lj
,
660 dvdlambda_q
, dvdlambda_lj
, cycles
);
663 } /***** end of quasi-loop, we stop with the break above */
666 walltime_accounting_end_time(walltime_accounting
);