[gmx-users] Re: Re: help: load imbalance

申昊 shenhao at mail.bnu.edu.cn
Thu Apr 11 07:07:55 CEST 2013


> > Hello,
> >    I wanna ask some questions about load imbalance.
> > 1> Here are the messages resulted from grompp -f md.mdp -p topol.top -c npt.gro -o md.tpr
> >
> >    NOTE 1 [file md.mdp]:
> >   The optimal PME mesh load for parallel simulations is below 0.5
> >   and for highly parallel simulations between 0.25 and 0.33,
> >   for higher performance, increase the cut-off and the PME grid spacing
> >
> > therefore, i changed the md.mdp as whrited below, then used the command grompp -f md.mdp -p topol.top -c npt.gro -o md.tpr , then there is no NOTE printed. So if i change the cut-offs to 2.0 nm and increase the grid spacing to 0.30, does the calculated results reasonable?
> 
> You can shift work between short- and long-range electrostatics by
> adjusting the *coulomb* cut-off freely, but *not* the VdW cut-off.
> 
> However, 2.0 nm sounds like a *very* long cut-off.
> 
Thank you for your reply. As mentioned above, what i need to do is either increase coulomb cut-off or increase the FFT grid spacing. However, the Vmd 
cut-off should longer than the coulomb cut-off. According to what justin had said, this kind of changes are nor reasonable. Now i trun the cut-offs to 
1.4 nm and grid spacing to 0.16. 


> >
> > ; Neighborsearching
> > ns_type         = grid          ; search neighboring grid cells
> > nstlist         = 5             ; 10 fs
> > rlist           = 2             ; short-range neighborlist cutoff (in nm)
> > rcoulomb        = 2             ; short-range electrostatic cutoff (in nm)
> > rvdw            = 2             ; short-range van der Waals cutoff (in nm)
> > ; Electrostatics
> > coulombtype     = PME           ; Particle Mesh Ewald for long-range electrostatics
> > pme_order       = 4             ; cubic interpolation
> > fourierspacing  = 0.3           ; grid spacing for FFT
> >
> > 2> and how about no changes, just simulate it with the original mdp. Is the results still reasonable?  Here are the messages without any changes:
> >
> > DD  load balancing is limited by minimum cell size in dimension X
> > DD  step 2933999  vol min/aver 0.189! load imb.: force 124.7%
> 
> You are simply pushing your simulation to the limit of how far it can
> be parallelized. As you can see from the above output, to compensate
> for the imbalance, the load balancing shrunk DD cells to the extent
> that the volume ratio of the smallest and average DD cell size is
> 0.189, meaning that the smallest DD cells are ~5.3x smaller than the
> average cell size - and the run is still *hugely* imbalanced. The "!"
> indicates what the line before says, that the DD load-balancing is
> limited and can't shrink cells further.
> 
> Some aspects that might be limiting your simulation are:
> i) running with just a few hundred atoms/core;
> ii) running on multiple very different cluster nodes;
> iii) using a very inhomogeneous system.
> 
> If you're using the group scheme (which i assume you are, otherwise
> the automated PP-PME balancing would have kicked in), you should be
> able to get better performance at high parallelization with the verlet
> scheme.
> 
> Cheers,
> --
> Szilard
> 
The system i simulated is about a short peptide with 2300 water molecules. It is absolute not the case i). Anyway, is seems not
a inhomogeneous system for case iii). I can sure that this simulation is running on totally one cluster node.
Does PP-PME helpful for the different cluster nodes? 

Best wishes!
Hao
> >
> >            Step           Time         Lambda
> >         2934000     5868.00000        0.00000
> >
> >    Energies (kJ/mol)
> >           Angle    Proper Dih. Ryckaert-Bell.          LJ-14     Coulomb-14
> >     2.99315e+02    2.13778e+01    1.74659e+02    2.22024e+02    2.02466e+03
> >         LJ (SR)  Disper. corr.   Coulomb (SR)   Coul. recip.      Potential
> >    -1.68074e+02   -2.09809e-01   -1.80294e+03   -3.28155e+03   -2.51074e+03
> >     Kinetic En.   Total Energy    Temperature Pres. DC (bar) Pressure (bar)
> >     1.69264e+04    1.44156e+04    2.95552e+02   -1.33866e-04    1.51489e+00
> >    Constr. rmsd
> >     2.60082e-05
> >
> >
> >
> > --
> > gmx-users mailing list    gmx-users at gromacs.org
> > http://lists.gromacs.org/mailman/listinfo/gmx-users
> > * Please search the archive at http://www.gromacs.org/Support/Mailing_Lists/Search before posting!
> > * Please don't post (un)subscribe requests to the list. Use the
> > www interface or send it to gmx-users-request at gromacs.org.
> > * Can't post? Read http://www.gromacs.org/Support/Mailing_Lists
> 
> 




More information about the gromacs.org_gmx-users mailing list