[gmx-users] AMD 32 core TR

paul buscemi pbuscemi at q.com
Thu Jan 10 02:15:51 CET 2019


Tomas,

Again, thanks for the response.

On re-reading the data Exxact Corp sent ( always helps to review) they did use a 2080ti and two zions.   

Your point on maxing out the GPU is interesting.  On the 8 core ( 16T) the GPU is maxed out as you inferred, but not on the 32 core (64T) with which the GPU runs at 60-85% depending on the model but still at greater efficiency than the 8 core ( ~ 60 vs 50 ns/day. I’ve never been able to max out the GTX 1080 TI on the 32 core system.  

The primary reason for the 32 core is to support two GPUs and there the efficiency increases from (e.g ) 70 ns/day to 90 -100 ns/day  with  ~ 150k atoms. I never expected a tremendous increase by increasing core number

I feel a little uncomfortable  citing such  generalities. So for now I  will state that am satisfied with the outcome and that users who use single workstations can expect to match professionally assembled systems.  


> On Jan 6, 2019, at 5:07 AM, Tamas Hegedus <tamas at hegelab.org> wrote:
> 
> It also comes into my mind that the AMD 32 core has not a double performance of AMD 16 core. Because of manufacturing the 4 dies (4x8 cpu units), it does not have the same bandwidth to the RAM. But this is told to affect only memory hungry applications. Thus at this moment I think that this does not effect much gmx runs. I hope to try an RT 2990WX (amd 32core) in 1-2 months.
> 
> On 2019. 01. 04. 4:18, paul buscemi wrote:
>> Tamas,  thanks for the response.
>> In previous posts I mention using a single gtx 1080ti, sorry for not making it clear in the last post.
>>    On the 8 core AMD and an Intel 6 core I am running Cuda 10 with Gromacs 18.3 with no issues.  I believe the larger factor in the slowness of the 32 core was in having the runtime Cuda 7 with the cuda 10 drivers.  On the 8 core, runtime cuda 9.1 and cuda 10 drivers work well together  - all with Gromacs 18.3.  Now with  Gromacs v19 , Cuda 10 and the 410 nvidia drivers, the 8 core and 32 core systems seem quite content.
>> I have been tracing results from the log, and you are correct in what it can tell you.  It was the log file that actually brought my attention to the Cuda 7 runtime issue. Also the PP PME distributions were noted with the ntomp/ntmpi arrangements. I have been experimenting with those as suggested in the Gromas acceleration hints.
>> By 10% I meant that the 32 core unit ( in my hands ) ran 10% faster  in ns/day than the 8 core AMD using the same model system and the the same 1080ti  GPU.  Gromacs points out that  150k to 300k atom systems are on the rather small side and so not to expect tremendous differences from the CPU.  The reason for using the 32 core is the eventual addition of a second GPU and the subsequent distribution of threads.
>> With a little OC and tweeking of the fourier spacing and vdw cutoffs in the npt I edged the 137k atom AHD model  to 57 ns/day,  but this falls short of  the Exxact corp benchmarks of 80-90 ns/d —  assuming they are using a 1080ti.  Schrodinger’s Maestro- with the 8 core AMD and 1080ti -  runs a 300k membrane model at about 15 ns/d  but a  60k atom model at 150 ns/day implying  30 ns/day for 300k atoms.  . In general, if I can indeed maintain 20-25 ns/day for 300k atoms I’d be satisfied.  The original posts were made because I was frustrated seeing 6 to 8 ns/d with the 32core machine and the 8 core was producing 20 ns/day.   As I mentioned the wounds were self inflicted  with the installation of Cuda runtime 7 and at one point compilation with g++-5.   As far as I am concerned it’s imperative that the latest drivers and Gromacs versions be used or at least the same  genre of drivers and versions be assembled.
>> Again, I’d like to point out that in using four different machines, 4 different Intel and AMD  CPU’s, 5 different MBs,  5 different GPU’s, now 4 progressive versions of Gromacs, and model systems of 200-300 k particles, I’ve not run across a single problem associated with the software or hardware per se but rather was caused by the my models or my compilation methods.
>> Hope this addresses your questions and helps any other users contemplating using a Ryzen TR.
>> Paul
>>> On Jan 3, 2019, at 2:09 PM, Tamas Hegedus <tamas at hegelab.org> wrote:
>>> 
>>> Please provide more information.
>>> 
>>> If you use gmx 2018 then I think that gmx limits the gcc version to 6 and not cuda 10.
>>> 
>>> You did not specify what type of and how many GPUs you use.
>>> 
>>> In addition, the choice of gmx for distributing computation could be also informative - you find this info in the log file.
>>> 
>>> It is also not clear what do you mean of 10% improvement: 8ns/day to 26ns/day are the only numbers but it corresponds to 3x faster simulations and not 1.1x
>>> 
>>> In addition, I think if you have 49.5 ns/day for 137K atoms than 26ns/day seems to be ok for 300K.
>>> 
>>> Bests, Tamas
>>> 
>>> 
>>> On 1/3/19 6:11 PM, pbuscemi at q.com wrote:
>>>> Dear users,
>>>> 
>>>>  I had trouble getting suitable performance from an AMD 32 core TR.  By
>>>> updating  all the cuda drivers and runtime to v10  and using gcc,g++ -6 from
>>>> v5  -- I did try gcc-7 but Cuda 10 did not appreciate the attempt  --  and
>>>> in particular removing  CUDA v7 runtime.), I was able to improve a 300k atom
>>>> nvt run from 8 ns/day to 26 ns/day .  I replicated  as far as possible the
>>>> Gromacs ADH benchmark with 137000 atoms-spc/e.  I could achieve an md of
>>>> 49.5 ns/day. I do not have a firm grasp if this is respectable or not (
>>>> comments ? )  but appears at least ok.   The input command was simply mdrun
>>>> ADH.md   -nb gpu  -pme gp   ( and not using -ntomp or ntmpi which in my
>>>> hands degraded performance ) .   To run the ADH  I replaced the two ZN ions
>>>> in  ADH file from PDB ( 2ieh.pdb ) with CA ions  since ZN was not found in
>>>> the OPLS data base in using pdb2gmx.
>>>> 
>>>>  The points being ( 1) Gromacs appears reasonably happy with  the 8 core and
>>>> 32 core Ryzen although ( again in my hands ) for these  smallish systems
>>>> there is only about a 10% improvement between the two, and  (2) , as often
>>>> suggested in the Gromacs literature, use the latest drivers possible
>>>> 
>>>>    
>>> -- 
>>> Tamas Hegedus, PhD
>>> Senior Research Fellow
>>> MTA-SE Molecular Biophysics Research Group
>>> Hungarian Academy of Sciences  | phone: (36) 1-459 1500/60233
>>> Semmelweis University          | fax:   (36) 1-266 6656
>>> Tuzolto utca 37-47             | mailto:tamas at hegelab.org
>>> Budapest, 1094, Hungary        | http://www.hegelab.org
>>> 
>>> -- 
>>> Gromacs Users mailing list
>>> 
>>> * Please search the archive at http://www.gromacs.org/Support/Mailing_Lists/GMX-Users_List before posting!
>>> 
>>> * Can't post? Read http://www.gromacs.org/Support/Mailing_Lists
>>> 
>>> * For (un)subscribe requests visit
>>> https://maillist.sys.kth.se/mailman/listinfo/gromacs.org_gmx-users or send a mail to gmx-users-request at gromacs.org.
> 
> -- 
> Tamas Hegedus, PhD
> Senior Research Fellow
> MTA-SE Molecular Biophysics Research Group
> Hungarian Academy of Sciences  | phone: (36) 1-459 1500/60233
> Semmelweis University          | fax:   (36) 1-266 6656
> Tuzolto utca 37-47             | mailto:tamas at hegelab.org
> Budapest, 1094, Hungary        | http://www.hegelab.org
> 
> ---
> This email has been checked for viruses by AVG.
> https://www.avg.com
> 
> -- 
> Gromacs Users mailing list
> 
> * Please search the archive at http://www.gromacs.org/Support/Mailing_Lists/GMX-Users_List before posting!
> 
> * Can't post? Read http://www.gromacs.org/Support/Mailing_Lists
> 
> * For (un)subscribe requests visit
> https://maillist.sys.kth.se/mailman/listinfo/gromacs.org_gmx-users or send a mail to gmx-users-request at gromacs.org.



More information about the gromacs.org_gmx-users mailing list