[Wien] omp vs mpi for LAPW1

Peter Blaha pblaha at theochem.tuwien.ac.at
Thu Aug 27 14:11:50 CEST 2020


Clearly, omp_global:12   is NOT efficient for lapw1. The diagonalization 
does not scale for so many cores.

Otherwise it means that your mpi-installation (what are you using ? 
ifort, gfortran, which mpi, blas, scalapack, elpa, ...) is VERY bad.


On 8/27/20 1:44 PM, karima Physique wrote:
> Dear Wien2k users,
> 
> A calculation with 38 equivalent atoms (96 inequivalent   ), we found 
> that using lapw1 with omp_global is very faster than lapw1_mpi
> machine file 1
> omp_global: 12
> 1: node1
> machine file 2
> 1: node1:12
> Have you any comment on this?
> 
> _______________________________________________
> Wien mailing list
> Wien at zeus.theochem.tuwien.ac.at
> http://zeus.theochem.tuwien.ac.at/mailman/listinfo/wien
> SEARCH the MAILING-LIST at:  http://www.mail-archive.com/wien@zeus.theochem.tuwien.ac.at/index.html
> 

-- 

                                       P.Blaha
--------------------------------------------------------------------------
Peter BLAHA, Inst.f. Materials Chemistry, TU Vienna, A-1060 Vienna
Phone: +43-1-58801-165300             FAX: +43-1-58801-165982
Email: blaha at theochem.tuwien.ac.at    WIEN2k: http://www.wien2k.at
WWW:   http://www.imc.tuwien.ac.at/TC_Blaha
--------------------------------------------------------------------------


More information about the Wien mailing list