[Pw_forum] parallel scaling in PWSCF

Andrea Ferretti ferretti.andrea at unimore.it
Wed Apr 26 22:28:21 CEST 2006


Hi everybody,

I am currently running a Copper surface with 140 Cu atoms + a molecule...
the system has 1642 electrons and (due to metallicity) the calculation is 
performed for 985 bands (few kpt, like 4)...
due to the 11 electrons for each Cu atom, I have a huge number of bands in 
a (relatively) small cell, and so a (relatively) low number of PWs respect 
to nbnd.
taking a look at the dimension of wfc, no problem with memory in 
principle, even if, due to the weird 
dimensions of the system, non-scalable memory is quite large, around 1Gb.

on a IBM Sp5 machine I observed a severe limit in the scaling passing from 
32 to 64 procs using both espresso 2.1.x and espresso 3.0...
( anyway, I succeeded in performing a "relax" calculation for the system 
!!!! )

as far as I know, this problem might be connected to a serial part in the 
diagonalization which has been parallelized in the current CVS version 
(as already pointed out by Axel)...
At the moment I am testing this CVS version against my system, I will let 
you know the results as soon as possible...

cheers
andrea  


-- 
Andrea Ferretti
Dipartimento di Fisica, Universita' di Modena e Reggio Emilia
Natl. Res. Center S3 INFM-CNR  ( http://s3.infm.it )
Via Campi 213/A I-41100 Modena, Italy
Tel:     +39 059 2055283
Fax:     +39 059 374794
E-mail:  ferretti.andrea at unimore.it
URL:     http://www.nanoscience.unimo.it

Please, if possible, don't  send me MS Word or PowerPoint attachments
Why? See:  http://www.gnu.org/philosophy/no-word-attachments.html


On Wed, 26 Apr 2006, Fernando A Reboredo wrote:

> Nichols
> The largest things I have run with pwscf are cobalt clusters with stuff on 
> the surface
> 55 Co atoms 12 CO molecules and 12au of vacuum around, 45 Ry ecut.
> ~560 electron including spin and PBE. Those calculation run in 48 nodes 90 
> processors. I did not study the scaling carefully. The scaling was 
> reasonably good provided the number of proccessors was a submultiple of the 
> FFT grid dimension.
>  Fernando Reboredo
> 
> ----- Original Message ----- 
> From: "Axel Kohlmeyer" <akohlmey at cmm.upenn.edu>
> To: <pw_forum at pwscf.org>
> Sent: Wednesday, April 26, 2006 1:38 PM
> Subject: Re: [Pw_forum] parallel scaling in PWSCF
> 
> 
> On 4/26/06, Nichols A. Romero <naromero at gmail.com> wrote:
> > Hi,
> 
> hi.
> 
> > Has anyone performed any very-large scale DFT calculations on PWSCF
> > using over 64 processors and over 1000 atoms? Does anyone know what
> > its current limits (system sizes and processors) are on parallel
> > computing environments with fast interconnects?
> 
> i've recently run a comparatively large job (272 atoms, 560 electrons,
> 4x4x4 k-points) across 768 processors on a Cray XT3. i had to hack
> some constants to make it work. however the scaling is not (yet)
> so good and depending on what kind of atoms you want to put into
> your system, i.e. if it is metallic, you may be better of with gamma
> only and car-parrinello dynamics. i ran the same system as above
> with a car-parrinello code (one that is not in quantum espresso) and
> would scale out at 512 nodes for gamma point only on an IBM BG/L.
> even though i had to use a much smaller time step, i did get much
> more trajectory that way.
> 
> judging from the CVS commit messages, efforts to optimize the
> quantum espresso codes for that kind of machine with large numbers
> of nodes are underway. note, that for systems as large as that,
> you might run into the scaling limitations of DFT with plane waves,
> so for a system that big using one of those 'linear scaling' DFT codes
> could be a more promising alternative.
> 
> best regards,
>      axel.
> 
> 
> >
> > Thanks,
> > --
> > Nichols A. Romero, Ph.D.
> > 1613 Denise Dr. Apt. D
> > Forest Hill, MD 21050
> > 443-567-8328 (C)
> > 410-306-0709 (O)
> > _______________________________________________
> > Pw_forum mailing list
> > Pw_forum at pwscf.org
> > http://www.democritos.it/mailman/listinfo/pw_forum
> >
> >
> 
> 
> --
> =======================================================================
> Axel Kohlmeyer   akohlmey at cmm.chem.upenn.edu   http://www.cmm.upenn.edu
>   Center for Molecular Modeling   --   University of Pennsylvania
> Department of Chemistry, 231 S.34th Street, Philadelphia, PA 19104-6323
> tel: 1-215-898-1582,  fax: 1-215-573-6233,  office-tel: 1-215-898-5425
> =======================================================================
> If you make something idiot-proof, the universe creates a better idiot.
> _______________________________________________
> Pw_forum mailing list
> Pw_forum at pwscf.org
> http://www.democritos.it/mailman/listinfo/pw_forum 
> 
> _______________________________________________
> Pw_forum mailing list
> Pw_forum at pwscf.org
> http://www.democritos.it/mailman/listinfo/pw_forum
> 



More information about the Pw_forum mailing list