[Pw_forum] FFT & MPI - issues with latency

Nicola Marzari marzari at MIT.EDU
Mon Jan 30 06:01:32 CET 2006


Dear Kostya, Alex,

this is our datapoint on reasonably good scaling for small
clusters (i.e. assuming that one should leave hardware
for massive parallelism to the experts):
http://nnn.mit.edu/ESPRESSO/CP90_tests/CP_2_1_3_half.pdf

We have a cluster of 32 Dell 1425 3.2 GHz dual-xeons, grouped in 4 nodes 
of 8 dual-xeons. Each node has its own private gigabit ethernet network,
so we can run on up to 16 cpus on each node (each dual-xeon in one node
is connected locally to the other dual-xeons on the same node with a
dedicated 8-port gigabit switch; all dual-xeon are also connected,
with their second gigabit port, to the master via a global master 
switch, that has no MPI traffic on it.)

The results for a medium sized job ("half", 243 bands of AgI
in a 30^ a.u. cell) are in the page above - more details on the job are 
in http://nnn.mit.edu/ESPRESSO/CP90_tests/, in the timings
file (it's the "half" job, i.e. half the electrons of large.j).

As you can see, scaling up to 8 blades works unexpectedly well
(another slightly older cluster of ours saturated at
around 6 blades). Also, the second CPU on a dual-xeon adds very
little, due to the limited memory bandwidth (Opterons would do much
better in scaling from 1 to 2 CPUs on the same motherboard, but
do not do as well in single-CPU performance).

So, if I were to purchase now, I would go with PIVs with the
fastest front side bus and ram (1066 MHz and 667-5300, resepctively),
grouped in small nodes of ~6 CPUs.


				nicola



Axel Kohlmeyer wrote:

> On Sun, 29 Jan 2006, Konstantin Kudin wrote:
> 
> KK>  Hi all,
> 
> hi kostya,
> 
---------------------------------------------------------------------
Prof Nicola Marzari   Department of Materials Science and Engineering
13-5066   MIT   77 Massachusetts Avenue   Cambridge MA 02139-4307 USA
tel 617.4522758 fax 2586534 marzari at mit.edu http://quasiamore.mit.edu



More information about the Pw_forum mailing list