HPC at CEA/DAM

22
CEA/DAM 1 HPC at CEA/DAM The TERA 1 super computer RFP 06/1999 Order to COMPAQ 02/2000 Intermediary machine .5 Tflops delivered 12/2000 • open to use 02/2001 • fully operational 06/2001 (CRAY T90 stoped) Final machine 5 peak teraflops power on 4/12/2001 • 1.13 TF sustained 13/12/2001 • 3.98 LINPACK 04/2002 4 on TOP500 • open to all users 09/2002 • fully operational 12/2002 (Intermediary machine stoped) Jean Gonnord Program Director for Numerical Simulation CEA/DAM Thanks to Pierre Leca, François Robin And all CEA/DAM/DSSI Team

description

HPC at CEA/DAM. The TERA 1 super computer RFP 06/1999 Order to COMPAQ 02/2000 Intermediary machine .5 Tflops delivered 12/2000 open to use 02/2001 fully operational 06/2001 (CRAY T90 stoped) Final machine 5 peak teraflops power on 4/12/2001 - PowerPoint PPT Presentation

Transcript of HPC at CEA/DAM

Page 1: HPC at CEA/DAM

CEA/DAM 1

HPC at CEA/DAM

The TERA 1 super computer• RFP 06/1999

• Order to COMPAQ 02/2000

• Intermediary machine .5 Tflops• delivered 12/2000• open to use 02/2001• fully operational 06/2001 (CRAY T90 stoped)

• Final machine 5 peak teraflops• power on 4/12/2001• 1.13 TF sustained 13/12/2001• 3.98 LINPACK 04/2002 4 on TOP500• open to all users 09/2002• fully operational 12/2002 (Intermediary machine stoped)

Jean GonnordProgram Director for Numerical

SimulationCEA/DAM

Thanks to Pierre Leca, François Robin And all CEA/DAM/DSSI Team

Page 2: HPC at CEA/DAM

CEA/DAM 2

TERA 1 Supercomputer

December 2001

608(+32)*4 (1 Ghz)

5 Tflops(1 Tflops)

3 TB

50 TB(7.5 GB/s)

600 kW

Installation

Nodes

Peak performance(sustained requested)

Main Memory

Disks(Bandwidth)

Power

Main features

170 frames, 90 km of cables

608 compute nodes

50 TB of disks

ComputeNode

ComputeNode

ComputeNode

ComputeNode

ComputeNode

ComputeNode

I/ONode

I/ONode

I/ONode

32 I/O nodes

Page 3: HPC at CEA/DAM

CEA/DAM 3

Key elements of the TERA computer

608 computational nodes – 8 Glops(4 processors per node)

800 disks72 Gb

36 switches2560 links at 400 Mb/s

Page 4: HPC at CEA/DAM

CEA/DAM 4

Each linkrepresents8 cables

(400 MB/s)

64 portsto nodes

80 portsto lower

level

64 portsto higher

level

QSW Networks (dual rail)

36 switches (128 ports) - 2560 links

Page 5: HPC at CEA/DAM

CEA/DAM 5

TERA Benchmark

Euler Equations of gaz dynamics solved in a 3D Eulerian mesh

December 12, 2001 : 1,32 Tflops on 2475 processors10 billion cells

A "demo applicaction" able to reach very high performances

Show that the whole systemcan be used by a single application

Efficacité

0

500

1000

1500

2000

2500

0 500 1000 1500 2000 2500 3000

Show what really means 1 sustained Tflops !

0)(

0)(

0

uu

uuu

uu

pt

pt

t

11 million cells16 proc EV735 hours

B. Meltz

Page 6: HPC at CEA/DAM

CEA/DAM 6

One year after - First Return of Experience

(key words : robustness and scalability)

Page 7: HPC at CEA/DAM

CEA/DAM 7

Production Codes (1)

o Four important codes requires a large part of computing power :o Currently : A,B,C, D (D under construction)o Old codes have been thrown away with the CRAY T90

o A, B and D have been designed from the start for parallel systems (MPI)

o C has been designed for vector machineso With minor modifications it has been possible to

produce an MPI version scalable to 8 processors

Page 8: HPC at CEA/DAM

CEA/DAM 8

Production Codes (2)

o Relative performance to T90 (1 proc to 1 proc) is between 0.5 and 2

o Good speed-up for A and B up to 64/128 procs depending on problem size

o Speed-up of C reaches 4 for 8 processorso D scales to 1000+ processors

o For users, it is now possible to work much faster than before :o Availability of a large number of processors + Good code

performances

o From weeks to days, from days to hourso More computations in parallel (parametric studies)

Page 9: HPC at CEA/DAM

CEA/DAM 9

Advances in weapon simulation

0

10

20

30

40

50

60

70

80

8 16 32 64 128

Number of processors

CP

U t

ime

(ho

urs

)TERA- I

TERA- F

Computation time (est.) CRAY T90 : more than 300 hoursTwo nights on 64 processors TERA = 1 month on CRAY T90

Page 10: HPC at CEA/DAM

CEA/DAM 10

Advances in simulation of instabilities

20000.5 millions of cells400h*12 processors IBM SP2

20006 millions of cells450h*16 processors COMPAQ

200115 millions of cells

300h*32 processors COMPAQ

2002150 millions of cells

180h*512 processors HP

Author : M. Boulet

Page 11: HPC at CEA/DAM

CEA/DAM 11

Advances in RCS computation

1990 20031997

ARLENE

ARLENEARLASCRAY

T3D/T3E Parallelism

New numerical methods

EID formulation Multipoles Domain

decomposition Direct/iterative solver

CRAYYMP

TERA

TERAARLENEARLAS

2000

30 000

500 000

30 000 000

Number ofUnknown

ODYSSEE

Processors 256 1024Time (s) 4781 1414Tflops 0.308 1.04% / peak perf. 60.15% 50.8%

o Cone shaped object at 2 Ghz :o 104 000 unknowno Matrix size : 85.5

GB

Authors : M. Sesques, M. Mandallena (CESTA)

Page 12: HPC at CEA/DAM

CEA/DAM 12

Pu239 --> U235 (88 keV) + He (5.2 MeV)

=> Default creation : type and number ?

Tool : Molecular dynamics (DEN/SRMP ; DAM/DSSI)

200 000 time steps 4 weeks of computation 32 processors of TERA

Also 20 106 atoms on 128 procs

5 nm

9 M atomes20 keV

85 ps

interstitiel remplacementlacune

Author : P. Pochet (CVA)

Physics code : Plutonium Ageing

Page 13: HPC at CEA/DAM

CEA/DAM 13

Physics codes

o Simulation of laser beam propagation in LMJ (D. Nassiet)o 2.5 day for 1 processoro Less than 1 hour for 64 procs

o Laser-Plasma interaction (E. Lefebvre)o 2D scalable to 256 procso 3D for the future

o And many others …in material simulation, wave propagation ...

00:00:00

00:07:12

00:14:24

00:21:36

00:28:48

00:36:00

00:43:12

00:50:24

00:57:36

64 128 256 512 1024Nombre de processeurs

Tem

ps

(H

H:M

M:S

S)

10

100

1000

10000

100000

1 10 100 1000

nombre de processeurs

tem

ps

calc

ul (

s)

TOTAL

mouvement

sources

diagnostics

maxw ell

conditions limites

tri particules

Page 14: HPC at CEA/DAM

CEA/DAM 14

HERA hydro computation : 15 µm mesh (15,5 cm x 15,5 cm)

Reference computation EULER

100 millions cells

AMR Computation ~3,5 millions cells

20 732 CPU hours256 processors

81 h

140 CPU hours1 processor

140 h

Auteurs : H. Jourdren - Ph. Ballereau - D. Dureau (DIF)

Page 15: HPC at CEA/DAM

CEA/DAM 15

Major technical issues

o Global File Systemo Reliabilityo Performances

o Hardware reliabilityo Adapt simulation codes and

environnement tools to :o Unreliable environmento Cluster Architecture

o System administrationo Network security

Page 16: HPC at CEA/DAM

CEA/DAM 16

Global File System (1)

o Reliability :

o Intermediate system (Sierra 2.0)• Since April 2002 a lot of file corruptions (at that time

parallel load had greatly increased on the machine, 1+TB/day)

• The major bug was fixed September 2002, others remain we are waiting for other bug fixes

– Very long time to get a fix– Problem difficult to reproduce + Galway/Bristol

o Same bugs exist on the final system (Sierra 2.5)• Apparently corrected with Sierra 2.5 • But re appearing (at a low level) with the very heavy load

of the machine during the last weeks

Page 17: HPC at CEA/DAM

CEA/DAM 17

Global File System (2)

o Performances

o For large IO, we are almost at contractual level (6.5 GB/s vs 7.5 GB/s).

• HP can solve the problem by adding more hardware or by having Ios use both rails

o For small IO, performances are VERY far from contractual level (0,07 GB/s vs 1.5 GB/s !)

• It is probably necessary to work a lot on PFS to solve this problem

Page 18: HPC at CEA/DAM

CEA/DAM 18

Hardware Reliability

o Interposer Change on EV68 (Final system)o Major problem with EV68 processorts starting end of 2001

• Up to 30 CPUs failure per dayo Interposer change : April 4 to April 9

• (2560 processors + spare)• Now : less than 2 CPU per week

o DIMM failures

o HSV110 failures

o Still some SOPF :o Administration networko Quadrics Rails

Page 19: HPC at CEA/DAM

CEA/DAM 19

CEA (640 Nodes, 2560 CPUs)

0

10

20

30

40

50

60

70

80

90

12/2

/200

1

12/9

/200

1

12/1

6/20

01

12/2

3/20

01

12/3

0/20

01

1/6/

2002

1/13

/200

2

1/20

/200

2

1/27

/200

2

2/3/

2002

2/10

/200

2

2/17

/200

2

2/24

/200

2

3/3/

2002

3/10

/200

2

3/17

/200

2

3/24

/200

2

3/31

/200

2

4/7/

2002

4/14

/200

2

4/21

/200

2

4/28

/200

2

5/5/

2002

5/12

/200

2

5/19

/200

2

5/26

/200

2

6/2/

2002

6/9/

2002

6/16

/200

2

6/23

/200

2

6/30

/200

2

7/7/

2002

7/14

/200

2

7/21

/200

2

7/28

/200

2

8/4/

2002

8/11

/200

2

8/18

/200

2

8/25

/200

2

9/1/

2002

9/8/

2002

System Run Time - Weeks

No

. Mo

du

le F

ailu

res

Weekly CPU Average Failure Rate

(For first 7 weeks = 4.4 Modules/Week)

(Total of 31 CPU Modules)

Predicted Weekly Failure Rate

for this System Configuration

1.25 Modules/WeekRev F

UpgradeCompleted

4/11

Page 20: HPC at CEA/DAM

CEA/DAM 20

Adapt tools to new environment

o Make CEA/DAM software as "bullet proof" as possibleo Check all IO operationso If problem, wait and retryo Use the "commit" paradigm as much as possibleo Log all messages on central logs (don't rely on users

for reporting the problems)o If everything goes wrong : try to inform the user as

efficiently as possible of the damages

Page 21: HPC at CEA/DAM

CEA/DAM 21

System Administration

o Boot, shutdown, system update (patches), …

o Too long

o Global commands, system updateo Unreliable

o Need tools to check software level on all nodeso Even maintain global date

o Need tools to monitor the load of the system

Page 22: HPC at CEA/DAM

CEA/DAM 22

A real success, but ...

o Necessity to adapt simulation codes and environnement software to Cluster Architecture (make it bullet proof !)

o IO and Global File Systemo Reliabilityo Performances (scalability)

o System administration

CONCLUSION