Designing Fault Resilient and Fault Tolerant Systems with ... · Designing Fault Resilient and...

35
Designing Fault Resilient and Fault Tolerant Designing Fault Resilient and Fault Tolerant Systems with InfiniBand Dhabaleswar K. (DK) Panda The Ohio State University E-mail: panda@cse ohio-state edu E mail: panda@cse.ohio state.edu http://www.cse.ohio-state.edu/~panda 1 HPC Resiliency '09

Transcript of Designing Fault Resilient and Fault Tolerant Systems with ... · Designing Fault Resilient and...

Page 1: Designing Fault Resilient and Fault Tolerant Systems with ... · Designing Fault Resilient and Fault Tolerant Systems with InfiniBand Dhabaleswar K. (DK) Panda ... – Pro-active

Designing Fault Resilient and Fault Tolerant Designing Fault Resilient and Fault Tolerant Systems with InfiniBand

Dhabaleswar K. (DK) PandaThe Ohio State University

E-mail: panda@cse ohio-state eduE mail: [email protected] state.eduhttp://www.cse.ohio-state.edu/~panda

1HPC Resiliency '09

Page 2: Designing Fault Resilient and Fault Tolerant Systems with ... · Designing Fault Resilient and Fault Tolerant Systems with InfiniBand Dhabaleswar K. (DK) Panda ... – Pro-active

Trends for Computing Clusters in the T 500 Li t

• Top 500 list of Supercomputers (www.top500.org)

Top 500 List

Jun. 2001: 33/500 (6.6%) Nov. 2005: 360/500 (72.0%)

Nov. 2001: 43/500 (8.6%) Jun. 2006: 364/500 (72.8%)

Jun. 2002: 80/500 (16%) Nov. 2006: 361/500 (72.2%)

Nov. 2002: 93/500 (18.6%) Jun. 2007: 373/500 (74.6%)

Jun. 2003: 149/500 (29.8%) Nov. 2007: 406/500 (81.2%)

Nov. 2003: 208/500 (41.6%) Jun. 2008: 400/500 (80.0%)

Jun. 2004: 291/500 (58.2%) Nov. 2008: 410/500 (82.0%)

Nov. 2004: 294/500 (58.8%) Jun. 2009: 410/500 (82.0%)

HPC Resiliency '09

Jun. 2005: 304/500 (60.8%) Nov. 2009: To be announced

2

Page 3: Designing Fault Resilient and Fault Tolerant Systems with ... · Designing Fault Resilient and Fault Tolerant Systems with InfiniBand Dhabaleswar K. (DK) Panda ... – Pro-active

InfiniBand in the Top500pSystems Performance

P t h f I fi iB d i t dil i i

HPC Resiliency '09

Percentage share of InfiniBand is steadily increasing

3

Page 4: Designing Fault Resilient and Fault Tolerant Systems with ... · Designing Fault Resilient and Fault Tolerant Systems with InfiniBand Dhabaleswar K. (DK) Panda ... – Pro-active

Large-scale InfiniBand Installationsg

• 151 IB clusters (30.2%) in the June ’09 TOP500 list (www.top500.org)

• Installations in the Top 30 (15 of them):

129,600 cores (RoadRunner) at LANL (1st) 12,288 cores at GENCI-CINES, France (20th)

51,200 cores (Pleiades) at NASA Ames (4th) 8,320 cores in UK (25th)

62,976 cores (Ranger) at TACC (8th) 8,320 cores in UK (26th)

th th26,304 cores (Juropa) at TACC (10th) 8,064 cores (DKRZ) in Germany (27th)

30,720 cores (Dawning) at Shanghai (15th) 12,032 cores at JAXA, Japan (28th)

14 336 cores at New Mexico (17th) 10 240 cores at TEP France (29th)14,336 cores at New Mexico (17 ) 10,240 cores at TEP, France (29 )

14,384 cores at Tata CRL, India (18th) 13,728 cores in Sweden (30th)

18 224 cores at LLNL (19th) More are getting installed !

HPC Resiliency '09

18,224 cores at LLNL (19 ) More are getting installed !

4

Page 5: Designing Fault Resilient and Fault Tolerant Systems with ... · Designing Fault Resilient and Fault Tolerant Systems with InfiniBand Dhabaleswar K. (DK) Panda ... – Pro-active

MVAPICH/MVAPICH2 Software

• High Performance MPI Library for IB and 10GE– MVAPICH (MPI-1) and MVAPICH2 (MPI-2)

– Used by more than 975 organizations in 51 countries

– More than 34,000 downloads from OSU site directly

– Empowering many TOP500 clusters• 8th ranked 62,976-core cluster (Ranger) at TACC

– Available with software stacks of many IB, 10GE and server vendors including Open Fabrics Enterprise Distribution (OFED)vendors including Open Fabrics Enterprise Distribution (OFED)

– Also supports uDAPL device to work with any network supporting uDAPL

– http://mvapich.cse.ohio-state.edu/HPC Resiliency '09 5

Page 6: Designing Fault Resilient and Fault Tolerant Systems with ... · Designing Fault Resilient and Fault Tolerant Systems with InfiniBand Dhabaleswar K. (DK) Panda ... – Pro-active

Presentation Overview

• Network-Level Fault Tolerance/Resiliency in MVAPICH/MVAPICH2

Presentation Overview

Network Level Fault Tolerance/Resiliency in MVAPICH/MVAPICH2

– Automatic Path Migration (APM)

– Mem-to-Mem Reliability

– Resiliency to Network Failures

• Process-Level Fault Tolerance in MVAPICH2

– BLCR-based systems-level Checkpoint-Restart (CR)

– Enhancing CR Performance with I/O Aggregation

– Fault-Tolerant Backplane (FTB) over InfiniBandFault Tolerant Backplane (FTB) over InfiniBand

– Pro-active Migration with Job-Suspend and Resume

• Virtualization and Fast Migration with InfiniBand

• Conclusion and Q&A6HPC Resiliency '09

Page 7: Designing Fault Resilient and Fault Tolerant Systems with ... · Designing Fault Resilient and Fault Tolerant Systems with InfiniBand Dhabaleswar K. (DK) Panda ... – Pro-active

Network-Level Fault Tolerance with Automatic Path Migration (APM)

Utili s R dund nt C mmuni ti n P ths• Utilizes Redundant Communication Paths– Multiple Ports

– LMC (LID Mask Control)

• Enables migrating connections to a different path

• Reliability guarantees for Service Type Maintained during Migration

• Support in both MVAPICH and MVAPICH2

A. Vishnu, A. Mamidala, S. Narravula and D. K. Panda, Automatic Path Migration over InfiniBand: Early Experiences, Third International Workshop on System Management Techniques, Processes, and Services, held in conjunction with IPDPS '07, March 2007.

HPC Resiliency '09 7

Page 8: Designing Fault Resilient and Fault Tolerant Systems with ... · Designing Fault Resilient and Fault Tolerant Systems with InfiniBand Dhabaleswar K. (DK) Panda ... – Pro-active

Screenshots: APM with OSU Bandwidth test

HPC Resiliency '09 8

Page 9: Designing Fault Resilient and Fault Tolerant Systems with ... · Designing Fault Resilient and Fault Tolerant Systems with InfiniBand Dhabaleswar K. (DK) Panda ... – Pro-active

Memory-to-Memory Reliability

• InfiniBand enforces HCA to HCA reliability using CRC

• No check to see if data is CPU Mem MemCPU

transmitted reliably over I/O Bus

• In different situations (high-altitudes or in hotter climates)

ErrorI/OBus

I/OBus

altitudes or in hotter climates), error rate increases sharply

• MVAPICH uses CRC-32 bit algorithm to ensure safe

HCA HCALink

CRC Protectedgmessage delivery CRC Protected

HPC Resiliency '09 9

Page 10: Designing Fault Resilient and Fault Tolerant Systems with ... · Designing Fault Resilient and Fault Tolerant Systems with InfiniBand Dhabaleswar K. (DK) Panda ... – Pro-active

N k L l R iliNetwork-Level Resiliency• Protection against various network failures

• Switch reboot/failure• HCA failureH f u• Severe congestion

• Can we stall a job instead of aborting it while the failed component is fixedfailed component is fixed

• Being designed and developed together with MellanoxWill b il bl i MV PICH 1 2• Will be available in MVAPICH 1.2

10HPC Resiliency '09

Page 11: Designing Fault Resilient and Fault Tolerant Systems with ... · Designing Fault Resilient and Fault Tolerant Systems with InfiniBand Dhabaleswar K. (DK) Panda ... – Pro-active

N k L l R l FlNetwork-Level Resiliency Flow

Fatal Event Restart Driver Migrate HCANormal

State

Error Event

Reconnect q

Reconnect RequestResend

• Recover from a fatal HCA failure (first restart, then migrate)• Recover from errors (intermittent switch failure etc)

This differs from Automatic Path Migration (APM) which can only recover from a sin le err r event (n n fatal) and cann t ait f r a specified time t retr

• Recover from errors (intermittent switch failure, etc) • Configurable retry settings

HPC Resiliency '0911

single error event (non-fatal) and cannot wait for a specified time to retry

Page 12: Designing Fault Resilient and Fault Tolerant Systems with ... · Designing Fault Resilient and Fault Tolerant Systems with InfiniBand Dhabaleswar K. (DK) Panda ... – Pro-active

P f IPerformance Impact

IS.CMG.CSP.C

Resiliency

BT CCG.CFT.C

ResiliencyNormal

0 0.5 1

BT.C

Normalized Time (256 cores DDR HCAs)Normalized Time (256 cores, DDR HCAs)

No performance change for application kernels

12HPC Resiliency '09

Page 13: Designing Fault Resilient and Fault Tolerant Systems with ... · Designing Fault Resilient and Fault Tolerant Systems with InfiniBand Dhabaleswar K. (DK) Panda ... – Pro-active

Presentation Overview

• Network-Level Fault Tolerance/Resiliency in MVAPICH/MVAPICH2

Presentation Overview

Network Level Fault Tolerance/Resiliency in MVAPICH/MVAPICH2

– Automatic Path Migration (APM)

– Mem-to-Mem Reliability

– Resiliency to Network Failures

• Process-Level Fault Tolerance in MVAPICH2

– BLCR-based systems-level Checkpoint-Restart (CR)

– Enhancing CR Performance with I/O Aggregation

– Fault-Tolerant Backplane (FTB) over InfiniBandFault Tolerant Backplane (FTB) over InfiniBand

– Pro-active Migration with Job-Suspend and Resume

• Virtualization and Fast Migration with InfiniBand

• Conclusion and Q&A13HPC Resiliency '09

Page 14: Designing Fault Resilient and Fault Tolerant Systems with ... · Designing Fault Resilient and Fault Tolerant Systems with InfiniBand Dhabaleswar K. (DK) Panda ... – Pro-active

Checkpoint/Restart Support for Checkpoint/Restart Support for MVAPICH2

• Process-level Fault Tolerance– User-transparent, system-level checkpointing– Based on BLCR from LBNL to take coordinated Based on BLCR from LBNL to take coordinated

checkpoints of entire program, including front end and individual processes

– Designed novel schemes to• Coordinate all MPI processes to drain all in flight messages

in IB connections • Store communication state and buffers, etc. while taking

checkpoint checkpoint • Restarting from the checkpoint

• Available for the last two years with MVAPICH2 and is being used by many organizations

HPC Resiliency '09 14

• Systems-level checkpoint can also be initiated from the application

Page 15: Designing Fault Resilient and Fault Tolerant Systems with ... · Designing Fault Resilient and Fault Tolerant Systems with InfiniBand Dhabaleswar K. (DK) Panda ... – Pro-active

Enhancing CR Performance

• Checkpoint time is dominated by writing the files to

Enhancing CR Performance

• Checkpoint time is dominated by writing the files to

storage

• Multi-core systems are emerging

– 8/16-cores per node

– a lot of data needs to be written

– affects scalability

• Can we reduce checkpoint time with I/O aggregation

of short messages?of short messages?

15HPC Resiliency '09

Page 16: Designing Fault Resilient and Fault Tolerant Systems with ... · Designing Fault Resilient and Fault Tolerant Systems with InfiniBand Dhabaleswar K. (DK) Panda ... – Pro-active

Profiled Results

Basic checkpoint writing information (class C, 64 processes, 8 processes/node)

16HPC Resiliency '09

Page 17: Designing Fault Resilient and Fault Tolerant Systems with ... · Designing Fault Resilient and Fault Tolerant Systems with InfiniBand Dhabaleswar K. (DK) Panda ... – Pro-active

Ch k i t W iti P fil f LU C 64Checkpoint Writing Profile for LU.C.64

17HPC Resiliency '09

Page 18: Designing Fault Resilient and Fault Tolerant Systems with ... · Designing Fault Resilient and Fault Tolerant Systems with InfiniBand Dhabaleswar K. (DK) Panda ... – Pro-active

Write Aggregation DesignWrite-Aggregation Design

Free bufferdata being writtendata ready to be flushed

Circular buffer

Presented atICPP ‘09

18HPC Resiliency '09

Page 19: Designing Fault Resilient and Fault Tolerant Systems with ... · Designing Fault Resilient and Fault Tolerant Systems with InfiniBand Dhabaleswar K. (DK) Panda ... – Pro-active

Time to Take One Checkpoint -Time to Take One Checkpoint -64 processes (8 nodes with 8 cores)

10000

12000

Original BLCR Aggregation

speedup=3.09

speedup=3.45

6000

8000

10000

t time (ms) speedup=3.18

2000

4000

6000

Checkpoint

speedup=1.31

0

LU.C.64 SP.C.64 BT.C.64 EP.C.64

• 64 MPI processes on 8 nodes, 8 processes/node• Checkpoint data is written to local disk files

19HPC Resiliency '09

Page 20: Designing Fault Resilient and Fault Tolerant Systems with ... · Designing Fault Resilient and Fault Tolerant Systems with InfiniBand Dhabaleswar K. (DK) Panda ... – Pro-active

Time to Take One Checkpoint -Time to Take One Checkpoint -64 processes (4 nodes with 16 cores)

40000

45000

t(ms)

Original BLCR Aggregation

speedup=11.57

25000

30000

35000

ne checkpoint

speedup=13.08

10000

15000

20000

me to take on

speedup=9.13

d 1 67

0

5000

LU.C.64 SP.C.64 BT.C.64 EP.C.64

Tim speedup=1.67

• 64 MPI processes on 4 nodes, 16 processes/node• Checkpoint data is written to local disk files

20HPC Resiliency '09

Will be available in theNext MVAPICH2 Release

Page 21: Designing Fault Resilient and Fault Tolerant Systems with ... · Designing Fault Resilient and Fault Tolerant Systems with InfiniBand Dhabaleswar K. (DK) Panda ... – Pro-active

Presentation Overview

• Network-Level Fault Tolerance/Resiliency in MVAPICH/MVAPICH2

Presentation Overview

Network Level Fault Tolerance/Resiliency in MVAPICH/MVAPICH2

– Automatic Path Migration (APM)

– Mem-to-Mem Reliability

– Resiliency to Network Failures

• Process-Level Fault Tolerance in MVAPICH2

– BLCR-based systems-level Checkpoint-Restart (CR)

– Enhancing CR Performance with I/O Aggregation

– Fault-Tolerant Backplane (FTB) over InfiniBandFault Tolerant Backplane (FTB) over InfiniBand

– Pro-active Migration with Job-Suspend and Resume

• Virtualization and Fast Migration with InfiniBand

• Conclusion and Q&A21HPC Resiliency '09

Page 22: Designing Fault Resilient and Fault Tolerant Systems with ... · Designing Fault Resilient and Fault Tolerant Systems with InfiniBand Dhabaleswar K. (DK) Panda ... – Pro-active

Coordinated Infrastructure for Fault Tolerant Systems (CIFTS) Framework

MPI

JOBLAUNCHER

&HIGH

PERFORMANCE

MVAPICH2

Checkpoint/Restart

INFINIBAND

IPMI

OSU

Collaborators

Migration

MPILIBRARY

&RESOURCEMANAGER

CHECKPOINTLIBRARY

PERFORMANCENETWORK

DIAGNOSTICTOOLS

IPMIMONITOR

• ANL• OSU• ORNL

FAULT TOLERANCE BACKPLANEOPERATING

SYSTEMFILE

SYSTEMS

EVENTHPC

ORNL• LBNL• IU• UT

APPLICATIONSMATHEMATICAL

LIBRARY

ANALYSISHPCHARDWARE

http://www.mcs.anl.gov/research/cifts/22HPC Resiliency '09

Page 23: Designing Fault Resilient and Fault Tolerant Systems with ... · Designing Fault Resilient and Fault Tolerant Systems with InfiniBand Dhabaleswar K. (DK) Panda ... – Pro-active

CIFTS Usage ScenarioCIFTS - Usage Scenario

J b S h dul

IO node failure. File system down

Parallel FS Job SchedulerLaunch jobs with NFS file system

File System sharesthis information

Migrates existingjobs

Application

MPI-IOPrints a coherent

Checkpoints itself

Application

error message

Checkpoints itself

Application

23HPC Resiliency '09

Page 24: Designing Fault Resilient and Fault Tolerant Systems with ... · Designing Fault Resilient and Fault Tolerant Systems with InfiniBand Dhabaleswar K. (DK) Panda ... – Pro-active

FTB IB (Overall Plan)FTB-IB (Overall Plan)

FTB-IB 1.0 version is available for download:http://nowlab.cse.ohio-state.edu/projects/ftb-ib/index.html

24HPC Resiliency '09

Page 25: Designing Fault Resilient and Fault Tolerant Systems with ... · Designing Fault Resilient and Fault Tolerant Systems with InfiniBand Dhabaleswar K. (DK) Panda ... – Pro-active

Comprehensive Solutionmp n u n(Putting All Components Together)

MVAPICH2Will be available in

future MVAPICH2 Release

Checkpoint/Restart(CR)

Job Suspend Resume

(Pro-active)

MV H a

FTB-IB IPMI Support

Available in MVAPICH2

InfiniBand

Introduced in MVAPICH2 1.4

25HPC Resiliency '09

Page 26: Designing Fault Resilient and Fault Tolerant Systems with ... · Designing Fault Resilient and Fault Tolerant Systems with InfiniBand Dhabaleswar K. (DK) Panda ... – Pro-active

Presentation Overview

• Network-Level Fault Tolerance/Resiliency in MVAPICH/MVAPICH2

Presentation Overview

Network Level Fault Tolerance/Resiliency in MVAPICH/MVAPICH2

– Automatic Path Migration (APM)

– Mem-to-Mem Reliability

– Resiliency to Network Failures

• Process-Level Fault Tolerance in MVAPICH2

– BLCR-based systems-level Checkpoint-Restart (CR)

– Enhancing CR Performance with I/O Aggregation

– Fault-Tolerant Backplane (FTB) over InfiniBandFault Tolerant Backplane (FTB) over InfiniBand

– Pro-active Migration with Job-Suspend and Resume

• Virtualization and Fast Migration with InfiniBand

• Conclusion and Q&A26HPC Resiliency '09

Page 27: Designing Fault Resilient and Fault Tolerant Systems with ... · Designing Fault Resilient and Fault Tolerant Systems with InfiniBand Dhabaleswar K. (DK) Panda ... – Pro-active

Problem with Current I/O Problem with Current I/O Virtualization

• Performance– Every I/O operation

involves the VMM and/or 0 6

0.81

1.21.4

xecu

tion

Tim

e VM Native

another VM– VMM may become a

performance bottleneckU i i l VM lt

00.2

0.40.6

BT CG EP IS SPNor

mal

ized

Ex

– Using a special VM results in expensive context switches between different VMs

BT CG EP IS SPN

Dom0 VMM DomU

CG 16.6% 10.7% 72.7%

IS 18 1% 13 1% 68 8%

– Undesirable for high end systems, especially those used in high performance

ti (HPC)

IS 18.1% 13.1% 68.8%

EP 00.6% 00.3% 99.0%

BT 06.1% 04.0% 89.9%

SP 09.7% 06.5% 83.8%computing (HPC)

27HPC Resiliency '09

Page 28: Designing Fault Resilient and Fault Tolerant Systems with ... · Designing Fault Resilient and Fault Tolerant Systems with InfiniBand Dhabaleswar K. (DK) Panda ... – Pro-active

Xen-IB and VMM-Bypass

Xen dom0 Xen domu

User Level IB Provider

Backend Virtual HCA

Core IB Module

Privileged Access

HCA Provider

Core IB ModuleProvider

VMM-Bypass Access

Privileged Access

HCA Hardware

J. Liu, W. Huang, B. Abali, D. K. Panda. High Performance VMM-Bypass I/O in g g ypVirtual Machines, USENIX Annual Technical Conference (USENIX’06), May, 2006

28HPC Resiliency '09

Page 29: Designing Fault Resilient and Fault Tolerant Systems with ... · Designing Fault Resilient and Fault Tolerant Systems with InfiniBand Dhabaleswar K. (DK) Panda ... – Pro-active

MPI Latency and Bandwidth MPI Latency and Bandwidth (MVAPICH)

Latency

20

25

30

us)

xennative

Bandwidth

600

800

1000

es/s

xennative

0

5

10

15

Late

ncy

(u

0

200

400

Mill

ionB

yte

• Only VMM Bypass operations are used

0 2 8 32

128 512 2k 8k

Msg size (Bytes)

1 4 16 64256 1k 4k16k 64k256

k 1M 4M

Msg size (Bytes)

Only VMM Bypass operations are used• Xen-IB performs similar to native InfiniBand• Numbers taken with MVAPICH

29HPC Resiliency '09

Page 30: Designing Fault Resilient and Fault Tolerant Systems with ... · Designing Fault Resilient and Fault Tolerant Systems with InfiniBand Dhabaleswar K. (DK) Panda ... – Pro-active

HPC Benchmarks (NAS)

Dom0 VMM DomU

BT 0.4% 0.2% 99.4%

CG 0.6% 0.3% 99.0%1

1.2

n Ti

me VM Native

EP 0.6% 0.3% 99.3%

FT 1.6% 0.5% 97.9%

IS 3.6% 1.9% 94.5%

LU 0.6% 0.3% 99.0%0 2

0.4

0.6

0.8

zed

Exec

utio

n

MG 1.8% 1.0% 97.3%

SP 0.3% 0.1% 99.6%0

0.2

BT CG EP FT IS LU MG SPNor

mal

iz

• NAS Parallel Benchmarks achieve similar performance in VM and native environment (8x2)

–J. Liu, W. Huang, B. Abali, D. K. Panda. High Performance VMM-Bypass I/O in Virtual Machines, USENIX Annual Technical Conference (USENIX’06), May, 2006Virtual Machines, USENIX Annual Technical Conference (USENIX 06), May, 2006–W. Huang, J. Liu, B. Abali, D. K. Panda. A Case for High Performance Computing with Virtual Machines, ACM International Conference on Supercomputing (ICS ’06), June, 2006

30HPC Resiliency '09

Page 31: Designing Fault Resilient and Fault Tolerant Systems with ... · Designing Fault Resilient and Fault Tolerant Systems with InfiniBand Dhabaleswar K. (DK) Panda ... – Pro-active

Optimizing VM migration through Optimizing VM migration through RDMA

VMHelper Process

Pre-allocate resources

Machine statesMachine statesMachine statesMachine states

Live VM migration:Physical host Physical host

• Step 1: Pre-allocate resource on target host• Step 2: Pre-copy machine states for multiple iterations• Step 3: Suspend VM and copy the latest updates to machine states• Step 3: Suspend VM and copy the latest updates to machine states• Step 4: Restart VM on the new host

31HPC Resiliency '09

Page 32: Designing Fault Resilient and Fault Tolerant Systems with ... · Designing Fault Resilient and Fault Tolerant Systems with InfiniBand Dhabaleswar K. (DK) Panda ... – Pro-active

Fast Migration over RDMA250 100%RDMA IPoIB

50

60

70

e (s

ec)

Native IPoIB RDMA

150

200

250

wid

th (M

B/s)

60%

70%

80%

90%

100%

zatio

n

CPU RDMA CPU IPoIB

10

20

30

40

Exec

utio

n Ti

me

50

100

ctiv

e Ba

ndw

20%

30%

40%

50%

CPU

Util

iz

0sp bt ft lu ep cg

Mi ti h d ith IP IB d ti ll i

0SP.A.9 BT.A.9 FT.B.8 LU.A.8 EP.B.9 CG.B.8

Effe

0%

10%

• Migration overhead with IPoIB drastically increases• RDMA achieves higher migration performance with less CPU usage

W. Huang, Q. Gao, J. Liu, D. K. Panda. High Performance Virtual Machine Migration with RDMA over Modern Interconnects. IEEE Conference on Cluster Computing (Cluster’07), September 2007 (Best Paper Award)

32HPC Resiliency '09

Page 33: Designing Fault Resilient and Fault Tolerant Systems with ... · Designing Fault Resilient and Fault Tolerant Systems with InfiniBand Dhabaleswar K. (DK) Panda ... – Pro-active

Xen-IB SoftwareXen-IB Software

• Initially designed jointly with IBMy g j y

• Taken up by Novell later on

• Available from OFED and Mellanox sites

• Integration with MVAPICH2 and other components are planned in future

33HPC Resiliency '09

Page 34: Designing Fault Resilient and Fault Tolerant Systems with ... · Designing Fault Resilient and Fault Tolerant Systems with InfiniBand Dhabaleswar K. (DK) Panda ... – Pro-active

Summary and ConclusionsSummary and Conclusions

• Fault-tolerance and resiliency issues are becoming extremely critical for next generation Exascale systems

• InfiniBand is an emerging interconnect which provides basic f ti liti f f lt t l t th t k l lfunctionalities for fault-tolerance at the network-level

• Presented how InfiniBand features can be used at the MPI layer to provide fault-tolerance and resiliencep

• Presented expanded solutions using virtualization

• Many open research challenges needing novel solutions for fault resiliency and fault tolerance in next generation Exascale systems

34HPC Resiliency '09

Page 35: Designing Fault Resilient and Fault Tolerant Systems with ... · Designing Fault Resilient and Fault Tolerant Systems with InfiniBand Dhabaleswar K. (DK) Panda ... – Pro-active

b PWeb Pointers

MVAPICH Web Pagehttp://mvapich.cse.ohio-state.edu/

E-mail: [email protected]

35HPC Resiliency '09