Steve LloydThe Data Deluge and the GridSlide 1 The Data Deluge and the Grid The Data Deluge The...

44
Steve Lloyd The Data Deluge and the G rid Slide 1 The Data Deluge and the Grid The Data Deluge The Large Hadron Collider The LHC Data Challenge The Grid Grid Applications • GridPP • Conclusion Steve Lloyd Queen Mary University of London [email protected]. uk

Transcript of Steve LloydThe Data Deluge and the GridSlide 1 The Data Deluge and the Grid The Data Deluge The...

Page 1: Steve LloydThe Data Deluge and the GridSlide 1 The Data Deluge and the Grid The Data Deluge The Large Hadron Collider The LHC Data Challenge The Grid Grid.

Steve Lloyd The Data Deluge and the Grid Slide 1

The Data Deluge and the Grid

• The Data Deluge• The Large Hadron Collider• The LHC Data Challenge• The Grid• Grid Applications• GridPP• Conclusion

Steve LloydQueen Mary

University of [email protected]

Page 2: Steve LloydThe Data Deluge and the GridSlide 1 The Data Deluge and the Grid The Data Deluge The Large Hadron Collider The LHC Data Challenge The Grid Grid.

Steve Lloyd The Data Deluge and the Grid Slide 2

The Data Deluge

Expect massive increases in amount of data being collected in several diverse fields over the next few years:– Astronomy - Massive sky surveys– Biology - Genome databases etc.– Earth Observing– Digitisation of paper, film, tape records etc to create

Digital Libraries, Museums . . .– Particle Physics - Large Hadron Collider– . . .1PByte ~1000 TBytes ~ 1M GBytes ~ 1.4M CDs [Petabyte Terabyte Gigabyte]

Page 3: Steve LloydThe Data Deluge and the GridSlide 1 The Data Deluge and the Grid The Data Deluge The Large Hadron Collider The LHC Data Challenge The Grid Grid.

Steve Lloyd The Data Deluge and the Grid Slide 3

Digital Sky Project

Federating new astronomical surveys: ~ 40,000 square degrees ~ 1/2 trillion pixels (1 arc second) ~ 1 TB x multi-wavelengths > 1 billion sources

Integrated catalogue and image database:– Digital Palomer Observatory Sky Survey– 2 All Sky Survey– NRAO VLA Sky Survey– VLA FIRST Radio Survey

Later:– ROSAT– IRAS– Westerbork 327 MHz Survey

Page 4: Steve LloydThe Data Deluge and the GridSlide 1 The Data Deluge and the Grid The Data Deluge The Large Hadron Collider The LHC Data Challenge The Grid Grid.

Steve Lloyd The Data Deluge and the Grid Slide 4

Sloan Digital Sky Survey

• ~ 1 million spectra• positions and images of 100 million objects• 5 wavelength bands• ~ 40 TB

Survey 10,000 square degrees of Northern Sky over 5 years

Page 5: Steve LloydThe Data Deluge and the GridSlide 1 The Data Deluge and the Grid The Data Deluge The Large Hadron Collider The LHC Data Challenge The Grid Grid.

Steve Lloyd The Data Deluge and the Grid Slide 5

VISTA

Visible and Infrared Survey Telescope for Astronomy

Page 6: Steve LloydThe Data Deluge and the GridSlide 1 The Data Deluge and the Grid The Data Deluge The Large Hadron Collider The LHC Data Challenge The Grid Grid.

Steve Lloyd The Data Deluge and the Grid Slide 6

Virtual Observatories

Crab Nebula

Optical

RadioInfra-red

X-ray

Jet in M87

HST optical

Gemini mid-IR

VLA radio

Chandra X-ray

Page 7: Steve LloydThe Data Deluge and the GridSlide 1 The Data Deluge and the Grid The Data Deluge The Large Hadron Collider The LHC Data Challenge The Grid Grid.

Steve Lloyd The Data Deluge and the Grid Slide 7

NASA’s Earth Observing System

1 TB/day

Galapagos Oil Spill:

Page 8: Steve LloydThe Data Deluge and the GridSlide 1 The Data Deluge and the Grid The Data Deluge The Large Hadron Collider The LHC Data Challenge The Grid Grid.

Steve Lloyd The Data Deluge and the Grid Slide 8

ESA EO Facilities

ESRIN

MATERA (I)

NEUSTREL.ITZ (D)

KIRUNA (S)- ESRANGE

MASPALOMAS (E)

TROMSO (N)

MATERA (I)

SEAWIFS

SPOT

IRS-P3

LANDSAT 7TERRA/MODIS

STANDARDPRODUCTION CHAINS

USERS

HISTORICALARCHIVES

USERS

PRODUCTS

AVHRR

GOME analysis detected ozone thinning over Europe 31 Jan 2002

Page 9: Steve LloydThe Data Deluge and the GridSlide 1 The Data Deluge and the Grid The Data Deluge The Large Hadron Collider The LHC Data Challenge The Grid Grid.

Steve Lloyd The Data Deluge and the Grid Slide 9

Species 2000

To enumerate all ~1.7 million known species of plants, animals, fungi and microbes on Earth

A federation of initially 18 taxonomic databases - eventually ~ 200 databases

Page 10: Steve LloydThe Data Deluge and the GridSlide 1 The Data Deluge and the Grid The Data Deluge The Large Hadron Collider The LHC Data Challenge The Grid Grid.

Steve Lloyd The Data Deluge and the Grid Slide 10

Genomics

Page 11: Steve LloydThe Data Deluge and the GridSlide 1 The Data Deluge and the Grid The Data Deluge The Large Hadron Collider The LHC Data Challenge The Grid Grid.

Steve Lloyd The Data Deluge and the Grid Slide 11

The LHC

The Large Hadron Collider (LHC) will be a 14 TeV centre of mass proton proton collider operating in the existing 26.7Km LEP tunnel at CERN. Due to start operation > 2006

– 1,232 superconducting main dipoles of 8.3Tesla

– 788 quadrupoles– 2,835 bunches of 1011 protons

per bunch spaced by 25ns

Page 12: Steve LloydThe Data Deluge and the GridSlide 1 The Data Deluge and the Grid The Data Deluge The Large Hadron Collider The LHC Data Challenge The Grid Grid.

Steve Lloyd The Data Deluge and the Grid Slide 12

Particle Physics Questions

• Need to discover (confirm) Higgs Particle– Study its properties– Prove that Higgs couplings depend on masses

• Other unanswered questions:– Does Supersymmetry exist?– How are quarks and leptons related?– Why are there 3 sets of quarks and leptons?– What about Gravity?– Anything unexpected?

Page 13: Steve LloydThe Data Deluge and the GridSlide 1 The Data Deluge and the Grid The Data Deluge The Large Hadron Collider The LHC Data Challenge The Grid Grid.

Steve Lloyd The Data Deluge and the Grid Slide 13

The LHC

Page 14: Steve LloydThe Data Deluge and the GridSlide 1 The Data Deluge and the Grid The Data Deluge The Large Hadron Collider The LHC Data Challenge The Grid Grid.

Steve Lloyd The Data Deluge and the Grid Slide 14

The LEP/LHC Tunnel

Page 15: Steve LloydThe Data Deluge and the GridSlide 1 The Data Deluge and the Grid The Data Deluge The Large Hadron Collider The LHC Data Challenge The Grid Grid.

Steve Lloyd The Data Deluge and the Grid Slide 15

LHC Experiments

LHC will house 4 experiments:– ATLAS and CMS are large 'General Purpose'

detectors designed to detect everything and anything

– LHCb is a specialised experiment designed to study CP violation in the b quark system

– ALICE is a dedicated Heavy Ion Physics Detector

Page 16: Steve LloydThe Data Deluge and the GridSlide 1 The Data Deluge and the Grid The Data Deluge The Large Hadron Collider The LHC Data Challenge The Grid Grid.

Steve Lloyd The Data Deluge and the Grid Slide 16

Schematic View of the LHC

Page 17: Steve LloydThe Data Deluge and the GridSlide 1 The Data Deluge and the Grid The Data Deluge The Large Hadron Collider The LHC Data Challenge The Grid Grid.

Steve Lloyd The Data Deluge and the Grid Slide 17

The ATLAS Experiment

ATLAS Consists of – An inner tracker to measures the

momentum of each charged particle – A calorimeter to measure the energies

carried by the particles – A muon spectrometer to identify and

measure muons – A huge magnet system for bending charged

particles for momentum measurement

A total of > 108 electronic channels

Page 18: Steve LloydThe Data Deluge and the GridSlide 1 The Data Deluge and the Grid The Data Deluge The Large Hadron Collider The LHC Data Challenge The Grid Grid.

Steve Lloyd The Data Deluge and the Grid Slide 18

The ATLAS Detector

Page 19: Steve LloydThe Data Deluge and the GridSlide 1 The Data Deluge and the Grid The Data Deluge The Large Hadron Collider The LHC Data Challenge The Grid Grid.

Steve Lloyd The Data Deluge and the Grid Slide 19

Simulated ATLAS Higgs Event

Page 20: Steve LloydThe Data Deluge and the GridSlide 1 The Data Deluge and the Grid The Data Deluge The Large Hadron Collider The LHC Data Challenge The Grid Grid.

Steve Lloyd The Data Deluge and the Grid Slide 20

LHC Event Rates

• The LHC proton bunches collide every 25ns and each collision yields ~20 proton proton interactions superimposed in the Detector i.e.– 40 MHz x 20 = 8x108 pp interactions/sec

• The (110 GeV) Higgs cross section is 24.2pb.• A good channel is H with a branching

ratio of 0.19% and a detector acceptance ~50%– At full (1034cm-2s-1) LHC luminosity this gives

1034 x 24.2x10-12 x 10-24 x 0.0019 x 0.5 = 2x10-4 H per second

A 2x10-4 needle in a 8x108 Haystack

Page 21: Steve LloydThe Data Deluge and the GridSlide 1 The Data Deluge and the Grid The Data Deluge The Large Hadron Collider The LHC Data Challenge The Grid Grid.

Steve Lloyd The Data Deluge and the Grid Slide 21

'Online' Data Reduction

Collision Rate 40 MHz

Level 1 Special Hardware Trigger

Level 2 Embedded Processor Trigger

Level 3 Processor Farm

Raw Data Storage

104 - 105 Hz

102 - 103 Hz

10 - 100 Hz

Offline Data Reconstruction

Selecting interesting events based on progressively more detector information

10-100 GB/sec

40 TB/sec

1-10 GB/sec

100-200 MB/sec

Page 22: Steve LloydThe Data Deluge and the GridSlide 1 The Data Deluge and the Grid The Data Deluge The Large Hadron Collider The LHC Data Challenge The Grid Grid.

Steve Lloyd The Data Deluge and the Grid Slide 22

Offline Analysis

Raw Data from Detector

Physics Analysis

1-2 MB/event @ 100-400 Hz

Data Reconstruction(Digits to

Energy/momentum etc)

Event Summary Data

0.5 MB/event

Analysis Event Selection

Analysis Object Data

10 kB/event

Total Data per year from one experiment

1 to 8 PBytes (1015 Bytes)

Page 23: Steve LloydThe Data Deluge and the GridSlide 1 The Data Deluge and the Grid The Data Deluge The Large Hadron Collider The LHC Data Challenge The Grid Grid.

Steve Lloyd The Data Deluge and the Grid Slide 23

Computing Resources Required

CPU Power (Reconstruction, Simulation, User Analysis etc)– 20 Million SpecInt2000– (A 1 GHz PC is rated at ~400 SpecInt2000)– i.e. 50,000 of yesterday/today's PCs

'Tape' Storage– 20,000 TB

Disk Storage– 2,500 TB

Analysis carried out throughout the world by hundreds of Physicists

Page 24: Steve LloydThe Data Deluge and the GridSlide 1 The Data Deluge and the Grid The Data Deluge The Large Hadron Collider The LHC Data Challenge The Grid Grid.

Steve Lloyd The Data Deluge and the Grid Slide 24

Worldwide Collaboration

CMS: 1800 physicists150 institutes32 countries

Page 25: Steve LloydThe Data Deluge and the GridSlide 1 The Data Deluge and the Grid The Data Deluge The Large Hadron Collider The LHC Data Challenge The Grid Grid.

Steve Lloyd The Data Deluge and the Grid Slide 25

Solutions

• Distributed solution:– exploit established computing expertise &

infrastructure in national labs and universities– reduce dependence on links to CERN– tap additional funding sources (spin off)

Is the Grid the solution?

• Centralised Solution:– Put all resources at CERN

• Funding agencies certainly won't place all their investment at CERN

• Sociological problems

Page 26: Steve LloydThe Data Deluge and the GridSlide 1 The Data Deluge and the Grid The Data Deluge The Large Hadron Collider The LHC Data Challenge The Grid Grid.

Steve Lloyd The Data Deluge and the Grid Slide 26

What is The Grid?

Analogy with the Electricity Power Grid:– Unlimited ubiquitous distributed computing– Transparent access to multipetabyte

distributed databases– Easy to plug in– Complexity of infrastructure hidden

Page 27: Steve LloydThe Data Deluge and the GridSlide 1 The Data Deluge and the Grid The Data Deluge The Large Hadron Collider The LHC Data Challenge The Grid Grid.

Steve Lloyd The Data Deluge and the Grid Slide 27

The Grid

Ian Foster and Carl Kesselman, editors, “The Grid: Blueprint for a New Computing Infrastructure,” Morgan Kaufmann, 1999, http://www.mkp.com/grids

Five emerging models:

•Distributed Computing

- synchronous processing

• High-Throughput Computing

- asynchronous processing

• On-Demand Computing

- dynamic resources

• Data-Intensive Computing

- databases

• Collaborative Computing

- scientists

Page 28: Steve LloydThe Data Deluge and the GridSlide 1 The Data Deluge and the Grid The Data Deluge The Large Hadron Collider The LHC Data Challenge The Grid Grid.

Steve Lloyd The Data Deluge and the Grid Slide 28

The Grid

Ian Foster / Carl Kesselman: "A computational Grid is a hardware

and software infrastructure that provides dependable, consistent, pervasive and inexpensive access to high-end computational capabilities."

Page 29: Steve LloydThe Data Deluge and the GridSlide 1 The Data Deluge and the Grid The Data Deluge The Large Hadron Collider The LHC Data Challenge The Grid Grid.

Steve Lloyd The Data Deluge and the Grid Slide 29

The Grid

• Dependable - Need to rely on remote equipment as much as the machine on your desk

• Consistency - Machines need to communicate so need consistent environments and interfaces

• Pervasive - The more resources that participate in the same system the more useful they all are

• Inexpensive - Important for pervasiveness - i.e. built using commodity PCs and disks

Page 30: Steve LloydThe Data Deluge and the GridSlide 1 The Data Deluge and the Grid The Data Deluge The Large Hadron Collider The LHC Data Challenge The Grid Grid.

Steve Lloyd The Data Deluge and the Grid Slide 30

The Grid

• You simply submit your job to the 'Grid'- you shouldn't have to know where the data you want is or where the job will run. The Grid software (Middleware) will take care of:– running the job where the data is or– moving the data to where there is CPU power

available

Page 31: Steve LloydThe Data Deluge and the GridSlide 1 The Data Deluge and the Grid The Data Deluge The Large Hadron Collider The LHC Data Challenge The Grid Grid.

Steve Lloyd The Data Deluge and the Grid Slide 31

@#%&*!

The Grid for the Scientist

E = mc2

Grid Middleware

“Putting the bottleneck back in the Scientist’s mind”

Page 32: Steve LloydThe Data Deluge and the GridSlide 1 The Data Deluge and the Grid The Data Deluge The Large Hadron Collider The LHC Data Challenge The Grid Grid.

Steve Lloyd The Data Deluge and the Grid Slide 32

Grid Tiers

• For the LHC we envisage a 'Hierarchical' structure based on several 'Tiers' since the data mostly originates at one place:– Tier-0 - CERN - the source of the data– Tier-1 - ~ 10 Major Regional Centres (inc

UK)– Tier-2 - smaller more specialised Regional

Centres (4 in UK?)– Tier-3 - University Groups– Tier-4 – My laptop? Mobile Phone?

• Doesn't need to be hierarchical e.g. for Biologists probably not desirable

Page 33: Steve LloydThe Data Deluge and the GridSlide 1 The Data Deluge and the Grid The Data Deluge The Large Hadron Collider The LHC Data Challenge The Grid Grid.

Steve Lloyd The Data Deluge and the Grid Slide 33

Grid Services

Resource-specific implementations of basic servicese.g., Transport protocols, name servers, differentiated services, CPU schedulers, public keyinfrastructure, site accounting, directory service, OS bypass

Resource-independent and application-independent services authentication, authorization, resource location, resource allocation, events, accounting,

remote data access, information, policy, fault detection

DistributedComputing

Toolkit

Grid Fabric(Resources)

Grid Services(Middleware)

ApplicationToolkits

Data-Intensive

ApplicationsToolkit

CollaborativeApplications

Toolkit

RemoteVisualizationApplications

Toolkit

ProblemSolving

ApplicationsToolkit

RemoteInstrumentation

ApplicationsToolkit

Applications Chemistry

Biology

Cosmology

Particle Physics

Environment

Page 34: Steve LloydThe Data Deluge and the GridSlide 1 The Data Deluge and the Grid The Data Deluge The Large Hadron Collider The LHC Data Challenge The Grid Grid.

Steve Lloyd The Data Deluge and the Grid Slide 34

Problems

• Scalability– Will it scale to thousands of processors,

thousands of disks, PetaBytes of data, Terabits/sec of IO?

• Wide-area distribution– How to distribute, replicate, cache,

synchronise, catalogue the data?– How to balance local ownership of resources

with the requirements of the whole?• Adaptability/Flexibility

– Need to adapt to rapidly changing hardware and costs, new analysis methods etc.

Page 35: Steve LloydThe Data Deluge and the GridSlide 1 The Data Deluge and the Grid The Data Deluge The Large Hadron Collider The LHC Data Challenge The Grid Grid.

Steve Lloyd The Data Deluge and the Grid Slide 35

SETI@home

• A distributed computing project - not really a Grid project

• You pull the data from them rather than they submit the job to you– total of 4,591,332 users– 963,646,331 results received– 1,545,634 years of cpu time– 3.3x1021 floating point operations– 125 different cpu types – 143 different operating systems

Arecibo telescope in Puerto Rico

Page 36: Steve LloydThe Data Deluge and the GridSlide 1 The Data Deluge and the Grid The Data Deluge The Large Hadron Collider The LHC Data Challenge The Grid Grid.

Steve Lloyd The Data Deluge and the Grid Slide 36

SETI@home

Page 37: Steve LloydThe Data Deluge and the GridSlide 1 The Data Deluge and the Grid The Data Deluge The Large Hadron Collider The LHC Data Challenge The Grid Grid.

Steve Lloyd The Data Deluge and the Grid Slide 37

Entropia

Uses idle cycles on Home PCs for profit and non-profit projects:

Mersenne Prime Search• 42,519 machines active• 560 years of cpu per day

FightAIDS@Home• 60,000 Machines• 1,400 years of cpu time

Page 38: Steve LloydThe Data Deluge and the GridSlide 1 The Data Deluge and the Grid The Data Deluge The Large Hadron Collider The LHC Data Challenge The Grid Grid.

Steve Lloyd The Data Deluge and the Grid Slide 38

NASA Information Power Grid

• Knit together widely distributed computing, data, instrumentation and human resources

• to address complex large scale computing and data analysis problems

Page 39: Steve LloydThe Data Deluge and the GridSlide 1 The Data Deluge and the Grid The Data Deluge The Large Hadron Collider The LHC Data Challenge The Grid Grid.

Steve Lloyd The Data Deluge and the Grid Slide 39

Collaborative Engineering

Real-timecollection

Multi-sourceData Analysis

Unitary Plan Wind Tunnel

Archival storage

Page 40: Steve LloydThe Data Deluge and the GridSlide 1 The Data Deluge and the Grid The Data Deluge The Large Hadron Collider The LHC Data Challenge The Grid Grid.

Steve Lloyd The Data Deluge and the Grid Slide 40

Other Grid Applications

• Distributed Supercomputing– Simultaneous execution across

multiple supercomputers• Smart Instruments

– Enhance the power of scientific instruments by providing access to data archives and online processing capabilities and visualisation e.g. coupling Argonne’s Photon

Source to a supercomputer

Page 41: Steve LloydThe Data Deluge and the GridSlide 1 The Data Deluge and the Grid The Data Deluge The Large Hadron Collider The LHC Data Challenge The Grid Grid.

Steve Lloyd The Data Deluge and the Grid Slide 41

GridPP

http://www.gridpp.ac.uk

Page 42: Steve LloydThe Data Deluge and the GridSlide 1 The Data Deluge and the Grid The Data Deluge The Large Hadron Collider The LHC Data Challenge The Grid Grid.

Steve Lloyd The Data Deluge and the Grid Slide 42

GridPP Overview

Provide architecture and middleware

Use the Grid with simulation data

Use the Grid with real data

Future LHC Experiments

Running US Experiments

Build prototype Tier-1 and Tier-2s in the UK and implement middleware in experiments

Page 43: Steve LloydThe Data Deluge and the GridSlide 1 The Data Deluge and the Grid The Data Deluge The Large Hadron Collider The LHC Data Challenge The Grid Grid.

Steve Lloyd The Data Deluge and the Grid Slide 43

The Prototype UK Tier-1

March 2003 •560 CPUs (450Mhz-1.4GHz)•50 TB Disk•35 TB Tape in use (theoretical

tape capacity 366 TB)

Page 44: Steve LloydThe Data Deluge and the GridSlide 1 The Data Deluge and the Grid The Data Deluge The Large Hadron Collider The LHC Data Challenge The Grid Grid.

Steve Lloyd The Data Deluge and the Grid Slide 44

Conclusions

• Enormous data challenges in next few years.

• The Grid is likely solution.• The Web gives ubiquitous access to

distributed information.• The Grid will give ubiquitous access to

computing resources and hence knowledge.• Many Grid projects and testbeds starting to

take off.• GridPP is building a UK Grid for Particle

Physicists to prepare for future LHC Data.