Jürgen Knobloch/CERN Slide 1 A Global Computer – the Grid Is Reality by Jürgen Knobloch October...
-
Upload
kristian-andrews -
Category
Documents
-
view
220 -
download
0
Transcript of Jürgen Knobloch/CERN Slide 1 A Global Computer – the Grid Is Reality by Jürgen Knobloch October...
Jürgen Knobloch/CERN Slide 1
A Global Computer – the Grid Is Realityby Jürgen Knobloch
October 31, 2007
Jürgen Knobloch/CERN Slide 2
LHC gets ready …
… what about Computing?
• The Challenge• Starting to grasp the scope• Going for the Grid• Are we there?• What other use is the Grid?• Where do we go from here?
Jürgen Knobloch/CERN Slide 3
Jürgen Knobloch/CERN Slide 4
The LHC Computing Challenge
• Signal/Noise 10-9
• Data volume– High rate * large number of
channels * 4 experiments
15 PetaBytes of new data each year
• Compute power– Event complexity * Nb. events *
thousands users
100 k of (today's) fastest CPUs
• Worldwide analysis & funding– Computing funding locally in
major regions & countries– Efficient analysis everywhere
GRID technology
Jürgen Knobloch/CERN Slide 5
Timeline LHC Computing
1994 1995 1996 1997 1998 1999 2000 2001 2002 2003 2004 2005 2006 2007 2008
LHC approved
ATLAS & CMS approved
ALICEapproved
LHCb approved
“Hoffmann”Review
7x107 MIPS1,900 TB disk
ATLAS (or CMS) requirementsfor first year at design luminosity
ATLAS&CMSCTP
107 MIPS100 TB disk
LHC start
ComputingTDRs
55x107 MIPS70,000 TB disk
(140 MSi2K)
CSC 2007
LCGEvolution of CPU Capacity at
CERN
SC (0.6GeV)
PS (28GeV)ISR (300GeV)
SPS (400GeV)
ppbar(540GeV)
LEP (100GeV)
LEP II (200GeV)
LHC (14 TeV)
Costs (2007Swiss Francs)
Includes infrastructurecosts (comp.centre,
power, cooling, ..) and physics tapes
Slide fro
m
Les Roberts
on
Requirements Match
Jürgen Knobloch/CERN Slide 7
Tape & diskrequirements:
>10 times CERNpossibility
Options as seen in 1996
Jürgen Knobloch/CERN Slide 8
Jürgen Knobloch/CERN Slide 9
Timeline Grids
Data Challenges
First physics
Cosmics
GRID 3
EGEE 1
LCG 1
Service Challenges
EU DataGrid
GriPhyN, iVDGL, PPDG
EGEE 2
OSG
LCG 2
EGEE 3
1994 1995 1996 1997 1998 1999 2000 2001 2002 2003 2004 2005 2006 2007 2008
WLCG Partially decentralized
model– replicate the event data at
about five regional centres
– data transfer via network ormovable media
RC2
CERN
RC1
Jürgen Knobloch/CERN Slide 10
WLCG Collaboration
• The Collaboration– 4 LHC experiments– ~250 computing centres– 12 large centres
(Tier-0, Tier-1)– 38 federations of smaller
“Tier-2” centres– Growing to ~40 countries– Grids: EGEE, OSG, Nordugrid
• Technical Design Reports– WLCG, 4 Experiments: June 2005
• Memorandum of Understanding– Agreed in October 2005
• Resources– 5-year forward look
Jürgen Knobloch/CERN Slide 11
CPU & Disk Requirements 2006
0
50
100
150
200
250
300
350
2007 2008 2009 2010
Year
MS
I200
0
LHCb-Tier-2
CMS-Tier-2
ATLAS-Tier-2
ALICE-Tier-2
LHCb-Tier-1
CMS-Tier-1
ATLAS-Tier-1
ALICE-Tier-1
LHCb-CERN
CMS-CERN
ATLAS-CERN
ALICE-CERN
0
20
40
60
80
100
120
140
2007 2008 2009 2010
Year
PB
CPU Disk
CERN: ~ 10%
Events at LHC
Jürgen Knobloch/CERN Slide 12
Trigger and Data Acquisition
Jürgen Knobloch/CERN Slide 13
Tier-0 Recording
Jürgen Knobloch/CERN Slide 14
Tier-0 -1 -2
Jürgen Knobloch/CERN Slide 15
Jürgen Knobloch/CERN Slide 16
Centers around the world form a Supercomputer
• The EGEE and OSG projects are the basis of the Worldwide LHC Computing Grid Project WLCG
Inter-operation between Grids is working!
Jürgen Knobloch/CERN Slide 17
Middleware
• Security – Virtual Organization Management (VOMS) – MyProxy
• Data management – File catalogue (LFC)– File transfer service (FTS)– Storage Element (SE)– Storage Resource Management (SRM)
• Job management – Work Load Management System(WMS)– Logging and Bookeeping (LB)– Computing Element (CE)– Worker Nodes (WN)
• Information System– Monitoring: BDII (Berkeley Database Information Index), RGMA
(Relational Grid Monitoring Architecture) aggregate service information from multiple Grid sites, now moved to SAM (Site Availability Monitoring)
– Monitoring & visualization (Griview, Dashboard, Gridmap etc.)
Moore Delivered for CPU & Disk
Jürgen Knobloch/CERN Slide 18
Slides from 1996
Expectations fulfilled!
Network was a Concern …
Jürgen Knobloch/CERN Slide 19
We arenow here
Jürgen Knobloch/CERN Slide 20
LHCOPN Architecture
Tier-2s and Tier-1s are inter-connected by the general
purpose research networks
Any Tier-2 mayaccess data at
any Tier-1
Tier-2 IN2P3
TRIUMF
ASCC
FNAL
BNL
Nordic
CNAF
SARAPIC
RAL
GridKa
Tier-2
Tier-2
Tier-2
Tier-2
Tier-2
Tier-2
Tier-2Tier-2
Tier-2
Jürgen Knobloch/CERN Slide 21
Networking
Data Transfer out of Tier-0
Jürgen Knobloch/CERN Slide 22
Jürgen Knobloch/CERN Slide 23
Site reliability
Site Reliability
Site ReliabilityTier-2 Sites
83 Tier-2 sites being monitored
Targets – CERN + Tier-1sBefore
July July 07 Dec 07 Avg.last 3 months
Each site 88% 91% 93% 89%
8 best sites 88% 93% 95% 93%
Jürgen Knobloch/CERN Slide 25
ARDA Dashboard
Gridmap
Jürgen Knobloch/CERN Slide 26
Increasing workloads
27EGEE'07; 2nd October 2007
32%non-LHC
Many Grid Applications
At present there are about 20 applications from more than 10 domains on the EGEE Grid infrastructure– Astronomy & Astrophysics - MAGIC, Planck – Computational Chemistry – Earth Sciences - Earth Observation, Solid Earth Physics,
Hydrology, Climate – Fusion – High Energy Physics - 4 LHC experiments (ALICE, ATLAS, CMS,
LHCb) BaBar, CDF, DØ, ZEUS – Life Sciences - Bioinformatics (Drug Discovery, GPS@,
Xmipp_MLrefine, etc.) – Condensed Matter Physics– Computational Fluid Dynamics – Computer Science/Tools – Civil Protection – Finance (through the Industry Task Force)
Jürgen Knobloch/CERN Slide 28
Grid Applications
Jürgen Knobloch/CERN Slide 29
Medical Metadata
Seismology
Chemistry
Astronomy
FusionParticle Physics
Available Infrastructure
EGEE: ~250 sites, >45000 CPUOSG: ~ 15 sites for LHC, > 10000 CPU
¼ of the resources are contributed by groups external to the project
~>25 k simultaneous jobs
30EGEE'07; 2nd October 2007
Ramp-up Needed for Startup
Jul Sep Apr -07 -07 -08
3.7 X Sep Jul Apr -06 -07 -08
Sep Jul Apr -06 -07 -08
3 X 2.9 X
Sep Jul Apr -06 -07 -08
Sep Jul Apr -06 -07 -08
2.3 X 3.7 X target usageusage
pledgeinstalled
Jürgen Knobloch/CERN Slide 32
3D - Distributed Deployment of Databases for LCG
ORACLE Streaming with Downstream Capture
(ATLAS, LHCb)
SQUID/FRONTIER Web caching
(CMS)
The Next Step
Jürgen Knobloch/CERN Slide 33
EGI – European Grid Initiative
Jürgen Knobloch/CERN Slide 34
www.eu-egi.orgwww.eu-egi.org EGI Design Study proposal to the European Commission (started Sept 07)
Supported by 37 National Grid Initiatives (NGIs)
2 year project to prepare the setup and operation of a new organizational model for a sustainable pan-European grid infrastructure after the end of EGEE-3
Jürgen Knobloch/CERN Slide 35
Tier-1 Centers: TRIUMF (Canada); GridKA(Germany); IN2P3 (France); CNAF (Italy); SARA/NIKHEF (NL); Nordic Data Grid Facility (NDGF); ASCC (Taipei); RAL (UK); BNL (US); FNAL (US); PIC (Spain)
The Grid is now in operation, working on: reliability, scaling up, sustainability