High-Luminosity Large Hadron Collider - ESFRI eu · High-Luminosity Large Hadron Collider Dr Tim...
Transcript of High-Luminosity Large Hadron Collider - ESFRI eu · High-Luminosity Large Hadron Collider Dr Tim...
CERN – IT Department CH-1211 Genève 23
Switzerland
www.cern.ch/it
High-Luminosity Large Hadron Collider
Dr Tim SMITH
ESFRI Workshop on 'Monitoring of RIs, periodic update of Landmarks, use of KPIs’
Milan – 2018/11/19
ESFRI Landmark
@TimSmithCH
CERN
High Luminosity
LHC
2026 - 2038
Large Hadron Collider
2010 - 2026
@TimSmithCH
50 PB/y Long Term Plan
150 PB/y 600 PB/y
320 PB EB ZB
Rate:
Total:
Run 1 Run 2 HL
LHC Run 4 Run 3 Run 5 Run 6
@TimSmithCH
Worldwide LHC Computing Grid
Tier 0: Data recording, reconstruction
and distribution
Tier 1: Permanent
storage, re-processing,
analysis
Tier 2: Simulation,
end-user analysis
167 Sites 42 Countries
63 MoUs
700 PB 2M jobs/day
WLCG: An International collaboration to distribute and analyse LHC data Integrates computer centres worldwide that provide computing and storage
resources into a single infrastructure accessible by all LHC physicists
@TimSmithCH
Keeping Data Accessible • Media Verification
– Hot / Cold Data
– 10% of production drive capacity for 2.6 years
• Media Migration – Drive and Media obsolescence
– 50% of current drive capacity for 2 years
Bit Rot
O(10-16) bit error rate
@TimSmithCH
CERN as a Trusted Digital Repository
• ISO 16363 certification - to implement best practices, for the long-term
Jamie Shiers (CERN)
Artefactual Systems
• Scope
– Scientific Data from CERN’s accelerator experiments
– Scientific publications, papers and articles;
– CERN’s “digital memory” • meeting minutes, photographs, video recordings and
scanned legacy material
• Timescale: complete prior to 2020
@TimSmithCH
Archiver Pre-commercial Procurement
2018 2019 2020 2021 2022 2023
Procured Resulting Services
Data Archiving in the cloud: Complement in-house data archiving/preservation with remote commercial cloud services
@TimSmithCH
Long-term Value of Data
LEP: 1989-2000
@TimSmithCH
Research Iceberg
kB 1,000
MB 1,000,000
GB 1,000,000,000
TB 1,000,000,000,000
PB 1,000,000,000,000,000
@TimSmithCH
Open Science Policies
Level 1: All, with publication
Level 2: As much as possible, as soon as possible
Level 3: 50% after 5 years, 100% after 10 years
Level 4: None (even collaboration dont have access)
@TimSmithCH
CERN Open Data
@TimSmithCH
Preservation ↔ Reusability
http://analysispreservation.cern.ch http://reana.io
@TimSmithCH
Task 2.2 Content Delivering and Caching
HTC/Grid
Cloud/commercial
HPC
citizen
Task 2.3 Efficient Access to Compute
Task 2.1 Storage Services
Task 2.1 Data transfer services
Task 2.4 Networking
Task 2.5 AAI
Task 2.2 Storage Orchestration Service
Goals: Prototype an infrastructure for the EOSC that is adapted to the Exabyte-scale needs of the large ESFRI science projects. Ensure that the science communities drive the development of the EOSC. Has to address FAIR data management, long term preservation, open access, open science, and contribute to the EOSC catalogue of services.
Work Packages WP2 – Data Infrastructure for Open Science WP3 – Open-source scientific Software and Service Repository WP4 – Connecting ESFRI projects to EOSC through VO framework WP5 – ESFRI Science Analysis Platform
Data centres (funded in WP2)
CERN, INFN, DESY, GSI, Nikhef, SURFSara, RUG, CCIN2P3, PIC, LAPP, INAF
@TimSmithCH
http://opendata.cern.ch
http://analysispreservation.cern.ch
http://reana.io
http://invenio-software.org
http://zenodo.org
0000-0002-1567-7116
http://cern.ch/tim.smith
Thank You ! Questions ?