Data Management Overview David M. Malon [email protected] U.S. ATLAS Computing Meeting Brookhaven, New...

30
Data Management Overview David M. Malon David M. Malon [email protected] [email protected] U.S. ATLAS Computing Meeting U.S. ATLAS Computing Meeting Brookhaven, New York Brookhaven, New York 28 August 2003 28 August 2003

Transcript of Data Management Overview David M. Malon [email protected] U.S. ATLAS Computing Meeting Brookhaven, New...

Page 1: Data Management Overview David M. Malon malon@anl.gov U.S. ATLAS Computing Meeting Brookhaven, New York 28 August 2003.

Data Management Overview

David M. MalonDavid M. Malon

[email protected]@anl.gov

U.S. ATLAS Computing MeetingU.S. ATLAS Computing Meeting

Brookhaven, New YorkBrookhaven, New York28 August 200328 August 2003

Page 2: Data Management Overview David M. Malon malon@anl.gov U.S. ATLAS Computing Meeting Brookhaven, New York 28 August 2003.

28 August 200328 August 2003David Malon, ANL U.S. ATLAS Computing MeetingDavid Malon, ANL U.S. ATLAS Computing Meeting

2

Outline

Technology and technology transitionTechnology and technology transition

POOL and POOL integrationPOOL and POOL integration

Detector description and primary numbersDetector description and primary numbers

Interval-of-validity databases and conditionsInterval-of-validity databases and conditions

Data Challenge databases: Magda, AMI, metadata, and virtual dataData Challenge databases: Magda, AMI, metadata, and virtual data

Near-term plansNear-term plans

Interactions with online, and with Technical CoordinationInteractions with online, and with Technical Coordination

Staffing Staffing

ConclusionsConclusions

Page 3: Data Management Overview David M. Malon malon@anl.gov U.S. ATLAS Computing Meeting Brookhaven, New York 28 August 2003.

28 August 200328 August 2003David Malon, ANL U.S. ATLAS Computing MeetingDavid Malon, ANL U.S. ATLAS Computing Meeting

3

Technology transition

ATLAS database strategy has been, consistently. to employ “LHC ATLAS database strategy has been, consistently. to employ “LHC

common solutions wherever possible”common solutions wherever possible” Until last year this meant Objectivity/DB as the baseline technology Objectivity/DB conversion services retained as a reference implementation

until developer releases leading to 6.0.0; retired at the end of 2002

Today’s baseline is LCG POOL (hybrid relational and ROOT-based Today’s baseline is LCG POOL (hybrid relational and ROOT-based

streaming layer)streaming layer) ATLAS is contributing staff to POOL development teams All ATLAS event store development is POOL-based

Transition period technology: AthenaROOT conversion serviceTransition period technology: AthenaROOT conversion service Deployed pre-POOL; provided input to persistence RTAG that led to POOL AthenaROOT service will, like Objectivity, be retired once POOL

infrastructure is sufficiently mature

Page 4: Data Management Overview David M. Malon malon@anl.gov U.S. ATLAS Computing Meeting Brookhaven, New York 28 August 2003.

4David Malon, ANL U.S. ATLAS Computing MeetingDavid Malon, ANL U.S. ATLAS Computing Meeting

28 August 200328 August 2003

What is POOL?

POOL is the LCG Persistence FrameworkPOOL is the LCG Persistence Framework Pool of persistent objects for LHC

Started by LCG-SC2 in April ’02Started by LCG-SC2 in April ’02 Common effort in which the experiments take over a major share of

the responsibility for defining the system architecture for development of POOL components

ramping up over the last year from 1.5 to ~10FTE

Dirk Duellmann is project leaderDirk Duellmann is project leader Information on this and the next several slides borrowed from him

See See http://pool.http://pool.cerncern..chch for details for details

Page 5: Data Management Overview David M. Malon malon@anl.gov U.S. ATLAS Computing Meeting Brookhaven, New York 28 August 2003.

28 August 200328 August 2003David Malon, ANL U.S. ATLAS Computing MeetingDavid Malon, ANL U.S. ATLAS Computing Meeting

5

POOL and the LCG Architecture Blueprint

POOL is a component-based systemPOOL is a component-based system A technology-neutral API

Abstract C++ interfaces Implemented reusing existing technology

ROOT I/O for object streaming complex data, simple consistency model

RDBMS for consistent metadata handling simple data, transactional consistency

POOL does not replace any of its components technologiesPOOL does not replace any of its components technologies It integrates them to provides higher level services Insulates physics applications from implementation details of

components and technologies used today

Page 6: Data Management Overview David M. Malon malon@anl.gov U.S. ATLAS Computing Meeting Brookhaven, New York 28 August 2003.

28 August 200328 August 2003David Malon, ANL U.S. ATLAS Computing MeetingDavid Malon, ANL U.S. ATLAS Computing Meeting

6

POOL Work Package breakdown

Storage ServiceStorage Service Stream transient C++ objects into/from storage Resolve a logical object reference into a physical object

File CatalogFile Catalog Track files (and physical/logical names) and their descriptions Resolve a logical file reference (FileID) into a physical file

Collections and MetadataCollections and Metadata Track (large, possibly distributed) collections of objects and their

descriptions (“tag” data); support object-level selection Object Cache (DataService)Object Cache (DataService)

Track and manage objects already in transient memory to speed access

Page 7: Data Management Overview David M. Malon malon@anl.gov U.S. ATLAS Computing Meeting Brookhaven, New York 28 August 2003.

28 August 200328 August 2003David Malon, ANL U.S. ATLAS Computing MeetingDavid Malon, ANL U.S. ATLAS Computing Meeting

7

POOL Internal Organization

POOL API

Storage Service FileCatalog Collections

ROOT I/OStorage Svc

XMLCatalog

MySQLCatalog

EDG Replica Location Service

ExplicitCollection

ImplicitCollection

RDBMSStorage Svc ?

Page 8: Data Management Overview David M. Malon malon@anl.gov U.S. ATLAS Computing Meeting Brookhaven, New York 28 August 2003.

28 August 200328 August 2003David Malon, ANL U.S. ATLAS Computing MeetingDavid Malon, ANL U.S. ATLAS Computing Meeting

8

POOL integration into ATLAS

AthenaPOOL conversion service prototype is available in current AthenaPOOL conversion service prototype is available in current

releasesreleases

Scheduled to be ready for early adopters in Release 7.0.0Scheduled to be ready for early adopters in Release 7.0.0 Based upon this month’s “user release” of POOL

POOL releases have been pretty much on schedulePOOL releases have been pretty much on schedule Current release, 1.2.1, incorporates most recent LCG SEAL release

Several integration issues are resolved in a stopgap fashion; much Several integration issues are resolved in a stopgap fashion; much

work remainswork remains

Further LCG dictionary work (SEAL) will be required to represent Further LCG dictionary work (SEAL) will be required to represent

ATLAS event modelATLAS event model

Page 9: Data Management Overview David M. Malon malon@anl.gov U.S. ATLAS Computing Meeting Brookhaven, New York 28 August 2003.

28 August 200328 August 2003David Malon, ANL U.S. ATLAS Computing MeetingDavid Malon, ANL U.S. ATLAS Computing Meeting

9

ATLAS POOL/SEAL integration

Many nuisance technical obstacles to POOL integration into ATLASMany nuisance technical obstacles to POOL integration into ATLAS

Not long-term concerns, but in the first year, they consume a great deal Not long-term concerns, but in the first year, they consume a great deal

of time of time Integration of POOL into ATLAS/Athena has not been trivial

ExamplesExamples Conflicts in how cmt/scram/ATLAS/SPI handle build environments,

compiler/linker settings, external packages and versions, … Conflicts between Gaudi/Athena dynamic loading infrastructure and SEAL

plug-in management Conflicts in lifetime management with multiple transient caches (Athena

StoreGate and POOL DataSvc) Issues in type identification handling between Gaudi/Athena and the

emerging SEAL dictionary Keeping up with moving targets (but rapid LCG development is good!)

Page 10: Data Management Overview David M. Malon malon@anl.gov U.S. ATLAS Computing Meeting Brookhaven, New York 28 August 2003.

28 August 200328 August 2003David Malon, ANL U.S. ATLAS Computing MeetingDavid Malon, ANL U.S. ATLAS Computing Meeting

10

ATLAS contributions to POOL

ATLAS has principal responsibility for POOL collections and metadata ATLAS has principal responsibility for POOL collections and metadata

work packagework package

Principal responsibility for POOL MySQL and related (e.g., MySQL++) Principal responsibility for POOL MySQL and related (e.g., MySQL++)

package and server configurationpackage and server configuration

Also contributing to foreign object persistence for ROOTAlso contributing to foreign object persistence for ROOT

Contributions to overall architecture, dataset ideas, requirements, …Contributions to overall architecture, dataset ideas, requirements, …

Related: participation in SEAL project’s dictionary requirements/design Related: participation in SEAL project’s dictionary requirements/design

efforteffort

Expect to contribute strongly to newly endorsed common project on Expect to contribute strongly to newly endorsed common project on

conditions data management when it is launchedconditions data management when it is launched

Page 11: Data Management Overview David M. Malon malon@anl.gov U.S. ATLAS Computing Meeting Brookhaven, New York 28 August 2003.

28 August 200328 August 2003David Malon, ANL U.S. ATLAS Computing MeetingDavid Malon, ANL U.S. ATLAS Computing Meeting

11

Notes on relational technology

POOL relational layer work is intended to be readily portable—to make POOL relational layer work is intended to be readily portable—to make

no deep assumptions about choice of relational technologyno deep assumptions about choice of relational technology

Collections work is currently implemented in MySQL; file catalog has Collections work is currently implemented in MySQL; file catalog has

MySQL and Oracle9i implementationsMySQL and Oracle9i implementations

Heterogeneous implementations are possible, e.g., with Oracle9i at Heterogeneous implementations are possible, e.g., with Oracle9i at

CERN and MySQL on small sitesCERN and MySQL on small sites CERN IT is an Oracle shop

Some planning afoot to put Oracle at Tier1s, and possibly beyond

Note that non-POOL ATLAS database work has tended to be Note that non-POOL ATLAS database work has tended to be

implemented in MySQL; like POOL, avoiding technology-specific implemented in MySQL; like POOL, avoiding technology-specific

design decisions design decisions

Page 12: Data Management Overview David M. Malon malon@anl.gov U.S. ATLAS Computing Meeting Brookhaven, New York 28 August 2003.

28 August 200328 August 2003David Malon, ANL U.S. ATLAS Computing MeetingDavid Malon, ANL U.S. ATLAS Computing Meeting

12

Detector description databases

““Primary numbers” (numbers that parameterize detector Primary numbers” (numbers that parameterize detector

description) database deployed based upon NOVA description) database deployed based upon NOVA Leverages externally developed software

Begun as a BNL LDRD project (Vaniachine, Nevski, Wenaus) Current implementation based upon MySQL

NOVA also used for other ATLAS purposes

Used increasingly in Athena directly (NovaConversionSvc), Used increasingly in Athena directly (NovaConversionSvc),

via GeoModel, and by standalone Geant3 and Geant4 via GeoModel, and by standalone Geant3 and Geant4

applicationsapplications

New data continually being addedNew data continually being added Most recently toroids/feet/rails, tiletest data, materials

Page 13: Data Management Overview David M. Malon malon@anl.gov U.S. ATLAS Computing Meeting Brookhaven, New York 28 August 2003.

28 August 200328 August 2003David Malon, ANL U.S. ATLAS Computing MeetingDavid Malon, ANL U.S. ATLAS Computing Meeting

13

NOVA work

Automatic generation of converters and object headers Automatic generation of converters and object headers

from database content integrated into nightly build from database content integrated into nightly build

infrastructure infrastructure

Work needed on input interfaces to NOVA, and on Work needed on input interfaces to NOVA, and on

consistent approaches to event/nonevent data object consistent approaches to event/nonevent data object

definitiondefinition

Sasha Vaniachine is the principal contact personSasha Vaniachine is the principal contact person

Page 14: Data Management Overview David M. Malon malon@anl.gov U.S. ATLAS Computing Meeting Brookhaven, New York 28 August 2003.

28 August 200328 August 2003David Malon, ANL U.S. ATLAS Computing MeetingDavid Malon, ANL U.S. ATLAS Computing Meeting

14

NOVA browser screenshots

Page 15: Data Management Overview David M. Malon malon@anl.gov U.S. ATLAS Computing Meeting Brookhaven, New York 28 August 2003.

28 August 200328 August 2003David Malon, ANL U.S. ATLAS Computing MeetingDavid Malon, ANL U.S. ATLAS Computing Meeting

15

Interval-of-validity (IOV) databases ATLAS was a principal contributor of requirements to an LHC conditions ATLAS was a principal contributor of requirements to an LHC conditions

database project organized under the aegis of RD45database project organized under the aegis of RD45

LCG SC2 in June endorsed a common project on conditions data, to begin LCG SC2 in June endorsed a common project on conditions data, to begin

soon, with this work as its starting pointsoon, with this work as its starting point

Lisbon TDAQ group has provided a MySQL implementation of common Lisbon TDAQ group has provided a MySQL implementation of common

project interfaceproject interface

ATLAS online/offline collaboration ATLAS online/offline collaboration Offline uses Lisbon-developed implementation in its releases as an interval-of-

validity database Offline provides an Athena service based upon this implementation

Prototype (with interval-of-validity access to data in NOVA) in releases nowPrototype (with interval-of-validity access to data in NOVA) in releases now

Due for early adopter use in Release 7.0.0Due for early adopter use in Release 7.0.0

Page 16: Data Management Overview David M. Malon malon@anl.gov U.S. ATLAS Computing Meeting Brookhaven, New York 28 August 2003.

28 August 200328 August 2003David Malon, ANL U.S. ATLAS Computing MeetingDavid Malon, ANL U.S. ATLAS Computing Meeting

16

IOV databases

ATLAS database architecture extends “usual” thinking about conditions ATLAS database architecture extends “usual” thinking about conditions

serviceservice

Interval-of-validity database acts as a Interval-of-validity database acts as a registration serviceregistration service and and mediatormediator

for many kinds of datafor many kinds of data

Example: Example: Geometry data is written in an appropriate technology (e.g., POOL), and

later “registered” in the IOV database with a range of runs as its interval of

validity

Similarly for calibrations produced offline

No need to figure out how to represent complex objects in another storage

technology (conditions database) when this is a problem already solved for

event store technology (POOL)

Page 17: Data Management Overview David M. Malon malon@anl.gov U.S. ATLAS Computing Meeting Brookhaven, New York 28 August 2003.

28 August 200328 August 2003David Malon, ANL U.S. ATLAS Computing MeetingDavid Malon, ANL U.S. ATLAS Computing Meeting

17

IOV Database

Conditions Data Writer

Conditions or other time-varying data

1. Store an instance of data that may vary with time or run or …

2. Return reference to data

3. Register reference,

assigning interval of

validity, tag, …

Page 18: Data Management Overview David M. Malon malon@anl.gov U.S. ATLAS Computing Meeting Brookhaven, New York 28 August 2003.

28 August 200328 August 2003David Malon, ANL U.S. ATLAS Computing MeetingDavid Malon, ANL U.S. ATLAS Computing Meeting

18

IOV Database

Athena Transient Conditions Store

Conditions data

1. Folder (data type), timestamp, tag, <version>

2. Ref to data (string)

3. Dereference via standard conversion services

4. Build transient conditions object

Conditions data client

Page 19: Data Management Overview David M. Malon malon@anl.gov U.S. ATLAS Computing Meeting Brookhaven, New York 28 August 2003.

28 August 200328 August 2003David Malon, ANL U.S. ATLAS Computing MeetingDavid Malon, ANL U.S. ATLAS Computing Meeting

19

Conditions and IOV people

LAL Orsay (Schaffer, Perus) is leading the IOV database integration LAL Orsay (Schaffer, Perus) is leading the IOV database integration

efforteffort

LBNL (Leggett) provides the transient infrastructure to handle time LBNL (Leggett) provides the transient infrastructure to handle time

validity of objects in the transient store (StoreGate) with respect to validity of objects in the transient store (StoreGate) with respect to

current event timestampcurrent event timestamp

Hong Ma (liquid argon) is an early adopter of the Athena-integrated IOV Hong Ma (liquid argon) is an early adopter of the Athena-integrated IOV

database servicedatabase service

Joe Rothberg (muons) is writing test beam conditions data to the Joe Rothberg (muons) is writing test beam conditions data to the

database database outsideoutside of Athena, for later Athena-based processing of Athena, for later Athena-based processing

Page 20: Data Management Overview David M. Malon malon@anl.gov U.S. ATLAS Computing Meeting Brookhaven, New York 28 August 2003.

28 August 200328 August 2003David Malon, ANL U.S. ATLAS Computing MeetingDavid Malon, ANL U.S. ATLAS Computing Meeting

20

Conditions Data Working Group

A Conditions Data Working Group has been newly A Conditions Data Working Group has been newly

commissioned, headed by Richard Hawkings (calibration commissioned, headed by Richard Hawkings (calibration

and alignment coordinator)and alignment coordinator)

Charged with articulating a model for conditions/calibration Charged with articulating a model for conditions/calibration

data flow between online/TDAQ and offline, including DCS, data flow between online/TDAQ and offline, including DCS,

for understanding and recording rates and requirements, for understanding and recording rates and requirements,

and more—not just conditions data persistenceand more—not just conditions data persistence

Contact Richard (or me, I guess) if you’d like to contributeContact Richard (or me, I guess) if you’d like to contribute

Page 21: Data Management Overview David M. Malon malon@anl.gov U.S. ATLAS Computing Meeting Brookhaven, New York 28 August 2003.

28 August 200328 August 2003David Malon, ANL U.S. ATLAS Computing MeetingDavid Malon, ANL U.S. ATLAS Computing Meeting

21

Production, bookkeeping, and metadata databases

Data Challenges have provided much of the impetus for development Data Challenges have provided much of the impetus for development

of production, bookkeeping, and metadata databasesof production, bookkeeping, and metadata databases

Strong leveraging of work done under external auspicesStrong leveraging of work done under external auspices

MAGDA (BNL) used for file/replica cataloging and transferMAGDA (BNL) used for file/replica cataloging and transfer Developed as an ATLAS activity funded by PPDG Magda/RLS integration/transition planned prior to DC2

AMI database (Grenoble) used for production metadataAMI database (Grenoble) used for production metadata Some grid integration of AMI (e.g., with EDG Spitfire)

Small focused production workshop held at CERN earlier this month to Small focused production workshop held at CERN earlier this month to

plan production infrastructure for Data Challenge 2plan production infrastructure for Data Challenge 2 Rich Baker, Kaushik De, Rob Gardner involved on the U.S. side Report due soon

Page 22: Data Management Overview David M. Malon malon@anl.gov U.S. ATLAS Computing Meeting Brookhaven, New York 28 August 2003.

28 August 200328 August 2003David Malon, ANL U.S. ATLAS Computing MeetingDavid Malon, ANL U.S. ATLAS Computing Meeting

22

Metadata handling

ATLAS Metadata workshop held 23-25 July in OxfordATLAS Metadata workshop held 23-25 July in Oxford

Issues:Issues: Metadata infrastructure to be deployed for Data Challenge 2

Integration of metadata at several levels from several sources Collection-level and event-level physics metadata Collection-level and event-level physical location metadata Provenance metadata …

Common recipe repository/transformation catalog to be shared

among components

Workshop report due soon (overdue, really)Workshop report due soon (overdue, really)

Page 23: Data Management Overview David M. Malon malon@anl.gov U.S. ATLAS Computing Meeting Brookhaven, New York 28 August 2003.

28 August 200328 August 2003David Malon, ANL U.S. ATLAS Computing MeetingDavid Malon, ANL U.S. ATLAS Computing Meeting

23

Virtual data catalogs

Data Challenges have been a testbed for virtual data catalog Data Challenges have been a testbed for virtual data catalog

prototyping, in AtCom, with VDC, and using the Chimera software from prototyping, in AtCom, with VDC, and using the Chimera software from

the GriPhyN (Grid Physics Networks) projectthe GriPhyN (Grid Physics Networks) project

Shared “recipe repository” (transformation catalog) discussions are Shared “recipe repository” (transformation catalog) discussions are

underwayunderway

Recent successes with Chimera-based ATLAS job execution on “CMS” Recent successes with Chimera-based ATLAS job execution on “CMS”

nodes on shared CMS/ATLAS grid testbedsnodes on shared CMS/ATLAS grid testbeds

More from the grid folks (Rob Gardner?) on thisMore from the grid folks (Rob Gardner?) on this

Work is needed on integration with ATLAS database infrastructure Work is needed on integration with ATLAS database infrastructure

Page 24: Data Management Overview David M. Malon malon@anl.gov U.S. ATLAS Computing Meeting Brookhaven, New York 28 August 2003.

28 August 200328 August 2003David Malon, ANL U.S. ATLAS Computing MeetingDavid Malon, ANL U.S. ATLAS Computing Meeting

24

Near-term plans

Focus in coming months: deploy and test a reasonable Focus in coming months: deploy and test a reasonable

prototype of ATLAS Computing Model in the time frame of prototype of ATLAS Computing Model in the time frame of

Data Challenge 2Data Challenge 2 Model is still being defined—Computing Model working group

preliminary report due in October(?) Note that DC2 is intended to provide an exercise of the Computing

Model sufficient to inform the writing of the Computing TDR

Ambitious development agenda requiredAmbitious development agenda required See database slides from July ATLAS Data Challenge workshop

Tier 0 reconstruction prototype is a principal focus, as is Tier 0 reconstruction prototype is a principal focus, as is

some infrastructure to support analysissome infrastructure to support analysis

Page 25: Data Management Overview David M. Malon malon@anl.gov U.S. ATLAS Computing Meeting Brookhaven, New York 28 August 2003.

28 August 200328 August 2003David Malon, ANL U.S. ATLAS Computing MeetingDavid Malon, ANL U.S. ATLAS Computing Meeting

25

Event persistence for Tier 0 reconstruction in DC2

Persistence for ESD, AOD, Tag data in POOL (8.0.0)Persistence for ESD, AOD, Tag data in POOL (8.0.0) ESD, AOD, Tag not yet defined by reconstruction group

Athena interfaces to POOL collection building and filtering infrastructure Athena interfaces to POOL collection building and filtering infrastructure

(7.5.0)(7.5.0)

Physical placement control, and placement metadata (to support selective Physical placement control, and placement metadata (to support selective

retrieval) (7.5.0)retrieval) (7.5.0)

Support for writing to multiple streams (e.g., by physics channel) (7.5.0)Support for writing to multiple streams (e.g., by physics channel) (7.5.0)

Support for concurrent processors contributing to common streams (8.0.0)Support for concurrent processors contributing to common streams (8.0.0)

Cataloging of database-resident event collections in son-of-AMI database, Cataloging of database-resident event collections in son-of-AMI database,

and other AMI integration with POOL (7.5.0)and other AMI integration with POOL (7.5.0)

MagdaMagdaRLSRLSPOOL integration (7.5.0++?)POOL integration (7.5.0++?)

Page 26: Data Management Overview David M. Malon malon@anl.gov U.S. ATLAS Computing Meeting Brookhaven, New York 28 August 2003.

28 August 200328 August 2003David Malon, ANL U.S. ATLAS Computing MeetingDavid Malon, ANL U.S. ATLAS Computing Meeting

26

Conditions database development

Extensions needed to Athena/StoreGate to support writing Extensions needed to Athena/StoreGate to support writing

calibrations/conditions from Athenacalibrations/conditions from Athena AthenaPOOL service should be capable of handling the persistence

aspects by Release 7.0.0/7.1.0

Work underway in the database group on organizing persistent Work underway in the database group on organizing persistent

calibration/conditions data, infrastructure for version tagging, for calibration/conditions data, infrastructure for version tagging, for

specifying in job options which data are neededspecifying in job options which data are needed Limited prototype capabilities in Release 6.3.0

Responsibility for model moving to new calibration/alignment coordinator

Some exercise of access to conditions data varying at the sub-run level Some exercise of access to conditions data varying at the sub-run level

by Tier 0 reconstruction is planned for inclusion in DC2by Tier 0 reconstruction is planned for inclusion in DC2

Page 27: Data Management Overview David M. Malon malon@anl.gov U.S. ATLAS Computing Meeting Brookhaven, New York 28 August 2003.

28 August 200328 August 2003David Malon, ANL U.S. ATLAS Computing MeetingDavid Malon, ANL U.S. ATLAS Computing Meeting

27

Conditions database futures

LCG conditions database common project will start soon LCG conditions database common project will start soon

ATLAS development agenda will be tied to thisATLAS development agenda will be tied to this Expect to contribute strongly to common project requirements and

development

Some DCS and muon test beam data are already going into Some DCS and muon test beam data are already going into

the ATLAS/Lisbon implementation of the common project the ATLAS/Lisbon implementation of the common project

interface that will be the LCG conditions project’s starting interface that will be the LCG conditions project’s starting

point; liquid argon testing of this infrastructure also point; liquid argon testing of this infrastructure also

underwayunderway

Page 28: Data Management Overview David M. Malon malon@anl.gov U.S. ATLAS Computing Meeting Brookhaven, New York 28 August 2003.

28 August 200328 August 2003David Malon, ANL U.S. ATLAS Computing MeetingDavid Malon, ANL U.S. ATLAS Computing Meeting

28

Coordination with online and TC

Historically, demonstrably good at the component level (cf. the joint Historically, demonstrably good at the component level (cf. the joint

conditions database work with Lisbon), but largely conditions database work with Lisbon), but largely ad hocad hoc

Now formalized with new ATLAS Database Coordination Group Now formalized with new ATLAS Database Coordination Group

commissioned by Dario Barberis, with representation from commissioned by Dario Barberis, with representation from

online/TDAQ, offline, and Technical Coordinationonline/TDAQ, offline, and Technical Coordination

Conditions Data Working Group also launched, with substantial Conditions Data Working Group also launched, with substantial

involvement from both offline and onlineinvolvement from both offline and online

Successful joint conditions data workshop organized in February in Successful joint conditions data workshop organized in February in

advance of thisadvance of this

Page 29: Data Management Overview David M. Malon malon@anl.gov U.S. ATLAS Computing Meeting Brookhaven, New York 28 August 2003.

28 August 200328 August 2003David Malon, ANL U.S. ATLAS Computing MeetingDavid Malon, ANL U.S. ATLAS Computing Meeting

29

Staffing

Current census: small groups (2-3 FTEs) at Argonne, Brookhaven, LAL Orsay; Current census: small groups (2-3 FTEs) at Argonne, Brookhaven, LAL Orsay;

about 1 FTE at Grenoble working on tag collector and AMI databases for about 1 FTE at Grenoble working on tag collector and AMI databases for

productionproduction Enlisting involvement from British GANGA team; cf. metadata workshop at

Oxford last month U.S. ATLAS computing management has worked hard to try to increase

support for database development, but we all know how tough the funding situation is

Trying to leverage grid projects wherever possible

Lack of database effort at CERN is conspicuous, and hurts us in several waysLack of database effort at CERN is conspicuous, and hurts us in several ways

Data Challenge production and support is a valuable source of requirements Data Challenge production and support is a valuable source of requirements

and experience, but it reduces development effortand experience, but it reduces development effort

Not clear that expected staffing levels will allow us to meet DC2 milestonesNot clear that expected staffing levels will allow us to meet DC2 milestones

More on this at the LHC manpower [sic] review next weekMore on this at the LHC manpower [sic] review next week

Page 30: Data Management Overview David M. Malon malon@anl.gov U.S. ATLAS Computing Meeting Brookhaven, New York 28 August 2003.

28 August 200328 August 2003David Malon, ANL U.S. ATLAS Computing MeetingDavid Malon, ANL U.S. ATLAS Computing Meeting

30

Conclusions

Work is underway on many fronts: common infrastructure, event store, Work is underway on many fronts: common infrastructure, event store,

conditions and IOV databases, primary numbers and geometry, conditions and IOV databases, primary numbers and geometry,

metadata, production databases, …metadata, production databases, …

Many things will be ready for early adopters soon (7.0.0/7.1.0)Many things will be ready for early adopters soon (7.0.0/7.1.0)

Development agenda for Data Challenge 2 is dauntingDevelopment agenda for Data Challenge 2 is daunting

We need help, but we cannot pay youWe need help, but we cannot pay you

If we survive DC2, look for persistence tutorials at the next U.S. ATLAS If we survive DC2, look for persistence tutorials at the next U.S. ATLAS

computing workshopcomputing workshop