* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project
Download slides - Indico
Survey
Document related concepts
Microsoft Access wikipedia , lookup
Serializability wikipedia , lookup
Entity–attribute–value model wikipedia , lookup
Microsoft SQL Server wikipedia , lookup
Open Database Connectivity wikipedia , lookup
Oracle Database wikipedia , lookup
Extensible Storage Engine wikipedia , lookup
Microsoft Jet Database Engine wikipedia , lookup
Relational model wikipedia , lookup
Functional Database Model wikipedia , lookup
Concurrency control wikipedia , lookup
ContactPoint wikipedia , lookup
Transcript
Database Access Patterns in ATLAS Computing Model G. Gieraltowski, J. Cranshaw, K. Karr, D. Malon, A. Vaniachine ANL P, Nevski, Yu. Smirnov, T. Wenaus BNL N. Barros, L. Goossens, R. Hawkings, A. Nairz, G. Poulard, Yu. Shapiro, F. Zema CERN XV International Conference on Computing in High Energy and Nuclear Physics T.I.F.R., Mumbai, India February 13-17, 2006 CHEP06, Mumbai, India February 13-17, 2006 Outline 1) Emphasis on the early days of LHC running: Calibration/Alignment is a priority Must be done before the reconstruction start ATLAS 2006 Computing System Commissioning: Calibration/Alignment procedures are included in acceptance tests 2) Real experience in prototypes and production systems General issues encountered: Increased fluctuations in database server load Connections count limitations 3) Development of the ATLAS distributed computing model: Server-side developments: Deployment: LCG3D Project and OSG Edge Services Framework Activity Technology: Grid-enabled server technology - Project DASH Application-side technology developments: Deployment: Integration with Production System database (Conditions data slices) Technology: ATLAS Database Client Library (now adopted by COOL/POOL/CORAL) Alexandre Vaniachine (ANL) 2 CHEP06, Mumbai, India February 13-17, 2006 ATLAS Computing Model In the ATLAS Computing Model widely distributed applications require access to terabytes of data stored in relational databases Realistic database services data flow – including Calibration & Alignment – is presented in the Computing Technical Design Report Preparations are on track towards Computing System Commissioning to exercise realistic database data flow Alexandre Vaniachine (ANL) 3 CHEP06, Mumbai, India February 13-17, 2006 ATLAS CSC Goals 2006 is the year of ATLAS CSC The first goal of the CSC is calibration and alignment procedures ConditionsDB is included in CSC acceptance tests WLCG SC4 Workshop - 12 February 2006 Computing System Commissioning Goals We have defined the high-level goals of the Computing System Commissioning operation during 2006 Formerly called “DC3” More a running-in of continuous operation than a stand-alone challenge Main aim of Computing System Commissioning will be to test the software and computing infrastructure that we will need at the beginning of 2007: Calibration and alignment procedures and conditions DB Full trigger chain Event reconstruction and data distribution Distributed access to the data for analysis At the end (autumn-winter 2006) we will have a working and operational system, ready to take data with cosmic rays at increasing rates Alexandre Vaniachine (ANL) 4 Dario Barberis: ATLAS SC4 Plans 4 CHEP06, Mumbai, India February 13-17, 2006 Towards the Early Days of LHC Running Calibration/Alignment is a priority Must be done before the reconstruction start Calibration/Alignment is a part of the overall Computing System Commissioning activity to: Demonstrate the calibration ‘closed loop’: Iterate and improve reconstruction Exercise the conditions DB access and distribution infrastructure Encourage development of subdetector calibration algorithms Initially focussed on ‘steady-state’ calibration Assuming required samples are available and can be selected Also want to look at initial 2007/2008 running at low luminosity Alexandre Vaniachine (ANL) 5 CHEP06, Mumbai, India February 13-17, 2006 Calibration Data Flow Alexandre Vaniachine (ANL) 6 CHEP06, Mumbai, India February 13-17, 2006 Prerequisites for Success Simulation Ability to simulate a realistic, misaligned, miscalibrated detector Reconstruction Use of calibration data in reconstruction; ability to handle timevarying calibration Calibration Algorithms Algorithms in Athena, running from standard ATLAS data Data Preparation Organisation and bookkeeping run number ranges, production system,… Alexandre Vaniachine (ANL) 7 CHEP06, Mumbai, India February 13-17, 2006 Production System Enhancements To prepare for new challenges first ATLAS Database Services Workshop was organized in December: http://agenda.cern.ch/fullAgenda.php?ida=a057425 Among the Workshop recommendations was: A tighter integration of the production system database, task definition, Distributed Data Management and conditions data tags Implementation opportunities: Distribute (push) snapshots via pacman Use of DDM for large payload files Try Oracle 10g file management for external files Expand existing ServersCatalog with top tags Alexandre Vaniachine (ANL) 8 CHEP06, Mumbai, India February 13-17, 2006 ATLAS DB Applications In preparation for data taking, the ATLAS experiment has run a series of large-scale computational exercises to test and validate multi-tier distributed data grid solutions under development Real experience in prototypes and production systems was collected with three ATLAS major database applications: Geometry DB Conditions DB TAG databases ATLAS computational exercises run on a world-wide federation of computational grids Alexandre Vaniachine (ANL) 9 CHEP06, Mumbai, India February 13-17, 2006 Data Mining of Operations The data-mining of the collected operations data reveals a striking feature – a very high degree of correlations between the failures: if the job submitted to some cluster failed, there is a high probability that a next job submitted to the cluster would fail too if the submit host failed, all the jobs scattered over different clusters will fail too Taking these correlations into account is not yet automated by the grid middleware That is why production databases and grid monitoring data that are providing immediate feedback on the Data Challenge operations to the production operators is very important for efficient utilization of the Grid capacities Alexandre Vaniachine (ANL) 10 CHEP06, Mumbai, India February 13-17, 2006 Production Rate Growth and Daily Fluctuations 14000 12000 10000 Jobs/day Rome Production (mix of jobs) LCG/CondorG LCG/Original NorduGrid Grid3 8000 2005 Database Capacities Bottleneck Data Challenge 2 (short jobs period) 6000 Data Challenge 2 (long jobs period) 4000 2000 0 Alexandre Vaniachine (ANL) Jul Aug Sep Oct Nov Dec 11 Jan Feb Mar Apr May CHEP06, Mumbai, India February 13-17, 2006 Lessons Learned Among the lessons learned is the increase in fluctuations in database server workloads due to the chaotic nature of grid computations The observed fluctuations in database access patterns are of a general nature and must be addressed through services enabling dynamic and flexibly managed provisioning of database resources In many cases the connections count happens to be the limiting resource Alexandre Vaniachine (ANL) 12 CHEP06, Mumbai, India February 13-17, 2006 Opportunistic Grids Campus computing grids like the GLOW http://osg-docdb.opensciencegrid.org/cgi-bin/ShowDocument?docid=361 utilize spare cycles to run jobs The priority has the owner of resource ATLAS jobs are often put to hibernate Thus optimal jobs are shorter, i.e. only few events Resulting in order of magnitude more frequent database access Jobs put to hibernation during the initialization phase overload CERN database resources by keeping database connections open for days This problem was resolved by deploying dedicated replica servers in US and CERN to support the GLOW grid In comparison to production grids opportunistic grids require extra development and support efforts not sustainable in the long run Alexandre Vaniachine (ANL) 13 CHEP06, Mumbai, India February 13-17, 2006 Client Library To improve robustness of database access in a data grid environment we developed the applicationside solution – a software component abstracting the database and/or middleware connectivity concerns in a generalized Database Client Library http://indico.cern.ch/contributionDisplay.py?contribId=32&sessionId=4&confId=048 Alexandre Vaniachine (ANL) 14 CHEP06, Mumbai, India February 13-17, 2006 Server Indirection One of lessons learnt in ATLAS Data Challenges is that the database server address should NOT be hardwired in data processing transformations The logical-physical indirection for database servers is now introduced in ATLAS Similar to the logical-physical file Replica Location Service indirection of the Grid file catalogs Supported by ATLAS Client Library Now adopted by LHC POOL project: http://indico.cern.ch/contributionDisplay.py?contribId=329&sessionId=4&confId=048 Alexandre Vaniachine (ANL) 15 CHEP06, Mumbai, India February 13-17, 2006 Tier-0 Operations CondDBB CTB DBs Online server (atlobk01) DB replication Data acquisition programs NOVA DBs POOL cat CondDB CTB DBs Offline server (atlobk02) Test DBs OBK DBs POOLcat NOVA DBs Browsing applications, Athena programs (Other Browsing applications) OBK DBs Alexandre Vaniachine (ANL) In addition to distributed operations, ATLAS database services are relevant to local CERN data taking operations including the conditions data flow of ATLAS Combined Test Beam operations, prototype Tier-0 scalability tests and event tag database operations 16 CHEP06, Mumbai, India February 13-17, 2006 TAG Database Access TAG Replication is a part of SC4 Tier-0 test Loading TAGs into the relational database at CERN Replicating it using Oracle streams from Tier-0 to Tier-1s and to Tier-2s Also as an independent test, using TAG files that are already available generated TAG Access • TAG is a keyed list of variables/event • Two roles • Direct access to event in file via pointer • Data collection definition function • Two formats, file and database • Now believe large queries require full database • Restricts it to Tier1s and large Tier2s/CAF • Ordinary Tier2s hold file-based TAG corresponding to locally-held datasets Alexandre Vaniachine (ANL) 17 CHEP06, Mumbai, India February 13-17, 2006 Participation in LCG 3D LCG 3D Service Architecture ATLAS is fully committed to use Distributed Database Deployment infrastructure developed in collaboration with the LCG 3D Project T0 M O M - autonomous reliable service T1- db back bone - all data replicated - reliable service O T2 - local db cache T3/4 -subset data -only local service M O M Oracle Streams Cross vendor copy MySQL/SQLight Files Proxy Cache Alexandre Vaniachine (ANL) 18 CHEP06, Mumbai, India February 13-17, 2006 Participation in OSG ESF US ATLAS is participating in OSG Edge Services Framework Activity to enhance traditional database services infrastructure deployed in 3D with dynamic database services deployment capabilities Open Science Grid Edge Services • Services executing on the edge of the public and private network CE CMS ATLAS CDF Guest VO SE Site Compute nodes and Storage nodes SC05 booth presentation OSG Edge Services Framework 1 http://indico.cern.ch/contributionDisplay.py?contribId=214&sessionId=7&confId=048 Alexandre Vaniachine (ANL) 19 CHEP06, Mumbai, India February 13-17, 2006 Project DASH To grid-enable MySQL database server ATLAS is participating in the project DASH: http://indico.cern.ch/contributionDisplay.py?contribId=36&sessionId=7&confId=048 A new collaborative project has just started at Argonne to grid-enable PostgreSQL database Both projects target integration with OSGA-DAI Please contact us if you are interested to contribute to these projects Alexandre Vaniachine (ANL) 20 CHEP06, Mumbai, India February 13-17, 2006 Conclusions As grid computing technologies mature, development must focus on database and grid integration New technologies are required to bridge the gap between data accessibility and the increasing power of grid computing used for distributed event production and processing Changes must happen both on the server side and on the client side Server technology Must support dynamic deployment of capacities Must support replication on a lower granularity level: Conditions DB slices Must be coordinated with production system Must support grid authorization (Project DASH) Client technology Must support database server indirection Must support coordinated client-side solution: ATLAS Database Client Library (now a part of COOL/POOL/CORAL) Alexandre Vaniachine (ANL) 21