c7cd8a838b6a81349d8cd134b803797e.ppt
- Количество слайдов: 26
Overview of the GLUE Project (Grid Laboratory Unified Environment) Author: Piotr Nowakowski, M. Sc. Cyfronet, Kraków
Presentation Summary • Goals of GLUE • Key GLUE contributors • GLUE schema • GLUE activities • Unresolved issues AT, Cyfronet, June 7, 2002
Goals of GLUE Ø Promote coordination between European and US Grid projects Ø Define, construct, test and deliver interoperable middleware to all Grid Projects Ø Experiment with intercontinental Grid deployment and operational issues Ø Establish procedures and policies regarding interoperability Once the GLUE collaboration establishes the necessary, minimum requirements for interoperability of middleware, any future software designed by the projects covered by the umbrella of the HICB and JTB must maintain the achieved interoperability. AT, Cyfronet, June 7, 2002
GLUE Organizationally Ø Management by i. VDGL and Data. TAG Ø Guidance and oversight by the High Energy Physics Intergrid Coordination Board (HICB) and Joint Technical Board (JTB) Ø Participating organizations (19 entities in all): • Grid Projects (EDG, Gri. Phy. N, Cross. Grid etc. ) • LHC experiments (Atlas, CMS etc. ) AT, Cyfronet, June 7, 2002
HENP Collaboration The HENP (High-Energy Nuclear Physics) Grid R&D projects (initially Data. Grid, Gri. Phy. N, and PPDG, as well as the national European Grid projects in UK, Italy, Netherlands and France) have agreed to coordinate their efforts to design, develop and deploy a consistent open source standards-based global Grid infrastructure. To that effect, their common efforts are organized in three major areas: • A HENP Inter. Grid Coordination Board (HICB) for high-level coordination • A Joint Technical Board (JTB) • Common Projects, and Task Forces to address needs in specific technical areas AT, Cyfronet, June 7, 2002
The Data. TAG Project Aim: Creation of an intercontinental Grid testbed using Data. Grid (EDG) and Gri. Phy. N components. Work packages: WP 1: WP 2: WP 3: WP 4: WP 5: WP 6: Establishment of an intercontinental testbed infrastructure High performance networking Bulk data transfer validations and performance monitoring Interoperability between Grid domains Information dissemination/exploitation Project management AT, Cyfronet, June 7, 2002
Data. TAG WP 4 Aims: • To produce an assessment of interoperability solutions, • To provide a test environment for LHC applications to extend existing use cases to test interoperability of Grid components, • To provide input to a common Grid LHC architecture, • To plan EU-US integrated Grid deployment. WP 4 Tasks: T 4. 1: Develop an intergrid resource discovery schema, T 4. 2: Develop intergrid Authentication, Authorization and Accounting (AAA) mechanisms, T 4. 3: Plan and deploy an „intergrid VO” in collaboration with i. VDGL. AT, Cyfronet, June 7, 2002
Data. TAG WP 4 Framework and Relationships AT, Cyfronet, June 7, 2002
The i. VDGL Project (International Virtual Data Grid Laboratory) Aim: To provide high-performance global computing infrastructure for keynote experiments in physics and astronomy (ATLAS, LIGO, SDSS etc. ) i. VDGL activities: • • • Establishing supercomputing sites throughout the U. S. and Europe; linking them with a multi-gigabit transatlantic link Establishing a Grid Operations Center (GOC) in Indiana Maintaining close cooperation with partnership projects in the EU and the Gri. Phy. N project. AT, Cyfronet, June 7, 2002
U. S. i. VDGL Network Selected participants: • Fermilab • Brookhaven National Laboratory • Argonne National Laboratory • Stanford LINAC Laboratory • University of Florida • University of Chicago • California Institute of Technology • Boston University • University of Wisconsin • Indiana University • Johns Hopkins University • Northwestern University • University of Texas • Pennsylvania State University • Hampton University • Salish Kootenai College AT, Cyfronet, June 7, 2002
i. VDGL Organization Plan • Project Steering Group – advises i. DVGL directors on important project decisions and issues. • Project Coordination Group – provides a forum for short-term planning and tracking of the project activities and schedules. The PCG includes representatives of related Grid projects, particularly EDT/EDG. • Facilities Team – identification of testbed sites, hardware procurement • Core Software Team – definitions of software suites and toolkits (Globus, VDT, operating systems etc. ) • Operations Team – performance monitoring, networking, coordination, security etc. • Applications Team – planning the deployment of applications and the related requirements • Outreach Team – Website maintenance, planning conferences, publishing research materials etc. Note: The GLUE effort is coordinated by the Interoperability Team (aka GLUE Team) AT, Cyfronet, June 7, 2002
The Gri. Phy. N Project Aims: • To provide the necessary IT solutions for petabyte-scale dataintensive science by advancing the Virtual Data concept, • To create Petascale Virtual Data Grids (PVDG) to meet the computational needs of thousands of scientists spread across the globe. Timescale: 5 years (2000 -2005) Gri. Phy. N applications: • • • The CMS and ATLAS LHC experiments at CERN LIGO (Laser Interferometer Gravitational Wave Observatory) SDSS (Sloan Digital Sky Survey) AT, Cyfronet, June 7, 2002
The Virtual Data Concept Virtual data: the definition and delivery to a large community of a (potentially unlimited) virtual space of data products derived from experimental data. In virtual data space, requests can be satisfied via direct access and/or computation, with local and global resource management, policy, and security constraints determining the strategy used. Gri. Phy. N IT targets: • • • Virtual Data technologies: new methods of cataloging, characterizing, validating, and archiving software components to implement virtual data manipulations Policy-driven request planning and scheduling of networked data and computational resources: mechanisms for representing and enforcing both local and global policy constraints and new policy-aware resource discovery techniques. Management of transactions and task execution across national-scale and worldwide virtual organizations: new mechanisms to meet user requirements for performance, reliability, and cost. AT, Cyfronet, June 7, 2002
Sample VDG Architecture AT, Cyfronet, June 7, 2002
Petascale Virtual Data Grids Petascale – both computationally intensive (Petaflops) and data intensive (Petabytes). Virtual – containing little ready-to-use information, instead focusing on methods of deiving this information from other data. The Tier Concept Developed for use by the most ambitious LHC experiments: ATLAS and CMS. • Tier 0: CERN HQ • Tier 1: National center • Tier 2: Regional center • Tier 3: HPC center • Tier 4: Desktop PC cluster AT, Cyfronet, June 7, 2002
The Data. Grid (EDG) Project Aim: To enable next-generation scientific exploration which requires sharing intensive computation and analysis of shared large-scale databases, from hundreds of terabytes to petabytes, across widely distributed scientific communities. Data. Grid Work Packages: WP 1: Workload Management WP 2: Data Management WP 3: Monitoring Services WP 4: Fabric Management WP 5: Storage Management WP 6: Integration (testbeds) WP 7: Network WP 8: Application – Particle Physics WP 9: Application – Biomedical Imaging WP 10: Application – Satellite surveys WP 11: Dissemination WP 12: Project Management AT, Cyfronet, June 7, 2002
GLUE Working Model The following actions take place once an interoperability issue is encountered: • The Data. TAG/i. VDGL managers define a plan and sub-tasks to address the relevant issue. This plan includes integrated tests and demonstrations which define overall success. • The Data. TAG/i. VDGL sub-task managers assemble all the input required to address the issue on hand. The HIJTB and other relevant experts would be strongly involved. • The Data. TAG/i. VFGL sub-task managers organize getting the work done using the identified solutions. • At appropriate points the work need is presented to the HICB, which discusses it on a technical level. Iterations take place. • At appropriate points the evolving solutions are presented to the HICB. • At an appropriate point the final solution is presented to the HICB with a recommendation that it be accepted by Grid projects. AT, Cyfronet, June 7, 2002
GLUE Working Model - example Issue: Data. GRID and i. VDGL use different data models for publishing resource information. Therefore RBs cannot work across domains. • The HIJTB recognizes this and proposes it as an early topic to address. The Data. TAG/i. VDGL management is advised to discuss this early on. • Data. TAG management has already identified this as a sub-task. • Data. TAG/i. VDGL employees are assigned to the problem. • Many possible solutions exist, from consolidation to translation on various levels (the information services level or even the RB level). The managers discuss the problem with clients in order to ascertain the optimal solution. • The group involved organizes its own meetings (regardless of the monthly HIJTB meetings). [this is taking place now] • A common resource model is proposed. Once it has been demonstrated to work within a limited test environment, the HIJTB/HICB will discuss if and when to deploy this generally, taking into account the ensuing modifications which will be needed to other components such as the resource broker. AT, Cyfronet, June 7, 2002
GLUE Schemas GLUE schemas: descriptions of objects and attributes needed to describe Grid resources and their mutual relations. GLUE schemas include: • Computing Element (CE) schema – in development • Storage Element (SE) schema – TBD • Network Element (NE) Schema – TBD The development of schemas is coordinated by JTB with collaboration from Globus, PPDG and EDG WP managers. AT, Cyfronet, June 7, 2002
CE Schema version 4 – 24/05/2002 • Computing Element: an entry point into a queuing system. Each queue points to one or more clusters. • Cluster: a group of subclusters or individual nodes. A cluster may be referenced by more than one computing element. • Subcluster: a homogenous group of individual computing nodes (all nodes must be represented by a predefined set of attributes). • Host: a physical computing element. No host may be part of more than one subcluster.
GLUE Schema Representation In existing MDS models, GLUE Schemas and their hierarchies can be represented through DITs (Directory Information Tree). Globus MDS v 2. 2 will be updated to handle the new schema. In future OGSA-based implementations (Globus v 3. 0) the structure can be converted to an XML document. AT, Cyfronet, June 7, 2002
GLUE Stage I Aims: Integration of US (i. VDGL) and European (EDG) testbeds; developing a permanent set of reference tests for new releases and services. Phase I In progress • Cross-organizational authentication • Unified service discovery and information infrastructure • Test of Phase I infrastructure Phase II • Data movement infrastructure • Test of Phase II infrastructure Phase III • Community authorization services • Test of the complete service AT, Cyfronet, June 7, 2002
Grid Middleware and Testbed The following middleware will be tested in Stage I of GLUE: • EDG Work Packages WP 1 (Workload management), WP 2 (Data management), WP 3 (Information and monitoring services), WP 5 (Storage management) • Gri. Phy. N middleware – Globus 2. 0, Condor v 6. 3. 1, VDT 1. 0, The GLUE testbed will consist of: • Computational resources: several CEs from Data. TAG and i. VDGL respectively. • Storage: access to mass storage systems at CERN and US Tier 1 sites. • Network: standard production networks should be sufficient. AT, Cyfronet, June 7, 2002
GLUE Stage I Schedule Feb 2002: May 2002: Jun 2002: Jul 2002: Sep 2002: Nov 2002: Dec 2002: Test interoperating certificates between US and EU – done Review of common resource discovery schema – in progress Full testbed proposal available for review. Review of common storage schema First version of common use cases (EDG WP 8) Refinement of testbed proposals through HICB feedback Intercontinental resource discovery infrastructure in test mode for production deployment in September Interoperating Community and VO authorization available Implementation of common use cases by the experiments Demonstrations planned Sites integrated into Grid executing all goals of Stage I AT, Cyfronet, June 7, 2002
Unresolved Issues • Ownership of GLUE schemas • Maintenance of GLUE schemas • Ownership (and maintenance) of MDS information providers AT, Cyfronet, June 7, 2002
Web Addresses • GLUE Homepage at HICB: http: //www. hicb. org/glue. html • GLUE-Schema site: http: //www. hicb. org/glue-schema/schema. htm • HENP Collaboration page: http: //www. hicb. org • The Data. TAG Project: http: //www. datatag. org • The i. VDGL Project: http: //www. ivdgl. org • The Gri. Phy. N Project: http: //www. griphyn. org • European Data. Grid: http: //www. eu-datagrid. org AT, Cyfronet, June 7, 2002
c7cd8a838b6a81349d8cd134b803797e.ppt