Survey
* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project
* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project
Future Scientific Infrastructure Ian Foster Mathematics and Computer Science Division Argonne National Laboratory and Department of Computer Science The University of Chicago http://www.mcs.anl.gov/~foster Keynote Talk, QUESTnet 2002 Conference, Gold Coast, July 4, 2002 2 Evolution of Infrastructure 1890: Local power generation – AC transmission => power Grid => economies of scale & revolutionary new devices 2002: Primarily local computing & storage – Internet & optical technologies => ??? [email protected] ARGONNE CHICAGO 3 A Computing Grid On-demand, ubiquitous access to computing, data, and services “We will perhaps see the spread of ‘computer utilities’, which, like present electric and telephone utilities, will service individual homes and offices across the country” (Len Kleinrock, 1969) New capabilities constructed dynamically and transparently from distributed services “When the network is as fast as the computer's internal links, the machine disintegrates across the net into a set of special purpose appliances” (George Gilder, 2001) [email protected] ARGONNE CHICAGO Distributed Computing+Visualization Remote Center Generates Tb+ datasets from simulation code WAN Transfer FLASH data transferred to ANL for visualization GridFTP parallelism utilizes high bandwidth (Capable of utilizing >Gb/s WAN links) Chiba City 4 Visualization code constructs and stores high-resolution visualization frames for display on many devices Job Submission Simulation code submitted to remote center for execution on 1000s of nodes ActiveMural Display Displays very high resolution large-screen dataset animations LAN/WAN Transfer User-friendly striped GridFTP application tiles the frames and stages tiles onto display nodes FUTURE (1-5 yrs) • 10s Gb/s LANs, WANs • End-to-end QoS • Automated replica management • Server-side data reduction & analysis • Interactive portals [email protected] ARGONNE CHICAGO eScience Application: Sloan Digital Sky Survey Analysis [email protected] 5 ARGONNE CHICAGO Cluster-finding Data Pipeline 6 catalog 5 cluster 4 core core 3 3 brg brg brg brg 2 2 2 2 field field field 1 1 1 field 1 tsObj [email protected] tsObj tsObj tsObj ARGONNE CHICAGO Chimera Application: Sloan Digital Sky Survey Analysis 7 Size distribution of galaxy clusters? Galaxy cluster size distribution 100000 Chimera Virtual Data System + iVDGL Data Grid (many CPUs) 10000 1000 100 10 [email protected] 1 1 10 Number of Galaxies 100 ARGONNE CHICAGO Grids at NASA: Aviation Safety 8 Wing Models •Lift Capabilities •Drag Capabilities •Responsiveness Airframe Models Stabilizer Models •Deflection capabilities •Responsiveness Crew Capabilities - accuracy - perception - stamina - re-action times - SOPs Engine Models Human Models •Braking performance •Steering capabilities •Traction •Dampening capabilities Landing Gear Models [email protected] •Thrust performance •Reverse Thrust performance •Responsiveness •Fuel Consumption ARGONNE CHICAGO 9 Life Sciences: Telemicroscopy DATA ACQUISITION PROCESSING, ANALYSIS ADVANCED VISUALIZATION NETWORK IMAGING INSTRUMENTS COMPUTATIONAL RESOURCES LARGE DATABASES [email protected] ARGONNE CHICAGO 10 Business Opportunities On-demand computing, storage, services – Significant savings due to reduced build-out, economies of scale, reduced admin costs – Greater flexibility => greater productivity Entirely new applications and services – Based on high-speed resource integration Solution to enterprise computing crisis – Render distributed infrastructures manageable [email protected] ARGONNE CHICAGO 11 Grid Evolution [email protected] ARGONNE CHICAGO 12 Grids and Industry: Early Examples Entropia: Distributed computing (BMS, Novartis, …) Butterfly.net: Grid for multi-player games [email protected] ARGONNE CHICAGO 13 Grid Infrastructure Resources A – Computing, storage, data Services – Authentication, discovery, … Connectivity – Reduce tyranny of distance A A Technologies – Build applications, services Communities – Operational procedures, … [email protected] ARGONNE CHICAGO 14 Example Grid Infrastructure Projects I-WAY (1995): 17 U.S. sites for one week GUSTO (1998): 80 sites worldwide, experim. NASA Information Power Grid (since 1999) – Production Grid linking NASA laboratories INFN Grid, EU DataGrid, iVDGL, … (2001+) – Grids for data-intensive science TeraGrid, DOE Science Grid (2002+) – Production Grids link supercomputer centers U.S. GRIDS Center – Software packaging, deployment, support [email protected] ARGONNE CHICAGO 15 Topics in Grid Infrastructure Regional, national, intl optical infrastructure – I-WIRE, StarLight, APAN TeraGrid: Deep infrastructure – High-end support for U.S. community iVDGL: Wide infrastructure – Building a (international) community Open Grid Services Architecture – Future service & technology infrastructure [email protected] ARGONNE CHICAGO 16 Topics in Grid Infrastructure Regional, national, intl optical infrastructure – I-WIRE, StarLight TeraGrid: Deep infrastructure – High-end support for U.S. community iVDGL: Wide infrastructure – Building a (international) community Open Grid Services Architecture – Future service & technology infrastructure [email protected] ARGONNE CHICAGO Targeted StarLight Optical Network Connections AsiaPacific CERN SURFnet CA*net4 Vancouver 17 Seattle NTON Portland U Wisconsin San Francisco Chicago NTON PSC IU NYC NCSA AsiaPacific DTF 40Gb Los Angeles San Diego (SDSC) NW Univ (Chicago) StarLight Hub I-WIRE ANL St Louis GigaPoP UICAtlanta Atlanta Chicago Cross connect Ill Inst of Tech Univ of Chicago Indianapolis AMPATH AMPATH (Abilene NOC) NCSA/UIUC [email protected] www.startap.net ARGONNE CHICAGO 19 I-Wire Fiber Topology Starlight (NU-Chicago) Argonne 18 4 Qwest 455 N. Cityfront UC Gleacher Ctr 450 N. Cityfront 4 UIC 10 4 12 12 • Fiber Providers: Qwest, Level(3), McLeodUSA, 360Networks • 10 segments • 190 route miles; 816 fiber miles •Longest segment: 140 miles • 4 strands minimum to each site [email protected] 4 McLeodUSA UIUC/NCSA 151/155 N. Michigan Doral Plaza Level(3) 111 N. Canal State/City Complex 2 2 James R. Thompson Ctr City Hall State of IL Bldg FNAL (est 4q2002) 2 IIT UChicago Numbers indicate fiber count (strands) ARGONNE CHICAGO 20 I-Wire Transport TeraGrid Linear 3x OC192 1x OC48 First light: 6/02 Starlight (NU-Chicago) Argonne Starlight Linear 4x OC192 4x OC48 (8x GbE) Operational UC Gleacher Ctr 450 N. Cityfront Qwest Metro Ring 1x OC48 per site First light: 8/02 455 N. Cityfront UIC McLeodUSA UIUC/NCSA 151/155 N. Michigan Doral Plaza State/City Complex James R. Thompson Ctr City Hall State of IL Bldg IIT UChicago • Each of these three ONI DWDM systems have capacity of up to 66 channels, up to 10 Gb/s per channel • Protection available in Metro Ring on a per-site basis [email protected] ARGONNE CHICAGO Illinois Distributed Optical Testbed 21 Northwestern Univ-Chicago “Starlight” I-290 DAS-2 UI-Chicago I-294 I-55 Illinois Inst. Tech Argonne Nat’l Lab Dan Ryan Expwy (I-90/94) (approx 25 miles SW) [email protected] U of Chicago UIUC/NCSA Urbana (approx 140 miles South) ARGONNE CHICAGO 22 Topics in Grid Infrastructure Regional, national, intl optical infrastructure – I-WIRE, StarLight TeraGrid: Deep infrastructure – High-end support for U.S. community iVDGL: Wide infrastructure – Building a (international) community Open Grid Services Architecture – Future service & technology infrastructure [email protected] ARGONNE CHICAGO TeraGrid: Deep Infrastructure www.teragrid.org [email protected] 23 ARGONNE CHICAGO 24 TeraGrid Objectives Create unprecedented capability – Integrated with extant PACI capabilities – Supporting a new class of scientific research Deploy a balanced, distributed system – Not a “distributed computer” but rather … – a distributed “system” using Grid technologies > Computing and data management > Visualization and scientific application analysis Define an open and extensible infrastructure – Enabling infrastructure for scientific research – Extensible beyond the original four sites > NCSA, SDSC, ANL, and Caltech [email protected] ARGONNE CHICAGO 26 TeraGrid Timelines Proposal Submitted To NSF TeraGrid Operational Jan ‘01 McKinley systems Jan ‘02 Initial apps On McKinley TeraGrid Prototypes Early access To McKinley At Intel TeraGrid prototype At SC2001, 60 Itanium Nodes, 10Gbs network Grid Services on Current Systems Networking Basic Grid svcs Linux clusters SDSC SP NCSA O2K Jan ‘03 Early McKinleys TeraGrid clusters at TG sites for Testing/benchmarking “TeraGrid Lite” Systems and Grids testbed Core Grid services deployment 10Gigabit Enet testing Operations Advanced Grid services testing TeraGrid Networking Deployment TeraGrid Operations Center Prototype Day Ops Production Applications [email protected] ARGONNE CHICAGO 27 Terascale Cluster Architecture (a) Terascale Architecture Overview (b) Example 320-node Clos Network Myrinet System Interconnect Spine Switches Clos mesh Interconnect Each line = 8 x 2Gb/s links 128-port Clos Switches 64 hosts 64 hosts 64 hosts 64 hosts 64 hosts Add’l Clusters, External Networks 64 inter-switch links 64 inter-switch links 64 inter-switch links = 4 links 64 TB RAID Local Display Networks for Remote Display 100Mb/s Switched Ethernet Management Network Rendered Image files (c) I/O - Storage [email protected] (d) Visualization •FCS Storage Network •GbE for external traffic (e) Compute ARGONNE CHICAGO 28 Initial TeraGrid Design 384 McKinley Processors (1.5 Teraflops, 96 nodes) 125 TB RAID storage 384 McKinley Processors (1.5 Teraflops, 96 nodes) 125 TB RAID storage Caltech ANL DWDM Optical Mesh SDSC 768 McKinley Processors (3 Teraflops, 192 nodes) 250 TB RAID storage [email protected] NCSA 2024 McKinley Processors (8 Teraflops, 512 nodes) 250 TB RAID storage ARGONNE CHICAGO 29 NSF TeraGrid: 14 TFLOPS, 750 TB 574p IA-32 Chiba City 256p HP X-Class 128p HP V2500 92p IA-32 Caltech: Data collection analysis HR Display & VR Facilities ANL: Visualization HPSS HPSS WAN Architecture Options: WAN Bandwidth Options: • • • • • • • • Myrinet-to-GbE; Myrinet as a WAN Layer2 design Wavelength Mesh Traditional IP Backbone Abilene (2.5 Gb/s, 10Gb/s late 2002) State and regional fiber initiatives plus CANARIE CA*Net Leased OC48 Dark Fiber, Dim Fiber, Wavelengths UniTree HPSS 1024p IA-32 320p IA-64 1176p IBM SP Blue Horizon Myrinet Myrinet 1500p Origin Sun E10K SDSC: Data-Intensive [email protected] NCSA: Compute-Intensive ARGONNE CHICAGO 34 NSF TeraGrid: 14 TFLOPS, 750 TB 574p IA-32 Chiba City 256p HP X-Class 128p Origin 128p HP V2500 HR Display & VR Facilities Caltech: Data collection analysis 92p IA-32 HPSS HPSS ANL: Visualization SDSC: Data-Intensive UniTree HPSS 1024p IA-32 320p IA-64 1176p IBM SP Blue Horizon Myrinet Myrinet 1500p Origin Sun E10K NCSA: Compute-Intensive [email protected] ARGONNE CHICAGO Defining Standard Services 35 Finite set of TeraGrid servicesapplications see standard services rather than particular implementations… Grid Applications …but sites also provide additional services that can be discovered and exploited. [email protected] IA-64 Linux Cluster Runtime Interactive Collection-Analysis Service File-based Data Service Volume-Render Service IA-64 Linux Cluster Interactive Development Collection-based Data Service ARGONNE CHICAGO 36 Standards Cyberinfrastructure • TeraGrid: focus on a finite set of service specifications applicable to TeraGrid resources. • If done well, other IA-64 cluster sites would adopt TeraGrid service specifications, increasing users’ leverage in writing to the specification, and others would adopt the framework for developing similar services (for Alpha, IA-32, etc.) • Note the specification should attempt to offer improvement over general Globus runtime environment without bogging down attempting to do everything (for which a user is better off running interactively!) Certificate Authority Certificate Authority Certificate Authority Certificate Authority TeraGrid Certificate Authority Grid Applications Grid Info Svces Visualization Services Alpha Clusters IA-64 Linux Clusters IA-32 Linux Clusters Data/Information File-based Data Service Collection-based Data Service Compute Analysis Interactive Development Interactive Collection-Analysis Service Runtime Visualization Services Relational dBase Data Service [email protected] ARGONNE CHICAGO 37 Strategy: Define Standard Services Finite number of TeraGrid Services – Defined as specifications, protocols, APIs – Separate from implementation Example: File-based Data Service – API/Protocol: Supports FTP and GridFTP, GSI authentication – SLA: All TeraGrid users have access to N TB storage, available 24/7 with M% availability, >= R Gb/s read, >= W Gb/s write, etc. [email protected] ARGONNE CHICAGO General TeraGrid Services 38 Authentication – GSI: Requires TeraGrid CA policy and services Resource Discovery and Monitoring – Define TeraGrid services/attributes to be published in Globus MDS-2 directory services – Require standard account information exchange to map use to allocation/individual – For many services, publish query interface > Scheduler: queue status > Compute, Visualization, etc. services: attribute details > Network Weather Service > Allocations/Accounting Database: for allocation status [email protected] ARGONNE CHICAGO 39 General TeraGrid Services Advanced Reservation – On-demand services – Staging data: coordination of storage+compute Communication and Data Movement – All services assume any TeraGrid cluster node can talk to any TeraGrid cluster node – All resources support GridFTP “Hosting environment” – Standard software environment – More sophisticated dynamic provisioning issues not yet addressed [email protected] ARGONNE CHICAGO 41 Topics in Grid Infrastructure Regional, national, intl optical infrastructure – I-WIRE, StarLight TeraGrid: Deep infrastructure – High-end support for U.S. community iVDGL: Wide infrastructure – Building a (international) community Open Grid Services Architecture – Future service & technology infrastructure [email protected] ARGONNE CHICAGO 42 iVDGL: A Global Grid Laboratory “We propose to create, operate and evaluate, over a sustained period of time, an international research laboratory for data-intensive science.” From NSF proposal, 2001 International Virtual-Data Grid Laboratory – A global Grid laboratory (US, Europe, Asia, South America, …) – A place to conduct Data Grid tests “at scale” – A mechanism to create common Grid infrastructure – A laboratory for other disciplines to perform Data Grid tests – A focus of outreach efforts to small institutions U.S. part funded by NSF (2001-2006) – $13.7M (NSF) + $2M (matching) [email protected] ARGONNE CHICAGO 43 Initial US-iVDGL Data Grid SKC BU Wisconsin PSU BNL Fermilab Indiana Caltech JHU UCSD Other sites to be added in 2002 [email protected] Hampton Florida Brownsville Tier1 (FNAL) Proto-Tier2 Tier3 university ARGONNE CHICAGO iVDGL: International Virtual Data Grid Laboratory 44 Tier0/1 facility Tier2 facility Tier3 facility 10 Gbps link 2.5 Gbps link 622 Mbps link Other link [email protected] CHICAGO U.S. PIs: Avery, Foster, Gardner, Newman, SzalayARGONNE www.ivdgl.org 45 iVDGL Architecture (from proposal) [email protected] ARGONNE CHICAGO 46 US iVDGL Interoperability US-iVDGL-1 Milestone (August 02) iGOC US-iVDGL1 Aug 2002 ATLAS SDSS/NVO CMS LIGO 1 1 2 2 1 2 [email protected] 1 2 ARGONNE CHICAGO 47 Transatlantic Interoperability iVDGL-2 Milestone (November 02) iGOC Outreach iVDGL-2 DataTAG Nov 2002 ATLAS SDSS/NVO CMS ANL HU IU LBL UC CS Research BNL BU LIGO FNAL CIT CIT FNAL CERN PSU INFN ANL UTB UK PPARC UC UWM U of A UCSD UF JHU UM UCB OU IU UTA ISI NU UW [email protected] ARGONNE CHICAGO 48 Topics in Grid Infrastructure Regional, national, intl optical infrastructure – I-WIRE, StarLight TeraGrid: Deep infrastructure – High-end support for U.S. community iVDGL: Wide infrastructure – Building a (international) community Open Grid Services Architecture – Future service & technology infrastructure [email protected] ARGONNE CHICAGO “Standard” Software Infrastructure: Globus ToolkitTM 49 Small, standards-based set of protocols for distributed system management – Authentication, delegation; resource discovery; reliable invocation; etc. Information-centric design – Data models; publication, discovery protocols Open source implementation – Large international user community Successful enabler of higher-level services and applications [email protected] ARGONNE CHICAGO 50 Example Grid Projects in eScience [email protected] ARGONNE CHICAGO 51 The Globus Toolkit in One Slide Grid protocols (GSI, GRAM, …) enable resource sharing within virtual orgs; toolkit provides reference implementation ( = Globus Toolkit services) MDS-2 (Monitor./Discov. Svc.) Reliable remote GSI User invocation Gatekeeper Reporter (Grid (registry + Authenticate & (factory) Security create proxy discovery) Create process Register Infrastruc- credential ture) User process #1 Proxy User process #2 Proxy #2 GRAM (Grid Resource Allocation & Management) Soft state registration; enquiry Other GSIauthenticated remote service requests GIIS: Grid Information Index Server (discovery) Other service (e.g. GridFTP) Protocols (and APIs) enable other tools and services for membership, discovery, data mgmt, workflow, … [email protected] ARGONNE CHICAGO 52 Globus Toolkit: Evaluation (+) Good technical solutions for key problems, e.g. – Authentication and authorization – Resource discovery and monitoring – Reliable remote service invocation – High-performance remote data access This & good engineering is enabling progress – Good quality reference implementation, multilanguage support, interfaces to many systems, large user base, industrial support – Growing community code base built on tools [email protected] ARGONNE CHICAGO 53 Globus Toolkit: Evaluation (-) Protocol deficiencies, e.g. – Heterogeneous basis: HTTP, LDAP, FTP – No standard means of invocation, notification, error propagation, authorization, termination, … Significant missing functionality, e.g. – Databases, sensors, instruments, workflow, … – Virtualization of end systems (hosting envs.) Little work on total system properties, e.g. – Dependability, end-to-end QoS, … – Reasoning about system properties [email protected] ARGONNE CHICAGO 54 Globus Toolkit Structure Service naming Soft state management Reliable invocation GRAM Notification MDS GSI GridFTP MDS ??? GSI GSI Data Resource Other Service or Application Job manager Job manager Compute Resource Lots of good mechanisms, but (with the exception of GSI) not that easily incorporated into other systems [email protected] ARGONNE CHICAGO Grid Evolution: Open Grid Services Architecture 55 Refactor Globus protocol suite to enable common base and expose key capabilities Service orientation to virtualize resources and unify resources/services/information Embrace key Web services technologies for standard IDL, leverage commercial efforts Result: standard interfaces & behaviors for distributed system management: the Grid service [email protected] ARGONNE CHICAGO Open Grid Services Architecture: Transient Service Instances 56 “Web services” address discovery & invocation of persistent services – Interface to persistent state of entire enterprise In Grids, must also support transient service instances, created/destroyed dynamically – Interfaces to the states of distributed activities – E.g. workflow, video conf., dist. data analysis Significant implications for how services are managed, named, discovered, and used – In fact, much of OGSA (and Grid) is concerned with the management of service instances [email protected] ARGONNE CHICAGO 57 Open Grid Services Architecture Defines fundamental (WSDL) interfaces and behaviors that define a Grid Service – Required + optional interfaces = WS “profile” – A unifying framework for interoperability & establishment of total system properties Defines WSDL extensibility elements – E.g., serviceType (a group of portTypes) Delivery via open source Globus Toolkit 3.0 – Leverage GT experience, code, community And commercial implementations [email protected] ARGONNE CHICAGO The Grid Service = Interfaces/Behaviors + Service Data Service data access Explicit destruction Soft-state lifetime Binding properties: - Reliable invocation - Authentication GridService (required) Service data element … other interfaces … (optional) Service data element 58 Standard: - Notification - Authorization - Service creation - Service registry - Manageability - Concurrency Service data element Implementation + applicationspecific interfaces Hosting environment/runtime (“C”, J2EE, .NET, …) [email protected] ARGONNE CHICAGO Grid Service Example: Database Service A DBaccess Grid service will support at least two portTypes Grid – GridService Service – DBaccess 59 Each has service data DBaccess Name, lifetime, etc. DB info – GridService: basic introspection information, lifetime, … – DBaccess: database type, query languages supported, current load, …, … Maybe other portTypes as well – E.g., NotificationSource (SDE = subscribers) [email protected] ARGONNE CHICAGO 60 Example: Data Mining for Bioinformatics Community Registry Mining Factory Database Service BioDB 1 User Application “I want to create a personal database containing data on e.coli metabolism” Compute Service Provider . . . . . . Database Service Database Factory BioDB n Storage Service Provider [email protected] ARGONNE CHICAGO 61 Example: Data Mining for Bioinformatics “Find me a data Community mining service, and Registry somewhere to store data” Mining Factory Database Service BioDB 1 User Application Compute Service Provider . . . . . . Database Service Database Factory BioDB n Storage Service Provider [email protected] ARGONNE CHICAGO 62 Example: Data Mining for Bioinformatics GSHs for Mining and Database factories User Application Community Registry Mining Factory Database Service BioDB 1 Compute Service Provider . . . . . . Database Service Database Factory BioDB n Storage Service Provider [email protected] ARGONNE CHICAGO 63 Example: Data Mining for Bioinformatics Community Registry “Create a data mining service with initial lifetime 10” User Application “Create a database with initial lifetime 1000” Mining Factory Database Service BioDB 1 Compute Service Provider . . . . . . Database Service Database Factory BioDB n Storage Service Provider [email protected] ARGONNE CHICAGO 64 Example: Data Mining for Bioinformatics Community Registry “Create a data mining service with initial lifetime 10” User Application “Create a database with initial lifetime 1000” Mining Factory Database Service Miner BioDB 1 Compute Service Provider . . . . . . Database Service Database Factory BioDB n Database Storage Service Provider [email protected] ARGONNE CHICAGO 65 Example: Data Mining for Bioinformatics Community Registry Mining Factory Query Miner User Application Database Service BioDB 1 Compute Service Provider . . Query . . . . Database Service Database Factory BioDB n Database Storage Service Provider [email protected] ARGONNE CHICAGO 66 Example: Data Mining for Bioinformatics Community Registry Keepalive Mining Factory Query Miner BioDB 1 Compute Service Provider . . Query . User Application Keepalive Database Service . . . Database Service Database Factory BioDB n Database Storage Service Provider [email protected] ARGONNE CHICAGO 67 Example: Data Mining for Bioinformatics Community Registry Keepalive User Application Keepalive Mining Factory Database Service Miner BioDB 1 Compute Service Provider . . . . . . Results Database Service Database Factory Results BioDB n Database Storage Service Provider [email protected] ARGONNE CHICAGO 68 Example: Data Mining for Bioinformatics Community Registry User Application Keepalive Mining Factory Database Service Miner BioDB 1 Compute Service Provider . . . . . . Database Service Database Factory BioDB n Database Storage Service Provider [email protected] ARGONNE CHICAGO 69 Example: Data Mining for Bioinformatics Community Registry Mining Factory Database Service BioDB 1 Compute Service Provider . . . User Application Keepalive . . . Database Service Database Factory BioDB n Database Storage Service Provider [email protected] ARGONNE CHICAGO GT3: An Open Source OGSACompliant Globus Toolkit 70 GT3 Core – Implements Grid service interfaces & behaviors – Reference impln of evolving standard – Multiple hosting envs: Java/J2EE, C, C#/.NET? GT3 Base Services GT3 Data Services Other Grid Services GT3 Base Services GT3 Core – Evolution of current Globus Toolkit capabilities Many other Grid services [email protected] ARGONNE CHICAGO OGSA Definition and Delivery (Very Approximate!!) 71 Other Systems T I M E GGF OGSI WG Grid Service Specification GGF WGs Other core specs: • Security • Res. Mgmt. • Etc. GGF WGs Other specs: • Databases • Etc. • Etc. [email protected] Prototype Feedback Globus OGSI Reference Impln Globus Toolkit Version 3 Other OGSAbased software ARGONNE CHICAGO Q2: What Higher-Level Services? [email protected] 72 ARGONNE CHICAGO Summary: Grid Infrastructure 73 Grid applications demand new infrastructure beyond traditional computers and networks – Network-accessible resources of all types – High-speed networks – Services and operational procedures – Software technology for building services (which must also be treated as infrastructure) TeraGrid, iVDGL, StarLight, DOT – Connections to international sites? [email protected] ARGONNE CHICAGO Summary: Open Grid Services Architecture 74 Open Grid Services Architecture represents (we hope!) next step in Grid evolution Service orientation enables unified treatment of resources, data, and services Standard interfaces and behaviors (the Grid service) for managing distributed state Deeply integrated information model for representing and disseminating service data Open source Globus Toolkit implementation (and commercial value adds) [email protected] ARGONNE CHICAGO For More Information 75 Survey + research articles – www.mcs.anl.gov/~foster I-WIRE: www.iwire.org TeraGrid: www.teragrid.org iVDGL: www.ivDGL.org The Globus Project™ – www.globus.org GriPhyN project – www.griphyn.org Global Grid Forum – www.gridforum.org – www.gridforum.org/ogsi-wg – Edinburgh, July 22-24 – Chicago, Oct 15-17 [email protected] ARGONNE CHICAGO