Survey
* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project
* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project
HENP Networks and Grids for Global Science Harvey B. Newman California Institute of Technology 3rd International Data Grid Workshop Daegu, Korea August 26, 2004 Challenges for Global HENP Experiments LHC Example- 2007 5000+ Physicists 250+ Institutes 60+ Countries BaBar/D0 Example - 2004 500+ Physicists 100+ Institutes 35+ Countries Major Challenges (Shared with Other Fields) Worldwide Communication and Collaboration Managing Globally Distributed Computing & Data Resources Cooperative Software Development and Data Analysis Large Hadron Collider (LHC) CERN, Geneva: 2007 Start pp s =14 TeV L=1034 cm-2 s-1 27 km Tunnel in Switzerland & France CMS TOTEM First Beams: Summer 2007 Physics Runs: from Fall 2007 pp, general purpose; HI ALICE : HI Atlas LHCb: B-physics Higgs, SUSY, QG Plasma, CP Violation, … the Unexpected Challenges of Next Generation Science in the Information Age Petabytes of complex data explored and analyzed by 1000s of globally dispersed scientists, in hundreds of teams Flagship Applications High Energy & Nuclear Physics, AstroPhysics Sky Surveys: TByte to PByte “block” transfers at 1-10+ Gbps Fusion Energy: Time Critical Burst-Data Distribution; Distributed Plasma Simulations, Visualization and Analysis; Preparations for Fusion Energy Experiment eVLBI: Many real time data streams at 1-10 Gbps BioInformatics, Clinical Imaging: GByte images on demand Provide results with rapid turnaround, coordinating large but limited computing and data handling resources, over networks of varying capability in different world regions Advanced integrated applications, such as Data Grids, rely on seamless operation of our LANs and WANs With reliable, quantifiable high performance LHC Data Grid Hierarchy: Developed at Caltech CERN/Outside Resource Ratio ~1:2 Tier0/( Tier1)/( Tier2) ~1:1:1 ~PByte/sec ~100-1500 MBytes/sec Online System Experiment CERN Center PBs of Disk; Tape Robot Tier 0 +1 Tier 1 10 - 40 Gbps IN2P3 Center INFN Center RAL Center FNAL Center ~10 Gbps ~10 Gbps Tier 2 Tier2 Center Tier2 Center Tier2 Center Tier2 CenterTier2 Center Tier 3 Institute Institute Physics data cache Workstations Institute Institute 1 to 10 Gbps Tens of Petabytes by 2007-8. An Exabyte ~5-7 Years later. Tier 4 Emerging Vision: A Richly Structured, Global Dynamic System ICFA and Global Networks for Collaborative Science National and International Networks, with sufficient (rapidly increasing) capacity and seamless end-to-end capability, are essential for The daily conduct of collaborative work in both experiment and theory Experiment development & construction on a global scale Grid systems supporting analysis involving physicists in all world regions The conception, design and implementation of next generation facilities as “global networks” “Collaborations on this scale would never have been attempted, if they could not rely on excellent networks” History of Bandwidth Usage – One Large Network; One Large Research Site ESnet Accepted Traffic 1/90 – 1/04 Exponential Growth Since ’92; ESnet Monthly Accepted Traffic 1/90-1/04 Annual Rate Increased from 1.7 to 2.0X Per Year In the Last 5 Years 300 200 150 100 50 Jul, 03 Oct, 02 Jan, 02 Apr, 01 Jul, 00 Oct, 99 Jan, 99 Apr, 98 Jul, 97 Oct, 96 Jan, 96 Apr, 95 Jul, 94 Oct,93 Jan, 93 Apr, 92 Jul,91 Oct, 90 0 Jan, 90 TByte/Month 250 SLAC Traffic ~300 Mbps; ESnet Limit Growth in Steps: ~ 10X/4 Years Projected: ~2 Terabits/s by ~2014 Int’l Networks BW on Major Links for HENP: US-CERN Example Rate of Progress >> Moore’s Law (US-CERN Example) 9.6 kbps Analog (1985) 64-256 kbps Digital (1989 - 1994) 1.5 Mbps Shared (1990-3; IBM) 2 -4 Mbps (1996-1998) 12-20 Mbps (1999-2000) 155-310 Mbps (2001-2) 622 Mbps (2002-3) 2.5 Gbps (2003-4) 10 Gbps (2005) 4x10 Gbps or 40 Gbps (2007-8) A factor of ~1M Bandwidth Improvement over 1985-2005 (a factor of ~5k during 1995-2005) A prime enabler of major HENP programs HENP has become a leading applications driver, and also a co-developer of global networks [X 7 – 27] [X 160] [X 200-400] [X 1.2k-2k] [X 16k – 32k] [X 65k] [X 250k] [X 1M] [X 4M] Internet Growth in the World At Large Amsterdam Internet Exchange Point Example 5 Minute Max 30 Gbps 20 Gbps Some Annual Growth Spurts; Typically In Summer-Fall The Rate of HENP Network Usage Growth (~100% Per Year) is Similar to the World at Large 11.08.04 Average http://www.guinnessworldrecords.com/ 6.6 Gbps 16500km LSR History – IPv4 single stream 80 5.4 Gbps 7067km 60 2.5 Gbps 0.9 Gbps 10037km 0.4 Gbps 10978km 12272km 40 20 Jun- 04 0 Apr 04 03 Nov- 03 Feb- Oct 03 02 Monitoring of the Abilene traffic in LA: Nov- 120 100 4.2 Gbps 5.6 Gbps 16343km 10949km Apr 02 Judged on product of transfer speed and distance end-to-end, using standard Internet (TCP/IP) protocols. IPv6 record: 4.0 Gbps between Geneva and Phoenix (SC2003) IPv4 Multi-stream record with Windows & Linux: 6.6 Gbps between Caltech and CERN (16 kkm; “Grand Tour d’Abilene”) June 2004 Exceeded 100 Petabit-m/sec Single Stream 7.5 Gbps X 16 kkm with Linux Achieved in July Concentrate now on reliable Terabyte-scale file transfers Note System Issues: CPU, PCI-X Bus, NIC, I/O Controllers, Drivers June 2004 Record Network Petabitmeter (10^15 bit*meter) Internet 2 Land Speed Record (LSR) Evolving Quantitative Science Requirements for Networks (DOE High Perf. Network Workshop) Science Areas Today End2End Throughput 5 years End2End Throughput 5-10 Years End2End Throughput Remarks High Energy Physics 0.5 Gb/s 100 Gb/s 1000 Gb/s High bulk throughput Climate (Data & Computation) 0.5 Gb/s 160-200 Gb/s N x 1000 Gb/s High bulk throughput SNS NanoScience Not yet started 1 Gb/s 1000 Gb/s + QoS for Control Channel Remote control and time critical throughput Fusion Energy 0.066 Gb/s (500 MB/s burst) 0.198 Gb/s (500MB/ 20 sec. burst) N x 1000 Gb/s Time critical throughput Astrophysics 0.013 Gb/s (1 TByte/week) N*N multicast 1000 Gb/s Computational steering and collaborations Genomics Data & Computation 0.091 Gb/s (1 TBy/day) 100s of users 1000 Gb/s + QoS for Control Channel High throughput and steering HENP Lambda Grids: Fibers for Physics Problem: Extract “Small” Data Subsets of 1 to 100 Terabytes from 1 to 1000 Petabyte Data Stores Survivability of the HENP Global Grid System, with hundreds of such transactions per day (circa 2007) requires that each transaction be completed in a relatively short time. Example: Take 800 secs to complete the transaction. Then Transaction Size (TB) Net Throughput (Gbps) 1 10 10 100 100 1000 (Capacity of Fiber Today) Summary: Providing Switching of 10 Gbps wavelengths within ~2-4 years; and Terabit Switching within 5-8 years would enable “Petascale Grids with Terabyte transactions”, to fully realize the discovery potential of major HENP programs, as well as other data-intensive research. SCIC in 2003-2004 http://cern.ch/icfa-scic Three 2004 Reports; Presented to ICFA in February Main Report: “Networking for HENP” [H. Newman et al.] Includes Brief Updates on Monitoring, the Digital Divide and Advanced Technologies [*] A World Network Overview (with 27 Appendices): Status and Plans for the Next Few Years of National & Regional Networks, and Optical Network Initiatives Monitoring Working Group Report [L. Cottrell] Digital Divide in Russia [V. Ilyin] August 2004 Update Reports at the SCIC Web Site: See http://icfa-scic.web.cern.ch/ICFA-SCIC/documents.htm Asia Pacific, Latin America, GLORIAD (US-Ru-Ko-China); Brazil, Korea, etc. SCIC Main Conclusion for 2003 Setting the Tone for 2004 The disparity among regions in HENP could increase even more sharply, as we learn to use advanced networks effectively, and we develop dynamic Grid systems in the most favored” regions We must take action, and work to Close the Digital Divide To make Physicists from All World Regions Full Partners in Their Experiments; and in the Process of Discovery This is essential for the health of our global experimental collaborations, our plans for future projects, and our field. ICFA Report: Networks for HENP General Conclusions (2) Reliable high End-to-end Performance of networked applications such as large file transfers and Data Grids is required. Achieving this requires: End-to-end monitoring extending to all regions serving our community. A coherent approach to monitoring that allows physicists throughout our community to extract clear information is required. Upgrading campus infrastructures. These are still not designed to support Gbps data transfers in most HEP centers. One reason for under-utilization of national and Int’l backbones, is the lack of bandwidth to end-user groups in the campus. Removing local, last mile, and nat’l and int’l bottlenecks end-to-end, whether technical or political in origin. While National and International backbones have reached 2.5 to 10 Gbps speeds in many countries, the bandwidths across borders, the countryside or the city may be much less. This problem is very widespread in our community, with examples stretching from the Asia Pacific to Latin America to the Northeastern U.S. Root causes for this vary, from lack of local infrastructure to unfavorable pricing policies. ICFA Report (2/2004) Update: Main Trends Continue, Some Accelerate Current generation of 2.5-10 Gbps network backbones and major Int’l links arrived in the last 2-3 Years [US+Europe+Japan; Now Korea and China] Capability: 4 to Hundreds of Times; Much Faster than Moore’s Law Proliferation of 10G links across the Atlantic Now; Will Begin use of Multiple 10G Links (e.g. US-CERN) Along Major Paths by Fall 2005 Direct result of Falling Network Prices: $ 0.5 – 1M Per Year for 10G Ability to fully use long 10G paths with TCP continues to advance: 7.5 Gbps X 16kkm (August 2004) Technological progress driving equipment costs in end-systems lower “Commoditization” of Gbit Ethernet (GbE) ~complete: ($20-50 per port) 10 GbE commoditization (e.g. < $ 2K per NIC with TOE) underway Some regions (US, Europe) moving to owned or leased dark fiber Emergence of the “Hybrid” Network Model: GNEW2004; UltraLight, GLIF Grid-based Analysis demands end-to-end high performance & management The rapid rate of progress is confined mostly to the US, Europe, Japan and Korea, as well as the major Transatlantic routes; this threatens to cause the Digital Divide to become a Chasm Work on the Digital Divide: Several Perspectives Work on Policies and/or Pricing: pk, in, br, cn, SE Europe, … Find Ways to work with vendors, NRENs, and/or Gov’ts Exploit Model Cases: e.g. Poland, Slovakia, Czech Republic Inter-Regional Projects GLORIAD, Russia-China-US Optical Ring South America: CHEPREO (US-Brazil); EU CLARA Project Virtual SILK Highway Project (DESY): FSU satellite links Workshops and Tutorials/Training Sessions For Example: Digital Divide and HEPGrid Workshop, UERJ Rio, February 2004; Next Daegu May 2005 Help with Modernizing the Infrastructure Design, Commissioning, Development Tools for Effective Use: Monitoring, Collaboration Participate in Standards Development; Open Tools Advanced TCP stacks; Grid systems Grid and Network Workshop at CERN March 15-16, 2004 WORKSHOP GOALS CONCLUDING STATEMENT Share and challenge the lessons learned by nat’l and international projects in the past years; Workshop "Following the 1st International Gridthree Networking Share the current engineering and by (GNEW2004) that wasstate heldof atnetwork CERN and co-organized infrastructureDANTE, and its likely near future; CERN/DataTAG, ESnet,evolution Internet2in&the TERENA, there is Examine our understanding of the networking needs of of awide consensus that hybrid network services capable Grid applications see the ICFA-SCIC reports); offering both packet-(e.g., and circuit/lambda-switching as well highly Develop a vision performance of how network engineering and as advanced measurements and a new infrastructure will (or should) support Grid computing generation distributed needs inof the next threesystem years. software, will be required in order to support emerging data intensive Grid applications, Such as High Energy Physics, Astrophysics, Climate and Supernova modeling, Genomics and Proteomics, requiring 10-100 Gbps and up over wide areas." Transition beginning now to optical, multiwavelength Community owned or leased “dark fiber” networks for R&E National Lambda Rail (NLR) SEA NLR POR Coming SAC NYC CHI OGD DEN SVL CLE FRE PIT KAN NAS STR LAX RAL PHO SDG WAL OLG ATL DAL JAC BOS WDC Up Now Initially 4 10G Wavelengths Northern Route Operation by 4Q04 Internet2 HOPI Initiative (w/HEP) To 40 10G Waves in Future nl, de, pl, cz,jp 18 US States 15808 Terminal, Regen or OADM site Fiber route JGN2: Japan Gigabit Network (4/04 – 3/08) 20 Gbps Backbone, 6 Optical Cross-Connects [Legends ] 20Gbps 10Gbps 1Gbps Optical testbeds Access points <10G> ・Ishikawa Hi-tech Exchange Center (Tatsunokuchi-machi, Ishikawa Prefecture) <100M> ・Toyama Institute of Information Systems (Toyama) ・Fukui Prefecture Data Super Highway AP * (Fukui) Core network nodes <1G> ・Teleport Okayama (Okayama) ・Hiroshima University (Higashi Hiroshima) <100M> ・Tottori University of Environmental Studies (Tottori) ・Techno Ark Shimane (Matsue) ・New Media Plaza Yamaguchi (Yamaguchi) <10G> ・Kyushu University (Fukuoka) <100M> ・NetCom Saga (Saga) ・Nagasaki University (Nagasaki) ・Kumamoto Prefectural Office (Kumamoto) ・Toyonokuni Hyper Network AP *(Oita) ・Miyazaki University (Miyazaki) ・Kagoshima University (Kagoshima) <10G> ・Kyoto University (Kyoto) ・Osaka University (Ibaraki) <1G> ・NICT Kansai Advanced Research Center (Kobe) <100M> ・Lake Biwa Data Highway AP * (Ohtsu) ・Nara Prefectural Institute of Industrial Technology (Nara) ・Wakayama University (Wakayama) ・Hyogo Prefecture Nishiharima Technopolis (Kamigori-cho, Hyogo Prefecture) Sapporo <100M> ・Niigata University (Niigata) ・Matsumoto Information Creation Center (Matsumoto, Nagano Prefecture) Sendai Kanazawa Nagano Osaka NICT Koganei Headquarters Okayama Kochi Okinawa <100M> ・Kagawa Prefecture Industry Promotion Center (Takamatsu) ・Tokushima University (Tokushima) ・Ehime University (Matsuyama) ・Kochi University of Technology (Tosayamada-cho, Kochi Prefecture) NICT Keihannna Human Info-Communications Research Center <1G> ・Tohoku University (Sendai) ・NICT Iwate IT Open Laboratory (Takizawa-mura, Iwate Prefecture) <100M> ・Hachinohe Institute of Technology (Hachinohe, Aomori Prefecture) ・Akita Regional IX * (Akita) ・Keio University Tsuruoka Campus (Tsuruoka, Yamagata Prefecture) ・Aizu University (Aizu Wakamatsu) <10G> ・Tokyo University Fukuoka NICT Kita Kyushu IT Open Laboratory <100M> ・Hokkaido Regional Network Association AP * (Sapporo) Nagoya <100M> ・Nagoya University (Nagoya) ・University of Shizuoka (Shizuoka) ・Softopia Japan (Ogaki, Gifu Prefecture) ・Mie Prefectural College of Nursing (Tsu) NICT Tsukuba Research Center (Bunkyo Ward, Tokyo) ・NICT Kashima Space Research Center (Kashima, Ibaraki Prefecture) <1G> ・Yokosuka Telecom Research Park (Yokosuka, Kanagawa Prefecture) <100M> ・Utsunomiya University (Utsunomiya) ・Gunma Industrial Technology Center (Maebashi) ・Reitaku University (Kashiwa, Chiba Prefecture) ・NICT Honjo Information and Communications Open Laboratory (Honjo, Saitama Prefecture) ・Yamanashi Prefecture Open R&D Center (Nakakoma-gun, Yamanashi Prefecture) Otemachi USA *IX:Internet eXchange AP:Access Point APAN-KR : KREONET/KREONet2 II UltraLight Collaboration: http://ultralight.caltech.edu Caltech, UF, FIU, UMich, SLAC,FNAL, MIT/Haystack, CERN, UERJ(Rio), NLR, CENIC, UCAID, Translight, UKLight, Netherlight, UvA, UCLondon, KEK, Taiwan Cisco, Level(3) Integrated hybrid experimental network, leveraging Transatlantic R&D network partnerships; packet-switched + dynamic optical paths 10 GbE across US and the Atlantic: NLR, DataTAG, TransLight, NetherLight, UKLight, etc.; Extensions to Japan, Taiwan, Brazil End-to-end monitoring; Realtime tracking and optimization; Dynamic bandwidth provisioning Agent-based services spanning all layers of the system, from the optical cross-connects to the applications. GLIF: Global Lambda Integrated Facility “GLIF is a World Scale Lambda based Lab for Application and Middleware development, where Grid applications ride on dynamically configured networks based on optical wavelengths ... Coexisting with more traditional packetswitched network traffic 4th GLIF Workshop: Nottingham UK Sept. 2004 10 Gbps Wavelengths For R&E Network Development Are Prolifering, Across Continents and Oceans PROGRESS in SE Europe (Sk, Pl, Cz, Hu, …) 1660 km of Dark Fiber CWDM Links, up to 112 km. 1 to 4 Gbps (GbE) August 2002: First NREN in Europe to establish Int’l GbE Dark Fiber Link, to Austria April 2003 to Czech Republic. Planning 10 Gbps Backbone; dark fiber link to Poland this year. Dark Fiber in Eastern Europe Poland: PIONIER Network 2650 km Fiber Connecting 16 MANs; 5200 km and 21 MANs by 2005 GDAŃS K KOS ZALIN OLS ZTYN S ZCZECIN BYDGOS ZCZ BIAŁYS TOK TORUŃ POZNAŃ Support Computational Grids Domain-Specific Grids Digital Libraries Interactive TV Add’l Fibers for WARS ZAWA GUBIN ZIELONA GÓRA S IEDLCE ŁÓDŹ PUŁAWY WROCŁAW RADOM CZĘS TOCHOWA KIELCE OPOLE GLIWICE KATOWICE KRAKÓW CIES ZYN BIELS KO-BIAŁA e-Regional Initiatives Ins ta lle d fibe r P IONIER node s Fibe rs pla nne d in 2004 P IONIER node s pla nne d in 2004 RZES ZÓW LUBLIN The Advantage of Dark Fiber CESNET Case Study (Czech Republic) 1 x 2,5G 2513 km Leased Fibers (Since 1999) about 150km (e.g. Ústí n.L. - Liberec) about 300km (e.g. Praha - Brno) * ** 4 x 2,5G Case Study Result Wavelength Service Vs. Fiber Lease: Cost Savings of 50-70% Over 4 Years for Long 2.5G or 10G Links about 150km (e.g. Ústí n.L. - Liberec) about 300km (e.g. Praha - Brno) * ** 1 x 10G about 150km (e.g. Ústí n.L. - Liberec) about 300km (e.g. Praha - Brno) * ** 4 x 10G about 150km (e.g. Ústí n.L. - Liberec) about 300km (e.g. Praha - Brno) * ** Leased 1 x 2,5G (EURO/Month) 7,000 8,000 Leased fibre with own equipment (EURO/Month) 5 000 * 7 000 ** 2 x booster 18dBm 2 x booster 27dBm + 2 x preamplifier + 6 x DCF Leased 4 x 2,5G (EURO/Month) 14,000 23,000 Leased fibre with own equipment (EURO/Month) 8 000 * 11 000 ** 2 x booster 24dBm, DWDM 2,5G 2 x (booster +In-line + preamplifier), 6 x DCF, DWDM 2,5G Leased 1 x 10G (EURO/Month) 14,000 16,000 Leased fibre with own equipment (EURO/Month) 5 000 * 8 000 ** 2 x booster 21dBm, 2 x DCF 2 x (booster 21dBm + in-line + preamplifier) + 6 x DCF Leased 4 x 10G (EURO/Month) 29,000 47,000 Leased fibre with own equipment (EURO/Month) 12 000 * 14 000 ** 2 x booster 24dBm, 2 x DCF, DWDM 10G 2 x (booster +In-line + preamplifier), 6 x DCF, DWDM 10G ICFA/SCIC Network Monitoring Prepared by Les Cottrell, SLAC, for ICFA www.slac.stanford.edu/grp/scs/net/talk03/icfa-aug04.ppt PingER: World View from SLAC Now monitoring 650 sites in 115 countries In last 9 months: Several sites in Russia (GLORIAD) Many hosts in Africa (27 of 54 Countries) Monitoring sites in Pakistan, Brazil TCP throughput measured from N. America From the PingER project, Aug 2004 to World Regions 10000 C. Asia (8) Latin America (37) 50% Improvement/year ~ factor of 10 in < 6 years 10000 Edu (141) 1000 1000 Europe(150) Canada (27) Mid East (16) S.E. Europe (21) 10 100 100 10 Caucasus (8) Important for policy makers Dec-04 Dec-03 1 Dec-02 Dec-01 Africa (30) Dec-00 Dec-99 India(7) Dec-98 Dec-97 Dec-96 Jan-96 China (13) Russia(17) 1 Jan-95 Derived TCP throughput in KBytes/sec View from CERN Confirms This View C. Asia, Russia, SE Europe, L. America, M. East, China: 45 yrs behind India, Africa: 7 yrs behind Research Networking in Latin America: August 2004 The only Countries with research network connectivity now in Latin America: Argentina, Brazil, Chile, Mexico, Venezuela AmPath Provided connectivity for some South American countries New Sao Paolo-Miami Link at 622 Mbps Starting This Month AmPath New: CLARA (Funded by EU) Regional Network Connecting 19 Countries: Argentina Brasil Bolivia Chile Colombia Costa Rica Cuba Dominican Republic Ecuador El Salvador Guatemala Honduras Mexico Panama Paraguay Peru Uruguay Venezuela Nicaragua 155 Mbps Backbone with 10-45 Mbps Spurs; 4 Mbps Satellite to Cuba; 622 Mbps to Europe Also NSF Proposals To Connect at 2.5G to US HEPGRID (CMS) in Brazil HEPGRID-CMS/BRAZIL is a project to build a Grid that At Regional Level will include CBPF,UFRJ,UFRGS,UFBA, UERJ & UNESP At International Level will be integrated with CMS Grid based at CERN; focal points include iVGDL/Grid3 and bilateral projects with Caltech Group Brazilian HEPGRID On line systems T0 +T1 2.5 - 10 Gbps CERN T1 France Germany UNESP/USP SPRACE-Working T3 T2 UFRGS UERJ: T2T1, 100500 Nodes; Plus T2s to 100 Nodes Italy BRAZIL 622 Mbps UERJ Regional Tier2 Ctr USA T2 T1 Gigabit UERJ CBPF UFBA UFRJ T4 Individual Machines Latin America Science Areas Interested in Improving Connectivity ( by Country) Subject Argentina Brazil Chile Colombia Costa Rica Equator Astrophysics e-VLBI High Energy Physics Geosciences Marine sciences Health and Biomedical applications Environmental studies Networks and Grids: The Potential to Spark a New Era of Science in the Region Mexico Asia Pacific Academic Network Connectivity APAN Status July 2004 RU Europe 200M 34M Connectivity to US from JP, KO, AU is Advancing Rapidly. Progress in the Region, and to Europe is Much Slower CN 2G KR 155M 1.2G 310M TW `722M HK TH LK 45M 90M 155M 932M (to 21 G) ID 2.5M Access Point Exchange Point Current Status 2004 (plan) 45M 7.5M PH 1.5M 155M 1.5M VN MY 2M 12M SG 2M US 9.1G 622M 777M IN 20.9G JP 16M AU Better North/South Linkages within Asia JP-SG link: 155Mbps in 2005 is proposed to NSF by CIREN JP- TH link: 2Mbps 45Mbps in 2004 is being studied. CIREN is studying an extension to India APAN Link Information Countries AU-US AU-US (PAIX) CN-HK CN-HK CN-JP CN-JP CN-US CN-US HK-US HK-TW IN-US/UK JP-ASIA JP-ID JP-KR JP-LA JP-MY JP-PH JP-PH JP-SG JP-TH JP-TH JP-US Network AARNet AARNet CERNET CSTNET CERNET CERNET CERNET CSTNET HARNET HARNET/TANET/ASNET ERNET UDL AI3(ITB) APII AI3 (NUOL) AI3(USM) AI3(ASTI) MAFFIN AI3(TP) AI3(AIT) SINET(ThaiSarn) TransPac Bandwidth (Mbps) 310 to 2 x 10 Gbps soon 622 622 155 155 45 155 155 45 100 16 9 0.5/1.5 2Gbps 0.128/0.128 1.5/0.5 1.5/0.5 6 1.5/0.5 (service interrupted) 2 5 Gbps to 2x10 Gbps soon 2004.7.7 [email protected] AUP/Remark R&E + Commodity R&E + Commodity R&E + Commodity R&E R&E Native IPv6 R&E + Commodity R&E R&E R&E R&E R&E R&E R&E R&E R&E R&E Research R&E R&E R&E R&E APAN Link Information Countries (JP)-US-EU JP-US JP-US JP-US JP-US JP-VN KR-FR KR-SG KR-US LK-JP MY-SG SG-US TH-US TW-HK TW-JP TW-SG TW-US (TW)-US-NL Network SINET SINET IEEAF IEEAF Japan-Hawaii AI3(IOIT) KOREN/RENATER APII KOREN/KREONet2 LEARN NRG/SICU SingaREN Uninet ASNET/TANET/TWAREN ASNET/TANET ASNET/SingAREN ASNET/TANET/TWAREN ASNET/TANET/TWAREN Bandwidth (Mbps) 155 5 Gbps 10Gbps 622 155 1.5/0.5 34 8 1.2Gbps 2.5 2 90 155 622 622 155 6.6 Gbps 2.5 Gbps 2004.7.7 [email protected] AUP/Remark R&E / No Transit R&E / No Transit R&E wave service R&E R&E R&E Research (TEIN) R&E R&E R&E Experiment (Down) R&E R&E R&E R&E R&E R&E R&E APAN Recommendations (at July 2004 Meeting in CAIRNS, Au) Central Issues for APAN this decade Stronger linkages between applications and infrastructure neither can exist independently Stronger application and infrastructure linkages among APAN members. Continuing focus on APAN as an organization that represents infrastructure interests in Asia Closer connection between APAN the infrastructure & applications organization and regional political organizations (e.g. APEC, ASEAN) New issues demand attention Application measurement, particularly end-to-end network performance measurement is increasingly critical (deterministic networking) Security must now be a consideration for every application and every network. KR-US/CA Transpacific connection Participation in Global-scale Lambda Networking Two STM-4 circuits (1.2G) : KR-CA-US Global lambda networking : North America, Europe, Asia Pacific, etc. Global Lambda Networking KREONET/SuperSIReN CA*Net4 StarLight STM-4 * 2 Chicago APII-testbed/KREONet2 Seattle PacWave New Record!!! 916 Mbps from CHEP to Caltech (22/06/’04) Subject: UDP test on KOREN-TransPAC-Caltech Date: Tue, 22 Jun 2004 13:47:25 +0900 From: "Kihwan Kwon" <[email protected]> To: <[email protected]> [root@sul Iperf]# ./iperf -c socrates.cacr.caltech.edu -u -b 1000m -----------------------------------------------------------Client connecting to socrates.cacr.caltech.edu, UDP port 5001 Sending 1470 byte datagrams; UDP buffer size: 64.0 KByte (default) -----------------------------------------------------------[ 5] local 155.230.20.20 port 33036 connected with 131.215.144.227 [ ID] Interval Transfer Bandwidth [ 5] 0.0-2595.2 sec 277 GBytes 916 Mbits/sec USA TransPAC KNU/Korea Max. 947.3Mbps G/H-Japan Global Ring Network for Advanced Applications Development OC3 circuits Moscow-Chicago- Beijing since January 2004 OC3 circuit Moscow-Beijing July 2004 (completes the ring) Korea (KISTI) joining US, Russia, China as full partner in GLORIAD Plans developing for Central Asian extension (w/Kyrgyz Government) Rapid traffic growth with heaviest US use from DOE (FermiLab), NASA, NOAA, NIH and Universities (UMD, IU, UCB, UNC, UMN, PSU, Harvard, Stanford, Wash., Oregon, 250+ others) Aug. 8 2004: P.K. Young, Korean IST Advisor to President Announces Korea Joining GLORIAD TEIN gradually to 10G, connected to GLORIAD Asia Pacific Info. InfraStructure (1G) will be backup net to GLORIAD > 5TBytes now transferred monthly via GLORIAD to US, Russia, China GLORIAD 5-year Proposal Pending (with US NSF) for expansion: 2.5G MoscowAmsterdam-Chicago-Seattle-Hong Kong-Pusan-Beijing circuits early 2005; 10G ring around northern hemisphere 2007; multiple wavelength service 2009 – providing hybrid circuit-switched (primarily Ethernet) and routed services Internet in China (J.P.Wu APAN July 2004) Internet users in China: from 6.8 Million to 78 Million within 6 months Total Wireline Dial Up ISDN 68M 23.4M 45.0M 4.9M IP Addresses: 32M(1A+233B+146C) Backbone:2.5-10G DWDM+Router International links:20G Exchange Points:> 30G(BJ,SH,GZ) Last Miles Broad band 9.8M Ethernet,WLAN,ADSL,CTV,CDMA,ISDN,GPRS, Dial-up Need IPv6 China: CERNET Update 1995, 64K Nation wide backbone connecting 8 cities, 100 Universities 1998, 2M Nation wide backbone connecting 20 cities, 300 Universities 2000, Own dark fiber crossing 30+ major cities and 30,000 kilometers 2001, CERNET DWDM/SDH network finished 2001, 2.5G/155M Backbone connecting 36 cities, 800 universities 2003,1300 + universities and institutes, over 15 million users CERNET2 and Key Technologies CERNET 2: Next Generation Education and Research Network in China CERNET 2 Backbone connecting 15-20 GigaPOPs at 2.5G-10Gbps (I2-like Model) Connecting 200 Universities and 100+ Research Institutes at 1Gbps-10Gbps Native IPv6 and Lambda Networking Support/Deployment of the following technologies: E2E performance monitoring Middleware and Advanced Applications Multicast AFRICA: Key Trends M. Jensen and P. Hamilton Infrastructure Report, March 2004 Growth in traffic and lack of infrastructure Predominance of Satellite; But these satellites are heavily subscribed Int’l Links: Only ~1% of traffic on links is for Internet connections; Most Internet traffic (for ~80% of countries) via satellite Flourishing Grey market for Internet & VOIP traffic using VSAT dishes Many Regional fiber projects in “planning phase” (some languished in the past); Only links from South Africa to Nimibia, Botswana done so far Int’l fiber Project: SAT-3/WASC/SAFE Cable from South Africa to Portugal Along West Coast of Africa Supplied by Alcatel to Worldwide Consortium of 35 Carriers 40 Gbps by Mid-2003; Heavily Subscribed. Ultimate Capacity 120 Gbps Extension to Interior Mostly by Satellite: < 1 Mbps to ~100 Mbps typical Note: World Conference on Physics and Sustainable Development, 10/31 – 11/2/05 in Durban South Africa; Part of World Year of Physics 2005. Sponsors: UNESCO, ICTP, IUPAP, APS, SAIP AFRICA: Nectar Net Initiative Growing Need to connect academic researchers, medical researchers & practitioners to many sites in Africa Examples: CDC & NIH: Global AIDS Project, Dept. of Parasitic Diseases, Nat’l Library of Medicine (Ghana, Nigeria) Gates $ 50M HIV/AIDS Center in Botswana; Project Coord at Harvard Africa Monsoon AMMA Project, Dakar Site [cf. East US Hurricanes] US Geological Survey: Global Spatial Data Infrastructure Distance Learning: Emory-Ibadan (Nigeria); Research Channel Content But Africa is Hard: 11M Sq. Miles, 600 M People, 54 Countries Little Telecommunications Infrastructure Approach: Use SAT-3/WASC Cable (to Portugal), GEANT Across Europe, Amsterdam-NY Link Across the Atlantic, then Peer with R&E Networks such as Abilene in NYC Cable Landings in 8 West African Countries and South Africa W. Matthews Pragmatic approach to reach end points: VSAT satellite, ADSL, microwave, etc. Georgia Tech Bandwidth prices in Africa vary dramatically; are in general many times what they could be if universities purchase in volume Sample Bandwidth Costs for African Universities Nigeria $20.00 Average $11.03 Uganda $9.84 Ghana $6.77 IBAUD Target USA $3.00 $0.27 $0.00 $5.00 $10.00 $15.00 $20.00 $25.00 $/kbps/month Sample size of 26 universities Average Cost for VSAT service: Quality, CIR, Rx, Tx not distinguished Roy Steiner Internet2 Workshop Grid2003: An Operational Production Grid, Since October 2003 27 sites (U.S., Korea) 2300-2800 CPUs 700-1100 Concurrent Jobs www.ivdgl.org/grid2003 Trillium: PPDG GriPhyN iVDGL Korea Prelude to Open Science Grid: www.opensciencegrid.org HENP Data Grids, and Now Services-Oriented Grids The original Computational and Data Grid concepts are largely stateless, open systems Analogous to the Web The classical Grid architecture had a number of implicit assumptions The ability to locate and schedule suitable resources, within a tolerably short time (i.e. resource richness) Short transactions with relatively simple failure modes HENP Grids are Data Intensive & Resource-Constrained Resource usage governed by local and global policies Long transactions; some long queues Analysis: 1000s of users competing for resources at dozens of sites: complex scheduling; management HENP Stateful, End-to-end Monitored and Tracked Paradigm Adopted in OGSA [Now WS Resource Framework] Increased functionality, standardization The Move to OGSA and then Managed Integrated Systems ~Integrated Systems Web services + … X.509, LDAP, FTP, … Custom solutions App-specific Services Open Grid Web Services Services Arch Resrc Framwk Stateful; Managed GGF: OGSI, … (+ OASIS, W3C) Globus Toolkit Multiple implementations, including Globus Toolkit Defacto standards GGF: GridFTP, GSI Time Managing Global Systems: Dynamic Scalable Services Architecture MonALISA: http://monalisa.cacr.caltech.edu 24 X 7 Operations Multiple Orgs. Grid2003 US CMS CMS-DC04 ALICE STAR VRVS ABILENE Soon: GEANT + GLORIAD “Station Server” Services-engines at sites host many “Dynamic Services” Scales to thousands of service-Instances Servers autodiscover and interconnect dynamically to form a robust fabric Autonomous agents + CLARENS: Web Services Fabric and Portal Architecture Grid Analysis Environment CLARENS: Web Services Architecture Analysis Client Analysis Client Analysis Clients talk Analysis Client HTTP, SOAP, XML/RPC Grid Services Web Server Scheduler Catalogs FullyAbstract Planner Metadata PartiallyAbstract Planner FullyConcrete Planner Data Management Virtual Data Monitoring Replica Execution Priority Manager Grid Wide Execution Service Caltech GAE Team Applications standard protocols to the CLARENS “Grid Services Web Server”, with a simple Web service API The secure Clarens portal hides the complexity of the Grid Services from the client Key features: Global Scheduler, Catalogs, Monitoring, and Gridwide Execution service; Clarens servers form a Global Peer to peer Network 52 World Summit on the Information Society (WSIS): Geneva 12/2003 and Tunis in 2005 The UN General Assembly adopted in 2001 a resolution endorsing the organization of the World Summit on the Information Society (WSIS), under UN Secretary-General, Kofi Annan, with the ITU and host governments taking the lead role in its preparation. GOAL: To Create an Information Society: A Common Definition was adopted in the “Tokyo Declaration” of January 2003: “… One in which highly developed ICT networks, equitable and ubiquitous access to information, appropriate content in accessible formats and effective communication can help people achieve their potential” Kofi Annan Challenged the Scientific Community to Help (3/03) CERN and ICFA SCIC have been quite active in the WSIS in Geneva (12/2003) Role of Science in the Information Society; WSIS 2003-2005 HENP Active in WSIS CERN RSIS Event SIS Forum & CERN/Caltech Online Stand at WSIS I (Geneva 12/03) Visitors at WSIS I Kofi Annan, UN Sec’y General John H. Marburger, Science Adviser to US President Ion Iliescu, President of Romania; and Dan Nica, Minister of ICT Jean-Paul Hubert, Ambassador of Canada in Switzerland … Planning Underway for WSIS II: Tunis 2005 HEPGRID and Digital Divide Workshop UERJ, Rio de Janeiro, Feb. 16-20 2004 Theme: Global Collaborations, Grids and Their Relationship to the Digital Divide NEWS: Bulletin: ONE TWO WELCOME BULLETIN General Information Registration Travel Information Hotel Registration Participant List How toTutorials Get UERJ/Hotel C++ Accounts Computer GridPhone Technologies Useful Numbers Program Grid-Enabled Contact us: Analysis Secretariat Networks Chairmen Collaborative ICFA, understanding the vital role of these issues for our field’s future, commissioned the Standing Committee on Inter-regional Connectivity (SCIC) in 1998, to survey and monitor the state of the networks used by our field, and identify problems. For the past three years the SCIC has focused on understanding and seeking the means of reducing or eliminating the Digital Divide, and proposed to ICFA that these issues, as they affect our field of High Energy Physics, be brought to our community for discussion. This led to ICFA’s approval, in July 2003, of the Digital Divide and HEP Grid Workshop. More Information: http://www.lishep.uerj.br SPONSORS Systems CLAF CNPQ FAPERJ UERJ Sessions & Tutorials Available (w/Video) on the Web International ICFA Workshop on HEP Networking, Grids and Digital Divide Issues for Global e-Science Proposed Workshop Dates: May 23-27, 2005 Venue: Daegu, Korea Dongchul Son Center for High Energy Physics Kyungpook National University ICFA, Beijing, China Aug. 2004 ICFA Approval Requested Today International ICFA Workshop on HEP Networking, Grids and Digital Divide Issues for Global e-Science Themes Networking, Grids, and Their Relationship to the Digital Divide for HEP as Global e-Science Focus on Key Issues of Inter-regional Connectivity Mission Statement ICFA, understanding the vital role of these issues for our field’s future, commissioned the Standing Committee on Inter-regional Connectivity (SCIC) in 1998, to survey and monitor the state of the networks used by our field, and identify problems. For the past three years the SCIC has focused on understanding and seeking the means of reducing or eliminating the Digital Divide, and proposed to ICFA that these issues, as they affect our field of High Energy Physics, be brought to our community for discussion. This workshop, the second in the series begun with the the 2004 Digital Divide and HEP Grid Workshop in Rio de Janeiro (approved by ICFA in July 2003) will carry forward this work while strengthening the involvement of scientists, technologists and governments in the Asia Pacific region. 고에너지물리연구센터 CENTER FOR HIGH ENERGY PHYSICS International ICFA Workshop on HEP Networking, Grids and Digital Divide Issues for Global e-Science Workshop Goals Review the current status, progress and barriers to the effective use of the major national, continental and transoceanic networks used by HEP Review progress, strengthen opportunities for collaboration, and explore the means to deal with key issues in Grid computing and Grid-enabled data analysis, for high energy physics and other fields of data intensive science, now and in the future Exchange information and ideas, and formulate plans to develop solutions to specific problems related to the Digital Divide in various regions, with a focus on Asia Pacific, Latin America, Russia and Africa Continue to advance a broad program of work on reducing or eliminating the Digital Divide, and ensuring global collaboration, as related to all of the above aspects. 고에너지물리연구센터 CENTER FOR HIGH ENERGY PHYSICS Networks and Grids, GLORIAD, ITER and HENP Network backbones and major links used by major experiments in HENP and other fields are advancing rapidly To the 10 G range in < 2 years; much faster than Moore’s Law New HENP and DOE Roadmaps: a factor ~1000 improvement per decade We are learning to use long distance 10 Gbps networks effectively 2003-2004 Developments: to 7.5 Gbps flows over 16 kkm Important advances in Asia-Pacific, notably Korea A transition to community-owned and operated R&E networks is beginning (us, ca, nl, pl, cz, sk …) or considered (de, ro, …) We Must Work to Close to Digital Divide Allowing Scientists and Students from All World Regions to Take Part in Discoveries at the Frontiers of Science Removing Regional, Last Mile, Local Bottlenecks and Compromises in Network Quality are now On the Critical Path GLORIAD is A Key Project to Achieve These Goals Synergies Between the Data-Intensive Missions of HENP & ITER Enhancing Partnership and Community among the US, Russia and China: both in Science and Education SC2004: HEP Network Layout Preview of Future Grid systems SLAC Australia Japan Brazil StarLight 2*10 Gbps NLR 10 Gbps NLR 2 Metro 10 Gbps Waves LA-Caltech Caltech CACR 3*10Gbps TeraGrid 10 Gbps Abilene FNAL LA UK 10 Gbps LHCNet CERN Geneva Joint Caltech, CERN, SLAC, FNAL, UKlight, HP, Cisco… Demo 6 to 8 10 Gbps waves to HEP setup on the show floor Bandwidth challenge: aggregate throughput goal of 40 to 60 Gbps SCIC in 2003-2004 http://cern.ch/icfa-scic Strong Focus on the Digital Divide Continues A Striking Picture Continues to Emerge: Remarkable Progress in Some Regions, and a Deepening Digital Divide Among Nations Intensive Work in the Field: > 60 Meetings and Workshops: E.g., Internet2, TERENA, AMPATH, APAN, CHEP2003, SC2003, Trieste, Telecom World 2003, SC2003, WSIS/RSIS, GLORIAD Launch, Digital Divide and HEPGrid Workshop (Feb. 16-20 in Rio), GNEW2004, GridNets2004, NASA ONT Workshop, … etc. 3rd Int’l Grid Workshop in Daegu (August 26-28, 2004); Plan for 2nd ICFA Digital Divide and Grid Workshop in Daegu (May 2005) HENP increasingly visible to governments; heads of state: Through Network advances (records), Grid developments, Work on the Digital Divide and issues of Global Collaboration Also through the World Summit on the Information Society Process. Next Step is WSIS II in TUNIS November 2005 Coverage Now monitoring 650 sites in 115 countries In last 9 months added: Several sites in Russia (thanks GLORIAD) Many hosts in Africa (5 36 now; in 27 out of 54 countries) Monitoring sites in Pakistan and Brazil (Sao Paolo and Rio) Working to install monitoring host in Bangalore, India Monitoring site Remote site Latin America: CLARA Network (2004-2006 EU Project) Significant contribution from European Comission and Dante through ALICE project NRENs in 18 LA countries forming a regional network for collaboration traffic Initial backbone ring bandwidth f 155 Mbps Spur links at 10 to 45 Mbps (Cuba at 4 Mbps by satellite) Initial connection to Europe at 622 Mbps from Brazil Tijuana (Mexico) PoP soon to be connected to US through dark fibre link (CUDI-CENIC) access to US, Canada and Asia - Pacific Rim NSF IRNC 2004: Two Proposals to Connect CLARA to the US (and Europe) 1st Proposal: FIU and CENIC To West Coast 2nd Proposal: Indiana and Internet2 To East to West Coast Coast to East Coast Note: CHEPREO (FIU, UF, FSU Caltech, UERJ, USP, RNP) 622 Mbps Sao Paolo – Miami Link Started in August GIGA Project: Experimental Gbps Network: Sites in Rio and Sao Paolo Universities IME PUC-Rio UERJ UFF UFRJ Unesp Unicamp USP R&D Centres CBPF - physics CPqD - telecom CPTEC - meteorology CTA - aerospace Fiocruz - health IMPA - mathematics INPE - space sciences LNCC - HPC LNLS - physics Slide from M. Stanton About 600 km extension - not to scale LNCC CTA INPE CPqD LNLS Unicam p CPTEC Fapesp telcos Unesp USP – Incor USP - C.Univ. CBPF LNCC Fiocruz IME IMPARNP PUC-Rio telcos UERJ UFRJ UFF Trans-Eurasia Information Network TEIN (2004-2007) Circuit between KOREN(Korea) and RENATER(France). AP Beneficiaries: China, Indonesia, Malaysia, Philippines, Thailand, Vietnam (Non-beneficiaries: Brunei, Japan, Korea, Singapore EU partners: NRENs of France, Netherlands, UK The scope expanded to South-East Asia and China recently. Upgraded to 34 Mbps in 11/2003. Upgrade to 155Mbps planned 12M Euro EU Funds Coordinating Partner DANTE Direct EU-AP Link; Other Links go Across the US APAN China Consortium Has been established in 1999. The China Education and Research Network (CERNET), the Natural Science Foundation of China Network (NSFCNET) and the China Science and Technology Network (CSTNET) are the main three advanced networks. CERNet NSFCnet 2.5 Gbps Tsinghua --- Tsinghua University PKU --- Peking University NSFC --- Natural Science Foundation of China CAS --- China Academy of Sciences BUPT --- Beijing Univ. of Posts and Telecom. BUAA --- Beijing Univ. of Aero- and Astronautics GLORIAD: Global Optical Ring (US-Ru-Cn; Korea Now Full Partner ) DOE: ITER Distributed Ops.; Fusion-HEP Cooperation NSF: Collaboration of Three Major R&E Communities Also Important for Intra-Russia Connectivity; Education and Outreach Aug. 8 2004: P.K. Young, Korean IST Advisor to President Announces Korea Joining GLORIAD TEIN gradually to 10G, connected to GLORIAD Asia Pacific Info. InfraStructure (1G) will be backup net to GLORIAD GLORIAD and HENP Example: Network Needs of IHEP Beijing ICFA SCIC Report: Appendix 18, on Network Needs for HEP in China (See http://cern.ch/icfa-scic) “IHEP is working with the Computer Network Information Center (CNIC) and other universities and institutes to build Grid applications for the experiments. The computing resources and storage management systems are being built or upgraded in the Institute. IHEP has a 100 Mbps link to CNIC, so it is quite easy to connect to GLORIAD and the link could be upgraded as needed.” Prospective Network Needs for IHEP Bejing Experiment LHC/LCG BES YICRO AMS Others Total (Sharing) Year 2004-2005 622Mbps 100Mbps 100Mbps 100Mbps 100Mbps 1Gbps Year 2006 and on 2.5Gbps 155Mbps 100Mbps 100Mbps 100Mbps 2.5Gbps Role of Sciences in Information Society. Palexpo, Geneva 12/2003 Demos at the CERN/Caltech RSIS Online Stand Advanced network and Grid-enabled analysis Monitoring very large scale Grid farms with MonALISA World Scale multisite multi-protocol videoconference with VRVS (Europe-US-Asia-South America) Distance diagnosis and surgery using Robots with “haptic” feedback (Geneva-Canada) Music Grid: live performances with bands at St. John’s, Canada and the Music Conservatory of Geneva on stage VRVS 37k hosts 106 Countries 2-3X Growth/Year Achieving throughput User can’t achieve throughput available (Wizard gap) TCP Stack, End-System and/or Local, Regional, Nat’l Network Issues Big step just to know what is achievable (e.g. 7.5 Gbps over 16 kkm Caltech-CERN)