Download CA*net II - Internet2

Survey
yes no Was this document useful for you?
   Thank you for your participation!

* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project

Document related concepts

Types of artificial neural networks wikipedia , lookup

Cracking of wireless networks wikipedia , lookup

Transcript
http://www.canarie.ca
CA*net 4
International Grid Testbed
[email protected]
Tel: +1.613.785.0426
Problem
1. TCP throughput over long fat pipes very susceptible to packet loss, MTU, TCP kernel,
Buffer memory, trail drop, AQM optimized for commodity Internet, etc
2. Packet loss can result from congestion, but also underlying BER
– achieve a gigabit per second with TCP on a coast-to-coast path (rtt = 40 msec),
with 1500 byte packets, the loss rate can not exceed 8.5x10^-8 packets
– “End to end” BER for optical networks 10^-12 to 10^-15 which means packet loss
rate of approximately 10^-8 to 10^-11
– The bigger the packet the greater the loss rate!!!
3. Cost of routers significantly greater than switches for 10 Gbps and higher (particularly for
large number of lambdas)
4. Lots of challenges maintaining consistent router performance across multiple
independent managed networks
– MTU, auto-negotiating Ethernet, insufficient buffer memory
5. Require consistent and similar throughput for multiple sites to maintain coherency for
grids and SANs and new “space” storage networks using erasure codes e.g. Oceanstore
6. For maximum throughput OS and kernel bypass may be required
7. Many commercial SAN/Grid products will only work with QoS network
Possible Solutions
1.
2.
3.
For point to point large file transfer a number of possible techniques such as
FAST, XCP, parallel TCP, UDP, etc
–
Very scalable and allows same process to be used for all sorts of file
transfer from large to small
–
But will it address other link problems?
Datagram QoS is a possibility to guarantee bandwidth
–
But requires costly routers and no proven approach across
independent managed networks (or campus)
–
Does not solve problem of MTU,link problems, etc
E2E lightpaths - all solutions are possible
–
Allows new TCP and non TCP file transfers
–
Allows parallel TCP with consistent skew on data striping
–
Allows protocols that support OS bypass, etc
–
Guarantees consistent throughput for distributed coherence and
enables news concepts of storing large data sets in “space”
–
Uses much lower cost switches and bypasses routers
What are E2E lightpaths?
>
>
>
>
Customer controlled E2E lightpaths are not about optical networking
– E2E lightpaths do not use GMPLS or ASON
The power of the Internet was that an overlay packet network controlled by end
user and ISPs could be built on top of telco switched network
– CA*net 4 is an optical overlay network on top of telco optical network
where switching will be controlled by end users
More akin to MAE-E “peermaker” but at a finer granularity
– “Do you have an e2e lightpath for file transfer terminating at a given IX?
Are you interested in peering with my e2e lightpath to enable big file
transfer?”
– Lightpath may be only from border router to border router
With OBGP can establish new BGP path that bypasses most (if not all) routers
– Allows lower cost remote peering and transit
– Allows e2e lightpaths for big file transfer
e2e Lightpaths
Of elephants and mice
Small mice traffic is
routed over normal IP
path
Normal IP/BGP path
Only y.y.y.1
advertised to
x.x.x.1 via
OBGP path
Only x.x.x.1
advertised to
y.y.y.1 via
OBGP path
Optical “Peermaker”
x.x.x.1
OBGP path
y.y.y.1
Application or end user
controls peering of BGP
optical paths to set up
dedicated route for
transfer of elephants
CA*net 4
Edmonton
Saskatoon
Calgary
Winnipeg
Vancouver
Halifax
Regina
St. John's
Victoria
Montreal
Ottawa
Seattle
Minneapolis
CA*net 4 Node
Charlottetown
Fredericton
Toronto
Boston
Existing CA*net 4 OC192
TransLight OC192
Chicago
New York
Halifax
Canada sets land speed record
Vancouver <-> Geneva
www.iGrid2002.org
for more info on iGrid2002
SAN land speed record
VANCOUVER
OTTAWA
8 x GE @ OC-12 (622Mb/s)
CHICAGO
1 x GE
loop-back
on OC-24
Sustained Throughput ~11.1 Gbps Ave. Utilization = 93%
ATLAS/CMS: Data Grid Hierarchy
Low level Trigger data
~PByte/sec
Online System ~100-1500
MBytes/sec
Experiment
Tier 0 +1
~2.5 Gbps
Tier 1
IN2P3 Center
Tier 3
INFN Center
RAL Center
~2.5 Gbps
Tier 2
InstituteInstitute Institute
~0.25TIPS
Physics data cache
Workstations
CERN 700k SI95
~1 PB Disk;
Tape Robot
Institute
0.1–10 Gbps
Tier 4
FNAL: 200k
SI95; 600 TB
2.5 Gbps
Tier2 Center
Tier2 Center
Tier2 Center
Tier2 Center
Tier2 Center
International Grid Testbed
> Joint CERN, SURFnet, STAR LIGHT, TransLight project
> Objectives:
1. To validate and test software for customer control and routing of lightpaths
2. Test remote processing of of low level trigger data from the ATLAS test
beam.
3. Develop and adapt grid applications which are designed to interact with a
LightPath Grid Service which treats networks and network elements as grid
resources which can be reserved, concatenated, consumed and released.
4. Characterize the performance of bulk data transfer over an end-to end
lightpath.
5. To investigate and test emerging technologies and its impact on high speed
long distance optical networks. These technologies include 10 Gbit Ethernet,
RDMA/IP, Fibre Channel/IP, serial SCSI, HyperSCSI over long distance
ethernet, etc.
6. Collaborate with the EU ESTA project which is developing 10 GbE equipment
with CERN, industrial and other academic partners.