What w e have learned from developing and running ABw E Jiri - - PowerPoint PPT Presentation

what w e have learned from developing and running abw e
SMART_READER_LITE
LIVE PREVIEW

What w e have learned from developing and running ABw E Jiri - - PowerPoint PPT Presentation

What w e have learned from developing and running ABw E Jiri Navratil, Les R.Cottrell (SLAC) Why E2E tools are needed The scientific community is increasingly dependent on networking as international cooperation grows. HEP users (needs


slide-1
SLIDE 1

What w e have learned from developing and running ABw E

Jiri Navratil, Les R.Cottrell (SLAC)

slide-2
SLIDE 2

Why E2E tools are needed

  • The scientific community is increasingly

dependent on networking as international cooperation grows. HEP users (needs transfer huge

amount of data between experimental sites as SLAC, FNAL, CERN, etc. (where data is created) and home institutes spread

  • ver the world)
  • What ISPs (as Abilene,Esnet,Geant..) can offer to

the users for getting information?

(Not too much because they are only in the middle of the path and they don’t cover all parts of connections)

slide-3
SLIDE 3

FZU LAN FZU LAN RAL LAN RAL LAN DL LAN DL LAN

CESNET CESNET JANET JANET

IN2P3 LAN IN2P3 LAN

CERN LAN CERN LAN

RENATER RENATER INFN INFN

FNAL-LAN FNAL-LAN

GEANT GEANT ABILENE ABILENE ESNET ESNET

SLAC LAN SLAC LAN

MichNET MichNET

NERSC- LAN NERSC- LAN

CALREN CALREN

Data sources Users

MIB LAN MIB LAN

slide-4
SLIDE 4

FZU LAN FZU LAN RAL LAN RAL LAN DL LAN DL LAN

CESNET CESNET JANET JANET

IN2P3 LAN IN2P3 LAN

CERN LAN CERN LAN

RENATER RENATER INFN INFN

FNAL-LAN FNAL-LAN

GEANT GEANT ABILENE ABILENE ESNET ESNET

SLAC LAN SLAC LAN

MichNET MichNET

NERSC- LAN NERSC- LAN

CALREN CALREN Data sources Users

MIB LAN MIB LAN

slide-5
SLIDE 5
  • There must be always somebody who gives complex

information to the users of the community

  • r

the users have to have a tool which give them such

information

  • How fast I can transfer 20 GB from my experimental site (SLAC,CERN) to

my home institute?

  • Can I run graphical 3D visualization program with data located 1000 miles

away?

  • How stable is line ? (Can I use it in the same conditions for 5 minutes or 2 hours or

whole day ?)

All such questions must be replied in few seconds doesn’t matter if for individual user or for Grid brokers

  • Global science has no day and night.

To reply this we needed the tools that could be used in continuous mode 24 hours a day 7 days a week which can non intrusively detect changes on multiple path or on demand by any user

slide-6
SLIDE 6

ABwE:Basic terminology:

  • Generally:

Available bandwidth = Capacity – Load

  • ABwE measure Td – Time dispersion P1-P2 (20x PP)

We are trying to distinguish two basic states in our results:

  • “Dominate (free)” – when Td ~= const
  • “loaded” with Td = other value

Td results from “Dominate” state are used to estimate

DBC - Dynamic Bottleneck Capacity

Td measured during the “loaded” state is used to estimate the level of

XTR (cross traffic)

ABw = DBC – XTR

f Td

slide-7
SLIDE 7

Dbc= Lpp/Td domin ”Dominating state” ”Dominating state”

(when sustained load or no load)

u = q/(q+1) CT=u*Dbc Abw= Dbc -CT

Abing: Estimation principles:

Td Tp (pairs)

q = Tx/Tn

(Tx=Td –Tp)

Tx – busy time (transmit time for cross trafic) Tn – transmit time for average packet q – relative queue increment (QDF) during decision interval Td (h-1)

Tn

Tx (cross traffic)

Td domin

Td i = Td i+1 = .. Td i+n

“Load state” “Load state”

(when load is changing) Td

Examples Td from different paths

f f Td

slide-8
SLIDE 8

What is DBC DBC

  • DBC

DBC characterize instant high capacity bottleneck that DOMINATE on the path

  • It covers situations when routers in the

path are overloaded and sending packets back to back with its maximal rates

  • We discovered that in most cases only
  • ne node dominates in the instant of our

measurements (in our decision interval)

slide-9
SLIDE 9

load

load 1000 622 622 622 622 1000 100 622 622

Empty pipes

No impact (in t1)

Light source Light beam

DBC DBC

No impact (in t1)

ABw E: Example of narrow link in the path narrow link in the path

ABW ABW

link that has domination effect

  • n bandwidth

DBC

(Pipes analogy w ith different diameter and aperture)

Ab Abw = DB DBC C – XTR XTR

ABW monitor SLAC to UFL ABW monitor SLAC to UFL

slide-10
SLIDE 10

load

load 1000 622 622 622 622 1000 415 622 622

Empty links (pipes)

No impact (in t1)

strong XTraffic -> Impact (in t1)

Light source Light beam

DBC DBC Example of heavy loaded link in the path heavy loaded link in the path

(Pipes analogy w ith different diameter and aperture)

Heavy load (strong cross traffic) appeared in the path It shows new DBC in the path because this load dominates in whole path ! Normal situation DBC~ 400 Mbits/s

Available bandwidth Abilene MRTG graph ATLA to UFL

Abw Abw = DBC DBC – XT XTR

ABW monitor SLAC to UFL ABW monitor SLAC to UFL

strong XTR (cross traffic)

slide-11
SLIDE 11

Heavy load (xtraffic) appeared in the path (defined new DBC in the path)

Normal situation

ABw E ABw E / MRTG match: / MRTG match: TCP test to TCP test to UFL UFL

IPLS shows traffic 800-900 Mbits/s CALREN shows sending traffic 600 Mbits/s

UFL UFL

slide-12
SLIDE 12

Confront ABw E results w ith other tools

Iperf,Pathload,Pathchirp

slide-13
SLIDE 13

Probe Probe Sender Sender

XT gen.

Pr Probe

  • be

Recei eceiver ver

XT rec.

DataTag SLAC

1 rtr-gsr-test 0.169 ms 0.176 ms 0.121 ms 2 rtr-dmz1-ger 0.318 ms 0.321 ms 0.340 ms 3 slac-rt4.es.net 0.339 ms 0.325 ms 0.345 ms 4 snv-pos-slac.es.net 0.685 ms 0.687 ms 0.693 ms 5 chicr1-oc192-snvcr1.es.net 48.777 ms 48.758 ms 48.766 ms 6 chirt1-ge0-chicr1.es.net 48.878 ms 48.778 ms 48.774 ms 7 chi-esnet.abilene.iu.edu 58.864 ms 58.851 ms 59.002 ms 8 r04chi-v-187.caltech.datatag.org 59.045 ms 59.060 ms 59.041 ms

ES.net path ES.net path (622 Mbits/s) (622 Mbits/s)

Chicago, Il Chicago, Il Menlo Park, Ca

To CERN (Ch)

Probing packets Injected Cross traffic Experimental path ES.net

NIC-1000Mbps NIC-1000Mbps NIC-1000Mbps NIC-1000Mbps

User traffic

User traffic (background)

SLAC-DataTAG-CERN test environment

(4 workstations with NIC1000Mbis/s + OC-12 ES.net path)

GbE GbE GbE GbE 2.5 Gbits/s

ES.net

User traffic

slide-14
SLIDE 14

Zoom

Level of background traffic

Injected CT (cross traffic by Iperf) Measured xt ( cross-traffic)

DBC (OC-12 )

The match of the cross traffic

(ABW – XT compare to injection traffic generated by Iperf) Available bandwidth Conlusion: Iperf measure own performance which can approach DBC (in best case)

slide-15
SLIDE 15
slide-16
SLIDE 16
slide-17
SLIDE 17

What w e learned from CAIDA testbed

slide-18
SLIDE 18

1 1 2 1 2 2

CT1 CT3

Packet Length ~ MTU

  • 1. Packet Pair
  • 2. Packet Pair

25 ms

Internet H Internet HOP/HOPS vers. Testbed P/HOPS vers. Testbed

CT2

TBedCT

I-HOP TBED

PP Internet cross traffic

  • Simul. cross traffic

PP Initial decision interval Decision interval (12 µs for Oc12) Cross traffic sources Probes I n t e r n e t P a t h Decision interval is changing (growing) If CT < 30% abw had detection problem !

.. 20 x

cause a dispersion Relevant packets

Not relevant packets

N

  • t

r e l e v a n t p a c k e t s

slide-19
SLIDE 19

1 1 1 1 2 2 6 1 2 2 4 3 2 5

CT CT CT

Packet Length ~ MTU

  • 1. Packet Pair
  • 2. Packet Pair

25 ms

How to improve “detection effectiveness” How to improve “detection effectiveness”

cause a dispersion Solution LP

Solution LP – Long packets (9k) (creates micro-bottlenecks) Solution nP – n dummy Packets (mini-train)

Solution nP

New initial decision interval

Relevant packets decision interval

.. 20 x .. 100 x Measurement time 0.5 s to 2.5 s

Solution X

slide-20
SLIDE 20

S2 (PP-Packet Pair) S10 (Mini-train with 8 dummy packets)

PP versus PP versus TRAIN: TRAIN: ABW and DBC merge

merge in TRAIN samples

(SLAC-CALTECH path)

slide-21
SLIDE 21

s2 s3 s4 s5 s7 s10

PP versus PP versus TRAIN: TRAIN: ABW and DBC merge

merge in TRAIN samples

(SLAC-CALTECH path)

slide-22
SLIDE 22

Compare long term Bandw idth statistics

  • n real paths

ESNET, Abilene, Europe

slide-23
SLIDE 23

SLAC - Rice.edu SLAC - Man.ac.uk SLAC - Mib.infn.it SLAC - ANL.gov

IEPM-Iperf vers. ABW (24 hours match)

IEPM (achievable throughput via Iperf) (red bars) IEPM (achievable throughput via Iperf) (red bars) ABW: Available bandwidth (blue lines) ABW: Available bandwidth (blue lines)

slide-24
SLIDE 24

Scatter plot graphs Achievable throughput via Iperf versus ABw

  • n different paths (range 20–800 Mbits/s)

(28 days history)

slide-25
SLIDE 25

ABw data Iperf data

28 days bandw idth history

During this time w e can see several different situations caused by different routing from SLAC to CALTECH

to 100 Mbits/s by error

Drop to 622 Mbits/s path back to new CENIC path

New CENIC path 1000 Mbits/s In all cases the match of results from Iperf and ABw is evident

slide-26
SLIDE 26

What we can detect with continues bandwidth monitoring

  • Immediate bandwidth on the path
  • Automatic routing changes when line is

broken (move to backup lines)

  • Unexpected Network changes (Routing

changes between networks, etc.)

  • Line updates (155 -> 1Giga, etc.)
  • Extreme heavy load
slide-27
SLIDE 27

Via Abilene Original path via CALREN/CENIC

(Example from SLAC – CENIC path) Problematic link discovered

Bandwidth problem discovered (14:00) BW problem resolved (17:00) Routing back on standard path

Results of traceroute analysis

Standard routing via CALREN/CENIC Available bandwidth Send alarm

ABw ABw as s Troubleshooting tool Troubleshooting tool

( Discovering Routing problems and initiate alarming ) ( Discovering Routing problems and initiate alarming )

DBC

User traffic

slide-28
SLIDE 28

SLAC SLAC – CENIC path upgrade from 1 to 10 Gigabit CENIC path upgrade from 1 to 10 Gigabit

(Current monitoring machines allow monitor traffic in range 1 < 1000 Mbits only) To backup Router (degrading line for while) Skip to new 10GBits/s link (our monitor is on 1GbE)

slide-29
SLIDE 29

Upgrade 155Mbits/s line to 1000Mbits/s at dl.uk

slide-30
SLIDE 30

via Abilene via ESNET

SLAC changed routing to CESNET

slide-31
SLIDE 31

Situation when the cross-traffic extreamly grows, BW decreased

SNVA-STTL (line broken) STTL-DNVR DNVR-STTL

Abilene – automatic rerouting – June 11,2003

Sending traffic from south branch receiving

slide-32
SLIDE 32

Transatlantic line to CERN (green=input) SLAC-ESNET (red output)

Seen at Chicago Seen at SLAC Seen at CERN

User traffic (bbftp to IN2p3.fr)

Additional traffic Iperf Seen by ABW at CERN Fig.12 Fig.12

Typical SLAC traffic (long data transfer when physical experiment ends)

MRTG shows only the traffic which pass to IN2p3.fr

Additional trafficIperf to Chicago seen also at CERN (common path)

slide-33
SLIDE 33
  • Interactive ( reply < 1 second)
  • Very low impact on the netw ork traffic (40

packets to get value for destination)

  • Simple and robust (responder can be installed
  • n any machine on the netw ork)
  • Keyw ord function for protecting the client-

server communication

  • Measurements in both directions
  • Same resolution as other similar methods

http://www-iepm.slac.stanford.edu/tools/abing

Abing new ABwE tool

slide-34
SLIDE 34

Thank you

References: http://moat.nlanr.net/PAM2003/PAM2003papers/3781.pdf http://w w w -iepm.slac.stanford.edu/tools/abing