Two days in The Life of The DNS Anycast Root Servers Ziqian Liu - - PowerPoint PPT Presentation

two days in the life of the dns anycast root servers
SMART_READER_LITE
LIVE PREVIEW

Two days in The Life of The DNS Anycast Root Servers Ziqian Liu - - PowerPoint PPT Presentation

Two days in The Life of The DNS Anycast Root Servers Ziqian Liu Beijing Jiaotong Univeristy Bradley Huffaker, Marina Fomenkov Nevil Brownlee, and kc claffy CAIDA PAM2007 Outline DNS root servers DNS anycast in root servers Data


slide-1
SLIDE 1

Two days in The Life of The DNS Anycast Root Servers

Ziqian Liu Beijing Jiaotong Univeristy Bradley Huffaker, Marina Fomenkov Nevil Brownlee, and kc claffy CAIDA

PAM2007

slide-2
SLIDE 2

Outline

  • DNS root servers
  • DNS anycast in root servers
  • Data
  • Traffic difference in anycast instances
  • Anycast coverage
  • Conclusion
slide-3
SLIDE 3

DNS Root Servers

  • Tree-structured distributed database

gov www edu com cn us google sun maps ucsd

  • rg

caida acm

Root Level Top Level Domain (TLD) Second Level Domain (SLD)

staff cse

generic TLD country code TLD

  • Root servers provide authoritative referrals to name

servers for gTDL and ccTLD domains.

  • Only 13 root servers world wide

[A-M].root-servers.net

slide-4
SLIDE 4

DNS Anycast in Root Servers

  • What is anycast

– Anycast group

  • A set of instances that are run by the same organization and

use the same IP address – the service address – but are physically different nodes.

  • e.g., k.root-servers.net – RIPE – 17 instances – 194.0.14.129

– For a DNS root servers, anycast provides a service where by clients send requests to the service address and the network delivers that request to at least one, preferably the closest, instance in the root server’s anycast group.

slide-5
SLIDE 5

DNS Anycast in Root Servers (2)

  • How to deploy

– Every instance in the anycast group announces reachability for the same prefix – service supernet – that covers the service address by BGP.

  • e.g k.root-servers.net

193.0.14.0/24

  • So, multiple AS paths are advertising the same prefix.

– Different BGP routing policies:

  • Local instance

– Limit the catchment area by using no-export community

  • Global instance

– Globally visible – Use AS-prepending to decrease the likelihood of their selection

  • ver a local instance
slide-6
SLIDE 6

DNS Anycast in Root Servers (3)

  • C, F, I, J, K and M roots (6/13) have deployed anycast.

– Over 120 root instances all togther (www.root-servers.org)

  • What’s the benefit?

– Allow the system to grow beyond the static 13 servers while avoiding a change to the existing protocol – Bring DNS service closer to the clients – Provide relatively reliable and stable service compared to a non- distributed structure.

  • Separate the server/network failure
  • Mitigate the impact of malicious traffic

– Oct 2002 DDoS attack against 13 root servers – Feb 2007 DDoS attack against root servers and gTLD servers

slide-7
SLIDE 7

http://dnsmon.ripe.net/ L G M F

slide-8
SLIDE 8

Data

  • ISC/ OARC (DNS Operations and Analysis Research Center)

/CAIDA have been conducting measurement at the DNS root servers

  • Three anycast root servers:

– C-root: 4 of 4 instances – F-root: 33 of 37 instances (40 up-to-date) – K-root: 16 of 17 instances

  • Time

– Tue Jan 10~Wed Jan 11, 2006, UTC, – 47.2 hour long (~2 days)

  • Data format

– Full record tcpdump traces

  • Our Focus: IPv4 UDP DNS requests
  • Available at http://imdc.datcat.org
slide-9
SLIDE 9

Traffic difference – Diurnal pattern

Tokyo noon back

slide-10
SLIDE 10

Traffic difference – Traffic load

10 20 30 40 50 60 2 4 6 8 10 x 10

5

Instance # IP addresses C-root F-root K-root jfk1* lax1*

  • rd1*

iad1* sfo2* pao1* linx* delhi* tokyo* denic nap* ams-ix* muc1 ams1 gru1 lga1 10 20 30 40 50 60 500 1000 1500 2000 2500 Instance # requests/sec) C-root F-root K-root iad1*

  • rd1*

lax1* jfk1* pao1* sfo2* linx* nap* denic tokyo* delhi* ams-ix* muc1 ams1 gru1 lga1

(a) Average request rate (b) Number of clients

  • Both plots have the same x-axis intance order – instances within each group

are arranged in an increasing request rate order

  • Global instances are marked with *

Note

slide-11
SLIDE 11

General statistics – Clients vs. Requests

10 10

2

10

4

10

6

10

8

10 10

2

10

4

10

6

# requests # clients

During the 2 days:

  • 80% of the 2.5M clients sent <100

reqs to the three roots

  • 15 clients sent > 10M reqs
  • Top client sent > 30M reqs

i.e. ~ 174reqs/sec !

slide-12
SLIDE 12

Anycast coverage

  • Geographic

– Client location: map client IP address to Geo info by using NetAcuity database – Instance location: coordinates of the closest airport – Distance: great circle distance

  • Topological:

– Route Views BGP table on Jan 10, 2006 for ASN and prefix

slide-13
SLIDE 13

Anycast coverage – geographic distr.

  • Client Continental distribution

25 50 75 100 % N.Amer S.Amer Europe Africa Asia Oceania Oceania Asia Africa Europe

  • S. Amer
  • N. Amer

F: San Francisco (US) F: New York (US) F: Sao Paulo (BR) F: Santiago (CL) K: Reykjavik (IS) K: Helsinki (FI) F:Johannesburg(SA) F: Tel Aviv (IL) K: Tokyo (JP) F: Brisbane (AU) F: Auckland (NZ) sfo2 pao1 delhi ams-ix jfk1 lax1

  • rd1

iad1 nap linx tokyo F: C: K:

slide-14
SLIDE 14

Anycast coverage – geographic distr.

  • DNS request continental distribution

25 50 75 100 % N.Amer S.Amer Europe Africa Asia Oceania Oceania Asia Africa Europe

  • S. Amer
  • N. Amer

F: San Francisco (US) F: New York (US) F: Santiago (CL) F: Sao Paulo (BR) K: Reykjavik (IS) K: Helsinki (FI) F:Johannesburg(SA) F: Tel Aviv (IL) K: Tokyo (JP) F: Brisbane (AU) F: Auckland (NZ) sfo2 pao1 lax1

  • rd1

iad1 jfk1 nap linx ams-ix delhi tokyo F: K: C:

slide-15
SLIDE 15

Anycast coverage – geographic distr.(2)

  • Distance distribution (instance client)

C-root

slide-16
SLIDE 16

Anycast coverage – geographic distr.(3)

  • Distance distribution (instance client)

F-root

slide-17
SLIDE 17

Anycast coverage – geographic distr.(4)

  • Distance distribution (instance client)

K-root

slide-18
SLIDE 18

Anycast coverage – geographic distr.(5)

  • Additional distance

= distance from the client to the instance it requests –

distance from the client to the closest instance

slide-19
SLIDE 19

Anycast coverage – topological coverage

Topological scope: observed 19,237 ASes, RouteViews 21,883 ASes (~88%)

10 20 30 40 50 60 10 20 30 40 50 60 70 Instance % C-root F-root K-root iad1* lax*

  • rd*

jfk* pao1* sfo2* linx* tokyo* denic delhi* nap* ams-ix* ams1 muc1 lga1 10 20 30 40 50 60 10 20 30 40 50 60 70 Instance % C-root F-root K-root iad1* lax1*

  • rd1*

jfk1* pao1* sfo2* linx* tokyo* denic delhi* nap* ams-ix* ams1 muc1 lga1

% = # seen by instance / # seen by all AS level Prefix level Both plots have the same x-axis intance order – instances within each group are arranged in an increasing AS coverage percentage order

slide-20
SLIDE 20

Anycast coverage – topological coverage (2)

  • denic – K-root local instance in Frankfurt, Germany

– AS paths observed from RouteViews: 193.0.14.0/24 “3292 8763 25152 i” “4513 8763 25152 i” “12956 8763 25152 i” – AS12956 belongs to Telefonica which has a large-scale coverage

25 50 75 100 % N.Amer S.Amer Europe Africa Asia Oceania Oceania Asia Africa Europe

  • S. Amer
  • N. Amer

F: San Francisco (US) F: New York (US) F: Sao Paulo (BR) F: Santiago (CL) K: Reykjavik (IS) K: Helsinki (FI) F:Johannesburg(SA) F: Tel Aviv (IL) K: Tokyo (JP) F: Brisbane (AU) F: Auckland (NZ) sfo2 pao1 delhi ams-ix jfk1 lax1

  • rd1

iad1 nap linx tokyo F: C: K:

denic

slide-21
SLIDE 21

Anycast coverage – topological coverage (3)

  • tokyo – K-root global instance in Tokyo, Japan

– AS paths observed from RouteViews: 193.0.14.0/24 “4713 25152 25152 25152 25152 i” – The longest among all five K-root global instances !

25 50 75 100 % N.Amer S.Amer Europe Africa Asia Oceania Oceania Asia Africa Europe

  • S. Amer
  • N. Amer

F: San Francisco (US) F: New York (US) F: Sao Paulo (BR) F: Santiago (CL) K: Reykjavik (IS) K: Helsinki (FI) F:Johannesburg(SA) F: Tel Aviv (IL) K: Tokyo (JP) F: Brisbane (AU) F: Auckland (NZ) sfo2 pao1 delhi ams-ix jfk1 lax1

  • rd1

iad1 nap linx tokyo F: C: K:

slide-22
SLIDE 22

Anycast coverage – topological coverage (4)

  • lax1 – F-root local instance in LA, U.S.

– AS paths observed from RouteViews: 192.5.5.0/24 “7660 2516 27318 3557 i” “7500 2497 27318 3557 i” “2497 27318 3557 i” Both AS7660 and AS2516 are in Japan!

25 50 75 100 % N.Amer S.Amer Europe Africa Asia Oceania Oceania Asia Africa Europe

  • S. Amer
  • N. Amer

F: San Francisco (US) F: New York (US) F: Sao Paulo (BR) F: Santiago (CL) K: Reykjavik (IS) K: Helsinki (FI) F:Johannesburg(SA) F: Tel Aviv (IL) K: Tokyo (JP) F: Brisbane (AU) F: Auckland (NZ) sfo2 pao1 delhi ams-ix jfk1 lax1

  • rd1

iad1 nap linx tokyo F: C: K:

f-lax1 go

slide-23
SLIDE 23

Anycast coverage – instance affinity

  • Anycast improves stability by shortening AS paths
  • Anycast increases chance of inconsistency among instances

and of clients’ transparent shifting to different instances.

  • Given DNS traffic is dominated by UDP, route flapping is

unimportant

  • But if DNS uses stateful transaction(TCP, fragmented

UDP)…[Barber,NANOG32]

  • Recent studies [Lorenzo] [Boothe,APNIC19]

[Karrenberg,NANOG34]suggests the impact of routing switches

  • n the query performance is rather minimal
slide-24
SLIDE 24

1 2 3 4 5 1 10 100 1K 10K # instances requested by clients # clients C-root F-root K-root

Anycast coverage – instance affinity (2)

  • 2 F-root global instances together saw 99.8% of total clients who switched

F-root instances

  • 5 K-root global instances together saw 86% of total clients who switched

K-root instances

  • Focusing on the clients who switched the most instances:

– 2 C-root clients are from Brazil and Bolivia – 3 K-root clients are from Uruguay – 27 F-root clients all from UK and they never requested lcy1 – the F-root local instance in London, instead they switch between ams1, lga1, pao1 and sfo2

All from South America, no C or K root instance there!

C-root F-root K-root 1 2 3 4 5 %

  • Pct. of clients switching instances
slide-25
SLIDE 25

Conclusion

  • Current method for limiting the catchment areas of local

instances appears to be generally successful (though exceptions do exist due to peculiar routing configuration)

  • A significant number of clients would benefit if routing

configurations of their local DNS root instances were optimized to route them to the closet instances

  • Instance selection by BGP is highly stable

– In 2-day period, <2% of both C and F-root clients and < 5% of K- root clients experienced an instance change.

slide-26
SLIDE 26

Questions?