diff options
author | Thomas Voss <mail@thomasvoss.com> | 2024-11-27 20:54:24 +0100 |
---|---|---|
committer | Thomas Voss <mail@thomasvoss.com> | 2024-11-27 20:54:24 +0100 |
commit | 4bfd864f10b68b71482b35c818559068ef8d5797 (patch) | |
tree | e3989f47a7994642eb325063d46e8f08ffa681dc /doc/rfc/rfc8483.txt | |
parent | ea76e11061bda059ae9f9ad130a9895cc85607db (diff) |
doc: Add RFC documents
Diffstat (limited to 'doc/rfc/rfc8483.txt')
-rw-r--r-- | doc/rfc/rfc8483.txt | 2187 |
1 files changed, 2187 insertions, 0 deletions
diff --git a/doc/rfc/rfc8483.txt b/doc/rfc/rfc8483.txt new file mode 100644 index 0000000..c654b90 --- /dev/null +++ b/doc/rfc/rfc8483.txt @@ -0,0 +1,2187 @@ + + + + + + +Independent Submission L. Song, Ed. +Request for Comments: 8483 D. Liu +Category: Informational Beijing Internet Institute +ISSN: 2070-1721 P. Vixie + TISF + A. Kato + Keio/WIDE + S. Kerr + October 2018 + + + Yeti DNS Testbed + +Abstract + + Yeti DNS is an experimental, non-production root server testbed that + provides an environment where technical and operational experiments + can safely be performed without risk to production root server + infrastructure. This document aims solely to document the technical + and operational experience of deploying a system that is similar to + but different from the Root Server system (on which the Internet's + Domain Name System is designed and built). + +Status of This Memo + + This document is not an Internet Standards Track specification; it is + published for informational purposes. + + This is a contribution to the RFC Series, independently of any other + RFC stream. The RFC Editor has chosen to publish this document at + its discretion and makes no statement about its value for + implementation or deployment. Documents approved for publication by + the RFC Editor are not candidates for any level of Internet Standard; + see Section 2 of RFC 7841. + + Information about the current status of this document, any errata, + and how to provide feedback on it may be obtained at + https://www.rfc-editor.org/info/rfc8483. + + + + + + + + + + + + + +Song, et al. Informational [Page 1] + +RFC 8483 Yeti DNS Testbed October 2018 + + +Copyright Notice + + Copyright (c) 2018 IETF Trust and the persons identified as the + document authors. All rights reserved. + + This document is subject to BCP 78 and the IETF Trust's Legal + Provisions Relating to IETF Documents + (https://trustee.ietf.org/license-info) in effect on the date of + publication of this document. Please review these documents + carefully, as they describe your rights and restrictions with respect + to this document. + +Table of Contents + + 1. Introduction . . . . . . . . . . . . . . . . . . . . . . . . 3 + 2. Requirements Notation and Conventions . . . . . . . . . . . . 5 + 3. Areas of Study . . . . . . . . . . . . . . . . . . . . . . . 5 + 3.1. Implementation of a Testbed like the Root Server System . 5 + 3.2. Yeti-Root Zone Distribution . . . . . . . . . . . . . . . 5 + 3.3. Yeti-Root Server Names and Addressing . . . . . . . . . . 5 + 3.4. IPv6-Only Yeti-Root Servers . . . . . . . . . . . . . . . 6 + 3.5. DNSSEC in the Yeti-Root Zone . . . . . . . . . . . . . . 6 + 4. Yeti DNS Testbed Infrastructure . . . . . . . . . . . . . . . 7 + 4.1. Root Zone Retrieval . . . . . . . . . . . . . . . . . . . 8 + 4.2. Transformation of Root Zone to Yeti-Root Zone . . . . . . 9 + 4.2.1. ZSK and KSK Key Sets Shared between DMs . . . . . . . 10 + 4.2.2. Unique ZSK per DM; No Shared KSK . . . . . . . . . . 10 + 4.2.3. Preserving Root Zone NSEC Chain and ZSK RRSIGs . . . 11 + 4.3. Yeti-Root Zone Distribution . . . . . . . . . . . . . . . 12 + 4.4. Synchronization of Service Metadata . . . . . . . . . . . 12 + 4.5. Yeti-Root Server Naming Scheme . . . . . . . . . . . . . 13 + 4.6. Yeti-Root Servers . . . . . . . . . . . . . . . . . . . . 14 + 4.7. Experimental Traffic . . . . . . . . . . . . . . . . . . 16 + 4.8. Traffic Capture and Analysis . . . . . . . . . . . . . . 16 + 5. Operational Experience with the Yeti DNS Testbed . . . . . . 17 + 5.1. Viability of IPv6-Only Operation . . . . . . . . . . . . 17 + 5.1.1. IPv6 Fragmentation . . . . . . . . . . . . . . . . . 18 + 5.1.2. Serving IPv4-Only End-Users . . . . . . . . . . . . . 19 + 5.2. Zone Distribution . . . . . . . . . . . . . . . . . . . . 19 + 5.2.1. Zone Transfers . . . . . . . . . . . . . . . . . . . 19 + 5.2.2. Delays in Yeti-Root Zone Distribution . . . . . . . . 20 + 5.2.3. Mixed RRSIGs from Different DM ZSKs . . . . . . . . . 21 + 5.3. DNSSEC KSK Rollover . . . . . . . . . . . . . . . . . . . 22 + 5.3.1. Failure-Case KSK Rollover . . . . . . . . . . . . . . 22 + 5.3.2. KSK Rollover vs. BIND9 Views . . . . . . . . . . . . 22 + 5.3.3. Large Responses during KSK Rollover . . . . . . . . . 23 + 5.4. Capture of Large DNS Response . . . . . . . . . . . . . . 24 + 5.5. Automated Maintenance of the Hints File . . . . . . . . . 24 + + + +Song, et al. Informational [Page 2] + +RFC 8483 Yeti DNS Testbed October 2018 + + + 5.6. Root Label Compression in Knot DNS Server . . . . . . . . 25 + 6. Conclusions . . . . . . . . . . . . . . . . . . . . . . . . . 26 + 7. Security Considerations . . . . . . . . . . . . . . . . . . . 28 + 8. IANA Considerations . . . . . . . . . . . . . . . . . . . . . 28 + 9. References . . . . . . . . . . . . . . . . . . . . . . . . . 29 + 9.1. Normative References . . . . . . . . . . . . . . . . . . 29 + 9.2. Informative References . . . . . . . . . . . . . . . . . 29 + Appendix A. Yeti-Root Hints File . . . . . . . . . . . . . . . . 33 + Appendix B. Yeti-Root Server Priming Response . . . . . . . . . 34 + Appendix C. Active IPv6 Prefixes in Yeti DNS Testbed . . . . . . 36 + Appendix D. Tools Developed for Yeti DNS Testbed . . . . . . . . 36 + Appendix E. Controversy . . . . . . . . . . . . . . . . . . . . 37 + Acknowledgments . . . . . . . . . . . . . . . . . . . . . . . . . 38 + Authors' Addresses . . . . . . . . . . . . . . . . . . . . . . . 39 + +1. Introduction + + The Domain Name System (DNS), as originally specified in [RFC1034] + and [RFC1035], has proved to be an enduring and important platform + upon which almost every end-user of the Internet relies. Despite its + longevity, extensions to the protocol, new implementations, and + refinements to DNS operations continue to emerge both inside and + outside the IETF. + + The Root Server system in particular has seen technical innovation + and development, for example, in the form of wide-scale anycast + deployment, the mitigation of unwanted traffic on a global scale, the + widespread deployment of Response Rate Limiting [RRL], the + introduction of IPv6 transport, the deployment of DNSSEC, changes in + DNSSEC key sizes, and preparations to roll the root zone's Key + Signing Key (KSK) and corresponding trust anchor. These projects + created tremendous qualitative operational change and required + impressive caution and study prior to implementation. They took + place in parallel with the quantitative expansion or delegations for + new TLDs (see <https://newgtlds.icann.org/>). + + Aspects of the operational structure of the Root Server system have + been described in such documents as [TNO2009], [ISC-TN-2003-1], + [RSSAC001], and [RFC7720]. Such references, considered together, + provide sufficient insight into the operations of the system as a + whole that it is straightforward to imagine structural changes to the + Root Server system's infrastructure and to wonder what the + operational implications of such changes might be. + + The Yeti DNS Project was conceived in May 2015 with the aim of + providing a non-production testbed that would be open for use by + anyone from the technical community to propose or run experiments + designed to answer these kinds of questions. Coordination for the + + + +Song, et al. Informational [Page 3] + +RFC 8483 Yeti DNS Testbed October 2018 + + + project was provided by BII, TISF, and the WIDE Project. Thus, Yeti + DNS is an independently coordinated project and is not affiliated + with the IETF, ICANN, IANA, or any Root Server Operator. The + objectives of the Yeti Project were set by the participants in the + project based on experiments that they considered would provide + valuable information. + + Many volunteers collaborated to build a distributed testbed that at + the time of writing includes 25 Yeti root servers with 16 operators + and handles experimental traffic from individual volunteers, + universities, DNS vendors, and distributed measurement networks. + + By design, the Yeti testbed system serves the root zone published by + the IANA with only those structural modifications necessary to ensure + that it is able to function usefully in the Yeti testbed system + instead of the production Root Server system. In particular, no + delegation for any top-level zone is changed, added, or removed from + the IANA-published root zone to construct the root zone served by the + Yeti testbed system, and changes in the root zone are reflected in + the testbed in near real-time. In this document, for clarity, we + refer to the zone derived from the IANA-published root zone as the + Yeti-Root zone. + + The Yeti DNS testbed serves a similar function to the Root Server + system in the sense that they both serve similar zones: the Yeti-Root + zone and the IANA-published root zone. However, the Yeti DNS testbed + only serves clients that are explicitly configured to participate in + the experiment, whereas the Root Server system serves the whole + Internet. Since the dependent end-users and systems of the Yeti DNS + testbed are known and their operations well-coordinated with those of + the Yeti project, it has been possible to deploy structural changes + in the Yeti DNS testbed with effective measurement and analysis, + something that is difficult or simply impractical in the production + Root Server system. + + This document describes the motivation for the Yeti project, + describes the Yeti testbed infrastructure, and provides the technical + and operational experiences of some users of the Yeti testbed. This + document neither addresses the relevant policies under which the Root + Server system is operated nor makes any proposal for changing any + aspect of its implementation or operation. + + + + + + + + + + +Song, et al. Informational [Page 4] + +RFC 8483 Yeti DNS Testbed October 2018 + + +2. Requirements Notation and Conventions + + Through the document, any mention of "Root" with an uppercase "R" and + without other prefix, refers to the "IANA Root" systems used in the + production Internet. Proper mentions of the Yeti infrastructure will + be prefixed with "Yeti", like "Yeti-Root zone", "Yeti DNS", and so + on. + +3. Areas of Study + + This section provides some examples of the topics that the developers + of the Yeti DNS testbed considered important to address. As noted in + Section 1, the Yeti DNS is an independently coordinated project and + is not affiliated with the IETF, ICANN, IANA, or any Root Server + Operator. Thus, the topics and areas for study were selected by (and + for) the proponents of the Yeti project to address their own concerns + and in the hope that the information and tools provided would be of + wider interest. + + Each example included below is illustrated with indicative questions. + +3.1. Implementation of a Testbed like the Root Server System + + o How can a testbed be constructed and deployed on the Internet, + allowing useful public participation without any risk of + disruption of the Root Server system? + + o How can representative traffic be introduced into such a testbed + such that insights into the impact of specific differences between + the testbed and the Root Server system can be observed? + +3.2. Yeti-Root Zone Distribution + + o What are the scaling properties of Yeti-Root zone distribution as + the number of Yeti-Root servers, Yeti-Root server instances, or + intermediate distribution points increases? + +3.3. Yeti-Root Server Names and Addressing + + o What naming schemes other than those closely analogous to the use + of ROOT-SERVERS.NET in the production root zone are practical, and + what are their respective advantages and disadvantages? + + o What are the risks and benefits of signing the zone that contains + the names of the Yeti-Root servers? + + + + + + +Song, et al. Informational [Page 5] + +RFC 8483 Yeti DNS Testbed October 2018 + + + o What automatic mechanisms might be useful to improve the rate at + which clients of Yeti-Root servers are able to react to a Yeti- + Root server renumbering event? + +3.4. IPv6-Only Yeti-Root Servers + + o Are there negative operational effects in the use of IPv6-only + Yeti-Root servers, compared to the use of servers that are dual- + stack? + + o What effect does the IPv6 fragmentation model have on the + operation of Yeti-Root servers, compared with that of IPv4? + +3.5. DNSSEC in the Yeti-Root Zone + + o Is it practical to sign the Yeti-Root zone using multiple, + independently operated DNSSEC signers and multiple corresponding + Zone Signing Keys (ZSKs)? + + o To what extent is [RFC5011] ("Automated Updates of DNS Security + (DNSSEC) Trust Anchors") supported by resolvers? + + o Does the KSK Rollover plan designed and in the process of being + implemented by ICANN work as expected on the Yeti testbed? + + o What is the operational impact of using much larger RSA key sizes + in the ZSKs used in a root? + + o What are the operational consequences of choosing DNSSEC + algorithms other than RSA to sign a root? + + + + + + + + + + + + + + + + + + + + + +Song, et al. Informational [Page 6] + +RFC 8483 Yeti DNS Testbed October 2018 + + +4. Yeti DNS Testbed Infrastructure + + The purpose of the testbed is to allow DNS queries from stub + resolvers, mediated by recursive resolvers, to be delivered to Yeti- + Root servers, and for corresponding responses generated on the Yeti- + Root servers to be returned, as illustrated in Figure 1. + + ,----------. ,-----------. ,------------. + | stub +------> | recursive +------> | Yeti-Root | + | resolver | <------+ resolver | <------+ nameserver | + `----------' `-----------' `------------' + ^ ^ ^ + | appropriate | Yeti-Root hints; | Yeti-Root zone + `- resolver `- Yeti-Root trust `- with DNSKEY RRset + configured anchor signed by + Yeti-Root KSK + + Figure 1: High-Level Testbed Components + + To use the Yeti DNS testbed, a recursive resolver must be configured + to use the Yeti-Root servers. That configuration consists of a list + of names and addresses for the Yeti-Root servers (often referred to + as a "hints file") that replaces the corresponding hints used for the + production Root Server system (Appendix A). If resolvers are + configured to validate DNSSEC, then they also need to be configured + with a DNSSEC trust anchor that corresponds to a KSK used in the Yeti + DNS Project, in place of the normal trust anchor set used for the + Root Zone. + + Since the Yeti root(s) are signed with Yeti keys, rather than those + used by the IANA Root, corresponding changes are needed in the + resolver trust anchors. Corresponding changes are required in the + Yeti-Root hints file Appendix A. Those changes would be properly + rejected as bogus by any validator using the production Root Server + system's root zone trust anchor set. + + Stub resolvers become part of the Yeti DNS testbed by their use of + recursive resolvers that are configured as described above. + + The data flow from IANA to stub resolvers through the Yeti testbed is + illustrated in Figure 2 and is described in more detail in the + sections that follow. + + + + + + + + + +Song, et al. Informational [Page 7] + +RFC 8483 Yeti DNS Testbed October 2018 + + + ,----------------. + ,-- / IANA Root Zone / ---. + | `----------------' | + | | | + | | | Root Zone + ,--------------. ,---V---. ,---V---. ,---V---. + | Yeti Traffic | | BII | | WIDE | | TISF | + | Collection | | DM | | DM | | DM | + `----+----+----' `---+---' `---+---' `---+---' + | | ,-----' ,-------' `----. + | | | | | Yeti-Root + ^ ^ | | | Zone + | | ,---V---. ,---V---. ,---V---. + | `---+ Yeti | | Yeti | . . . . . . . | Yeti | + | | Root | | Root | | Root | + | `---+---' `---+---' `---+---' + | | | | DNS + | | | | Response + | ,--V----------V-------------------------V--. + `---------+ Yeti Resolvers | + `--------------------+---------------------' + | DNS + | Response + ,--------------------V---------------------. + | Yeti Stub Resolvers | + `------------------------------------------' + + The three coordinators of the Yeti DNS testbed: + BII : Beijing Internet Institute + WIDE: Widely Integrated Distributed Environment Project + TISF: A collaborative engineering and security project by Paul Vixie + + Figure 2: Testbed Data Flow + + Note that the roots are not bound to Distribution Masters (DMs). DMs + update their zone on a schedule described in Section 4.1. Each DM + that updates the latest zone can notify all roots, so the zone + transfer can happen between any DM and any root. + +4.1. Root Zone Retrieval + + The Yeti-Root zone is distributed within the Yeti DNS testbed through + a set of internal master servers that are referred to as Distribution + Masters (DMs). These server elements distribute the Yeti-Root zone + to all Yeti-Root servers. The means by which the Yeti DMs construct + the Yeti-Root zone for distribution is described below. + + + + + +Song, et al. Informational [Page 8] + +RFC 8483 Yeti DNS Testbed October 2018 + + + Since Yeti DNS DMs do not receive DNS NOTIFY [RFC1996] messages from + the Root Server system, a polling approach is used to determine when + new revisions of the root zone are available from the production Root + Server system. Each Yeti DM requests the Root Zone SOA record from a + Root server that permits unauthenticated zone transfers of the root + zone, and performs a zone transfer from that server if the retrieved + value of SOA.SERIAL is greater than that of the last retrieved zone. + + At the time of writing, unauthenticated zone transfers of the Root + Zone are available directly from B-Root, C-Root, F-Root, G-Root, + K-Root, and L-Root; two servers XFR.CJR.DNS.ICANN.ORG and + XFR.LAX.DNS.ICANN.ORG; and via FTP from sites maintained by the Root + Zone Maintainer and the IANA Functions Operator. The Yeti DNS + testbed retrieves the Root Zone using zone transfers from F-Root. + The schedule on which F-Root is polled by each Yeti DM is as follows: + + +-------------+-----------------------+ + | DM Operator | Time | + +-------------+-----------------------+ + | BII | UTC hour + 00 minutes | + | WIDE | UTC hour + 20 minutes | + | TISF | UTC hour + 40 minutes | + +-------------+-----------------------+ + + The Yeti DNS testbed uses multiple DMs, each of which acts + autonomously and equivalently to its siblings. Any single DM can act + to distribute new revisions of the Yeti-Root zone and is also + responsible for signing the RRsets that are changed as part of the + transformation of the Root Zone into the Yeti-Root zone described in + Section 4.2. This multiple DM model intends to provide a basic + structure to implement the idea of shared zone control as proposed in + [ITI2014]. + +4.2. Transformation of Root Zone to Yeti-Root Zone + + Two distinct approaches have been deployed in the Yeti DNS testbed, + separately, to transform the Root Zone into the Yeti-Root zone. At a + high level, the approaches are equivalent in the sense that they + replace a minimal set of information in the root zone with + corresponding data for the Yeti DNS testbed; the mechanisms by which + the transforms are executed are different, however. The approaches + are discussed in Sections 4.2.1 and 4.2.2. + + A third approach has also been proposed, but not yet implemented. + The motivations and changes implied by that approach are described in + Section 4.2.3. + + + + + +Song, et al. Informational [Page 9] + +RFC 8483 Yeti DNS Testbed October 2018 + + +4.2.1. ZSK and KSK Key Sets Shared between DMs + + The approach described here was the first to be implemented. It + features entirely autonomous operation of each DM, but also requires + secret key material (the private key in each of the Yeti-Root KSK and + ZSK key pairs) to be distributed and maintained on each DM in a + coordinated way. + + The Root Zone is transformed as follows to produce the Yeti-Root + zone. This transformation is carried out autonomously on each Yeti + DNS Project DM. Each DM carries an authentic copy of the current set + of Yeti KSK and ZSK key pairs, synchronized between all DMs (see + Section 4.4). + + 1. SOA.MNAME is set to www.yeti-dns.org. + + 2. SOA.RNAME is set to <dm-operator>.yeti-dns.org, where + <dm-operator> is currently one of "wide", "bii", or "tisf". + + 3. All DNSKEY, RRSIG, and NSEC records are removed. + + 4. The apex Name Server (NS) RRset is removed, with the + corresponding root server glue (A and AAAA) RRsets. + + 5. A Yeti DNSKEY RRset is added to the apex, comprising the public + parts of all Yeti KSK and ZSKs. + + 6. A Yeti NS RRset is added to the apex that includes all Yeti-Root + servers. + + 7. Glue records (AAAA only, since Yeti-Root servers are v6-only) for + all Yeti-Root servers are added. + + 8. The Yeti-Root zone is signed: the NSEC chain is regenerated; the + Yeti KSK is used to sign the DNSKEY RRset; and the shared ZSK is + used to sign every other RRset. + + Note that the SOA.SERIAL value published in the Yeti-Root zone is + identical to that found in the root zone. + +4.2.2. Unique ZSK per DM; No Shared KSK + + The approach described here was the second to be implemented and + maintained as stable state. Each DM is provisioned with its own, + dedicated ZSK key pairs that are not shared with other DMs. A Yeti- + Root DNSKEY RRset is constructed and signed upstream of all DMs as + the union of the set of active Yeti-Root KSKs and the set of active + ZSKs for every individual DM. Each DM now only requires the secret + + + +Song, et al. Informational [Page 10] + +RFC 8483 Yeti DNS Testbed October 2018 + + + part of its own dedicated ZSK key pairs to be available locally, and + no other secret key material is shared. The high-level approach is + illustrated in Figure 3. + + ,----------. ,-----------. + .--------> BII ZSK +---------> Yeti-Root | + | signs `----------' signs `-----------' + | + ,-----------. | ,----------. ,-----------. + | Yeti KSK +-+--------> TISF ZSK +---------> Yeti-Root | + `-----------' | signs `----------' signs `-----------' + | + | ,----------. ,-----------. + `--------> WIDE ZSK +---------> Yeti-Root | + signs `----------' signs `-----------' + + Figure 3: Unique ZSK per DM + + The process of retrieving the Root Zone from the Root Server system + and replacing and signing the apex DNSKEY RRset no longer takes place + on the DMs; instead, it takes place on a central Hidden Master. The + production of signed DNSKEY RRsets is analogous to the use of Signed + Key Responses (SKRs) produced during ICANN KSK key ceremonies + [ICANN2010]. + + Each DM now retrieves source data (with a premodified and Yeti-signed + DNSKEY RRset, but otherwise unchanged) from the Yeti DNS Hidden + Master instead of from the Root Server system. + + Each DM carries out a similar transformation to that described in + Section 4.2.1, except that DMs no longer need to modify or sign the + DNSKEY RRset, and the DM's unique local ZSK is used to sign every + other RRset. + +4.2.3. Preserving Root Zone NSEC Chain and ZSK RRSIGs + + A change to the transformation described in Section 4.2.2 has been + proposed as a Yeti experiment called PINZ [PINZ], which would + preserve the NSEC chain from the Root Zone and all RRSIG RRs + generated using the Root Zone's ZSKs. The DNSKEY RRset would + continue to be modified to replace the Root Zone KSKs, but Root Zone + ZSKs would be kept intact, and the Yeti KSK would be used to generate + replacement signatures over the apex DNSKEY and NS RRsets. Source + data would continue to flow from the Root Server system through the + Hidden Master to the set of DMs, but no DNSSEC operations would be + required on the DMs, and the source NSEC and most RRSIGs would remain + intact. + + + + +Song, et al. Informational [Page 11] + +RFC 8483 Yeti DNS Testbed October 2018 + + + This approach has been suggested in order to keep minimal changes + from the IANA Root zone and provide cryptographically verifiable + confidence that no owner name in the root zone had been changed in + the process of producing the Yeti-Root zone from the Root Zone, + thereby addressing one of the concerns described in Appendix E in a + way that can be verified automatically. + +4.3. Yeti-Root Zone Distribution + + Each Yeti DM is configured with a full list of Yeti-Root server + addresses to send NOTIFY [RFC1996] messages to. This also forms the + basis for an address-based access-control list for zone transfers. + Authentication by address could be replaced with more rigorous + mechanisms (e.g., using Transaction Signatures (TSIGs) [RFC2845]). + This has not been done at the time of writing since the use of + address-based controls avoids the need for the distribution of shared + secrets amongst the Yeti-Root server operators. + + Individual Yeti-Root servers are configured with a full set of Yeti + DM addresses to which SOA and AXFR queries may be sent in the + conventional manner. + +4.4. Synchronization of Service Metadata + + Changes in the Yeti DNS testbed infrastructure such as the addition + or removal of Yeti-Root servers, renumbering Yeti-Root servers, or + DNSSEC key rollovers require coordinated changes to take place on all + DMs. The Yeti DNS testbed is subject to more frequent changes than + are observed in the Root Server system and includes substantially + more Yeti-Root servers than there are IANA Root Servers, and hence a + manual change process in the Yeti testbed would be more likely to + suffer from human error. An automated and cooperative process was + consequently implemented. + + The theory of this operation is that each DM operator runs a Git + repository locally, containing all service metadata involved in the + operation of each DM. When a change is desired and approved among + all Yeti coordinators, one DM operator (usually BII) updates the + local Git repository. A serial number in the future (in two days) is + chosen for when the changes become active. The DM operator then + pushes the changes to the Git repositories of the other two DM + operators who have a chance to check and edit the changes. When the + serial number of the root zone passes the number chosen, the changes + are pulled automatically to individual DMs and promoted to + production. + + + + + + +Song, et al. Informational [Page 12] + +RFC 8483 Yeti DNS Testbed October 2018 + + + The three Git repositories are synchronized by configuring them as + remote servers. For example, at BII we push to all three DMs' + repositories as follows: + + $ git remote -v + origin yeticonf@yeti-conf.dns-lab.net:dm (fetch) + origin yeticonf@yeti-conf.dns-lab.net:dm (push) + origin yeticonf@yeti-dns.tisf.net:dm (push) + origin yeticonf@yeti-repository.wide.ad.jp:dm (push) + + For more detailed information on DM synchronization, please see this + document in Yeti's GitHub repository: <https://github.com/BII-Lab/ + Yeti-Project/blob/master/doc/Yeti-DM-Sync.md>. + +4.5. Yeti-Root Server Naming Scheme + + The current naming scheme for Root Servers was normalized to use + single-character host names ("A" through "M") under the domain ROOT- + SERVERS.NET, as described in [RSSAC023]. The principal benefit of + this naming scheme was that DNS label compression could be used to + produce a priming response that would fit within 512 bytes at the + time it was introduced, where 512 bytes is the maximum DNS message + size using UDP transport without EDNS(0) [RFC6891]. + + Yeti-Root servers do not use this optimization, but rather use free- + form nameserver names chosen by their respective operators -- in + other words, no attempt is made to minimize the size of the priming + response through the use of label compression. This approach aims to + challenge the need to minimize the priming response in a modern DNS + ecosystem where EDNS(0) is prevalent. + + Priming responses from Yeti-Root servers (unlike those from Root + Servers) do not always include server addresses in the additional + section. In particular, Yeti-Root servers running BIND9 return an + empty additional section if the configuration parameter "minimum- + responses" is set, forcing resolvers to complete the priming process + with a set of conventional recursive lookups in order to resolve + addresses for each Yeti-Root server. The Yeti-Root servers running + NSD were observed to return a fully populated additional section + (depending, of course, on the EDNS buffer size in use). + + Various approaches to normalize the composition of the priming + response were considered, including: + + o Require use of DNS implementations that exhibit a desired behavior + in the priming response. + + + + + +Song, et al. Informational [Page 13] + +RFC 8483 Yeti DNS Testbed October 2018 + + + o Modify nameserver software or configuration as used by Yeti-Root + servers. + + o Isolate the names of Yeti-Root servers in one or more zones that + could be slaved on each Yeti-Root server, renaming servers as + necessary, giving each a source of authoritative data with which + the authority section of a priming response could be fully + populated. This is the approach used in the Root Server system + with the ROOT-SERVERS.NET zone. + + The potential mitigation of renaming all Yeti-Root servers using a + scheme that would allow their names to exist directly in the root + zone was not considered because that approach implies the invention + of new top-level labels not present in the Root Zone. + + Given the relative infrequency of priming queries by individual + resolvers and the additional complexity or other compromises implied + by each of those mitigations, the decision was made to make no effort + to ensure that the composition of priming responses was identical + across servers. Even the empty additional sections generated by + Yeti-Root servers running BIND9 seem to be sufficient for all + resolver software tested; resolvers simply perform a new recursive + lookup for each authoritative server name they need to resolve. + +4.6. Yeti-Root Servers + + Various volunteers have donated authoritative servers to act as Yeti- + Root servers. At the time of writing, there are 25 Yeti-Root servers + distributed globally, one of which is named using a label as + specified in IDNA2008 [RFC5890] (it is shown in the following list in + punycode). + + + + + + + + + + + + + + + + + + + + +Song, et al. Informational [Page 14] + +RFC 8483 Yeti DNS Testbed October 2018 + + + +-------------------------------------+---------------+-------------+ + | Name | Operator | Location | + +-------------------------------------+---------------+-------------+ + | bii.dns-lab.net | BII | CHINA | + | yeti-ns.tsif.net | TSIF | USA | + | yeti-ns.wide.ad.jp | WIDE Project | Japan | + | yeti-ns.as59715.net | as59715 | Italy | + | dahu1.yeti.eu.org | Dahu Group | France | + | ns-yeti.bondis.org | Bond Internet | Spain | + | | Systems | | + | yeti-ns.ix.ru | Russia | MSK-IX | + | yeti.bofh.priv.at | CERT Austria | Austria | + | yeti.ipv6.ernet.in | ERNET India | India | + | yeti-dns01.dnsworkshop.org | dnsworkshop | Germany | + | | /informnis | | + | dahu2.yeti.eu.org | Dahu Group | France | + | yeti.aquaray.com | Aqua Ray SAS | France | + | yeti-ns.switch.ch | SWITCH | Switzerland | + | yeti-ns.lab.nic.cl | NIC Chile | Chile | + | yeti-ns1.dns-lab.net | BII | China | + | yeti-ns2.dns-lab.net | BII | China | + | yeti-ns3.dns-lab.net | BII | China | + | ca...a23dc.yeti-dns.net | Yeti-ZA | South | + | | | Africa | + | 3f...374cd.yeti-dns.net | Yeti-AU | Australia | + | yeti1.ipv6.ernet.in | ERNET India | India | + | xn--r2bi1c.xn--h2bv6c0a.xn--h2brj9c | ERNET India | India | + | yeti-dns02.dnsworkshop.org | dnsworkshop | USA | + | | /informnis | | + | yeti.mind-dns.nl | Monshouwer | Netherlands | + | | Internet | | + | | Diensten | | + | yeti-ns.datev.net | DATEV | Germany | + | yeti.jhcloos.net. | jhcloos | USA | + +-------------------------------------+---------------+-------------+ + + The current list of Yeti-Root servers is made available to a + participating resolver first using a substitute hints file Appendix A + and subsequently by the usual resolver priming process [RFC8109]. + All Yeti-Root servers are IPv6-only, because of the IPv6-only + Internet of the foreseeable future, and hence the Yeti-Root hints + file contains no IPv4 addresses and the Yeti-Root zone contains no + IPv4 glue records. Note that the rationale of an IPv6-only testbed + is to test whether an IPv6-only root can survive any problem or + impact when IPv4 is turned off, much like the context of the IETF + SUNSET4 WG [SUNSET4]. + + + + + +Song, et al. Informational [Page 15] + +RFC 8483 Yeti DNS Testbed October 2018 + + + At the time of writing, all root servers within the Root Server + system serve the ROOT-SERVERS.NET zone in addition to the root zone, + and all but one also serve the ARPA zone. Yeti-Root servers serve + the Yeti-Root zone only. + + Significant software diversity exists across the set of Yeti-Root + servers, as reported by their volunteer operators at the time of + writing: + + o Platform: 18 of 25 Yeti-Root servers are implemented on a Virtual + Private Server (VPS) rather than bare metal. + + o Operating System: 15 Yeti-Root servers run on Linux (Ubuntu, + Debian, CentOS, Red Hat, and ArchLinux); 4 run on FreeBSD; 1 on + NetBSD; and 1 on Windows Server 2016. + + o DNS software: 16 of 25 Yeti-Root servers use BIND9 (versions + varying between 9.9.7 and 9.10.3); 4 use NSD (4.10 and 4.15); 2 + use Knot (2.0.1 and 2.1.0); 1 uses Bundy (1.2.0); 1 uses PowerDNS + (4.1.3); and 1 uses MS DNS (10.0.14300.1000). + +4.7. Experimental Traffic + + For the Yeti DNS testbed to be useful as a platform for + experimentation, it needs to carry statistically representative + traffic. Several approaches have been taken to load the system with + traffic, including both real-world traffic triggered by end-users and + synthetic traffic. + + Resolvers that have been explicitly configured to participate in the + testbed, as described in Section 4, are a source of real-world, end- + user traffic. Due to an efficient cache mechanism, the mean query + rate is less than 100 qps in the Yeti testbed, but a variety of + sources were observed as active during 2017, as summarized in + Appendix C. + + Synthetic traffic has been introduced to the system from time to time + in order to increase traffic loads. Approaches include the use of + distributed measurement platforms such as RIPE ATLAS to send DNS + queries to Yeti-Root servers and the capture of traffic (sent from + non-Yeti resolvers to the Root Server system) that was subsequently + modified and replayed towards Yeti-Root servers. + +4.8. Traffic Capture and Analysis + + Traffic capture of queries and responses is available in the testbed + in both Yeti resolvers and Yeti-Root servers in anticipation of + experiments that require packet-level visibility into DNS traffic. + + + +Song, et al. Informational [Page 16] + +RFC 8483 Yeti DNS Testbed October 2018 + + + Traffic capture is performed on Yeti-Root servers using either + + o dnscap <https://www.dns-oarc.net/tools/dnscap> or + + o pcapdump, part of the pcaputils Debian package + <https://packages.debian.org/sid/pcaputils>, with a patch to + facilitate triggered file upload (see <https://bugs.debian.org/ + cgi-bin/bugreport.cgi?bug=545985>). + + PCAP-format files containing packet captures are uploaded using rsync + to central storage. + +5. Operational Experience with the Yeti DNS Testbed + + The following sections provide commentary on the operation and impact + analyses of the Yeti DNS testbed described in Section 4. More + detailed descriptions of observed phenomena are available in the Yeti + DNS mailing list archives <http://lists.yeti-dns.org/pipermail/ + discuss/> and on the Yeti DNS blog <https://yeti-dns.org/blog.html>. + +5.1. Viability of IPv6-Only Operation + + All Yeti-Root servers were deployed with IPv6 connectivity, and no + IPv4 addresses for any Yeti-Root server were made available (e.g., in + the Yeti hints file or in the DNS itself). This implementation + decision constrained the Yeti-Root system to be v6 only. + + DNS implementations are generally adept at using both IPv4 and IPv6 + when both are available. Servers that cannot be reliably reached + over one protocol might be better queried over the other, to the + benefit of end-users in the common case where DNS resolution is on + the critical path for end-users' perception of performance. However, + this optimization also means that systemic problems with one protocol + can be masked by the other. By forcing all traffic to be carried + over IPv6, the Yeti DNS testbed aimed to expose any such problems and + make them easier to identify and understand. Several examples of + IPv6-specific phenomena observed during the operation of the testbed + are described in the sections that follow. + + Although the Yeti-Root servers themselves were only reachable using + IPv6, real-world end-users often have no IPv6 connectivity. The + testbed was also able to explore the degree to which IPv6-only Yeti- + Root servers were able to serve single-stack, IPv4-only end-user + populations through the use of dual-stack Yeti resolvers. + + + + + + + +Song, et al. Informational [Page 17] + +RFC 8483 Yeti DNS Testbed October 2018 + + +5.1.1. IPv6 Fragmentation + + In the Root Server system, structural changes with the potential to + increase response sizes (and hence fragmentation, fallback to TCP + transport, or both) have been exercised with great care, since the + impact on clients has been difficult to predict or measure. The Yeti + DNS testbed is experimental and has the luxury of a known client + base, making it far easier to make such changes and measure their + impact. + + Many of the experimental design choices described in this document + were expected to trigger larger responses. For example, the choice + of naming scheme for Yeti-Root servers described in Section 4.5 + defeats label compression. It makes a large priming response (up to + 1754 octets with 25 NS records and their corresponding glue records); + the Yeti-Root zone transformation approach described in Section 4.2.2 + greatly enlarges the apex DNSKEY RRset especially during the KSK + rollover (up to 1975 octets with 3 ZSKs and 2 KSKs). Therefore, an + increased incidence of fragmentation was expected. + + The Yeti DNS testbed provides service on IPv6 only. However, + middleboxes (such as firewalls and some routers) are not friendly on + IPv6 fragments. There are reports of a notable packet drop rate due + to the mistreatment of middleboxes on IPv6 fragments [FRAGDROP] + [RFC7872]. One APNIC study [IPv6-frag-DNS] reported that 37% of + endpoints using IPv6-capable DNS resolvers cannot receive a + fragmented IPv6 response over UDP. + + To study the impact, RIPE Atlas probes were used. For each Yeti-Root + server, an Atlas measurement was set up using 100 IPv6-enabled probes + from five regions, sending a DNS query for "./IN/DNSKEY" using UDP + transport with DO=1. This measurement, when carried out concurrently + with a Yeti KSK rollover, further exacerbating the potential for + fragmentation, identified a 7% failure rate compared with a non- + fragmented control. A failure rate of 2% was observed with response + sizes of 1414 octets, which was surprising given the expected + prevalence of 1500-octet (Ethernet-framed) MTUs. + + The consequences of fragmentation were not limited to failures in + delivering DNS responses over UDP transport. There were two cases + where a Yeti-Root server failed when using TCP to transfer the Yeti- + Root zone from a DM. DM log files revealed "socket is not connected" + errors corresponding to zone transfer requests. Further + experimentation revealed that combinations of NetBSD 6.1, NetBSD + 7.0RC1, FreeBSD 10.0, Debian 3.2, and VMWare ESXI 5.5 resulted in a + high TCP Maximum Segment Size (MSS) value of 1440 octets being + negotiated between client and server despite the presence of the + IPV6_USE_MIN_MTU socket option, as described in [USE_MIN_MTU]. The + + + +Song, et al. Informational [Page 18] + +RFC 8483 Yeti DNS Testbed October 2018 + + + mismatch appears to cause outbound segments of a size greater than + 1280 octets to be dropped before sending. Setting the local TCP MSS + to 1220 octets (chosen as 1280 - 60, the size of the IPv6 TCP header + with no other extension headers) was observed to be a pragmatic + mitigation. + +5.1.2. Serving IPv4-Only End-Users + + Yeti resolvers have been successfully used by real-world end-users + for general name resolution within a number of participant + organizations, including resolution of names to IPv4 addresses and + resolution by IPv4-only end-user devices. + + Some participants, recognizing the operational importance of + reliability in resolver infrastructure and concerned about the + stability of their IPv6 connectivity, chose to deploy Yeti resolvers + in parallel to conventional resolvers, making both available to end- + users. While the viability of this approach provides a useful data + point, end-users using Yeti resolvers exclusively provided a better + opportunity to identify and understand any failures in the Yeti DNS + testbed infrastructure. + + Resolvers deployed in IPv4-only environments were able to join the + Yeti DNS testbed by way of upstream, dual-stack Yeti resolvers. In + one case (CERNET2), this was done by assigning IPv4 addresses to + Yeti-Root servers and mapping them in dual-stack IVI translation + devices [RFC6219]. + +5.2. Zone Distribution + + The Yeti DNS testbed makes use of multiple DMs to distribute the + Yeti-Root zone, an approach that would allow the number of Yeti-Root + servers to scale to a higher number than could be supported by a + single distribution source and that provided redundancy. The use of + multiple DMs introduced some operational challenges, however, which + are described in the following sections. + +5.2.1. Zone Transfers + + Yeti-Root servers were configured to serve the Yeti-Root zone as + slaves. Each slave had all DMs configured as masters, providing + redundancy in zone synchronization. + + Each DM in the Yeti testbed served a Yeti-Root zone that was + functionally equivalent but not congruent to that served by every + other DM (see Section 4.3). The differences included variations in + the SOA.MNAME field and, more critically, in the RRSIGs for + everything other than the apex DNSKEY RRset, since signatures for all + + + +Song, et al. Informational [Page 19] + +RFC 8483 Yeti DNS Testbed October 2018 + + + other RRsets are generated using a private key that is only available + to the DM serving its particular variant of the zone (see Sections + 4.2.1 and 4.2.2). + + Incremental Zone Transfer (IXFR), as described in [RFC1995], is a + viable mechanism to use for zone synchronization between any Yeti- + Root server and a consistent, single DM. However, if that Yeti-Root + server ever selected a different DM, IXFR would no longer be a safe + mechanism; structural changes between the incongruent zones on + different DMs would not be included in any transferred delta, and the + result would be a zone that was not internally self-consistent. For + this reason, the first transfer after a change of DM would require + AXFR not IXFR. + + None of the DNS software in use on Yeti-Root servers supports this + mixture of IXFR/AXFR according to the master server in use. This is + unsurprising, given that the environment described above in the Yeti- + Root system is idiosyncratic; conventional zone transfer graphs + involve zones that are congruent between all nodes. For this reason, + all Yeti-Root servers are configured to use AXFR at all times, and + never IXFR, to ensure that zones being served are internally self- + consistent. + +5.2.2. Delays in Yeti-Root Zone Distribution + + Each Yeti DM polled the Root Server system for a new revision of the + root zone on an interleaved schedule, as described in Section 4.1. + Consequently, different DMs were expected to retrieve each revision + of the root zone, and make a corresponding revision of the Yeti-Root + zone available, at different times. The availability of a new + revision of the Yeti-Root zone on the first DM would typically + precede that of the last by 40 minutes. + + Given this distribution mechanism, it might be expected that the + maximum latency between the publication of a new revision of the root + zone and the availability of the corresponding Yeti-Root zone on any + Yeti-Root server would be 20 minutes, since in normal operation at + least one DM should serve that Yeti-Zone within 20 minutes of root + zone publication. In practice, this was not observed. + + In one case, a Yeti-Root server running Bundy 1.2.0 on FreeBSD + 10.2-RELEASE was found to lag root zone publication by as much as ten + hours. Upon investigation, this was found to be due to software + defects that were subsequently corrected. + + More generally, Yeti-Root servers were observed routinely to lag root + zone publication by more than 20 minutes, and relatively often by + more than 40 minutes. Whilst in some cases this might be assumed to + + + +Song, et al. Informational [Page 20] + +RFC 8483 Yeti DNS Testbed October 2018 + + + be a result of connectivity problems, perhaps suppressing the + delivery of NOTIFY messages, it was also observed that Yeti-Root + servers receiving a NOTIFY from one DM would often send SOA queries + and AXFR requests to a different DM. If that DM were not yet serving + the new revision of the Yeti-Root zone, a delay in updating the Yeti- + Root server would naturally result. + +5.2.3. Mixed RRSIGs from Different DM ZSKs + + The second approach for doing the transformation of Root Zone to + Yeti-Root zone (Section 4.2.2) introduces a situation where mixed + RRSIGs from different DM ZSKs are cached in one resolver. + + It is observed that the Yeti-Root zone served by any particular Yeti- + Root server will include signatures generated using the ZSK from the + DM that served the Yeti-Root zone to that Yeti-Root server. + Signatures cached at resolvers might be retrieved from any Yeti-Root + server, and hence are expected to be a mixture of signatures + generated by different ZSKs. Since all ZSKs can be trusted through + the signature by the Yeti KSK over the DNSKEY RRset, which includes + all ZSKs, the mixture of signatures was predicted not to be a threat + to reliable validation. + + It was first tested in BII's lab environment as a proof of concept. + It was observed in the resolver's DNSSEC log that the process of + verifying an RDATA set shows "success" with a key (keyid) in the + DNSKEY RRset. It was implemented later in three DMs that were + carefully coordinated and made public to all Yeti resolver operators + and participants in Yeti's mailing list. At least 45 Yeti resolvers + (deployed by Yeti operators) were being monitored and had set a + reporting trigger if anything was wrong. In addition, the Yeti + mailing list is open for error reports from other participants. So + far, the Yeti testbed has been operated in this configuration (with + multiple ZSKs) for 2 years. This configuration has proven workable + and reliable, even when rollovers of individual ZSKs are on different + schedules. + + Another consequence of this approach is that the apex DNSKEY RRset in + the Yeti-Root zone is much larger than the corresponding DNSKEY RRset + in the Root Zone. This requires more space and produces a larger + response to the query for the DNSKEY RRset especially during the KSK + rollover. + + + + + + + + + +Song, et al. Informational [Page 21] + +RFC 8483 Yeti DNS Testbed October 2018 + + +5.3. DNSSEC KSK Rollover + + At the time of writing, the Root Zone KSK is expected to undergo a + carefully orchestrated rollover as described in [ICANN2016]. ICANN + has commissioned various tests and has published an external test + plan [ICANN2017]. + + Three related DNSSEC KSK rollover exercises were carried out on the + Yeti DNS testbed, somewhat concurrent with the planning and execution + of the rollover in the root zone. Brief descriptions of these + exercises are included below. + +5.3.1. Failure-Case KSK Rollover + + The first KSK rollover that was executed on the Yeti DNS testbed + deliberately ignored the 30-day hold-down timer specified in + [RFC5011] before retiring the outgoing KSK. + + It was confirmed that clients of some (but not all) validating Yeti + resolvers experienced resolution failures (received SERVFAIL + responses) following this change. Those resolvers required + administrator intervention to install a functional trust anchor + before resolution was restored. + +5.3.2. KSK Rollover vs. BIND9 Views + + The second Yeti KSK rollover was designed with similar phases to the + ICANN's KSK rollover, although with modified timings to reduce the + time required to complete the process. The "slot" used in this + rollover was ten days long, as follows: + + +-----------------+----------------+----------+ + | | Old Key: 19444 | New Key | + +-----------------+----------------+----------+ + | slot 1 | pub+sign | | + | slot 2, 3, 4, 5 | pub+sign | pub | + | slot 6, 7 | pub | pub+sign | + | slot 8 | revoke | pub+sign | + | slot 9 | | pub+sign | + +-----------------+----------------+----------+ + + During this rollover exercise, a problem was observed on one Yeti + resolver that was running BIND 9.10.4-p2 [KROLL-ISSUE]. That + resolver was configured with multiple views serving clients in + different subnets at the time that the KSK rollover began. DNSSEC + validation failures were observed following the completion of the KSK + rollover, triggered by the addition of a new view that was intended + to serve clients from a new subnet. + + + +Song, et al. Informational [Page 22] + +RFC 8483 Yeti DNS Testbed October 2018 + + + BIND 9.10 requires "managed-keys" configuration to be specified in + every view, a detail that was apparently not obvious to the operator + in this case and that was subsequently highlighted by the Internet + Systems Consortium (ISC) in their general advice relating to KSK + rollover in the root zone to users of BIND 9 [ISC-BIND]. When the + "managed-keys" configuration is present in every view that is + configured to perform validation, trust anchors for all views are + updated during a KSK rollover. + +5.3.3. Large Responses during KSK Rollover + + Since a KSK rollover necessarily involves the publication of outgoing + and incoming public keys simultaneously, an increase in the size of + DNSKEY responses is expected. The third KSK rollover carried out on + the Yeti DNS testbed was accompanied by a concerted effort to observe + response sizes and their impact on end-users. + + As described in Section 4.2.2, in the Yeti DNS testbed each DM can + maintain control of its own set of ZSKs, which can undergo rollover + independently. During a KSK rollover where concurrent ZSK rollovers + are executed by each of three DMs, the maximum number of apex DNSKEY + RRs present is eight (incoming and outgoing KSK, plus incoming and + outgoing of each of three ZSKs). In practice, however, such + concurrency did not occur; only the BII ZSK was rolled during the KSK + rollover, and hence only three DNSKEY RRset configurations were + observed: + + o 3 ZSKs and 2 KSKs, DNSKEY response of 1975 octets; + + o 3 ZSKs and 1 KSK, DNSKEY response of 1414 octets; and + + o 2 ZSKs and 1 KSK, DNSKEY response of 1139 octets. + + RIPE Atlas probes were used as described in Section 5.1.1 to send + DNSKEY queries directly to Yeti-Root servers. The numbers of queries + and failures were recorded and categorized according to the response + sizes at the time the queries were sent. A summary of the results + ([YetiLR]) is as follows: + + +---------------+----------+---------------+--------------+ + | Response Size | Failures | Total Queries | Failure Rate | + +---------------+----------+---------------+--------------+ + | 1139 | 274 | 64252 | 0.0042 | + | 1414 | 3141 | 126951 | 0.0247 | + | 1975 | 2920 | 42529 | 0.0687 | + +---------------+----------+---------------+--------------+ + + + + + +Song, et al. Informational [Page 23] + +RFC 8483 Yeti DNS Testbed October 2018 + + + The general approach illustrated briefly here provides a useful + example of how the design of the Yeti DNS testbed, separate from the + Root Server system but constructed as a live testbed on the Internet, + facilitates the use of general-purpose active measurement facilities + (such as RIPE Atlas probes) as well as internal passive measurement + (such as packet capture). + +5.4. Capture of Large DNS Response + + Packet capture is a common approach in production DNS systems where + operators require fine-grained insight into traffic in order to + understand production traffic. For authoritative servers, capture of + inbound query traffic is often sufficient, since responses can be + synthesized with knowledge of the zones being served at the time the + query was received. Queries are generally small enough not to be + fragmented, and even with TCP transport are generally packed within a + single segment. + + The Yeti DNS testbed has different requirements; in particular, there + is a desire to compare responses obtained from the Yeti + infrastructure with those received from the Root Server system in + response to a single query stream (e.g., using the "Yeti Many Mirror + Verifier" (YmmV) as described in Appendix D). Some Yeti-Root servers + were capable of recovering complete DNS messages from within + nameservers, e.g., using dnstap; however, not all servers provided + that functionality, and a consistent approach was desirable. + + The requirement to perform passive capture of responses from the wire + together with experiments that were expected (and in some cases + designed) to trigger fragmentation and use of TCP transport led to + the development of a new tool, PcapParser, to perform fragment and + TCP stream reassembly from raw packet capture data. A brief + description of PcapParser is included in Appendix D. + +5.5. Automated Maintenance of the Hints File + + Renumbering events in the Root Server system are relatively rare. + Although each such event is accompanied by the publication of an + updated hints file in standard locations, the task of updating local + copies of that file used by DNS resolvers is manual, and the process + has an observably long tail. For example, in 2015 J-Root was still + receiving traffic at its old address some thirteen years after + renumbering [Wessels2015]. + + The observed impact of these old, deployed hints files is minimal, + likely due to the very low frequency of such renumbering events. + Even the oldest of hints files would still contain some accurate root + server addresses from which priming responses could be obtained. + + + +Song, et al. Informational [Page 24] + +RFC 8483 Yeti DNS Testbed October 2018 + + + By contrast, due to the experimental nature of the system and the + fact that it is operated mainly by volunteers, Yeti-Root servers are + added, removed, and renumbered with much greater frequency. A tool + to facilitate automatic maintenance of hints files was therefore + created: [hintUpdate]. + + The automated procedure followed by the hintUpdate tool is as + follows. + + 1. Use the local resolver to obtain a response to the query + "./IN/NS". + + 2. Use the local resolver to obtain a set of IPv4 and IPv6 addresses + for each name server. + + 3. Validate all signatures obtained from the local resolvers and + confirm that all data is signed. + + 4. Compare the data obtained to that contained within the currently + active hints file; if there are differences, rotate the old one + away and replace it with a new one. + + This tool would not function unmodified when used in the Root Server + system, since the names of individual Root Servers (e.g., A.ROOT- + SERVERS.NET) are not DNSSEC signed. All Yeti-Root server names are + DNSSEC signed, however, and hence this tool functions as expected in + that environment. + +5.6. Root Label Compression in Knot DNS Server + + [RFC1035] specifies that domain names can be compressed when encoded + in DNS messages, and can be represented as one of + + 1. a sequence of labels ending in a zero octet; + + 2. a pointer; or + + 3. a sequence of labels ending with a pointer. + + The purpose of this flexibility is to reduce the size of domain names + encoded in DNS messages. + + It was observed that Yeti-Root servers running Knot 2.0 would + compress the zero-length label (the root domain, often represented as + ".") using a pointer to an earlier example. Although legal, this + encoding increases the encoded size of the root label from one octet + to two; it was also found to break some client software -- in + + + + +Song, et al. Informational [Page 25] + +RFC 8483 Yeti DNS Testbed October 2018 + + + particular, the Go DNS library. Bug reports were filed against both + Knot and the Go DNS library, and both were resolved in subsequent + releases. + +6. Conclusions + + Yeti DNS was designed and implemented as a live DNS root system + testbed. It serves a root zone ("Yeti-Root" in this document) + derived from the root zone published by the IANA with only those + structural modifications necessary to ensure its function in the + testbed system. The Yeti DNS testbed has proven to be a useful + platform to address many questions that would be challenging to + answer using the production Root Server system, such as those + included in Section 3. + + Indicative findings following from the construction and operation of + the Yeti DNS testbed include: + + o Operation in a pure IPv6-only environment; confirmation of a + significant failure rate in the transmission of large responses + (~7%), but no other persistent failures observed. Two cases in + which Yeti-Root servers failed to retrieve the Yeti-Root zone due + to fragmentation of TCP segments; mitigated by setting a TCP MSS + of 1220 octets (see Section 5.1.1). + + o Successful operation with three autonomous Yeti-Root zone signers + and 25 Yeti-Root servers, and confirmation that IXFR is not an + appropriate transfer mechanism of zones that are structurally + incongruent across different transfer paths (see Section 5.2). + + o ZSK size increased to 2048 bits and multiple KSK rollovers + executed to exercise support of RFC 5011 in validating resolvers; + identification of pitfalls relating to views in BIND9 when + configured with "managed-keys" (see Section 5.3). + + o Use of natural (non-normalized) names for Yeti-Root servers + exposed some differences between implementations in the inclusion + of additional-section glue in responses to priming queries; + however, despite this inefficiency, Yeti resolvers were observed + to function adequately (see Section 4.5). + + o It was observed that Knot 2.0 performed label compression on the + root (empty) label. This resulted in an increased encoding size + for references to the root label, since a pointer is encoded as + two octets whilst the root label itself only requires one (see + Section 5.6). + + + + + +Song, et al. Informational [Page 26] + +RFC 8483 Yeti DNS Testbed October 2018 + + + o Some tools were developed in response to the operational + experience of running and using the Yeti DNS testbed: DNS fragment + and DNS Additional Truncated Response (ATR) for large DNS + responses, a BIND9 patch for additional-section glue, YmmV, and + IPv6 defrag for capturing and mirroring traffic. In addition, a + tool to facilitate automatic maintenance of hints files was + created (see Appendix D). + + The Yeti DNS testbed was used only by end-users whose local + infrastructure providers had made the conscious decision to do so, as + is appropriate for an experimental, non-production system. So far, + no serious user complaints have reached Yeti's mailing list during + Yeti normal operation. Adding more instances into the Yeti root + system may help to enhance the quality of service, but it is + generally accepted that Yeti DNS performance is good enough to serve + the purpose of DNS Root testbed. + + The experience gained during the operation of the Yeti DNS testbed + suggested several topics worthy of further study: + + o Priming truncation and TCP-only Yeti-Root servers: observe and + measure the worst-possible case for priming truncation by + responding with TC=1 to all priming queries received over UDP + transport, forcing clients to retry using TCP. This should also + give some insight into the usefulness of TCP-only DNS in general. + + o KSK ECDSA Rollover: one possible way to reduce DNSKEY response + sizes is to change to an elliptic curve signing algorithm. While + in principle this can be done separately for the KSK and the ZSK, + the RIPE NCC has done research recently and discovered that some + resolvers require that both KSK and ZSK use the same algorithm. + This means that an algorithm roll also involves a KSK roll. + Performing an algorithm roll at the root would be an interesting + challenge. + + o Sticky Notify for zone transfer: the non-applicability of IXFR as + a zone transfer mechanism in the Yeti DNS testbed could be + mitigated by the implementation of a sticky preference for master + server for each slave. This would be so that an initial AXFR + response could be followed up with IXFR requests without + compromising zone integrity in the case (as with Yeti) that + equivalent but incongruent versions of a zone are served by + different masters. + + + + + + + + +Song, et al. Informational [Page 27] + +RFC 8483 Yeti DNS Testbed October 2018 + + + o Key distribution for zone transfer credentials: the use of a + shared secret between slave and master requires key distribution + and management whose scaling properties are not ideally suited to + systems with large numbers of transfer clients. Other approaches + for key distribution and authentication could be considered. + + o DNS is a tree-based hierarchical database. Mathematically, it has + a root node and dependency between parent and child nodes. So, + any failures and instability of parent nodes (Root in Yeti's case) + may impact their child nodes if there is a human mistake, a + malicious attack, or even an earthquake. It is proposed to define + technology and practices to allow any organization, from the + smallest company to nations, to be self-sufficient in their DNS. + + o In Section 3.12 of [RFC8324], a "Centrally Controlled Root" is + viewed as an issue of DNS. In future work, it would be + interesting to test some technical tools like blockchain [BC] to + either remove the technical requirement for a central authority + over the root or enhance the security and stability of the + existing Root. + +7. Security Considerations + + As introduced in Section 4.4, service metadata is synchronized among + 3 DMs using Git tool. Any security issue around Git may affect Yeti + DM operation. For example, a hacker may compromise one DM's Git + repository and push unwanted changes to the Yeti DM system; this may + introduce a bad root server or bad key for a period of time. + + The Yeti resolver needs the bootstrapping files to join the testbed, + like the hints file and trust anchor of Yeti. All required + information is published on <yeti-dns.org> and <github.com>. If a + hacker tampers with those websites by creating a fake page, a new + resolver may lose its way and be configured with a bad root. + + DNSSEC is an important research goal in the Yeti DNS testbed. To + reduce the central function of DNSSEC for Root zone, we sign the + Yeti-Root zone using multiple, independently operated DNSSEC signers + and multiple corresponding ZSKs (see Section 4.2). To verify ICANN's + KSK rollover, we rolled the Yeti KSK three times according to RFC + 5011, and we do have some observations (see Section 5.3). In + addition, larger RSA key sizes were used in the testbed before + 2048-bit keys were used in the ZSK signing process of the IANA Root + zone. + +8. IANA Considerations + + This document has no IANA actions. + + + +Song, et al. Informational [Page 28] + +RFC 8483 Yeti DNS Testbed October 2018 + + +9. References + +9.1. Normative References + + [RFC1034] Mockapetris, P., "Domain names - concepts and facilities", + STD 13, RFC 1034, DOI 10.17487/RFC1034, November 1987, + <https://www.rfc-editor.org/info/rfc1034>. + + [RFC1035] Mockapetris, P., "Domain names - implementation and + specification", STD 13, RFC 1035, DOI 10.17487/RFC1035, + November 1987, <https://www.rfc-editor.org/info/rfc1035>. + + [RFC1995] Ohta, M., "Incremental Zone Transfer in DNS", RFC 1995, + DOI 10.17487/RFC1995, August 1996, + <https://www.rfc-editor.org/info/rfc1995>. + + [RFC1996] Vixie, P., "A Mechanism for Prompt Notification of Zone + Changes (DNS NOTIFY)", RFC 1996, DOI 10.17487/RFC1996, + August 1996, <https://www.rfc-editor.org/info/rfc1996>. + + [RFC5011] StJohns, M., "Automated Updates of DNS Security (DNSSEC) + Trust Anchors", STD 74, RFC 5011, DOI 10.17487/RFC5011, + September 2007, <https://www.rfc-editor.org/info/rfc5011>. + + [RFC5890] Klensin, J., "Internationalized Domain Names for + Applications (IDNA): Definitions and Document Framework", + RFC 5890, DOI 10.17487/RFC5890, August 2010, + <https://www.rfc-editor.org/info/rfc5890>. + +9.2. Informative References + + [ATR] Song, L., "ATR: Additional Truncation Response for Large + DNS Response", Work in Progress, draft-song-atr-large- + resp-02, August 2018. + + [BC] Wikipedia, "Blockchain", September 2018, + <https://en.wikipedia.org/w/ + index.php?title=Blockchain&oldid=861681529>. + + [FRAGDROP] Jaeggli, J., Colitti, L., Kumari, W., Vyncke, E., Kaeo, + M., and T. Taylor, "Why Operators Filter Fragments and + What It Implies", Work in Progress, draft-taylor-v6ops- + fragdrop-02, December 2013. + + [FRAGMENTS] + Sivaraman, M., Kerr, S., and D. Song, "DNS message + fragments", Work in Progress, draft-muks-dns-message- + fragments-00, July 2015. + + + +Song, et al. Informational [Page 29] + +RFC 8483 Yeti DNS Testbed October 2018 + + + [hintUpdate] + "Hintfile Auto Update", commit de428c0, October 2015, + <https://github.com/BII-Lab/Hintfile-Auto-Update>. + + [HOW_ATR_WORKS] + Huston, G., "How well does ATR actually work?", + APNIC blog, April 2018, + <https://blog.apnic.net/2018/04/16/ + how-well-does-atr-actually-work/>. + + [ICANN2010] + Schlyter, J., Lamb, R., and R. Balasubramanian, "DNSSEC + Key Management Implementation for the Root Zone (DRAFT)", + May 2010, <http://www.root-dnssec.org/wp-content/ + uploads/2010/05/draft-icann-dnssec-keymgmt-01.txt>. + + [ICANN2016] + Design Team, "Root Zone KSK Rollover Plan", March 2016, + <https://www.iana.org/reports/2016/ + root-ksk-rollover-design-20160307.pdf>. + + [ICANN2017] + ICANN, "2017 KSK Rollover External Test Plan", July 2016, + <https://www.icann.org/en/system/files/files/ + ksk-rollover-external-test-plan-22jul16-en.pdf>. + + [IPv6-frag-DNS] + Huston, G., "Dealing with IPv6 fragmentation in the DNS", + APNIC blog, August 2017, + <https://blog.apnic.net/2017/08/22/ + dealing-ipv6-fragmentation-dns>. + + [ISC-BIND] Risk, V., "2017 Root Key Rollover - What Does it Mean for + BIND Users?", Internet Systems Consortium, December 2016, + <https://www.isc.org/blogs/2017-root-key-rollover-what- + does-it-mean-for-bind-users/>. + + [ISC-TN-2003-1] + Abley, J., "Hierarchical Anycast for Global Service + Distribution", March 2003, + <http://ftp.isc.org/isc/pubs/tn/isc-tn-2003-1.txt>. + + [ITI2014] ICANN, "Identifier Technology Innovation Report", May + 2014, <https://www.icann.org/en/system/files/files/ + iti-report-15may14-en.pdf>. + + + + + + +Song, et al. Informational [Page 30] + +RFC 8483 Yeti DNS Testbed October 2018 + + + [KROLL-ISSUE] + Song, D., "A DNSSEC issue during Yeti KSK rollover", Yeti + DNS blog, October 2016, <http://yeti-dns.org/yeti/blog/ + 2016/10/26/A-DNSSEC-issue-during-Yeti-KSK-rollover.html>. + + [PINZ] Song, D., "Yeti experiment plan for PINZ", Yeti DNS blog, + May 2018, <http://yeti-dns.org/yeti/blog/2018/05/01/ + Experiment-plan-for-PINZ.html>. + + [RFC2826] Internet Architecture Board, "IAB Technical Comment on the + Unique DNS Root", RFC 2826, DOI 10.17487/RFC2826, May + 2000, <https://www.rfc-editor.org/info/rfc2826>. + + [RFC2845] Vixie, P., Gudmundsson, O., Eastlake 3rd, D., and B. + Wellington, "Secret Key Transaction Authentication for DNS + (TSIG)", RFC 2845, DOI 10.17487/RFC2845, May 2000, + <https://www.rfc-editor.org/info/rfc2845>. + + [RFC6219] Li, X., Bao, C., Chen, M., Zhang, H., and J. Wu, "The + China Education and Research Network (CERNET) IVI + Translation Design and Deployment for the IPv4/IPv6 + Coexistence and Transition", RFC 6219, + DOI 10.17487/RFC6219, May 2011, + <https://www.rfc-editor.org/info/rfc6219>. + + [RFC6891] Damas, J., Graff, M., and P. Vixie, "Extension Mechanisms + for DNS (EDNS(0))", STD 75, RFC 6891, + DOI 10.17487/RFC6891, April 2013, + <https://www.rfc-editor.org/info/rfc6891>. + + [RFC7720] Blanchet, M. and L-J. Liman, "DNS Root Name Service + Protocol and Deployment Requirements", BCP 40, RFC 7720, + DOI 10.17487/RFC7720, December 2015, + <https://www.rfc-editor.org/info/rfc7720>. + + [RFC7872] Gont, F., Linkova, J., Chown, T., and W. Liu, + "Observations on the Dropping of Packets with IPv6 + Extension Headers in the Real World", RFC 7872, + DOI 10.17487/RFC7872, June 2016, + <https://www.rfc-editor.org/info/rfc7872>. + + [RFC8109] Koch, P., Larson, M., and P. Hoffman, "Initializing a DNS + Resolver with Priming Queries", BCP 209, RFC 8109, + DOI 10.17487/RFC8109, March 2017, + <https://www.rfc-editor.org/info/rfc8109>. + + + + + + +Song, et al. Informational [Page 31] + +RFC 8483 Yeti DNS Testbed October 2018 + + + [RFC8324] Klensin, J., "DNS Privacy, Authorization, Special Uses, + Encoding, Characters, Matching, and Root Structure: Time + for Another Look?", RFC 8324, DOI 10.17487/RFC8324, + February 2018, <https://www.rfc-editor.org/info/rfc8324>. + + [RRL] Vixie, P. and V. Schryver, "Response Rate Limiting in the + Domain Name System (DNS RRL)", June 2012, + <http://www.redbarn.org/dns/ratelimits>. + + [RSSAC001] Root Server System Advisory Committee (RSSAC), "Service + Expectations of Root Servers", RSSAC001 Version 1, + December 2015, + <https://www.icann.org/en/system/files/files/ + rssac-001-root-service-expectations-04dec15-en.pdf>. + + [RSSAC023] Root Server System Advisory Committee (RSSAC), "History of + the Root Server System", November 2016, + <https://www.icann.org/en/system/files/files/ + rssac-023-04nov16-en.pdf>. + + [SUNSET4] IETF, "Sunsetting IPv4 (sunset4) Concluded WG", + <https://datatracker.ietf.org/wg/sunset4/about/>. + + [TNO2009] Gijsen, B., Jamakovic, A., and F. Roijers, "Root Scaling + Study: Description of the DNS Root Scaling Model", + TNO report, September 2009, + <https://www.icann.org/en/system/files/files/ + root-scaling-model-description-29sep09-en.pdf>. + + [USE_MIN_MTU] + Andrews, M., "TCP Fails To Respect IPV6_USE_MIN_MTU", Work + in Progress, draft-andrews-tcp-and-ipv6-use-minmtu-04, + October 2015. + + [Wessels2015] + Wessels, D., Castonguay, J., and P. Barber, "Thirteen + Years of 'Old J-Root'", DNS-OARC Fall 2015 Workshop, + October 2015, <https://indico.dns-oarc.net/event/24/ + session/10/contribution/10/material/slides/0.pdf>. + + [YetiLR] "Observation on Large response issue during Yeti KSK + rollover", Yeti DNS blog, August 2017, + <https://yeti-dns.org/yeti/blog/2017/08/02/ + large-packet-impact-during-yeti-ksk-rollover.html>. + + + + + + + +Song, et al. Informational [Page 32] + +RFC 8483 Yeti DNS Testbed October 2018 + + +Appendix A. Yeti-Root Hints File + + The following hints file (complete and accurate at the time of + writing) causes a DNS resolver to use the Yeti DNS testbed in place + of the production Root Server system and hence participate in + experiments running on the testbed. + + Note that some lines have been wrapped in the text that follows in + order to fit within the production constraints of this document. + Wrapped lines are indicated with a blackslash character ("\"), + following common convention. + + . 3600000 IN NS bii.dns-lab.net + bii.dns-lab.net 3600000 IN AAAA 240c:f:1:22::6 + . 3600000 IN NS yeti-ns.tisf.net + yeti-ns.tisf.net 3600000 IN AAAA 2001:559:8000::6 + . 3600000 IN NS yeti-ns.wide.ad.jp + yeti-ns.wide.ad.jp 3600000 IN AAAA 2001:200:1d9::35 + . 3600000 IN NS yeti-ns.as59715.net + yeti-ns.as59715.net 3600000 IN AAAA \ + 2a02:cdc5:9715:0:185:5:203:53 + . 3600000 IN NS dahu1.yeti.eu.org + dahu1.yeti.eu.org 3600000 IN AAAA \ + 2001:4b98:dc2:45:216:3eff:fe4b:8c5b + . 3600000 IN NS ns-yeti.bondis.org + ns-yeti.bondis.org 3600000 IN AAAA 2a02:2810:0:405::250 + . 3600000 IN NS yeti-ns.ix.ru + yeti-ns.ix.ru 3600000 IN AAAA 2001:6d0:6d06::53 + . 3600000 IN NS yeti.bofh.priv.at + yeti.bofh.priv.at 3600000 IN AAAA 2a01:4f8:161:6106:1::10 + . 3600000 IN NS yeti.ipv6.ernet.in + yeti.ipv6.ernet.in 3600000 IN AAAA 2001:e30:1c1e:1::333 + . 3600000 IN NS yeti-dns01.dnsworkshop.org + yeti-dns01.dnsworkshop.org \ + 3600000 IN AAAA 2001:1608:10:167:32e::53 + . 3600000 IN NS yeti-ns.conit.co + yeti-ns.conit.co 3600000 IN AAAA \ + 2604:6600:2000:11::4854:a010 + . 3600000 IN NS dahu2.yeti.eu.org + dahu2.yeti.eu.org 3600000 IN AAAA 2001:67c:217c:6::2 + . 3600000 IN NS yeti.aquaray.com + yeti.aquaray.com 3600000 IN AAAA 2a02:ec0:200::1 + . 3600000 IN NS yeti-ns.switch.ch + yeti-ns.switch.ch 3600000 IN AAAA 2001:620:0:ff::29 + . 3600000 IN NS yeti-ns.lab.nic.cl + yeti-ns.lab.nic.cl 3600000 IN AAAA 2001:1398:1:21::8001 + . 3600000 IN NS yeti-ns1.dns-lab.net + + + + +Song, et al. Informational [Page 33] + +RFC 8483 Yeti DNS Testbed October 2018 + + + yeti-ns1.dns-lab.net 3600000 IN AAAA 2001:da8:a3:a027::6 + . 3600000 IN NS yeti-ns2.dns-lab.net + yeti-ns2.dns-lab.net 3600000 IN AAAA 2001:da8:268:4200::6 + . 3600000 IN NS yeti-ns3.dns-lab.net + yeti-ns3.dns-lab.net 3600000 IN AAAA 2400:a980:30ff::6 + . 3600000 IN NS \ + ca978112ca1bbdcafac231b39a23dc.yeti-dns.net + ca978112ca1bbdcafac231b39a23dc.yeti-dns.net \ + 3600000 IN AAAA 2c0f:f530::6 + . 3600000 IN NS \ + 3e23e8160039594a33894f6564e1b1.yeti-dns.net + 3e23e8160039594a33894f6564e1b1.yeti-dns.net \ + 3600000 IN AAAA 2803:80:1004:63::1 + . 3600000 IN NS \ + 3f79bb7b435b05321651daefd374cd.yeti-dns.net + 3f79bb7b435b05321651daefd374cd.yeti-dns.net \ + 3600000 IN AAAA 2401:c900:1401:3b:c::6 + . 3600000 IN NS \ + xn--r2bi1c.xn--h2bv6c0a.xn--h2brj9c + xn--r2bi1c.xn--h2bv6c0a.xn--h2brj9c \ + 3600000 IN AAAA 2001:e30:1c1e:10::333 + . 3600000 IN NS yeti1.ipv6.ernet.in + yeti1.ipv6.ernet.in 3600000 IN AAAA 2001:e30:187d::333 + . 3600000 IN NS yeti-dns02.dnsworkshop.org + yeti-dns02.dnsworkshop.org \ + 3600000 IN AAAA 2001:19f0:0:1133::53 + . 3600000 IN NS yeti.mind-dns.nl + yeti.mind-dns.nl 3600000 IN AAAA 2a02:990:100:b01::53:0 + +Appendix B. Yeti-Root Server Priming Response + + Here is the reply of a Yeti root name server to a priming request. + The authoritative server runs NSD. + + ... + ;; Got answer: + ;; ->>HEADER<<- opcode: QUERY, status: NOERROR, id: 62391 + ;; flags: qr aa rd; QUERY: 1, ANSWER: 26, AUTHORITY: 0, ADDITIONAL: 7 + ;; WARNING: recursion requested but not available + + ;; OPT PSEUDOSECTION: + ; EDNS: version: 0, flags: do; udp: 1460 + ;; QUESTION SECTION: + ;. IN NS + + ;; ANSWER SECTION: + . 86400 IN NS bii.dns-lab.net. + . 86400 IN NS yeti.bofh.priv.at. + + + +Song, et al. Informational [Page 34] + +RFC 8483 Yeti DNS Testbed October 2018 + + + . 86400 IN NS yeti.ipv6.ernet.in. + . 86400 IN NS yeti.aquaray.com. + . 86400 IN NS yeti.jhcloos.net. + . 86400 IN NS yeti.mind-dns.nl. + . 86400 IN NS dahu1.yeti.eu.org. + . 86400 IN NS dahu2.yeti.eu.org. + . 86400 IN NS yeti1.ipv6.ernet.in. + . 86400 IN NS ns-yeti.bondis.org. + . 86400 IN NS yeti-ns.ix.ru. + . 86400 IN NS yeti-ns.lab.nic.cl. + . 86400 IN NS yeti-ns.tisf.net. + . 86400 IN NS yeti-ns.wide.ad.jp. + . 86400 IN NS yeti-ns.datev.net. + . 86400 IN NS yeti-ns.switch.ch. + . 86400 IN NS yeti-ns.as59715.net. + . 86400 IN NS yeti-ns1.dns-lab.net. + . 86400 IN NS yeti-ns2.dns-lab.net. + . 86400 IN NS yeti-ns3.dns-lab.net. + . 86400 IN NS xn--r2bi1c.xn--h2bv6c0a.xn--h2brj9c. + . 86400 IN NS yeti-dns01.dnsworkshop.org. + . 86400 IN NS yeti-dns02.dnsworkshop.org. + . 86400 IN NS 3f79bb7b435b05321651daefd374cd.yeti-dns.net. + . 86400 IN NS ca978112ca1bbdcafac231b39a23dc.yeti-dns.net. + . 86400 IN RRSIG NS 8 0 86400 ( + 20171121050105 20171114050105 26253 . + FUvezvZgKtlLzQx2WKyg+D6dw/pITcbuZhzStZfg+LNa + DjLJ9oGIBTU1BuqTujKHdxQn0DcdFh9QE68EPs+93bZr + VlplkmObj8f0B7zTQgGWBkI/K4Tn6bZ1I7QJ0Zwnk1mS + BmEPkWmvo0kkaTQbcID+tMTodL6wPAgW1AdwQUInfy21 + p+31GGm3+SU6SJsgeHOzPUQW+dUVWmdj6uvWCnUkzW9p + +5en4+85jBfEOf+qiyvaQwUUe98xZ1TOiSwYvk5s/qiv + AMjG6nY+xndwJUwhcJAXBVmGgrtbiR8GiGZfGqt748VX + 4esLNtD8vdypucffem6n0T0eV1c+7j/eIA== ) + + ;; ADDITIONAL SECTION: + bii.dns-lab.net. 86400 IN AAAA 240c:f:1:22::6 + yeti.bofh.priv.at. 86400 IN AAAA 2a01:4f8:161:6106:1::10 + yeti.ipv6.ernet.in. 86400 IN AAAA 2001:e30:1c1e:1::333 + yeti.aquaray.com. 86400 IN AAAA 2a02:ec0:200::1 + yeti.jhcloos.net. 86400 IN AAAA 2001:19f0:5401:1c3::53 + yeti.mind-dns.nl. 86400 IN AAAA 2a02:990:100:b01::53:0 + + ;; Query time: 163 msec + ;; SERVER: 2001:4b98:dc2:45:216:3eff:fe4b:8c5b#53 + ;; WHEN: Tue Nov 14 16:45:37 +08 2017 + ;; MSG SIZE rcvd: 1222 + + + + + +Song, et al. Informational [Page 35] + +RFC 8483 Yeti DNS Testbed October 2018 + + +Appendix C. Active IPv6 Prefixes in Yeti DNS Testbed + + The following table shows the prefixes that were active during 2017. + + +----------------------+---------------------------------+----------+ + | Prefix | Originator | Location | + +----------------------+---------------------------------+----------+ + | 240c::/28 | BII | CN | + | 2001:6d0:6d06::/48 | MSK-IX | RU | + | 2001:1488::/32 | CZ.NIC | CZ | + | 2001:620::/32 | SWITCH | CH | + | 2001:470::/32 | Hurricane Electric, Inc. | US | + | 2001:0DA8:0202::/48 | BUPT6-CERNET2 | CN | + | 2001:19f0:6c00::/38 | Choopa, LLC | US | + | 2001:da8:205::/48 | BJTU6-CERNET2 | CN | + | 2001:62a::/31 | Vienna University Computer | AT | + | | Center | | + | 2001:67c:217c::/48 | AFNIC | FR | + | 2a02:2478::/32 | Profitbricks GmbH | DE | + | 2001:1398:1::/48 | NIC Chile | CL | + | 2001:4490:dc4c::/46 | NIB (National Internet | IN | + | | Backbone) | | + | 2001:4b98::/32 | Gandi | FR | + | 2a02:aa8:0:2000::/52 | T-Systems-Eltec | ES | + | 2a03:b240::/32 | Netskin GmbH | CH | + | 2801:1a0::/42 | Universidad de Ibague | CO | + | 2a00:1cc8::/40 | ICT Valle Umbra s.r.l. | IT | + | 2a02:cdc0::/29 | ORG-CdSB1-RIPE | IT | + +----------------------+---------------------------------+----------+ + +Appendix D. Tools Developed for Yeti DNS Testbed + + Various tools were developed to support the Yeti DNS testbed, a + selection of which are described briefly below. + + YmmV ("Yeti Many Mirror Verifier") is designed to make it easy and + safe for a DNS administrator to capture traffic sent from a resolver + to the Root Server system and to replay it towards Yeti-Root servers. + Responses from both systems are recorded and compared, and + differences are logged. See <https://github.com/BII-Lab/ymmv>. + + PcapParser is a module used by YmmV which reassembles fragmented IPv6 + datagrams and TCP segments from a PCAP archive and extracts DNS + messages contained within them. See <https://github.com/RunxiaWan/ + PcapParser>. + + + + + + +Song, et al. Informational [Page 36] + +RFC 8483 Yeti DNS Testbed October 2018 + + + DNS-layer-fragmentation implements DNS proxies that perform + application-level fragmentation of DNS messages, based on + [FRAGMENTS]. The idea with these proxies is to explore splitting DNS + messages in the protocol itself, so they will not by fragmented by + the IP layer. See <https://github.com/BII-Lab/DNS-layer- + Fragmentation>. + + DNS_ATR is an implementation of DNS Additional Truncated Response + (ATR), as described in [ATR] and [HOW_ATR_WORKS]. DNS_ATR acts as a + proxy between resolver and authoritative servers, forwarding queries + and responses as a silent and transparent listener. Responses that + are larger than a nominated threshold (1280 octets by default) + trigger additional truncated responses to be sent immediately + following the large response. See <https://github.com/songlinjian/ + DNS_ATR>. + +Appendix E. Controversy + + The Yeti DNS Project, its infrastructure and the various experiments + that have been carried out using that infrastructure, have been + described by people involved in the project in many public meetings + at technical venues since its inception. The mailing lists using + which the operation of the infrastructure has been coordinated are + open to join, and their archives are public. The project as a whole + has been the subject of robust public discussion. + + Some commentators have expressed concern that the Yeti DNS Project + is, in effect, operating an alternate root, challenging the IAB's + comments published in [RFC2826]. Other such alternate roots are + considered to have caused end-user confusion and instability in the + namespace of the DNS by the introduction of new top-level labels or + the different use of top-level labels present in the Root Server + system. The coordinators of the Yeti DNS Project do not consider the + Yeti DNS Project to be an alternate root in this sense, since by + design the namespace enabled by the Yeti-Root zone is identical to + that of the Root Zone. + + Some commentators have expressed concern that the Yeti DNS Project + seeks to influence or subvert administrative policy relating to the + Root Server system, in particular in the use of DNSSEC trust anchors + not published by the IANA and the use of Yeti-Root servers in regions + where governments or other organizations have expressed interest in + operating a Root Server. The coordinators of the Yeti-Root project + observe that their mandate is entirely technical and has no ambition + to influence policy directly; they do hope, however, that technical + findings from the Yeti DNS Project might act as a useful resource for + the wider technical community. + + + + +Song, et al. Informational [Page 37] + +RFC 8483 Yeti DNS Testbed October 2018 + + +Acknowledgments + + Firstly, the authors would like to acknowledge the contributions from + the people who were involved in the implementation and operation of + the Yeti DNS by donating their time and resources. They are: + + Tomohiro Ishihara, Antonio Prado, Stephane Bortzmeyer, Mickael + Jouanne, Pierre Beyssac, Joao Damas, Pavel Khramtsov, Dmitry + Burkov, Dima Burkov, Kovalenko Dmitry, Otmar Lendl, Praveen Misra, + Carsten Strotmann, Edwin Gomez, Daniel Stirnimann, Andreas + Schulze, Remi Gacogne, Guillaume de Lafond, Yves Bovard, Hugo + Salgado, Kees Monshouwer, Li Zhen, Daobiao Gong, Andreas Schulze, + James Cloos, and Runxia Wan. + + Thanks to all people who gave important advice and comments to Yeti, + either in face-to-face meetings or virtually via phone or mailing + list. Some of the individuals are as follows: + + Wu Hequan, Zhou Hongren, Cheng Yunqing, Xia Chongfeng, Tang + Xiongyan, Li Yuxiao, Feng Ming, Zhang Tongxu, Duan Xiaodong, Wang + Yang, Wang JiYe, Wang Lei, Zhao Zhifeng, Chen Wei, Wang Wei, Wang + Jilong, Du Yuejing, Tan XiaoSheng, Chen Shangyi, Huang Chenqing, + Ma Yan, Li Xing, Cui Yong, Bi Jun, Duan Haixing, Marc Blanchet, + Andrew Sullivan, Suzanne Wolf, Terry Manderson, Geoff Huston, Jaap + Akkerhuis, Kaveh Ranjbar, Jun Murai, Paul Wilson, and Kilnam + Chonm. + + The authors also acknowledge the assistance of the Independent + Submissions Editorial Board, and of the following reviewers whose + opinions helped improve the clarity of this document: + + Joe Abley, Paul Mockapetris, and Subramanian Moonesamy. + + + + + + + + + + + + + + + + + + + +Song, et al. Informational [Page 38] + +RFC 8483 Yeti DNS Testbed October 2018 + + +Authors' Addresses + + Linjian Song (editor) + Beijing Internet Institute + 2nd Floor, Building 5, No.58 Jing Hai Wu Lu, BDA + Beijing 100176 + China + Email: songlinjian@gmail.com + URI: http://www.biigroup.com/ + + + Dong Liu + Beijing Internet Institute + 2nd Floor, Building 5, No.58 Jing Hai Wu Lu, BDA + Beijing 100176 + China + Email: dliu@biigroup.com + URI: http://www.biigroup.com/ + + + Paul Vixie + TISF + 11400 La Honda Road + Woodside, California 94062 + United States of America + Email: vixie@tisf.net + URI: http://www.redbarn.org/ + + + Akira Kato + Keio University/WIDE Project + Graduate School of Media Design, 4-1-1 Hiyoshi, Kohoku + Yokohama 223-8526 + Japan + Email: kato@wide.ad.jp + URI: http://www.kmd.keio.ac.jp/ + + + Shane Kerr + Antoon Coolenlaan 41 + Uithoorn 1422 GN + The Netherlands + Email: shane@time-travellers.org + + + + + + + + +Song, et al. Informational [Page 39] + |