diff options
author | Thomas Voss <mail@thomasvoss.com> | 2024-11-27 20:54:24 +0100 |
---|---|---|
committer | Thomas Voss <mail@thomasvoss.com> | 2024-11-27 20:54:24 +0100 |
commit | 4bfd864f10b68b71482b35c818559068ef8d5797 (patch) | |
tree | e3989f47a7994642eb325063d46e8f08ffa681dc /doc/rfc/rfc4063.txt | |
parent | ea76e11061bda059ae9f9ad130a9895cc85607db (diff) |
doc: Add RFC documents
Diffstat (limited to 'doc/rfc/rfc4063.txt')
-rw-r--r-- | doc/rfc/rfc4063.txt | 619 |
1 files changed, 619 insertions, 0 deletions
diff --git a/doc/rfc/rfc4063.txt b/doc/rfc/rfc4063.txt new file mode 100644 index 0000000..8655707 --- /dev/null +++ b/doc/rfc/rfc4063.txt @@ -0,0 +1,619 @@ + + + + + + +Network Working Group V. Manral +Request for Comments: 4063 SiNett Corp. +Category: Informational R. White + Cisco Systems + A. Shaikh + AT&T Labs (Research) + April 2005 + + + Considerations When Using Basic OSPF Convergence Benchmarks + +Status of This Memo + + This memo provides information for the Internet community. It does + not specify an Internet standard of any kind. Distribution of this + memo is unlimited. + +Copyright Notice + + Copyright (C) The Internet Society (2005). + +Abstract + + This document discusses the applicability of various tests for + measuring single router control plane convergence, specifically in + regard to the Open Shortest First (OSPF) protocol. There are two + general sections in this document, the first discusses advantages and + limitations of specific OSPF convergence tests, and the second + discusses more general pitfalls to be considered when routing + protocol convergence is tested. + +1. Introduction + + There is a growing interest in testing single router control plane + convergence for routing protocols, and many people are looking at + testing methodologies that can provide information on how long it + takes for a network to converge after various network events occur. + It is important to consider the framework within which any given + convergence test is executed when one attempts to apply the results + of the testing, since the framework can have a major impact on the + results. For instance, determining when a network is converged, what + parts of the router's operation are considered within the testing, + and other such things will have a major impact on the apparent + performance that routing protocols provide. + + + + + + + +Manral, et al. Informational [Page 1] + +RFC 4063 Considerations in OSPF Benchmarking April 2005 + + + This document describes in detail various benefits and pitfalls of + tests described in [BENCHMARK]. It also explains how such + measurements can be useful for providers and the research community. + + NOTE: In this document, the word "convergence" refers to single + router control plane convergence [TERM]. + +2. Advantages of Such Measurement + + o To be able to compare the iterations of a protocol + implementation. It is often useful to be able to compare the + performance of two iterations of a given implementation of a + protocol in order to determine where improvements have been made + and where further improvements can be made. + + o To understand, given a set of parameters (network conditions), + how a particular implementation on a particular device will + perform. For instance, if you were trying to decide the + processing power (size of device) required in a certain location + within a network, you could emulate the conditions that will + exist at that point in the network and use the test described to + measure the performance of several different routers. The + results of these tests can provide one possible data point for + an intelligent decision. + + If the device being tested is to be deployed in a running + network, using routes taken from the network where the equipment + is to be deployed rather than some generated topology in these + tests will yield results that are closer to the real performance + of the device. Care should be taken to emulate or take routes + from the actual location in the network where the device will be + (or would be) deployed. For instance, one set of routes may be + taken from an ABR, one set from an area 0 only router, various + sets from stub area, another set from various normal areas, etc. + + o To measure the performance of an OSPF implementation in a wide + variety of scenarios. + + o To be used as parameters in OSPF simulations by researchers. It + may sometimes be required for certain kinds of research to + measure the individual delays of each parameter within an OSPF + implementation. These delays can be measured using the methods + defined in [BENCHMARK]. + + o To help optimize certain configurable parameters. It may + sometimes be helpful for operators to know the delay required + for individual tasks in order to optimize the resource usage in + the network. For example, if the processing time on a router is + + + +Manral, et al. Informational [Page 2] + +RFC 4063 Considerations in OSPF Benchmarking April 2005 + + + found to be x seconds, determining the rate at which to flood + LSAs to that router would be helpful so as not to overload the + network. + +3. Assumptions Made and Limitations of Such Measurements + + o The interactions of convergence and forwarding; testing is + restricted to events occurring within the control plane. + Forwarding performance is the primary focus in [INTERCONNECT], + and it is expected to be dealt with in work that ensues from + [FIB-TERM]. + + o Duplicate LSAs are Acknowledged Immediately. A few tests rely + on the property that duplicate LSA Acknowledgements are not + delayed but are done immediately. However, if an implementation + does not acknowledge duplicate LSAs immediately on receipt, the + testing methods presented in [BENCHMARK] could give inaccurate + measurements. + + o It is assumed that SPF is non-preemptive. If SPF is implemented + so that it can (and will be) preempted, the SPF measurements + taken in [BENCHMARK] would include the times that the SPF + process is not running, thus giving inaccurate measurements. + ([BENCHMARK] measures the total time taken for SPF to run, not + the amount of time that SPF actually spends on the device's + processor.) + + o Some implementations may be multithreaded or use a + multiprocess/multirouter model of OSPF. If because of this any + of the assumptions made during measurement are violated in such + a model, measurements could be inaccurate. + + o The measurements resulting from the tests in [BENCHMARK] may not + provide the information required to deploy a device in a large- + scale network. The tests described focus on individual + components of an OSPF implementation's performance, and it may + be difficult to combine the measurements in a way that + accurately depicts a device's performance in a large-scale + network. Further research is required in this area. + + o The measurements described in [BENCHMARK] should be used with + great care when comparing two different implementations of OSPF + from two different vendors. For instance, there are many other + factors than convergence speed that need to be taken into + consideration when comparing different vendors' products. One + difficulty is aligning the resources available on one device to + the resources available on another. + + + + +Manral, et al. Informational [Page 3] + +RFC 4063 Considerations in OSPF Benchmarking April 2005 + + +4. Observations on the Tests Described in [BENCHMARK] + + Some observations recorded while implementing the tests described in + [BENCHMARK] are noted in this section. + +4.1. Measuring the SPF Processing Time Externally + + The most difficult test to perform is the external measurement of the + time required to perform an SPF calculation because the amount of + time between the first LSA that indicates a topology change and the + duplicate LSA is critical. If the duplicate LSA is sent too quickly, + it may be received before the device being tested actually begins + running SPF on the network change information. If the delay between + the two LSAs is too long, the device may finish SPF processing before + receiving the duplicate LSA. It is important to closely investigate + any delays between the receipt of an LSA and the beginning of an SPF + calculation in the tested device; multiple tests with various delays + might be required to determine what delay needs to be used to measure + the SPF calculation time accurately. + + Some implementations may force two intervals, the SPF hold time and + the SPF delay, between successive SPF calculations. If an SPF hold + time exists, it should be subtracted from the total SPF execution + time. If an SPF delay exists, it should be noted in the test + results. + +4.2. Noise in the Measurement Device + + The device on which measurements are taken (not the device being + tested) also adds noise to the test results, primarily in the form of + delay in packet processing and measurement output. The largest + source of noise is generally the delay between the receipt of packets + by the measuring device and the receipt of information about the + packet by the device's output, where the event can be measured. The + following steps may be taken to reduce this sampling noise: + + o Increasing the number of samples taken will generally improve + the tester's ability to determine what is noise, and to remove + it from the results. This applies to the DUT as well. + + o Try to take time-stamp for a packet as early as possible. + Depending on the operating system being used on the box, one can + instrument the kernel to take the time-stamp when the interrupt + is processed. This does not eliminate the noise completely, but + at least reduces it. + + o Keep the measurement box as lightly loaded as possible. This + applies to the DUT as well. + + + +Manral, et al. Informational [Page 4] + +RFC 4063 Considerations in OSPF Benchmarking April 2005 + + + o Having an estimate of noise can also be useful. + + The DUT also adds noise to the measurement. + +4.3. Gaining an Understanding of the Implementation Improves + Measurements + + Although the tester will (generally) not have access to internal + information about the OSPF implementation being tested using + [BENCHMARK], the more thorough the tester's knowledge of the + implementation is, the more accurate the results of the tests will + be. For instance, in some implementations, the installation of + routes in local routing tables may occur while the SPF is being + calculated, dramatically impacting the time required to calculate the + SPF. + +4.4. Gaining an Understanding of the Tests Improves Measurements + + One method that can be used to become familiar with the tests + described in [BENCHMARK] is to perform the tests on an OSPF + implementation for which all the internal details are available. + Although there is no assurance that any two implementations will be + similar, this will provide a better understanding of the tests + themselves. + +5. LSA and Destination Mix + + In many OSPF benchmark tests, a generator injecting a number of LSAs + is called for. There are several areas in which injected LSAs can be + varied in testing: + + o The number of destinations represented by the injected LSAs + + Each destination represents a single reachable IP network; these + will be leaf nodes on the shortest path tree. The primary + impact to performance should be the time required to insert + destinations in the local routing table and handling the memory + required to store the data. + + o The types of LSAs injected + + There are several types of LSAs that would be acceptable under + different situations; within an area, for instance, types 1, 2, + 3, 4, and 5 are likely to be received by a router. Within a + not-so-stubby area, however, type-7 LSAs would replace the + type-5 LSAs received. These sorts of characterizations are + important to note in any test results. + + + + +Manral, et al. Informational [Page 5] + +RFC 4063 Considerations in OSPF Benchmarking April 2005 + + + o The number of LSAs injected + + Within any injected set of information, the number of each type + of LSA injected is also important. This will impact the + shortest path algorithm's ability to handle large numbers of + nodes, large shortest path first trees, etc. + + o The order of LSA injection + + The order in which LSAs are injected should not favor any given + data structure used for storing the LSA database on the device + being tested. For instance, AS-External LSAs have AS wide + flooding scope; any type-5 LSA originated is immediately flooded + to all neighbors. However, the type-4 LSA, which announces the + ASBR as a border router, is originated in an area at SPF time + (by ABRs on the edge of the area in which the ASBR is). If SPF + isn't scheduled immediately on the ABRs originating the type-4 + LSA, the type-4 LSA is sent after the type-5 LSA's reach a + router in the adjacent area. Therefore, routes to the external + destinations aren't immediately added to the routers in the + other areas. When the routers that already have the type 5s + receive the type-4 LSA, all the external routes are added to the + tree at the same time. This timing could produce different + results than a router receiving a type 4 indicating the presence + of a border router, followed by the type 5s originated by that + border router. + + The ordering can be changed in various tests to provide insight + into the efficiency of storage within the DUT. Any such changes + in ordering should be noted in test results. + +6. Tree Shape and the SPF Algorithm + + The complexity of Dijkstra's algorithm depends on the data structure + used for storing vertices with their current minimum distances from + the source; the simplest structure is a list of vertices currently + reachable from the source. In a simple list of vertices, finding the + minimum cost vertex would then take O(size of the list). There will + be O(n) such operations if we assume that all the vertices are + ultimately reachable from the source. Moreover, after the vertex + with minimum cost is found, the algorithm iterates through all the + edges of the vertex and updates the cost of other vertices. With an + adjacency list representation, this step, when iterated over all the + vertices, would take O(E) time, with E being the number of edges in + the graph. Thus, the overall running time is: + + O(sum(i:1, n)(size(list at level i) + E). + + + + +Manral, et al. Informational [Page 6] + +RFC 4063 Considerations in OSPF Benchmarking April 2005 + + + So everything boils down to the size(list at level i). + + If the graph is linear, + + root + | + 1 + | + 2 + | + 3 + | + 4 + | + 5 + | + 6 + + and source is a vertex on the end, then size(list at level i) = 1 for + all i. Moreover, E = n - 1. Therefore, running time is O(n). + + If the graph is a balanced binary tree, + + root + / \ + 1 2 + / \ / \ + 3 4 5 6 + + size(list at level i) is a little complicated. First, it increases + by 1 at each level up to a certain number, and then it goes down by + 1. If we assume that the tree is a complete tree (as shown above) + with k levels (1 to k), then size(list) goes on like this: 1, 2, 3, + + Then the number of edges E is still n - 1. It then turns out that + the run-time is O(n^2) for such a tree. + + If the graph is a complete graph (fully-connected mesh), then + size(list at level i) = n - i. Number of edges E = O(n^2). + Therefore, run-time is O(n^2). + + Therefore, the performance of the shortest path first algorithm used + to compute the best paths through the network is dependent on the + construction of the tree. The best practice would be to try to make + any emulated network look as much like a real network as possible, + especially in the area of the tree depth, the meshiness of the + + + + + +Manral, et al. Informational [Page 7] + +RFC 4063 Considerations in OSPF Benchmarking April 2005 + + + network, the number of stub links versus transit links, and the + number of connections and nodes to process at each level within the + original tree. + +7. Topology Generation + + As the size of networks grows, it becomes more and more difficult to + actually create a large-scale network on which to test the properties + of routing protocols and their implementations. In general, network + emulators are used to provide emulated topologies that can be + advertised to a device with varying conditions. Route generators + tend to be either a specialized device, a piece of software which + runs on a router, or a process that runs on another operating system, + such as Linux or another variant of Unix. + + Some of the characteristics of this device should be as follows: + + o The ability to connect to several devices using both point-to- + point and broadcast high-speed media. Point-to-point links can + be emulated with high-speed Ethernet as long as there is no hub + or other device between the DUT and the route generator, and the + link is configured as a point-to-point link within OSPF + [BROADCAST-P2P]. + + o The ability to create a set of LSAs that appear to be a logical, + realistic topology. For instance, the generator should be able + to mix the number of point-to-point and broadcast links within + the emulated topology and to inject varying numbers of + externally reachable destinations. + + o The ability to withdraw and add routing information into and + from the emulated topology to emulate flapping links. + + o The ability to randomly order the LSAs representing the emulated + topology as they are advertised. + + o The ability to log or otherwise measure the time between packets + transmitted and received. + + o The ability to change the rate at which OSPF LSAs are + transmitted. + + o The generator and the collector should be fast enough that they + are not bottlenecks. The devices should also have a degree of + granularity of measurement at least as small as is desired from + the test results. + + + + + +Manral, et al. Informational [Page 8] + +RFC 4063 Considerations in OSPF Benchmarking April 2005 + + +8. Security Considerations + + This document does not modify the underlying security considerations + in [OSPF]. + +9. Acknowledgements + + Thanks to Howard Berkowitz (hcb@clark.net) and the rest of the BGP + benchmarking team for their support and to Kevin Dubray + (kdubray@juniper.net), who realized the need for this document. + +10. Normative References + + [BENCHMARK] Manral, V., White, R., and A. Shaikh, "Benchmarking + Basic OSPF Single Router Control Plane Convergence", + RFC 4061, April 2005. + + [TERM] Manral, V., White, R., and A. Shaikh, "OSPF + Benchmarking Terminology and Concepts", RFC 4062, + April 2005. + + [OSPF] Moy, J., "OSPF Version 2", STD 54, RFC 2328, April + 1998. + +11. Informative References + + [INTERCONNECT] Bradner, S. and J. McQuaid, "Benchmarking Methodology + for Network Interconnect Devices", RFC 2544, March + 1999. + + [FIB-TERM] Trotter, G., "Terminology for Forwarding Information + Base (FIB) based Router Performance", RFC 3222, + December 2001. + + [BROADCAST-P2P] Shen, Naiming, et al., "Point-to-point operation over + LAN in link-state routing protocols", Work in + Progress, August, 2003. + + + + + + + + + + + + + + +Manral, et al. Informational [Page 9] + +RFC 4063 Considerations in OSPF Benchmarking April 2005 + + +Authors' Addresses + + Vishwas Manral + SiNett Corp, + Ground Floor, + Embassy Icon Annexe, + 2/1, Infantry Road, + Bangalore, India + + EMail: vishwas@sinett.com + + + Russ White + Cisco Systems, Inc. + 7025 Kit Creek Rd. + Research Triangle Park, NC 27709 + + EMail: riw@cisco.com + + + Aman Shaikh + AT&T Labs (Research) + 180 Park Av, PO Box 971 + Florham Park, NJ 07932 + + EMail: ashaikh@research.att.com + + + + + + + + + + + + + + + + + + + + + + + + + +Manral, et al. Informational [Page 10] + +RFC 4063 Considerations in OSPF Benchmarking April 2005 + + +Full Copyright Statement + + Copyright (C) The Internet Society (2005). + + This document is subject to the rights, licenses and restrictions + contained in BCP 78, and except as set forth therein, the authors + retain all their rights. + + This document and the information contained herein are provided on an + "AS IS" basis and THE CONTRIBUTOR, THE ORGANIZATION HE/SHE REPRESENTS + OR IS SPONSORED BY (IF ANY), THE INTERNET SOCIETY AND THE INTERNET + ENGINEERING TASK FORCE DISCLAIM ALL WARRANTIES, EXPRESS OR IMPLIED, + INCLUDING BUT NOT LIMITED TO ANY WARRANTY THAT THE USE OF THE + INFORMATION HEREIN WILL NOT INFRINGE ANY RIGHTS OR ANY IMPLIED + WARRANTIES OF MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. + +Intellectual Property + + The IETF takes no position regarding the validity or scope of any + Intellectual Property Rights or other rights that might be claimed to + pertain to the implementation or use of the technology described in + this document or the extent to which any license under such rights + might or might not be available; nor does it represent that it has + made any independent effort to identify any such rights. Information + on the procedures with respect to rights in RFC documents can be + found in BCP 78 and BCP 79. + + Copies of IPR disclosures made to the IETF Secretariat and any + assurances of licenses to be made available, or the result of an + attempt made to obtain a general license or permission for the use of + such proprietary rights by implementers or users of this + specification can be obtained from the IETF on-line IPR repository at + http://www.ietf.org/ipr. + + The IETF invites any interested party to bring to its attention any + copyrights, patents or patent applications, or other proprietary + rights that may cover technology that may be required to implement + this standard. Please address the information to the IETF at ietf- + ipr@ietf.org. + +Acknowledgement + + Funding for the RFC Editor function is currently provided by the + Internet Society. + + + + + + + +Manral, et al. Informational [Page 11] + |