diff options
Diffstat (limited to 'doc/rfc/rfc2544.txt')
-rw-r--r-- | doc/rfc/rfc2544.txt | 1739 |
1 files changed, 1739 insertions, 0 deletions
diff --git a/doc/rfc/rfc2544.txt b/doc/rfc/rfc2544.txt new file mode 100644 index 0000000..9d8372a --- /dev/null +++ b/doc/rfc/rfc2544.txt @@ -0,0 +1,1739 @@ + + + + + + +Network Working Group S. Bradner +Request for Comments: 2544 Harvard University +Obsoletes: 1944 J. McQuaid +Category: Informational NetScout Systems + March 1999 + + + Benchmarking Methodology for Network Interconnect Devices + +Status of this Memo + + This memo provides information for the Internet community. It does + not specify an Internet standard of any kind. Distribution of this + memo is unlimited. + +Copyright Notice + + Copyright (C) The Internet Society (1999). All Rights Reserved. + +IESG Note + + This document is a republication of RFC 1944 correcting the values + for the IP addresses which were assigned to be used as the default + addresses for networking test equipment. (See section C.2.2 ). This + RFC replaces and obsoletes RFC 1944. + +Abstract + + This document discusses and defines a number of tests that may be + used to describe the performance characteristics of a network + interconnecting device. In addition to defining the tests this + document also describes specific formats for reporting the results of + the tests. Appendix A lists the tests and conditions that we believe + should be included for specific cases and gives additional + information about testing practices. Appendix B is a reference + listing of maximum frame rates to be used with specific frame sizes + on various media and Appendix C gives some examples of frame formats + to be used in testing. + +1. Introduction + + Vendors often engage in "specsmanship" in an attempt to give their + products a better position in the marketplace. This often involves + "smoke & mirrors" to confuse the potential users of the products. + + + + + + + +Bradner & McQuaid Informational [Page 1] + +RFC 2544 Benchmarking Methodology March 1999 + + + This document defines a specific set of tests that vendors can use to + measure and report the performance characteristics of network + devices. The results of these tests will provide the user comparable + data from different vendors with which to evaluate these devices. + + A previous document, "Benchmarking Terminology for Network + Interconnect Devices" (RFC 1242), defined many of the terms that are + used in this document. The terminology document should be consulted + before attempting to make use of this document. + +2. Real world + + In producing this document the authors attempted to keep in mind the + requirement that apparatus to perform the described tests must + actually be built. We do not know of "off the shelf" equipment + available to implement all of the tests but it is our opinion that + such equipment can be constructed. + +3. Tests to be run + + There are a number of tests described in this document. Not all of + the tests apply to all types of devices under test (DUTs). Vendors + should perform all of the tests that can be supported by a specific + type of product. The authors understand that it will take a + considerable period of time to perform all of the recommended tests + nder all of the recommended conditions. We believe that the results + are worth the effort. Appendix A lists some of the tests and + conditions that we believe should be included for specific cases. + +4. Evaluating the results + + Performing all of the recommended tests will result in a great deal + of data. Much of this data will not apply to the evaluation of the + devices under each circumstance. For example, the rate at which a + router forwards IPX frames will be of little use in selecting a + router for an environment that does not (and will not) support that + protocol. Evaluating even that data which is relevant to a + particular network installation will require experience which may not + be readily available. Furthermore, selection of the tests to be run + and evaluation of the test data must be done with an understanding of + generally accepted testing practices regarding repeatability, + variance and statistical significance of small numbers of trials. + + + + + + + + + +Bradner & McQuaid Informational [Page 2] + +RFC 2544 Benchmarking Methodology March 1999 + + +5. Requirements + + In this document, the words that are used to define the significance + of each particular requirement are capitalized. These words are: + + * "MUST" This word, or the words "REQUIRED" and "SHALL" mean that + the item is an absolute requirement of the specification. + + * "SHOULD" This word or the adjective "RECOMMENDED" means that + there may exist valid reasons in particular circumstances to + ignore this item, but the full implications should be + understood and the case carefully weighed before choosing a + different course. + + * "MAY" This word or the adjective "OPTIONAL" means that this + item is truly optional. One vendor may choose to include the + item because a particular marketplace requires it or because it + enhances the product, for example; another vendor may omit the + same item. + + An implementation is not compliant if it fails to satisfy one or more + of the MUST requirements for the protocols it implements. An + implementation that satisfies all the MUST and all the SHOULD + requirements for its protocols is said to be "unconditionally + compliant"; one that satisfies all the MUST requirements but not all + the SHOULD requirements for its protocols is said to be + "conditionally compliant". + +6. Test set up + + The ideal way to implement this series of tests is to use a tester + with both transmitting and receiving ports. Connections are made + from the sending ports of the tester to the receiving ports of the + DUT and from the sending ports of the DUT back to the tester. (see + Figure 1) Since the tester both sends the test traffic and receives + it back, after the traffic has been forwarded but the DUT, the tester + can easily determine if all of the transmitted packets were received + and verify that the correct packets were received. The same + functionality can be obtained with separate transmitting and + receiving devices (see Figure 2) but unless they are remotely + controlled by some computer in a way that simulates the single + tester, the labor required to accurately perform some of the tests + (particularly the throughput test) can be prohibitive. + + + + + + + + +Bradner & McQuaid Informational [Page 3] + +RFC 2544 Benchmarking Methodology March 1999 + + + +------------+ + | | + +------------| tester |<-------------+ + | | | | + | +------------+ | + | | + | +------------+ | + | | | | + +----------->| DUT |--------------+ + | | + +------------+ + Figure 1 + + +--------+ +------------+ +----------+ + | | | | | | + | sender |-------->| DUT |--------->| receiver | + | | | | | | + +--------+ +------------+ +----------+ + Figure 2 + +6.1 Test set up for multiple media types + + Two different setups could be used to test a DUT which is used in + real-world networks to connect networks of differing media type, + local Ethernet to a backbone FDDI ring for example. The tester could + support both media types in which case the set up shown in Figure 1 + would be used. + + Two identical DUTs are used in the other test set up. (see Figure 3) + In many cases this set up may more accurately simulate the real + world. For example, connecting two LANs together with a WAN link or + high speed backbone. This set up would not be as good at simulating + a system where clients on a Ethernet LAN were interacting with a + server on an FDDI backbone. + + +-----------+ + | | + +---------------------| tester |<---------------------+ + | | | | + | +-----------+ | + | | + | +----------+ +----------+ | + | | | | | | + +------->| DUT 1 |-------------->| DUT 2 |---------+ + | | | | + +----------+ +----------+ + + Figure 3 + + + +Bradner & McQuaid Informational [Page 4] + +RFC 2544 Benchmarking Methodology March 1999 + + + +7. DUT set up + + Before starting to perform the tests, the DUT to be tested MUST be + configured following the instructions provided to the user. + Specifically, it is expected that all of the supported protocols will + be configured and enabled during this set up (See Appendix A). It is + expected that all of the tests will be run without changing the + configuration or setup of the DUT in any way other than that required + to do the specific test. For example, it is not acceptable to change + the size of frame handling buffers between tests of frame handling + rates or to disable all but one transport protocol when testing the + throughput of that protocol. It is necessary to modify the + configuration when starting a test to determine the effect of filters + on throughput, but the only change MUST be to enable the specific + filter. The DUT set up SHOULD include the normally recommended + routing update intervals and keep alive frequency. The specific + version of the software and the exact DUT configuration, including + what functions are disabled, used during the tests MUST be included + as part of the report of the results. + +8. Frame formats + + The formats of the test frames to use for TCP/IP over Ethernet are + shown in Appendix C: Test Frame Formats. These exact frame formats + SHOULD be used in the tests described in this document for this + protocol/media combination and that these frames will be used as a + template for testing other protocol/media combinations. The specific + formats that are used to define the test frames for a particular test + series MUST be included in the report of the results. + +9. Frame sizes + + All of the described tests SHOULD be performed at a number of frame + sizes. Specifically, the sizes SHOULD include the maximum and minimum + legitimate sizes for the protocol under test on the media under test + and enough sizes in between to be able to get a full characterization + of the DUT performance. Except where noted, at least five frame + sizes SHOULD be tested for each test condition. + + Theoretically the minimum size UDP Echo request frame would consist + of an IP header (minimum length 20 octets), a UDP header (8 octets) + and whatever MAC level header is required by the media in use. The + theoretical maximum frame size is determined by the size of the + length field in the IP header. In almost all cases the actual + maximum and minimum sizes are determined by the limitations of the + media. + + + + +Bradner & McQuaid Informational [Page 5] + +RFC 2544 Benchmarking Methodology March 1999 + + + In theory it would be ideal to distribute the frame sizes in a way + that would evenly distribute the theoretical frame rates. These + recommendations incorporate this theory but specify frame sizes which + are easy to understand and remember. In addition, many of the same + frame sizes are specified on each of the media types to allow for + easy performance comparisons. + + Note: The inclusion of an unrealistically small frame size on some of + the media types (i.e. with little or no space for data) is to help + characterize the per-frame processing overhead of the DUT. + +9.1 Frame sizes to be used on Ethernet + + 64, 128, 256, 512, 1024, 1280, 1518 + + These sizes include the maximum and minimum frame sizes permitted by + the Ethernet standard and a selection of sizes between these extremes + with a finer granularity for the smaller frame sizes and higher frame + rates. + +9.2 Frame sizes to be used on 4Mb and 16Mb token ring + + 54, 64, 128, 256, 1024, 1518, 2048, 4472 + + The frame size recommendations for token ring assume that there is no + RIF field in the frames of routed protocols. A RIF field would be + present in any direct source route bridge performance test. The + minimum size frame for UDP on token ring is 54 octets. The maximum + size of 4472 octets is recommended for 16Mb token ring instead of the + theoretical size of 17.9Kb because of the size limitations imposed by + many token ring interfaces. The reminder of the sizes are selected + to permit direct comparisons with other types of media. An IP (i.e. + not UDP) frame may be used in addition if a higher data rate is + desired, in which case the minimum frame size is 46 octets. + +9.3 Frame sizes to be used on FDDI + + 54, 64, 128, 256, 1024, 1518, 2048, 4472 + + The minimum size frame for UDP on FDDI is 53 octets, the minimum size + of 54 is recommended to allow direct comparison to token ring + performance. The maximum size of 4472 is recommended instead of the + theoretical maximum size of 4500 octets to permit the same type of + comparison. An IP (i.e. not UDP) frame may be used in addition if a + higher data rate is desired, in which case the minimum frame size is + 45 octets. + + + + + +Bradner & McQuaid Informational [Page 6] + +RFC 2544 Benchmarking Methodology March 1999 + + +9.4 Frame sizes in the presence of disparate MTUs + + When the interconnect DUT supports connecting links with disparate + MTUs, the frame sizes for the link with the *larger* MTU SHOULD be + used, up to the limit of the protocol being tested. If the + interconnect DUT does not support the fragmenting of frames in the + presence of MTU mismatch, the forwarding rate for that frame size + shall be reported as zero. + + For example, the test of IP forwarding with a bridge or router that + joins FDDI and Ethernet should use the frame sizes of FDDI when going + from the FDDI to the Ethernet link. If the bridge does not support IP + fragmentation, the forwarding rate for those frames too large for + Ethernet should be reported as zero. + +10. Verifying received frames + + The test equipment SHOULD discard any frames received during a test + run that are not actual forwarded test frames. For example, keep- + alive and routing update frames SHOULD NOT be included in the count + of received frames. In any case, the test equipment SHOULD verify + the length of the received frames and check that they match the + expected length. + + Preferably, the test equipment SHOULD include sequence numbers in the + transmitted frames and check for these numbers on the received + frames. If this is done, the reported results SHOULD include in + addition to the number of frames dropped, the number of frames that + were received out of order, the number of duplicate frames received + and the number of gaps in the received frame numbering sequence. + This functionality is required for some of the described tests. + +11. Modifiers + + It might be useful to know the DUT performance under a number of + conditions; some of these conditions are noted below. The reported + results SHOULD include as many of these conditions as the test + equipment is able to generate. The suite of tests SHOULD be first + run without any modifying conditions and then repeated under each of + the conditions separately. To preserve the ability to compare the + results of these tests any frames that are required to generate the + modifying conditions (management queries for example) will be + included in the same data stream as the normal test frames in place + of one of the test frames and not be supplied to the DUT on a + separate network port. + + + + + + +Bradner & McQuaid Informational [Page 7] + +RFC 2544 Benchmarking Methodology March 1999 + + +11.1 Broadcast frames + + In most router designs special processing is required when frames + addressed to the hardware broadcast address are received. In bridges + (or in bridge mode on routers) these broadcast frames must be flooded + to a number of ports. The stream of test frames SHOULD be augmented + with 1% frames addressed to the hardware broadcast address. The + frames sent to the broadcast address should be of a type that the + router will not need to process. The aim of this test is to + determine if there is any effect on the forwarding rate of the other + data in the stream. The specific frames that should be used are + included in the test frame format document. The broadcast frames + SHOULD be evenly distributed throughout the data stream, for example, + every 100th frame. + + The same test SHOULD be performed on bridge-like DUTs but in this + case the broadcast packets will be processed and flooded to all + outputs. + + It is understood that a level of broadcast frames of 1% is much + higher than many networks experience but, as in drug toxicity + evaluations, the higher level is required to be able to gage the + effect which would otherwise often fall within the normal variability + of the system performance. Due to design factors some test equipment + will not be able to generate a level of alternate frames this low. + In these cases the percentage SHOULD be as small as the equipment can + provide and that the actual level be described in the report of the + test results. + +11.2 Management frames + + Most data networks now make use of management protocols such as SNMP. + In many environments there can be a number of management stations + sending queries to the same DUT at the same time. + + The stream of test frames SHOULD be augmented with one management + query as the first frame sent each second during the duration of the + trial. The result of the query must fit into one response frame. The + response frame SHOULD be verified by the test equipment. One example + of the specific query frame that should be used is shown in Appendix + C. + +11.3 Routing update frames + + The processing of dynamic routing protocol updates could have a + significant impact on the ability of a router to forward data frames. + The stream of test frames SHOULD be augmented with one routing update + frame transmitted as the first frame transmitted during the trial. + + + +Bradner & McQuaid Informational [Page 8] + +RFC 2544 Benchmarking Methodology March 1999 + + + Routing update frames SHOULD be sent at the rate specified in + Appendix C for the specific routing protocol being used in the test. + Two routing update frames are defined in Appendix C for the TCP/IP + over Ethernet example. The routing frames are designed to change the + routing to a number of networks that are not involved in the + forwarding of the test data. The first frame sets the routing table + state to "A", the second one changes the state to "B". The frames + MUST be alternated during the trial. + + The test SHOULD verify that the routing update was processed by the + DUT. + +11.4 Filters + + Filters are added to routers and bridges to selectively inhibit the + forwarding of frames that would normally be forwarded. This is + usually done to implement security controls on the data that is + accepted between one area and another. Different products have + different capabilities to implement filters. + + The DUT SHOULD be first configured to add one filter condition and + the tests performed. This filter SHOULD permit the forwarding of the + test data stream. In routers this filter SHOULD be of the form: + + forward input_protocol_address to output_protocol_address + + In bridges the filter SHOULD be of the form: + + forward destination_hardware_address + + The DUT SHOULD be then reconfigured to implement a total of 25 + filters. The first 24 of these filters SHOULD be of the form: + + block input_protocol_address to output_protocol_address + + The 24 input and output protocol addresses SHOULD not be any that are + represented in the test data stream. The last filter SHOULD permit + the forwarding of the test data stream. By "first" and "last" we + mean to ensure that in the second case, 25 conditions must be checked + before the data frames will match the conditions that permit the + forwarding of the frame. Of course, if the DUT reorders the filters + or does not use a linear scan of the filter rules the effect of the + sequence in which the filters are input is properly lost. + + The exact filters configuration command lines used SHOULD be included + with the report of the results. + + + + + +Bradner & McQuaid Informational [Page 9] + +RFC 2544 Benchmarking Methodology March 1999 + + +11.4.1 Filter Addresses + + Two sets of filter addresses are required, one for the single filter + case and one for the 25 filter case. + + The single filter case should permit traffic from IP address + 198.18.1.2 to IP address 198.19.65.2 and deny all other traffic. + + The 25 filter case should follow the following sequence. + + deny aa.ba.1.1 to aa.ba.100.1 + deny aa.ba.2.2 to aa.ba.101.2 + deny aa.ba.3.3 to aa.ba.103.3 + ... + deny aa.ba.12.12 to aa.ba.112.12 + allow aa.bc.1.2 to aa.bc.65.1 + deny aa.ba.13.13 to aa.ba.113.13 + deny aa.ba.14.14 to aa.ba.114.14 + ... + deny aa.ba.24.24 to aa.ba.124.24 + deny all else + + + All previous filter conditions should be cleared from the router + before this sequence is entered. The sequence is selected to test to + see if the router sorts the filter conditions or accepts them in the + order that they were entered. Both of these procedures will result + in a greater impact on performance than will some form of hash + coding. + +12. Protocol addresses + + It is easier to implement these tests using a single logical stream + of data, with one source protocol address and one destination + protocol address, and for some conditions like the filters described + above, a practical requirement. Networks in the real world are not + limited to single streams of data. The test suite SHOULD be first run + with a single protocol (or hardware for bridge tests) source and + destination address pair. The tests SHOULD then be repeated with + using a random destination address. While testing routers the + addresses SHOULD be random and uniformly distributed over a range of + 256 networks and random and uniformly distributed over the full MAC + range for bridges. The specific address ranges to use for IP are + shown in Appendix C. + + + + + + + +Bradner & McQuaid Informational [Page 10] + +RFC 2544 Benchmarking Methodology March 1999 + + +13. Route Set Up + + It is not reasonable that all of the routing information necessary to + forward the test stream, especially in the multiple address case, + will be manually set up. At the start of each trial a routing update + MUST be sent to the DUT. This routing update MUST include all of the + network addresses that will be required for the trial. All of the + addresses SHOULD resolve to the same "next-hop". Normally this will + be the address of the receiving side of the test equipment. This + routing update will have to be repeated at the interval required by + the routing protocol being used. An example of the format and + repetition interval of the update frames is given in Appendix C. + +14. Bidirectional traffic + + Normal network activity is not all in a single direction. To test + the bidirectional performance of a DUT, the test series SHOULD be run + with the same data rate being offered from each direction. The sum of + the data rates should not exceed the theoretical limit for the media. + +15. Single stream path + + The full suite of tests SHOULD be run along with whatever modifier + conditions that are relevant using a single input and output network + port on the DUT. If the internal design of the DUT has multiple + distinct pathways, for example, multiple interface cards each with + multiple network ports, then all possible types of pathways SHOULD be + tested separately. + +16. Multi-port + + Many current router and bridge products provide many network ports in + the same module. In performing these tests first half of the ports + are designated as "input ports" and half are designated as "output + ports". These ports SHOULD be evenly distributed across the DUT + architecture. For example if a DUT has two interface cards each of + which has four ports, two ports on each interface card are designated + as input and two are designated as output. The specified tests are + run using the same data rate being offered to each of the input + ports. The addresses in the input data streams SHOULD be set so that + a frame will be directed to each of the output ports in sequence so + that all "output" ports will get an even distribution of packets from + this input. The same configuration MAY be used to perform a + bidirectional multi-stream test. In this case all of the ports are + considered both input and output ports and each data stream MUST + consist of frames addressed to all of the other ports. + + + + + +Bradner & McQuaid Informational [Page 11] + +RFC 2544 Benchmarking Methodology March 1999 + + + Consider the following 6 port DUT: + + -------------- + ---------| in A out X|-------- + ---------| in B out Y|-------- + ---------| in C out Z|-------- + -------------- + + The addressing of the data streams for each of the inputs SHOULD be: + + stream sent to input A: + packet to out X, packet to out Y, packet to out Z + stream sent to input B: + packet to out X, packet to out Y, packet to out Z + stream sent to input C + packet to out X, packet to out Y, packet to out Z + + Note that these streams each follow the same sequence so that 3 + packets will arrive at output X at the same time, then 3 packets at + Y, then 3 packets at Z. This procedure ensures that, as in the real + world, the DUT will have to deal with multiple packets addressed to + the same output at the same time. + +17. Multiple protocols + + This document does not address the issue of testing the effects of a + mixed protocol environment other than to suggest that if such tests + are wanted then frames SHOULD be distributed between all of the test + protocols. The distribution MAY approximate the conditions on the + network in which the DUT would be used. + +18. Multiple frame sizes + + This document does not address the issue of testing the effects of a + mixed frame size environment other than to suggest that if such tests + are wanted then frames SHOULD be distributed between all of the + listed sizes for the protocol under test. The distribution MAY + approximate the conditions on the network in which the DUT would be + used. The authors do not have any idea how the results of such a test + would be interpreted other than to directly compare multiple DUTs in + some very specific simulated network. + +19. Testing performance beyond a single DUT. + + In the performance testing of a single DUT, the paradigm can be + described as applying some input to a DUT and monitoring the output. + The results of which can be used to form a basis of characterization + of that device under those test conditions. + + + +Bradner & McQuaid Informational [Page 12] + +RFC 2544 Benchmarking Methodology March 1999 + + + This model is useful when the test input and output are homogenous + (e.g., 64-byte IP, 802.3 frames into the DUT; 64 byte IP, 802.3 + frames out), or the method of test can distinguish between dissimilar + input/output. (E.g., 1518 byte IP, 802.3 frames in; 576 byte, + fragmented IP, X.25 frames out.) + + By extending the single DUT test model, reasonable benchmarks + regarding multiple DUTs or heterogeneous environments may be + collected. In this extension, the single DUT is replaced by a system + of interconnected network DUTs. This test methodology would support + the benchmarking of a variety of device/media/service/protocol + combinations. For example, a configuration for a LAN-to-WAN-to-LAN + test might be: + + (1) 802.3-> DUT 1 -> X.25 @ 64kbps -> DUT 2 -> 802.3 + + Or a mixed LAN configuration might be: + + (2) 802.3 -> DUT 1 -> FDDI -> DUT 2 -> FDDI -> DUT 3 -> 802.3 + + In both examples 1 and 2, end-to-end benchmarks of each system could + be empirically ascertained. Other behavior may be characterized + through the use of intermediate devices. In example 2, the + configuration may be used to give an indication of the FDDI to FDDI + capability exhibited by DUT 2. + + Because multiple DUTs are treated as a single system, there are + limitations to this methodology. For instance, this methodology may + yield an aggregate benchmark for a tested system. That benchmark + alone, however, may not necessarily reflect asymmetries in behavior + between the DUTs, latencies introduce by other apparatus (e.g., + CSUs/DSUs, switches), etc. + + Further, care must be used when comparing benchmarks of different + systems by ensuring that the DUTs' features/configuration of the + tested systems have the appropriate common denominators to allow + comparison. + +20. Maximum frame rate + + The maximum frame rates that should be used when testing LAN + connections SHOULD be the listed theoretical maximum rate for the + frame size on the media. + + + + + + + + +Bradner & McQuaid Informational [Page 13] + +RFC 2544 Benchmarking Methodology March 1999 + + + The maximum frame rate that should be used when testing WAN + connections SHOULD be greater than the listed theoretical maximum + rate for the frame size on that speed connection. The higher rate + for WAN tests is to compensate for the fact that some vendors employ + various forms of header compression. + + A list of maximum frame rates for LAN connections is included in + Appendix B. + +21. Bursty traffic + + It is convenient to measure the DUT performance under steady state + load but this is an unrealistic way to gauge the functioning of a DUT + since actual network traffic normally consists of bursts of frames. + Some of the tests described below SHOULD be performed with both + steady state traffic and with traffic consisting of repeated bursts + of frames. The frames within a burst are transmitted with the + minimum legitimate inter-frame gap. + + The objective of the test is to determine the minimum interval + between bursts which the DUT can process with no frame loss. During + each test the number of frames in each burst is held constant and the + inter-burst interval varied. Tests SHOULD be run with burst sizes of + 16, 64, 256 and 1024 frames. + +22. Frames per token + + Although it is possible to configure some token ring and FDDI + interfaces to transmit more than one frame each time that the token + is received, most of the network devices currently available transmit + only one frame per token. These tests SHOULD first be performed + while transmitting only one frame per token. + + Some current high-performance workstation servers do transmit more + than one frame per token on FDDI to maximize throughput. Since this + may be a common feature in future workstations and servers, + interconnect devices with FDDI interfaces SHOULD be tested with 1, 4, + 8, and 16 frames per token. The reported frame rate SHOULD be the + average rate of frame transmission over the total trial period. + +23. Trial description + + A particular test consists of multiple trials. Each trial returns + one piece of information, for example the loss rate at a particular + input frame rate. Each trial consists of a number of phases: + + a) If the DUT is a router, send the routing update to the "input" + port and pause two seconds to be sure that the routing has settled. + + + +Bradner & McQuaid Informational [Page 14] + +RFC 2544 Benchmarking Methodology March 1999 + + + b) Send the "learning frames" to the "output" port and wait 2 + seconds to be sure that the learning has settled. Bridge learning + frames are frames with source addresses that are the same as the + destination addresses used by the test frames. Learning frames for + other protocols are used to prime the address resolution tables in + the DUT. The formats of the learning frame that should be used are + shown in the Test Frame Formats document. + + c) Run the test trial. + + d) Wait for two seconds for any residual frames to be received. + + e) Wait for at least five seconds for the DUT to restabilize. + +24. Trial duration + + The aim of these tests is to determine the rate continuously + supportable by the DUT. The actual duration of the test trials must + be a compromise between this aim and the duration of the benchmarking + test suite. The duration of the test portion of each trial SHOULD be + at least 60 seconds. The tests that involve some form of "binary + search", for example the throughput test, to determine the exact + result MAY use a shorter trial duration to minimize the length of the + search procedure, but the final determination SHOULD be made with + full length trials. + +25. Address resolution + + The DUT SHOULD be able to respond to address resolution requests sent + by the DUT wherever the protocol requires such a process. + +26. Benchmarking tests: + + Note: The notation "type of data stream" refers to the above + modifications to a frame stream with a constant inter-frame gap, for + example, the addition of traffic filters to the configuration of the + DUT. + +26.1 Throughput + + Objective: To determine the DUT throughput as defined in RFC 1242. + + Procedure: Send a specific number of frames at a specific rate + through the DUT and then count the frames that are transmitted by the + DUT. If the count of offered frames is equal to the count of received + frames, the fewer frames are received than were transmitted, the rate + of the offered stream is reduced and the test is rerun. + + + + +Bradner & McQuaid Informational [Page 15] + +RFC 2544 Benchmarking Methodology March 1999 + + + The throughput is the fastest rate at which the count of test frames + transmitted by the DUT is equal to the number of test frames sent to + it by the test equipment. + + Reporting format: The results of the throughput test SHOULD be + reported in the form of a graph. If it is, the x coordinate SHOULD be + the frame size, the y coordinate SHOULD be the frame rate. There + SHOULD be at least two lines on the graph. There SHOULD be one line + showing the theoretical frame rate for the media at the various frame + sizes. The second line SHOULD be the plot of the test results. + Additional lines MAY be used on the graph to report the results for + each type of data stream tested. Text accompanying the graph SHOULD + indicate the protocol, data stream format, and type of media used in + the tests. + + We assume that if a single value is desired for advertising purposes + the vendor will select the rate for the minimum frame size for the + media. If this is done then the figure MUST be expressed in frames + per second. The rate MAY also be expressed in bits (or bytes) per + second if the vendor so desires. The statement of performance MUST + include a/ the measured maximum frame rate, b/ the size of the frame + used, c/ the theoretical limit of the media for that frame size, and + d/ the type of protocol used in the test. Even if a single value is + used as part of the advertising copy, the full table of results + SHOULD be included in the product data sheet. + +26.2 Latency + + Objective: To determine the latency as defined in RFC 1242. + + Procedure: First determine the throughput for DUT at each of the + listed frame sizes. Send a stream of frames at a particular frame + size through the DUT at the determined throughput rate to a specific + destination. The stream SHOULD be at least 120 seconds in duration. + An identifying tag SHOULD be included in one frame after 60 seconds + with the type of tag being implementation dependent. The time at + which this frame is fully transmitted is recorded (timestamp A). The + receiver logic in the test equipment MUST recognize the tag + information in the frame stream and record the time at which the + tagged frame was received (timestamp B). + + The latency is timestamp B minus timestamp A as per the relevant + definition frm RFC 1242, namely latency as defined for store and + forward devices or latency as defined for bit forwarding devices. + + The test MUST be repeated at least 20 times with the reported value + being the average of the recorded values. + + + + +Bradner & McQuaid Informational [Page 16] + +RFC 2544 Benchmarking Methodology March 1999 + + + This test SHOULD be performed with the test frame addressed to the + same destination as the rest of the data stream and also with each of + the test frames addressed to a new destination network. + + Reporting format: The report MUST state which definition of latency + (from RFC 1242) was used for this test. The latency results SHOULD + be reported in the format of a table with a row for each of the + tested frame sizes. There SHOULD be columns for the frame size, the + rate at which the latency test was run for that frame size, for the + media types tested, and for the resultant latency values for each + type of data stream tested. + +26.3 Frame loss rate + + Objective: To determine the frame loss rate, as defined in RFC 1242, + of a DUT throughout the entire range of input data rates and frame + sizes. + + Procedure: Send a specific number of frames at a specific rate + through the DUT to be tested and count the frames that are + transmitted by the DUT. The frame loss rate at each point is + calculated using the following equation: + + ( ( input_count - output_count ) * 100 ) / input_count + + + The first trial SHOULD be run for the frame rate that corresponds to + 100% of the maximum rate for the frame size on the input media. + Repeat the procedure for the rate that corresponds to 90% of the + maximum rate used and then for 80% of this rate. This sequence + SHOULD be continued (at reducing 10% intervals) until there are two + successive trials in which no frames are lost. The maximum + granularity of the trials MUST be 10% of the maximum rate, a finer + granularity is encouraged. + + Reporting format: The results of the frame loss rate test SHOULD be + plotted as a graph. If this is done then the X axis MUST be the + input frame rate as a percent of the theoretical rate for the media + at the specific frame size. The Y axis MUST be the percent loss at + the particular input rate. The left end of the X axis and the bottom + of the Y axis MUST be 0 percent; the right end of the X axis and the + top of the Y axis MUST be 100 percent. Multiple lines on the graph + MAY used to report the frame loss rate for different frame sizes, + protocols, and types of data streams. + + Note: See section 18 for the maximum frame rates that SHOULD be used. + + + + + +Bradner & McQuaid Informational [Page 17] + +RFC 2544 Benchmarking Methodology March 1999 + + +26.4 Back-to-back frames + + Objective: To characterize the ability of a DUT to process back-to- + back frames as defined in RFC 1242. + + Procedure: Send a burst of frames with minimum inter-frame gaps to + the DUT and count the number of frames forwarded by the DUT. If the + count of transmitted frames is equal to the number of frames + forwarded the length of the burst is increased and the test is rerun. + If the number of forwarded frames is less than the number + transmitted, the length of the burst is reduced and the test is + rerun. + + The back-to-back value is the number of frames in the longest burst + that the DUT will handle without the loss of any frames. The trial + length MUST be at least 2 seconds and SHOULD be repeated at least 50 + times with the average of the recorded values being reported. + + Reporting format: The back-to-back results SHOULD be reported in the + format of a table with a row for each of the tested frame sizes. + There SHOULD be columns for the frame size and for the resultant + average frame count for each type of data stream tested. The + standard deviation for each measurement MAY also be reported. + +26.5 System recovery + + Objective: To characterize the speed at which a DUT recovers from an + overload condition. + + Procedure: First determine the throughput for a DUT at each of the + listed frame sizes. + + Send a stream of frames at a rate 110% of the recorded throughput + rate or the maximum rate for the media, whichever is lower, for at + least 60 seconds. At Timestamp A reduce the frame rate to 50% of the + above rate and record the time of the last frame lost (Timestamp B). + The system recovery time is determined by subtracting Timestamp B + from Timestamp A. The test SHOULD be repeated a number of times and + the average of the recorded values being reported. + + Reporting format: The system recovery results SHOULD be reported in + the format of a table with a row for each of the tested frame sizes. + There SHOULD be columns for the frame size, the frame rate used as + the throughput rate for each type of data stream tested, and for the + measured recovery time for each type of data stream tested. + + + + + + +Bradner & McQuaid Informational [Page 18] + +RFC 2544 Benchmarking Methodology March 1999 + + +26.6 Reset + + Objective: To characterize the speed at which a DUT recovers from a + device or software reset. + + Procedure: First determine the throughput for the DUT for the + minimum frame size on the media used in the testing. + + Send a continuous stream of frames at the determined throughput rate + for the minimum sized frames. Cause a reset in the DUT. Monitor the + output until frames begin to be forwarded and record the time that + the last frame (Timestamp A) of the initial stream and the first + frame of the new stream (Timestamp B) are received. A power + interruption reset test is performed as above except that the power + to the DUT should be interrupted for 10 seconds in place of causing a + reset. + + This test SHOULD only be run using frames addressed to networks + directly connected to the DUT so that there is no requirement to + delay until a routing update is received. + + The reset value is obtained by subtracting Timestamp A from Timestamp + B. + + Hardware and software resets, as well as a power interruption SHOULD + be tested. + + Reporting format: The reset value SHOULD be reported in a simple set + of statements, one for each reset type. + +27. Security Considerations + + Security issues are not addressed in this document. + + + + + + + + + + + + + + + + + + +Bradner & McQuaid Informational [Page 19] + +RFC 2544 Benchmarking Methodology March 1999 + + +28. Editors' Addresses + + Scott Bradner + Harvard University + 1350 Mass. Ave, room 813 + Cambridge, MA 02138 + + Phone: +1 617 495-3864 + Fax: +1 617 496-8500 + EMail: sob@harvard.edu + + + Jim McQuaid + NetScout Systems + 4 Westford Tech Park Drive + Westford, MA 01886 + + Phone: +1 978 614-4116 + Fax: +1 978 614-4004 + EMail: mcquaidj@netscout.com + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +Bradner & McQuaid Informational [Page 20] + +RFC 2544 Benchmarking Methodology March 1999 + + +Appendix A: Testing Considerations + +A.1 Scope Of This Appendix + + This appendix discusses certain issues in the benchmarking + methodology where experience or judgment may play a role in the tests + selected to be run or in the approach to constructing the test with a + particular DUT. As such, this appendix MUST not be read as an + amendment to the methodology described in the body of this document + but as a guide to testing practice. + + 1. Typical testing practice has been to enable all protocols to be + tested and conduct all testing with no further configuration of + protocols, even though a given set of trials may exercise only one + protocol at a time. This minimizes the opportunities to "tune" a + DUT for a single protocol. + + 2. The least common denominator of the available filter functions + should be used to ensure that there is a basis for comparison + between vendors. Because of product differences, those conducting + and evaluating tests must make a judgment about this issue. + + 3. Architectural considerations may need to be considered. For + example, first perform the tests with the stream going between + ports on the same interface card and the repeat the tests with the + stream going into a port on one interface card and out of a port + on a second interface card. There will almost always be a best + case and worst case configuration for a given DUT architecture. + + 4. Testing done using traffic streams consisting of mixed protocols + has not shown much difference between testing with individual + protocols. That is, if protocol A testing and protocol B testing + give two different performance results, mixed protocol testing + appears to give a result which is the average of the two. + + 5. Wide Area Network (WAN) performance may be tested by setting up + two identical devices connected by the appropriate short- haul + versions of the WAN modems. Performance is then measured between + a LAN interface on one DUT to a LAN interface on the other DUT. + + The maximum frame rate to be used for LAN-WAN-LAN configurations is a + judgment that can be based on known characteristics of the overall + system including compression effects, fragmentation, and gross link + speeds. Practice suggests that the rate should be at least 110% of + the slowest link speed. Substantive issues of testing compression + itself are beyond the scope of this document. + + + + + +Bradner & McQuaid Informational [Page 21] + +RFC 2544 Benchmarking Methodology March 1999 + + +Appendix B: Maximum frame rates reference + + (Provided by Roger Beeman, Cisco Systems) + + Size Ethernet 16Mb Token Ring FDDI + (bytes) (pps) (pps) (pps) + + 64 14880 24691 152439 + 128 8445 13793 85616 + 256 4528 7326 45620 + 512 2349 3780 23585 + 768 1586 2547 15903 + 1024 1197 1921 11996 + 1280 961 1542 9630 + 1518 812 1302 8138 + + Ethernet size + Preamble 64 bits + Frame 8 x N bits + Gap 96 bits + + 16Mb Token Ring size + SD 8 bits + AC 8 bits + FC 8 bits + DA 48 bits + SA 48 bits + RI 48 bits ( 06 30 00 12 00 30 ) + SNAP + DSAP 8 bits + SSAP 8 bits + Control 8 bits + Vendor 24 bits + Type 16 bits + Data 8 x ( N - 18) bits + FCS 32 bits + ED 8 bits + FS 8 bits + + Tokens or idles between packets are not included + + FDDI size + Preamble 64 bits + SD 8 bits + FC 8 bits + DA 48 bits + SA 48 bits + SNAP + + + +Bradner & McQuaid Informational [Page 22] + +RFC 2544 Benchmarking Methodology March 1999 + + + DSAP 8 bits + SSAP 8 bits + Control 8 bits + Vendor 24 bits + Type 16 bits + Data 8 x ( N - 18) bits + FCS 32 bits + ED 4 bits + FS 12 bits + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +Bradner & McQuaid Informational [Page 23] + +RFC 2544 Benchmarking Methodology March 1999 + + +Appendix C: Test Frame Formats + + This appendix defines the frame formats that may be used with these + tests. It also includes protocol specific parameters for TCP/IP over + Ethernet to be used with the tests as an example. + +C.1. Introduction + + The general logic used in the selection of the parameters and the + design of the frame formats is explained for each case within the + TCP/IP section. The same logic has been used in the other sections. + Comments are used in these sections only if there is a protocol + specific feature to be explained. Parameters and frame formats for + additional protocols can be defined by the reader by using the same + logic. + +C.2. TCP/IP Information + + The following section deals with the TCP/IP protocol suite. + +C.2.1 Frame Type. + + An application level datagram echo request is used for the test data + frame in the protocols that support such a function. A datagram + protocol is used to minimize the chance that a router might expect a + specific session initialization sequence, as might be the case for a + reliable stream protocol. A specific defined protocol is used because + some routers verify the protocol field and refuse to forward unknown + protocols. + + For TCP/IP a UDP Echo Request is used. + +C.2.2 Protocol Addresses + + Two sets of addresses must be defined: first the addresses assigned + to the router ports, and second the address that are to be used in + the frames themselves and in the routing updates. + + The network addresses 192.18.0.0 through 198.19.255.255 are have been + assigned to the BMWG by the IANA for this purpose. This assignment + was made to minimize the chance of conflict in case a testing device + were to be accidentally connected to part of the Internet. The + specific use of the addresses is detailed below. + + + + + + + + +Bradner & McQuaid Informational [Page 24] + +RFC 2544 Benchmarking Methodology March 1999 + + +C.2.2.1 Router port protocol addresses + + Half of the ports on a multi-port router are referred to as "input" + ports and the other half as "output" ports even though some of the + tests use all ports both as input and output. A contiguous series of + IP Class C network addresses from 198.18.1.0 to 198.18.64.0 have been + assigned for use on the "input" ports. A second series from + 198.19.1.0 to 198.19.64.0 have been assigned for use on the "output" + ports. In all cases the router port is node 1 on the appropriate + network. For example, a two port DUT would have an IP address of + 198.18.1.1 on one port and 198.19.1.1 on the other port. + + Some of the tests described in the methodology memo make use of an + SNMP management connection to the DUT. The management access address + for the DUT is assumed to be the first of the "input" ports + (198.18.1.1). + +C.2.2.2 Frame addresses + + Some of the described tests assume adjacent network routing (the + reboot time test for example). The IP address used in the test frame + is that of node 2 on the appropriate Class C network. (198.19.1.2 for + example) + + If the test involves non-adjacent network routing the phantom routers + are located at node 10 of each of the appropriate Class C networks. + A series of Class C network addresses from 198.18.65.0 to + 198.18.254.0 has been assigned for use as the networks accessible + through the phantom routers on the "input" side of DUT. The series + of Class C networks from 198.19.65.0 to 198.19.254.0 have been + assigned to be used as the networks visible through the phantom + routers on the "output" side of the DUT. + +C.2.3 Routing Update Frequency + + The update interval for each routing protocol is may have to be + determined by the specifications of the individual protocol. For IP + RIP, Cisco IGRP and for OSPF a routing update frame or frames should + precede each stream of test frames by 5 seconds. This frequency is + sufficient for trial durations of up to 60 seconds. Routing updates + must be mixed with the stream of test frames if longer trial periods + are selected. The frequency of updates should be taken from the + following table. + + IP-RIP 30 sec + IGRP 90 sec + OSPF 90 sec + + + + +Bradner & McQuaid Informational [Page 25] + +RFC 2544 Benchmarking Methodology March 1999 + + +C.2.4 Frame Formats - detailed discussion + +C.2.4.1 Learning Frame + + In most protocols a procedure is used to determine the mapping + between the protocol node address and the MAC address. The Address + Resolution Protocol (ARP) is used to perform this function in TCP/IP. + No such procedure is required in XNS or IPX because the MAC address + is used as the protocol node address. + + In the ideal case the tester would be able to respond to ARP requests + from the DUT. In cases where this is not possible an ARP request + should be sent to the router's "output" port. This request should be + seen as coming from the immediate destination of the test frame + stream. (i.e. the phantom router (Figure 2) or the end node if + adjacent network routing is being used.) It is assumed that the + router will cache the MAC address of the requesting device. The ARP + request should be sent 5 seconds before the test frame stream starts + in each trial. Trial lengths of longer than 50 seconds may require + that the router be configured for an extended ARP timeout. + + +--------+ +------------+ + | | | phantom |------ P LAN + A + IN A------| DUT |------------| |------ P LAN + B + | | OUT A | router |------ P LAN + C + +--------+ +------------+ + + Figure 2 + + In the case where full routing is being used + +C.2.4.2 Routing Update Frame + + If the test does not involve adjacent net routing the tester must + supply proper routing information using a routing update. A single + routing update is used before each trial on each "destination" port + (see section C.24). This update includes the network addresses that + are reachable through a phantom router on the network attached to the + port. For a full mesh test, one destination network address is + present in the routing update for each of the "input" ports. The + test stream on each "input" port consists of a repeating sequence of + frames, one to each of the "output" ports. + + + + + + +Bradner & McQuaid Informational [Page 26] + +RFC 2544 Benchmarking Methodology March 1999 + + +C.2.4.3 Management Query Frame + + The management overhead test uses SNMP to query a set of variables + that should be present in all DUTs that support SNMP. The variables + for a single interface only are read by an NMS at the appropriate + intervals. The list of variables to retrieve follow: + + sysUpTime + ifInOctets + ifOutOctets + ifInUcastPkts + ifOutUcastPkts + +C.2.4.4 Test Frames + + The test frame is an UDP Echo Request with enough data to fill out + the required frame size. The data should not be all bits off or all + bits on since these patters can cause a "bit stuffing" process to be + used to maintain clock synchronization on WAN links. This process + will result in a longer frame than was intended. + +C.2.4.5 Frame Formats - TCP/IP on Ethernet + + Each of the frames below are described for the 1st pair of DUT ports, + i.e. "input" port #1 and "output" port #1. Addresses must be changed + if the frame is to be used for other ports. + +C.2.6.1 Learning Frame + + ARP Request on Ethernet + + -- DATAGRAM HEADER + offset data (hex) description + 00 FF FF FF FF FF FF dest MAC address send to + broadcast address + 06 xx xx xx xx xx xx set to source MAC address + 12 08 06 ARP type + 14 00 01 hardware type Ethernet = 1 + 16 08 00 protocol type IP = 800 + 18 06 hardware address length 48 bits + on Ethernet + 19 04 protocol address length 4 octets + for IP + 20 00 01 opcode request = 1 + 22 xx xx xx xx xx xx source MAC address + 28 xx xx xx xx source IP address + 32 FF FF FF FF FF FF requesting DUT's MAC address + 38 xx xx xx xx DUT's IP address + + + +Bradner & McQuaid Informational [Page 27] + +RFC 2544 Benchmarking Methodology March 1999 + + +C.2.6.2 Routing Update Frame + + -- DATAGRAM HEADER + offset data (hex) description + 00 FF FF FF FF FF FF dest MAC address is broadcast + 06 xx xx xx xx xx xx source hardware address + 12 08 00 type + + -- IP HEADER + 14 45 IP version - 4, header length (4 + byte units) - 5 + 15 00 service field + 16 00 EE total length + 18 00 00 ID + 20 40 00 flags (3 bits) 4 (do not + fragment), + fragment offset-0 + 22 0A TTL + 23 11 protocol - 17 (UDP) + 24 C4 8D header checksum + 26 xx xx xx xx source IP address + 30 xx xx xx destination IP address + 33 FF host part = FF for broadcast + + -- UDP HEADER + 34 02 08 source port 208 = RIP + 36 02 08 destination port 208 = RIP + 38 00 DA UDP message length + 40 00 00 UDP checksum + + -- RIP packet + 42 02 command = response + 43 01 version = 1 + 44 00 00 0 + + -- net 1 + 46 00 02 family = IP + 48 00 00 0 + 50 xx xx xx net 1 IP address + 53 00 net not node + 54 00 00 00 00 0 + 58 00 00 00 00 0 + 62 00 00 00 07 metric 7 + + -- net 2 + 66 00 02 family = IP + 68 00 00 0 + 70 xx xx xx net 2 IP address + + + +Bradner & McQuaid Informational [Page 28] + +RFC 2544 Benchmarking Methodology March 1999 + + + 73 00 net not node + 74 00 00 00 00 0 + 78 00 00 00 00 0 + 82 00 00 00 07 metric 7 + + -- net 3 + 86 00 02 family = IP + 88 00 00 0 + 90 xx xx xx net 3 IP address + 93 00 net not node + 94 00 00 00 00 0 + 98 00 00 00 00 0 + 102 00 00 00 07 metric 7 + + -- net 4 + 106 00 02 family = IP + 108 00 00 0 + 110 xx xx xx net 4 IP address + 113 00 net not node + 114 00 00 00 00 0 + 118 00 00 00 00 0 + 122 00 00 00 07 metric 7 + + -- net 5 + 126 00 02 family = IP + 128 00 00 0 + 130 00 net 5 IP address + 133 00 net not node + 134 00 00 00 00 0 + 138 00 00 00 00 0 + 142 00 00 00 07 metric 7 + + -- net 6 + 146 00 02 family = IP + 148 00 00 0 + 150 xx xx xx net 6 IP address + 153 00 net not node + 154 00 00 00 00 0 + 158 00 00 00 00 0 + 162 00 00 00 07 metric 7 + +C.2.4.6 Management Query Frame + + To be defined. + +C.2.6.4 Test Frames + + UDP echo request on Ethernet + + + +Bradner & McQuaid Informational [Page 29] + +RFC 2544 Benchmarking Methodology March 1999 + + + -- DATAGRAM HEADER + offset data (hex) description + 00 xx xx xx xx xx xx set to dest MAC address + 06 xx xx xx xx xx xx set to source MAC address + 12 08 00 type + + -- IP HEADER + 14 45 IP version - 4 header length 5 4 + byte units + 15 00 TOS + 16 00 2E total length* + 18 00 00 ID + 20 00 00 flags (3 bits) - 0 fragment + offset-0 + 22 0A TTL + 23 11 protocol - 17 (UDP) + 24 C4 8D header checksum* + 26 xx xx xx xx set to source IP address** + 30 xx xx xx xx set to destination IP address** + + -- UDP HEADER + 34 C0 20 source port + 36 00 07 destination port 07 = Echo + 38 00 1A UDP message length* + 40 00 00 UDP checksum + + -- UDP DATA + 42 00 01 02 03 04 05 06 07 some data*** + 50 08 09 0A 0B 0C 0D 0E 0F + + * - change for different length frames + + ** - change for different logical streams + + *** - fill remainder of frame with incrementing octets, + repeated if required by frame length + + Values to be used in Total Length and UDP message length fields: + + frame size total length UDP message length + 64 00 2E 00 1A + 128 00 6E 00 5A + 256 00 EE 00 9A + 512 01 EE 01 9A + 768 02 EE 02 9A + 1024 03 EE 03 9A + 1280 04 EE 04 9A + 1518 05 DC 05 C8 + + + +Bradner & McQuaid Informational [Page 30] + +RFC 2544 Benchmarking Methodology March 1999 + + +Full Copyright Statement + + Copyright (C) The Internet Society (1999). All Rights Reserved. + + This document and translations of it may be copied and furnished to + others, and derivative works that comment on or otherwise explain it + or assist in its implementation may be prepared, copied, published + and distributed, in whole or in part, without restriction of any + kind, provided that the above copyright notice and this paragraph are + included on all such copies and derivative works. However, this + document itself may not be modified in any way, such as by removing + the copyright notice or references to the Internet Society or other + Internet organizations, except as needed for the purpose of + developing Internet standards in which case the procedures for + copyrights defined in the Internet Standards process must be + followed, or as required to translate it into languages other than + English. + + The limited permissions granted above are perpetual and will not be + revoked by the Internet Society or its successors or assigns. + + This document and the information contained herein is provided on an + "AS IS" basis and THE INTERNET SOCIETY AND THE INTERNET ENGINEERING + TASK FORCE DISCLAIMS ALL WARRANTIES, EXPRESS OR IMPLIED, INCLUDING + BUT NOT LIMITED TO ANY WARRANTY THAT THE USE OF THE INFORMATION + HEREIN WILL NOT INFRINGE ANY RIGHTS OR ANY IMPLIED WARRANTIES OF + MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. + + + + + + + + + + + + + + + + + + + + + + + + +Bradner & McQuaid Informational [Page 31] + |