summaryrefslogtreecommitdiff
path: root/doc/rfc/rfc1944.txt
diff options
context:
space:
mode:
authorThomas Voss <mail@thomasvoss.com> 2024-11-27 20:54:24 +0100
committerThomas Voss <mail@thomasvoss.com> 2024-11-27 20:54:24 +0100
commit4bfd864f10b68b71482b35c818559068ef8d5797 (patch)
treee3989f47a7994642eb325063d46e8f08ffa681dc /doc/rfc/rfc1944.txt
parentea76e11061bda059ae9f9ad130a9895cc85607db (diff)
doc: Add RFC documents
Diffstat (limited to 'doc/rfc/rfc1944.txt')
-rw-r--r--doc/rfc/rfc1944.txt1683
1 files changed, 1683 insertions, 0 deletions
diff --git a/doc/rfc/rfc1944.txt b/doc/rfc/rfc1944.txt
new file mode 100644
index 0000000..29defc1
--- /dev/null
+++ b/doc/rfc/rfc1944.txt
@@ -0,0 +1,1683 @@
+
+
+
+
+
+
+Network Working Group S. Bradner
+Request for Comments: 1944 Harvard University
+Category: Informational J. McQuaid
+ Bay Networks
+ May 1996
+
+
+ Benchmarking Methodology for Network Interconnect Devices
+
+Status of This Memo
+
+ This memo provides information for the Internet community. This memo
+ does not specify an Internet standard of any kind. Distribution of
+ this memo is unlimited.
+
+Abstract
+
+ This document discusses and defines a number of tests that may be
+ used to describe the performance characteristics of a network
+ interconnecting device. In addition to defining the tests this
+ document also describes specific formats for reporting the results of
+ the tests. Appendix A lists the tests and conditions that we believe
+ should be included for specific cases and gives additional
+ information about testing practices. Appendix B is a reference
+ listing of maximum frame rates to be used with specific frame sizes
+ on various media and Appendix C gives some examples of frame formats
+ to be used in testing.
+
+1. Introduction
+
+ Vendors often engage in "specsmanship" in an attempt to give their
+ products a better position in the marketplace. This often involves
+ "smoke & mirrors" to confuse the potential users of the products.
+
+ This document defines a specific set of tests that vendors can use to
+ measure and report the performance characteristics of network
+ devices. The results of these tests will provide the user comparable
+ data from different vendors with which to evaluate these devices.
+
+ A previous document, "Benchmarking Terminology for Network
+ Interconnect Devices" (RFC 1242), defined many of the terms that are
+ used in this document. The terminology document should be consulted
+ before attempting to make use of this document.
+
+
+
+
+
+
+
+
+Bradner & McQuaid Informational [Page 1]
+
+RFC 1944 Benchmarking Methodology May 1996
+
+
+2. Real world
+
+ In producing this document the authors attempted to keep in mind the
+ requirement that apparatus to perform the described tests must
+ actually be built. We do not know of "off the shelf" equipment
+ available to implement all of the tests but it is our opinion that
+ such equipment can be constructed.
+
+3. Tests to be run
+
+ There are a number of tests described in this document. Not all of
+ the tests apply to all types of devices under test (DUTs). Vendors
+ should perform all of the tests that can be supported by a specific
+ type of product. The authors understand that it will take a
+ considerable period of time to perform all of the recommended tests
+ nder all of the recommended conditions. We believe that the results
+ are worth the effort. Appendix A lists some of the tests and
+ conditions that we believe should be included for specific cases.
+
+4. Evaluating the results
+
+ Performing all of the recommended tests will result in a great deal
+ of data. Much of this data will not apply to the evaluation of the
+ devices under each circumstance. For example, the rate at which a
+ router forwards IPX frames will be of little use in selecting a
+ router for an environment that does not (and will not) support that
+ protocol. Evaluating even that data which is relevant to a
+ particular network installation will require experience which may not
+ be readily available. Furthermore, selection of the tests to be run
+ and evaluation of the test data must be done with an understanding of
+ generally accepted testing practices regarding repeatability,
+ variance and statistical significance of small numbers of trials.
+
+5. Requirements
+
+ In this document, the words that are used to define the significance
+ of each particular requirement are capitalized. These words are:
+
+ * "MUST" This word, or the words "REQUIRED" and "SHALL" mean that
+ the item is an absolute requirement of the specification.
+
+ * "SHOULD" This word or the adjective "RECOMMENDED" means that there
+ may exist valid reasons in particular circumstances to ignore this
+ item, but the full implications should be understood and the case
+ carefully weighed before choosing a different course.
+
+ * "MAY" This word or the adjective "OPTIONAL" means that this item
+ is truly optional. One vendor may choose to include the item because
+
+
+
+Bradner & McQuaid Informational [Page 2]
+
+RFC 1944 Benchmarking Methodology May 1996
+
+
+ a particular marketplace requires it or because it enhances the
+ product, for example; another vendor may omit the same item.
+
+ An implementation is not compliant if it fails to satisfy one or more
+ of the MUST requirements for the protocols it implements. An
+ implementation that satisfies all the MUST and all the SHOULD
+ requirements for its protocols is said to be "unconditionally
+ compliant"; one that satisfies all the MUST requirements but not all
+ the SHOULD requirements for its protocols is said to be
+ "conditionally compliant".
+
+6. Test set up
+
+ The ideal way to implement this series of tests is to use a tester
+ with both transmitting and receiving ports. Connections are made
+ from the sending ports of the tester to the receiving ports of the
+ DUT and from the sending ports of the DUT back to the tester. (see
+ Figure 1) Since the tester both sends the test traffic and receives
+ it back, after the traffic has been forwarded but the DUT, the tester
+ can easily determine if all of the transmitted packets were received
+ and verify that the correct packets were received. The same
+ functionality can be obtained with separate transmitting and
+ receiving devices (see Figure 2) but unless they are remotely
+ controlled by some computer in a way that simulates the single
+ tester, the labor required to accurately perform some of the tests
+ (particularly the throughput test) can be prohibitive.
+
+ +------------+
+ | |
+ +------------| tester |<-------------+
+ | | | |
+ | +------------+ |
+ | |
+ | +------------+ |
+ | | | |
+ +----------->| DUT |--------------+
+ | |
+ +------------+
+ Figure 1
+
+ +--------+ +------------+ +----------+
+ | | | | | |
+ | sender |-------->| DUT |--------->| receiver |
+ | | | | | |
+ +--------+ +------------+ +----------+
+ Figure 2
+
+
+
+
+
+Bradner & McQuaid Informational [Page 3]
+
+RFC 1944 Benchmarking Methodology May 1996
+
+
+6.1 Test set up for multiple media types
+
+ Two different setups could be used to test a DUT which is used in
+ real-world networks to connect networks of differing media type,
+ local Ethernet to a backbone FDDI ring for example. The tester could
+ support both media types in which case the set up shown in Figure 1
+ would be used.
+
+ Two identical DUTs are used in the other test set up. (see Figure 3)
+ In many cases this set up may more accurately simulate the real
+ world. For example, connecting two LANs together with a WAN link or
+ high speed backbone. This set up would not be as good at simulating
+ a system where clients on a Ethernet LAN were interacting with a
+ server on an FDDI backbone.
+
+ +-----------+
+ | |
+ +---------------------| tester |<---------------------+
+ | | | |
+ | +-----------+ |
+ | |
+ | +----------+ +----------+ |
+ | | | | | |
+ +------->| DUT 1 |-------------->| DUT 2 |---------+
+ | | | |
+ +----------+ +----------+
+
+ Figure 3
+
+7. DUT set up
+
+ Before starting to perform the tests, the DUT to be tested MUST be
+ configured following the instructions provided to the user.
+ Specifically, it is expected that all of the supported protocols will
+ be configured and enabled during this set up (See Appendix A). It is
+ expected that all of the tests will be run without changing the
+ configuration or setup of the DUT in any way other than that required
+ to do the specific test. For example, it is not acceptable to change
+ the size of frame handling buffers between tests of frame handling
+ rates or to disable all but one transport protocol when testing the
+ throughput of that protocol. It is necessary to modify the
+ configuration when starting a test to determine the effect of filters
+ on throughput, but the only change MUST be to enable the specific
+ filter. The DUT set up SHOULD include the normally recommended
+ routing update intervals and keep alive frequency. The specific
+ version of the software and the exact DUT configuration, including
+ what functions are disabled, used during the tests MUST be included
+ as part of the report of the results.
+
+
+
+Bradner & McQuaid Informational [Page 4]
+
+RFC 1944 Benchmarking Methodology May 1996
+
+
+8. Frame formats
+
+ The formats of the test frames to use for TCP/IP over Ethernet are
+ shown in Appendix C: Test Frame Formats. These exact frame formats
+ SHOULD be used in the tests described in this document for this
+ protocol/media combination and that these frames will be used as a
+ template for testing other protocol/media combinations. The specific
+ formats that are used to define the test frames for a particular test
+ series MUST be included in the report of the results.
+
+9. Frame sizes
+
+ All of the described tests SHOULD be performed at a number of frame
+ sizes. Specifically, the sizes SHOULD include the maximum and minimum
+ legitimate sizes for the protocol under test on the media under test
+ and enough sizes in between to be able to get a full characterization
+ of the DUT performance. Except where noted, at least five frame
+ sizes SHOULD be tested for each test condition.
+
+ Theoretically the minimum size UDP Echo request frame would consist
+ of an IP header (minimum length 20 octets), a UDP header (8 octets)
+ and whatever MAC level header is required by the media in use. The
+ theoretical maximum frame size is determined by the size of the
+ length field in the IP header. In almost all cases the actual
+ maximum and minimum sizes are determined by the limitations of the
+ media.
+
+ In theory it would be ideal to distribute the frame sizes in a way
+ that would evenly distribute the theoretical frame rates. These
+ recommendations incorporate this theory but specify frame sizes which
+ are easy to understand and remember. In addition, many of the same
+ frame sizes are specified on each of the media types to allow for
+ easy performance comparisons.
+
+ Note: The inclusion of an unrealistically small frame size on some of
+ the media types (i.e. with little or no space for data) is to help
+ characterize the per-frame processing overhead of the DUT.
+
+ 9.1 Frame sizes to be used on Ethernet
+
+ 64, 128, 256, 512, 1024, 1280, 1518
+
+ These sizes include the maximum and minimum frame sizes permitted
+ by the Ethernet standard and a selection of sizes between these
+ extremes with a finer granularity for the smaller frame sizes and
+ higher frame rates.
+
+
+
+
+
+Bradner & McQuaid Informational [Page 5]
+
+RFC 1944 Benchmarking Methodology May 1996
+
+
+ 9.2 Frame sizes to be used on 4Mb and 16Mb token ring
+
+ 54, 64, 128, 256, 1024, 1518, 2048, 4472
+
+ The frame size recommendations for token ring assume that there is
+ no RIF field in the frames of routed protocols. A RIF field would
+ be present in any direct source route bridge performance test.
+ The minimum size frame for UDP on token ring is 54 octets. The
+ maximum size of 4472 octets is recommended for 16Mb token ring
+ instead of the theoretical size of 17.9Kb because of the size
+ limitations imposed by many token ring interfaces. The reminder
+ of the sizes are selected to permit direct comparisons with other
+ types of media. An IP (i.e. not UDP) frame may be used in
+ addition if a higher data rate is desired, in which case the
+ minimum frame size is 46 octets.
+
+ 9.3 Frame sizes to be used on FDDI
+
+ 54, 64, 128, 256, 1024, 1518, 2048, 4472
+
+ The minimum size frame for UDP on FDDI is 53 octets, the minimum
+ size of 54 is recommended to allow direct comparison to token ring
+ performance. The maximum size of 4472 is recommended instead of
+ the theoretical maximum size of 4500 octets to permit the same
+ type of comparison. An IP (i.e. not UDP) frame may be used in
+ addition if a higher data rate is desired, in which case the
+ minimum frame size is 45 octets.
+
+ 9.4 Frame sizes in the presence of disparate MTUs
+
+ When the interconnect DUT supports connecting links with disparate
+ MTUs, the frame sizes for the link with the *larger* MTU SHOULD be
+ used, up to the limit of the protocol being tested. If the
+ interconnect DUT does not support the fragmenting of frames in the
+ presence of MTU mismatch, the forwarding rate for that frame size
+ shall be reported as zero.
+
+ For example, the test of IP forwarding with a bridge or router
+ that joins FDDI and Ethernet should use the frame sizes of FDDI
+ when going from the FDDI to the Ethernet link. If the bridge does
+ not support IP fragmentation, the forwarding rate for those frames
+ too large for Ethernet should be reported as zero.
+
+10. Verifying received frames
+
+ The test equipment SHOULD discard any frames received during a test
+ run that are not actual forwarded test frames. For example, keep-
+ alive and routing update frames SHOULD NOT be included in the count
+
+
+
+Bradner & McQuaid Informational [Page 6]
+
+RFC 1944 Benchmarking Methodology May 1996
+
+
+ of received frames. In any case, the test equipment SHOULD verify
+ the length of the received frames and check that they match the
+ expected length.
+
+ Preferably, the test equipment SHOULD include sequence numbers in the
+ transmitted frames and check for these numbers on the received
+ frames. If this is done, the reported results SHOULD include in
+ addition to the number of frames dropped, the number of frames that
+ were received out of order, the number of duplicate frames received
+ and the number of gaps in the received frame numbering sequence.
+ This functionality is required for some of the described tests.
+
+11. Modifiers
+
+ It might be useful to know the DUT performance under a number of
+ conditions; some of these conditions are noted below. The reported
+ results SHOULD include as many of these conditions as the test
+ equipment is able to generate. The suite of tests SHOULD be first
+ run without any modifying conditions and then repeated under each of
+ the conditions separately. To preserve the ability to compare the
+ results of these tests any frames that are required to generate the
+ modifying conditions (management queries for example) will be
+ included in the same data stream as the normal test frames in place
+ of one of the test frames and not be supplied to the DUT on a
+ separate network port.
+
+ 11.1 Broadcast frames
+
+ In most router designs special processing is required when frames
+ addressed to the hardware broadcast address are received. In
+ bridges (or in bridge mode on routers) these broadcast frames must
+ be flooded to a number of ports. The stream of test frames SHOULD
+ be augmented with 1% frames addressed to the hardware broadcast
+ address. The frames sent to the broadcast address should be of a
+ type that the router will not need to process. The aim of this
+ test is to determine if there is any effect on the forwarding rate
+ of the other data in the stream. The specific frames that should
+ be used are included in the test frame format document. The
+ broadcast frames SHOULD be evenly distributed throughout the data
+ stream, for example, every 100th frame.
+
+ The same test SHOULD be performed on bridge-like DUTs but in this
+ case the broadcast packets will be processed and flooded to all
+ outputs.
+
+ It is understood that a level of broadcast frames of 1% is much
+ higher than many networks experience but, as in drug toxicity
+ evaluations, the higher level is required to be able to gage the
+
+
+
+Bradner & McQuaid Informational [Page 7]
+
+RFC 1944 Benchmarking Methodology May 1996
+
+
+ effect which would otherwise often fall within the normal
+ variability of the system performance. Due to design factors some
+ test equipment will not be able to generate a level of alternate
+ frames this low. In these cases the percentage SHOULD be as small
+ as the equipment can provide and that the actual level be
+ described in the report of the test results.
+
+ 11.2 Management frames
+
+ Most data networks now make use of management protocols such as
+ SNMP. In many environments there can be a number of management
+ stations sending queries to the same DUT at the same time.
+
+ The stream of test frames SHOULD be augmented with one management
+ query as the first frame sent each second during the duration of
+ the trial. The result of the query must fit into one response
+ frame. The response frame SHOULD be verified by the test
+ equipment. One example of the specific query frame that should be
+ used is shown in Appendix C.
+
+ 11.3 Routing update frames
+
+ The processing of dynamic routing protocol updates could have a
+ significant impact on the ability of a router to forward data
+ frames. The stream of test frames SHOULD be augmented with one
+ routing update frame transmitted as the first frame transmitted
+ during the trial. Routing update frames SHOULD be sent at the
+ rate specified in Appendix C for the specific routing protocol
+ being used in the test. Two routing update frames are defined in
+ Appendix C for the TCP/IP over Ethernet example. The routing
+ frames are designed to change the routing to a number of networks
+ that are not involved in the forwarding of the test data. The
+ first frame sets the routing table state to "A", the second one
+ changes the state to "B". The frames MUST be alternated during
+ the trial.
+
+ The test SHOULD verify that the routing update was processed by
+ the DUT.
+
+ 11.4 Filters
+
+ Filters are added to routers and bridges to selectively inhibit
+ the forwarding of frames that would normally be forwarded. This
+ is usually done to implement security controls on the data that is
+ accepted between one area and another. Different products have
+ different capabilities to implement filters.
+
+
+
+
+
+Bradner & McQuaid Informational [Page 8]
+
+RFC 1944 Benchmarking Methodology May 1996
+
+
+ The DUT SHOULD be first configured to add one filter condition and
+ the tests performed. This filter SHOULD permit the forwarding of
+ the test data stream. In routers this filter SHOULD be of the
+ form:
+
+ forward input_protocol_address to output_protocol_address
+
+ In bridges the filter SHOULD be of the form:
+
+ forward destination_hardware_address
+
+ The DUT SHOULD be then reconfigured to implement a total of 25
+ filters. The first 24 of these filters SHOULD be of the form:
+
+ block input_protocol_address to output_protocol_address
+
+ The 24 input and output protocol addresses SHOULD not be any that
+ are represented in the test data stream. The last filter SHOULD
+ permit the forwarding of the test data stream. By "first" and
+ "last" we mean to ensure that in the second case, 25 conditions
+ must be checked before the data frames will match the conditions
+ that permit the forwarding of the frame. Of course, if the DUT
+ reorders the filters or does not use a linear scan of the filter
+ rules the effect of the sequence in which the filters are input is
+ properly lost.
+
+ The exact filters configuration command lines used SHOULD be
+ included with the report of the results.
+
+ 11.4.1 Filter Addresses
+
+ Two sets of filter addresses are required, one for the single
+ filter case and one for the 25 filter case.
+
+ The single filter case should permit traffic from IP address
+ 198.18.1.2 to IP address 198.19.65.2 and deny all other
+ traffic.
+
+ The 25 filter case should follow the following sequence.
+
+ deny aa.ba.1.1 to aa.ba.100.1
+ deny aa.ba.2.2 to aa.ba.101.2
+ deny aa.ba.3.3 to aa.ba.103.3
+ ...
+ deny aa.ba.12.12 to aa.ba.112.12
+ allow aa.bc.1.2 to aa.bc.65.1
+ deny aa.ba.13.13 to aa.ba.113.13
+ deny aa.ba.14.14 to aa.ba.114.14
+
+
+
+Bradner & McQuaid Informational [Page 9]
+
+RFC 1944 Benchmarking Methodology May 1996
+
+
+ ...
+ deny aa.ba.24.24 to aa.ba.124.24
+ deny all else
+
+ All previous filter conditions should be cleared from the
+ router before this sequence is entered. The sequence is
+ selected to test to see if the router sorts the filter
+ conditions or accepts them in the order that they were entered.
+ Both of these procedures will result in a greater impact on
+ performance than will some form of hash coding.
+
+12. Protocol addresses
+
+ It is easier to implement these tests using a single logical stream
+ of data, with one source protocol address and one destination
+ protocol address, and for some conditions like the filters described
+ above, a practical requirement. Networks in the real world are not
+ limited to single streams of data. The test suite SHOULD be first run
+ with a single protocol (or hardware for bridge tests) source and
+ destination address pair. The tests SHOULD then be repeated with
+ using a random destination address. While testing routers the
+ addresses SHOULD be random and uniformly distributed over a range of
+ 256 networks and random and uniformly distributed over the full MAC
+ range for bridges. The specific address ranges to use for IP are
+ shown in Appendix C.
+
+13. Route Set Up
+
+ It is not reasonable that all of the routing information necessary to
+ forward the test stream, especially in the multiple address case,
+ will be manually set up. At the start of each trial a routing update
+ MUST be sent to the DUT. This routing update MUST include all of the
+ network addresses that will be required for the trial. All of the
+ addresses SHOULD resolve to the same "next-hop". Normally this will
+ be the address of the receiving side of the test equipment. This
+ routing update will have to be repeated at the interval required by
+ the routing protocol being used. An example of the format and
+ repetition interval of the update frames is given in Appendix C.
+
+14. Bidirectional traffic
+
+ Normal network activity is not all in a single direction. To test
+ the bidirectional performance of a DUT, the test series SHOULD be run
+ with the same data rate being offered from each direction. The sum of
+ the data rates should not exceed the theoretical limit for the media.
+
+
+
+
+
+
+Bradner & McQuaid Informational [Page 10]
+
+RFC 1944 Benchmarking Methodology May 1996
+
+
+15. Single stream path
+
+ The full suite of tests SHOULD be run along with whatever modifier
+ conditions that are relevant using a single input and output network
+ port on the DUT. If the internal design of the DUT has multiple
+ distinct pathways, for example, multiple interface cards each with
+ multiple network ports, then all possible types of pathways SHOULD be
+ tested separately.
+
+16. Multi-port
+
+ Many current router and bridge products provide many network ports in
+ the same module. In performing these tests first half of the ports
+ are designated as "input ports" and half are designated as "output
+ ports". These ports SHOULD be evenly distributed across the DUT
+ architecture. For example if a DUT has two interface cards each of
+ which has four ports, two ports on each interface card are designated
+ as input and two are designated as output. The specified tests are
+ run using the same data rate being offered to each of the input
+ ports. The addresses in the input data streams SHOULD be set so that
+ a frame will be directed to each of the output ports in sequence so
+ that all "output" ports will get an even distribution of packets from
+ this input. The same configuration MAY be used to perform a
+ bidirectional multi-stream test. In this case all of the ports are
+ considered both input and output ports and each data stream MUST
+ consist of frames addressed to all of the other ports.
+
+ Consider the following 6 port DUT:
+
+ --------------
+ ---------| in A out X|--------
+ ---------| in B out Y|--------
+ ---------| in C out Z|--------
+ --------------
+
+ The addressing of the data streams for each of the inputs SHOULD be:
+
+ stream sent to input A:
+ packet to out X, packet to out Y, packet to out Z
+ stream sent to input B:
+ packet to out X, packet to out Y, packet to out Z
+ stream sent to input C
+ packet to out X, packet to out Y, packet to out Z
+
+ Note that these streams each follow the same sequence so that 3
+ packets will arrive at output X at the same time, then 3 packets at
+ Y, then 3 packets at Z. This procedure ensures that, as in the real
+ world, the DUT will have to deal with multiple packets addressed to
+
+
+
+Bradner & McQuaid Informational [Page 11]
+
+RFC 1944 Benchmarking Methodology May 1996
+
+
+ the same output at the same time.
+
+17. Multiple protocols
+
+ This document does not address the issue of testing the effects of a
+ mixed protocol environment other than to suggest that if such tests
+ are wanted then frames SHOULD be distributed between all of the test
+ protocols. The distribution MAY approximate the conditions on the
+ network in which the DUT would be used.
+
+18. Multiple frame sizes
+
+ This document does not address the issue of testing the effects of a
+ mixed frame size environment other than to suggest that if such tests
+ are wanted then frames SHOULD be distributed between all of the
+ listed sizes for the protocol under test. The distribution MAY
+ approximate the conditions on the network in which the DUT would be
+ used. The authors do not have any idea how the results of such a test
+ would be interpreted other than to directly compare multiple DUTs in
+ some very specific simulated network.
+
+19. Testing performance beyond a single DUT.
+
+ In the performance testing of a single DUT, the paradigm can be
+ described as applying some input to a DUT and monitoring the output.
+ The results of which can be used to form a basis of characterization
+ of that device under those test conditions.
+
+ This model is useful when the test input and output are homogenous
+ (e.g., 64-byte IP, 802.3 frames into the DUT; 64 byte IP, 802.3
+ frames out), or the method of test can distinguish between dissimilar
+ input/output. (E.g., 1518 byte IP, 802.3 frames in; 576 byte,
+ fragmented IP, X.25 frames out.)
+
+ By extending the single DUT test model, reasonable benchmarks
+ regarding multiple DUTs or heterogeneous environments may be
+ collected. In this extension, the single DUT is replaced by a system
+ of interconnected network DUTs. This test methodology would support
+ the benchmarking of a variety of device/media/service/protocol
+ combinations. For example, a configuration for a LAN-to-WAN-to-LAN
+ test might be:
+
+ (1) 802.3-> DUT 1 -> X.25 @ 64kbps -> DUT 2 -> 802.3
+
+ Or a mixed LAN configuration might be:
+
+ (2) 802.3 -> DUT 1 -> FDDI -> DUT 2 -> FDDI -> DUT 3 -> 802.3
+
+
+
+
+Bradner & McQuaid Informational [Page 12]
+
+RFC 1944 Benchmarking Methodology May 1996
+
+
+ In both examples 1 and 2, end-to-end benchmarks of each system could
+ be empirically ascertained. Other behavior may be characterized
+ through the use of intermediate devices. In example 2, the
+ configuration may be used to give an indication of the FDDI to FDDI
+ capability exhibited by DUT 2.
+
+ Because multiple DUTs are treated as a single system, there are
+ limitations to this methodology. For instance, this methodology may
+ yield an aggregate benchmark for a tested system. That benchmark
+ alone, however, may not necessarily reflect asymmetries in behavior
+ between the DUTs, latencies introduce by other apparatus (e.g.,
+ CSUs/DSUs, switches), etc.
+
+ Further, care must be used when comparing benchmarks of different
+ systems by ensuring that the DUTs' features/configuration of the
+ tested systems have the appropriate common denominators to allow
+ comparison.
+
+20. Maximum frame rate
+
+ The maximum frame rates that should be used when testing LAN
+ connections SHOULD be the listed theoretical maximum rate for the
+ frame size on the media.
+
+ The maximum frame rate that should be used when testing WAN
+ connections SHOULD be greater than the listed theoretical maximum
+ rate for the frame size on that speed connection. The higher rate
+ for WAN tests is to compensate for the fact that some vendors employ
+ various forms of header compression.
+
+ A list of maximum frame rates for LAN connections is included in
+ Appendix B.
+
+21. Bursty traffic
+
+ It is convenient to measure the DUT performance under steady state
+ load but this is an unrealistic way to gauge the functioning of a DUT
+ since actual network traffic normally consists of bursts of frames.
+ Some of the tests described below SHOULD be performed with both
+ steady state traffic and with traffic consisting of repeated bursts
+ of frames. The frames within a burst are transmitted with the
+ minimum legitimate inter-frame gap.
+
+ The objective of the test is to determine the minimum interval
+ between bursts which the DUT can process with no frame loss. During
+ each test the number of frames in each burst is held constant and the
+ inter-burst interval varied. Tests SHOULD be run with burst sizes of
+ 16, 64, 256 and 1024 frames.
+
+
+
+Bradner & McQuaid Informational [Page 13]
+
+RFC 1944 Benchmarking Methodology May 1996
+
+
+22. Frames per token
+
+ Although it is possible to configure some token ring and FDDI
+ interfaces to transmit more than one frame each time that the token
+ is received, most of the network devices currently available transmit
+ only one frame per token. These tests SHOULD first be performed
+ while transmitting only one frame per token.
+
+ Some current high-performance workstation servers do transmit more
+ than one frame per token on FDDI to maximize throughput. Since this
+ may be a common feature in future workstations and servers,
+ interconnect devices with FDDI interfaces SHOULD be tested with 1, 4,
+ 8, and 16 frames per token. The reported frame rate SHOULD be the
+ average rate of frame transmission over the total trial period.
+
+23. Trial description
+
+ A particular test consists of multiple trials. Each trial returns
+ one piece of information, for example the loss rate at a particular
+ input frame rate. Each trial consists of a number of phases:
+
+ a) If the DUT is a router, send the routing update to the "input"
+ port and pause two seconds to be sure that the routing has settled.
+
+ b) Send the "learning frames" to the "output" port and wait 2
+ seconds to be sure that the learning has settled. Bridge learning
+ frames are frames with source addresses that are the same as the
+ destination addresses used by the test frames. Learning frames for
+ other protocols are used to prime the address resolution tables in
+ the DUT. The formats of the learning frame that should be used are
+ shown in the Test Frame Formats document.
+
+ c) Run the test trial.
+
+ d) Wait for two seconds for any residual frames to be received.
+
+ e) Wait for at least five seconds for the DUT to restabilize.
+
+24. Trial duration
+
+ The aim of these tests is to determine the rate continuously
+ supportable by the DUT. The actual duration of the test trials must
+ be a compromise between this aim and the duration of the benchmarking
+ test suite. The duration of the test portion of each trial SHOULD be
+ at least 60 seconds. The tests that involve some form of "binary
+ search", for example the throughput test, to determine the exact
+ result MAY use a shorter trial duration to minimize the length of the
+ search procedure, but the final determination SHOULD be made with
+
+
+
+Bradner & McQuaid Informational [Page 14]
+
+RFC 1944 Benchmarking Methodology May 1996
+
+
+ full length trials.
+
+25. Address resolution
+
+ The DUT SHOULD be able to respond to address resolution requests sent
+ by the DUT wherever the protocol requires such a process.
+
+26. Benchmarking tests:
+
+ Note: The notation "type of data stream" refers to the above
+ modifications to a frame stream with a constant inter-frame gap, for
+ example, the addition of traffic filters to the configuration of the
+ DUT.
+
+ 26.1 Throughput
+
+ Objective:
+ To determine the DUT throughput as defined in RFC 1242.
+
+ Procedure:
+ Send a specific number of frames at a specific rate through the
+ DUT and then count the frames that are transmitted by the DUT. If
+ the count of offered frames is equal to the count of received
+ frames, the rate of the offered stream is raised and the test
+ rerun. If fewer frames are received than were transmitted, the
+ rate of the offered stream is reduced and the test is rerun.
+
+ The throughput is the fastest rate at which the count of test
+ frames transmitted by the DUT is equal to the number of test
+ frames sent to it by the test equipment.
+
+ Reporting format:
+ The results of the throughput test SHOULD be reported in the form
+ of a graph. If it is, the x coordinate SHOULD be the frame size,
+ the y coordinate SHOULD be the frame rate. There SHOULD be at
+ least two lines on the graph. There SHOULD be one line showing
+ the theoretical frame rate for the media at the various frame
+ sizes. The second line SHOULD be the plot of the test results.
+ Additional lines MAY be used on the graph to report the results
+ for each type of data stream tested. Text accompanying the graph
+ SHOULD indicate the protocol, data stream format, and type of
+ media used in the tests.
+
+ We assume that if a single value is desired for advertising
+ purposes the vendor will select the rate for the minimum frame
+ size for the media. If this is done then the figure MUST be
+ expressed in frames per second. The rate MAY also be expressed in
+ bits (or bytes) per second if the vendor so desires. The
+
+
+
+Bradner & McQuaid Informational [Page 15]
+
+RFC 1944 Benchmarking Methodology May 1996
+
+
+ statement of performance MUST include a/ the measured maximum
+ frame rate, b/ the size of the frame used, c/ the theoretical
+ limit of the media for that frame size, and d/ the type of
+ protocol used in the test. Even if a single value is used as part
+ of the advertising copy, the full table of results SHOULD be
+ included in the product data sheet.
+
+ 26.2 Latency
+
+ Objective:
+ To determine the latency as defined in RFC 1242.
+
+ Procedure:
+ First determine the throughput for DUT at each of the listed frame
+ sizes. Send a stream of frames at a particular frame size through
+ the DUT at the determined throughput rate to a specific
+ destination. The stream SHOULD be at least 120 seconds in
+ duration. An identifying tag SHOULD be included in one frame
+ after 60 seconds with the type of tag being implementation
+ dependent. The time at which this frame is fully transmitted is
+ recorded (timestamp A). The receiver logic in the test equipment
+ MUST recognize the tag information in the frame stream and record
+ the time at which the tagged frame was received (timestamp B).
+
+ The latency is timestamp B minus timestamp A as per the relevant
+ definition frm RFC 1242, namely latency as defined for store and
+ forward devices or latency as defined for bit forwarding devices.
+
+ The test MUST be repeated at least 20 times with the reported
+ value being the average of the recorded values.
+
+ This test SHOULD be performed with the test frame addressed to the
+ same destination as the rest of the data stream and also with each
+ of the test frames addressed to a new destination network.
+
+ Reporting format:
+ The report MUST state which definition of latency (from RFC 1242)
+ was used for this test. The latency results SHOULD be reported
+ in the format of a table with a row for each of the tested frame
+ sizes. There SHOULD be columns for the frame size, the rate at
+ which the latency test was run for that frame size, for the media
+ types tested, and for the resultant latency values for each
+ type of data stream tested.
+
+
+
+
+
+
+
+
+Bradner & McQuaid Informational [Page 16]
+
+RFC 1944 Benchmarking Methodology May 1996
+
+
+ 26.3 Frame loss rate
+
+ Objective:
+ To determine the frame loss rate, as defined in RFC 1242, of a DUT
+ throughout the entire range of input data rates and frame sizes.
+
+ Procedure:
+ Send a specific number of frames at a specific rate through the
+ DUT to be tested and count the frames that are transmitted by the
+ DUT. The frame loss rate at each point is calculated using the
+ following equation:
+
+ ( ( input_count - output_count ) * 100 ) / input_count
+
+ The first trial SHOULD be run for the frame rate that corresponds
+ to 100% of the maximum rate for the frame size on the input media.
+ Repeat the procedure for the rate that corresponds to 90% of the
+ maximum rate used and then for 80% of this rate. This sequence
+ SHOULD be continued (at reducing 10% intervals) until there are
+ two successive trials in which no frames are lost. The maximum
+ granularity of the trials MUST be 10% of the maximum rate, a finer
+ granularity is encouraged.
+
+ Reporting format:
+ The results of the frame loss rate test SHOULD be plotted as a
+ graph. If this is done then the X axis MUST be the input frame
+ rate as a percent of the theoretical rate for the media at the
+ specific frame size. The Y axis MUST be the percent loss at the
+ particular input rate. The left end of the X axis and the bottom
+ of the Y axis MUST be 0 percent; the right end of the X axis and
+ the top of the Y axis MUST be 100 percent. Multiple lines on the
+ graph MAY used to report the frame loss rate for different frame
+ sizes, protocols, and types of data streams.
+
+ Note: See section 18 for the maximum frame rates that SHOULD be
+ used.
+
+ 26.4 Back-to-back frames
+
+ Objective:
+ To characterize the ability of a DUT to process back-to-back
+ frames as defined in RFC 1242.
+
+ Procedure:
+ Send a burst of frames with minimum inter-frame gaps to the DUT
+ and count the number of frames forwarded by the DUT. If the count
+ of transmitted frames is equal to the number of frames forwarded
+ the length of the burst is increased and the test is rerun. If
+
+
+
+Bradner & McQuaid Informational [Page 17]
+
+RFC 1944 Benchmarking Methodology May 1996
+
+
+ the number of forwarded frames is less than the number
+ transmitted, the length of the burst is reduced and the test is
+ rerun.
+
+ The back-to-back value is the number of frames in the longest
+ burst that the DUT will handle without the loss of any frames.
+ The trial length MUST be at least 2 seconds and SHOULD be
+ repeated at least 50 times with the average of the recorded values
+ being reported.
+
+ Reporting format:
+ The back-to-back results SHOULD be reported in the format of a
+ table with a row for each of the tested frame sizes. There SHOULD
+ be columns for the frame size and for the resultant average frame
+ count for each type of data stream tested. The standard deviation
+ for each measurement MAY also be reported.
+
+ 26.5 System recovery
+
+ Objective:
+ To characterize the speed at which a DUT recovers from an overload
+ condition.
+
+ Procedure:
+ First determine the throughput for a DUT at each of the listed
+ frame sizes.
+
+ Send a stream of frames at a rate 110% of the recorded throughput
+ rate or the maximum rate for the media, whichever is lower, for at
+ least 60 seconds. At Timestamp A reduce the frame rate to 50% of
+ the above rate and record the time of the last frame lost
+ (Timestamp B). The system recovery time is determined by
+ subtracting Timestamp B from Timestamp A. The test SHOULD be
+ repeated a number of times and the average of the recorded values
+ being reported.
+
+ Reporting format:
+ The system recovery results SHOULD be reported in the format of a
+ table with a row for each of the tested frame sizes. There SHOULD
+ be columns for the frame size, the frame rate used as the
+ throughput rate for each type of data stream tested, and for the
+ measured recovery time for each type of data stream tested.
+
+ 26.6 Reset
+
+ Objective:
+ To characterize the speed at which a DUT recovers from a device or
+ software reset.
+
+
+
+Bradner & McQuaid Informational [Page 18]
+
+RFC 1944 Benchmarking Methodology May 1996
+
+
+ Procedure:
+ First determine the throughput for the DUT for the minimum frame
+ size on the media used in the testing.
+
+ Send a continuous stream of frames at the determined throughput
+ rate for the minimum sized frames. Cause a reset in the DUT.
+ Monitor the output until frames begin to be forwarded and record
+ the time that the last frame (Timestamp A) of the initial stream
+ and the first frame of the new stream (Timestamp B) are received.
+ A power interruption reset test is performed as above except that
+ the power to the DUT should be interrupted for 10 seconds in place
+ of causing a reset.
+
+ This test SHOULD only be run using frames addressed to networks
+ directly connected to the DUT so that there is no requirement to
+ delay until a routing update is received.
+
+ The reset value is obtained by subtracting Timestamp A from
+ Timestamp B.
+
+ Hardware and software resets, as well as a power interruption
+ SHOULD be tested.
+
+ Reporting format:
+ The reset value SHOULD be reported in a simple set of statements,
+ one for each reset type.
+
+27. Security Considerations
+
+ Security issues are not addressed in this document.
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+Bradner & McQuaid Informational [Page 19]
+
+RFC 1944 Benchmarking Methodology May 1996
+
+
+28. Editors' Addresses
+
+ Scott Bradner
+ Harvard University
+ 1350 Mass. Ave, room 813
+ Cambridge, MA 02138
+
+ Phone +1 617 495-3864
+ Fax +1 617 496-8500
+ EMail: sob@harvard.edu
+
+
+ Jim McQuaid
+ Bay Networks
+ 3 Federal Street
+ Billerica, MA 01821
+
+ Phone +1 508 436-3915
+ Fax: +1 508 670-8145
+ EMail: jmcquaid@baynetworks.com
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+Bradner & McQuaid Informational [Page 20]
+
+RFC 1944 Benchmarking Methodology May 1996
+
+
+Appendix A: Testing Considerations
+
+A.1 Scope Of This Appendix
+
+ This appendix discusses certain issues in the benchmarking
+ methodology where experience or judgment may play a role in the tests
+ selected to be run or in the approach to constructing the test with a
+ particular DUT. As such, this appendix MUST not be read as an
+ amendment to the methodology described in the body of this document
+ but as a guide to testing practice.
+
+ 1. Typical testing practice has been to enable all protocols to be
+ tested and conduct all testing with no further configuration of
+ protocols, even though a given set of trials may exercise only one
+ protocol at a time. This minimizes the opportunities to "tune" a
+ DUT for a single protocol.
+
+ 2. The least common denominator of the available filter functions
+ should be used to ensure that there is a basis for comparison
+ between vendors. Because of product differences, those conducting
+ and evaluating tests must make a judgment about this issue.
+
+ 3. Architectural considerations may need to be considered. For
+ example, first perform the tests with the stream going between
+ ports on the same interface card and the repeat the tests with the
+ stream going into a port on one interface card and out of a port
+ on a second interface card. There will almost always be a best
+ case and worst case configuration for a given DUT architecture.
+
+ 4. Testing done using traffic streams consisting of mixed protocols
+ has not shown much difference between testing with individual
+ protocols. That is, if protocol A testing and protocol B testing
+ give two different performance results, mixed protocol testing
+ appears to give a result which is the average of the two.
+
+ 5. Wide Area Network (WAN) performance may be tested by setting up
+ two identical devices connected by the appropriate short- haul
+ versions of the WAN modems. Performance is then measured between
+ a LAN interface on one DUT to a LAN interface on the other DUT.
+
+ The maximum frame rate to be used for LAN-WAN-LAN configurations is a
+ judgment that can be based on known characteristics of the overall
+ system including compression effects, fragmentation, and gross link
+ speeds. Practice suggests that the rate should be at least 110% of
+ the slowest link speed. Substantive issues of testing compression
+ itself are beyond the scope of this document.
+
+
+
+
+
+Bradner & McQuaid Informational [Page 21]
+
+RFC 1944 Benchmarking Methodology May 1996
+
+
+Appendix B: Maximum frame rates reference
+
+ (Provided by Roger Beeman, Cisco Systems)
+
+ Size Ethernet 16Mb Token Ring FDDI
+ (bytes) (pps) (pps) (pps)
+
+ 64 14880 24691 152439
+ 128 8445 13793 85616
+ 256 4528 7326 45620
+ 512 2349 3780 23585
+ 768 1586 2547 15903
+ 1024 1197 1921 11996
+ 1280 961 1542 9630
+ 1518 812 1302 8138
+
+ Ethernet size
+ Preamble 64 bits
+ Frame 8 x N bits
+ Gap 96 bits
+
+ 16Mb Token Ring size
+ SD 8 bits
+ AC 8 bits
+ FC 8 bits
+ DA 48 bits
+ SA 48 bits
+ RI 48 bits ( 06 30 00 12 00 30 )
+ SNAP
+ DSAP 8 bits
+ SSAP 8 bits
+ Control 8 bits
+ Vendor 24 bits
+ Type 16 bits
+ Data 8 x ( N - 18) bits
+ FCS 32 bits
+ ED 8 bits
+ FS 8 bits
+
+ Tokens or idles between packets are not included
+
+ FDDI size
+ Preamble 64 bits
+ SD 8 bits
+ FC 8 bits
+ DA 48 bits
+ SA 48 bits
+ SNAP
+
+
+
+Bradner & McQuaid Informational [Page 22]
+
+RFC 1944 Benchmarking Methodology May 1996
+
+
+ DSAP 8 bits
+ SSAP 8 bits
+ Control 8 bits
+ Vendor 24 bits
+ Type 16 bits
+ Data 8 x ( N - 18) bits
+ FCS 32 bits
+ ED 4 bits
+ FS 12 bits
+
+Appendix C: Test Frame Formats
+
+ This appendix defines the frame formats that may be used with these
+ tests. It also includes protocol specific parameters for TCP/IP over
+ Ethernet to be used with the tests as an example.
+
+C.1. Introduction
+
+ The general logic used in the selection of the parameters and the
+ design of the frame formats is explained for each case within the
+ TCP/IP section. The same logic has been used in the other sections.
+ Comments are used in these sections only if there is a protocol
+ specific feature to be explained. Parameters and frame formats for
+ additional protocols can be defined by the reader by using the same
+ logic.
+
+C.2. TCP/IP Information
+
+ The following section deals with the TCP/IP protocol suite.
+
+ C.2.1 Frame Type.
+ An application level datagram echo request is used for the test
+ data frame in the protocols that support such a function. A
+ datagram protocol is used to minimize the chance that a router
+ might expect a specific session initialization sequence, as might
+ be the case for a reliable stream protocol. A specific defined
+ protocol is used because some routers verify the protocol field
+ and refuse to forward unknown protocols.
+
+ For TCP/IP a UDP Echo Request is used.
+
+ C.2.2 Protocol Addresses
+ Two sets of addresses must be defined: first the addresses
+ assigned to the router ports, and second the address that are to
+ be used in the frames themselves and in the routing updates.
+
+ The network addresses 192.18.0.0 through 192.19.255.255 are have
+ been assigned to the BMWG by the IANA for this purpose. This
+
+
+
+Bradner & McQuaid Informational [Page 23]
+
+RFC 1944 Benchmarking Methodology May 1996
+
+
+ assignment was made to minimize the chance of conflict in case a
+ testing device were to be accidentally connected to part of the
+ Internet. The specific use of the addresses is detailed below.
+
+ C.2.2.1 Router port protocol addresses
+ Half of the ports on a multi-port router are referred to as
+ "input" ports and the other half as "output" ports even though
+ some of the tests use all ports both as input and output. A
+ contiguous series of IP Class C network addresses from
+ 198.18.1.0 to 198.18.64.0 have been assigned for use on the
+ "input" ports. A second series from 198.19.1.0 to 198.19.64.0
+ have been assigned for use on the "output" ports. In all cases
+ the router port is node 1 on the appropriate network. For
+ example, a two port DUT would have an IP address of 198.18.1.1
+ on one port and 198.19.1.1 on the other port.
+
+ Some of the tests described in the methodology memo make use of
+ an SNMP management connection to the DUT. The management
+ access address for the DUT is assumed to be the first of the
+ "input" ports (198.18.1.1).
+
+ C.2.2.2 Frame addresses
+ Some of the described tests assume adjacent network routing
+ (the reboot time test for example). The IP address used in the
+ test frame is that of node 2 on the appropriate Class C
+ network. (198.19.1.2 for example)
+
+ If the test involves non-adjacent network routing the phantom
+ routers are located at node 10 of each of the appropriate Class
+ C networks. A series of Class C network addresses from
+ 198.18.65.0 to 198.18.254.0 has been assigned for use as the
+ networks accessible through the phantom routers on the "input"
+ side of DUT. The series of Class C networks from 198.19.65.0
+ to 198.19.254.0 have been assigned to be used as the networks
+ visible through the phantom routers on the "output" side of the
+ DUT.
+
+ C.2.3 Routing Update Frequency
+
+ The update interval for each routing protocol is may have to be
+ determined by the specifications of the individual protocol. For
+ IP RIP, Cisco IGRP and for OSPF a routing update frame or frames
+ should precede each stream of test frames by 5 seconds. This
+ frequency is sufficient for trial durations of up to 60 seconds.
+ Routing updates must be mixed with the stream of test frames if
+ longer trial periods are selected. The frequency of updates
+ should be taken from the following table.
+
+
+
+
+Bradner & McQuaid Informational [Page 24]
+
+RFC 1944 Benchmarking Methodology May 1996
+
+
+ IP-RIP 30 sec
+ IGRP 90 sec
+ OSPF 90 sec
+
+ C.2.4 Frame Formats - detailed discussion
+
+ C.2.4.1 Learning Frame
+ In most protocols a procedure is used to determine the mapping
+ between the protocol node address and the MAC address. The
+ Address Resolution Protocol (ARP) is used to perform this
+ function in TCP/IP. No such procedure is required in XNS or
+ IPX because the MAC address is used as the protocol node
+ address.
+
+ In the ideal case the tester would be able to respond to ARP
+ requests from the DUT. In cases where this is not possible an
+ ARP request should be sent to the router's "output" port. This
+ request should be seen as coming from the immediate destination
+ of the test frame stream. (i.e. the phantom router (Figure 2)
+ or the end node if adjacent network routing is being used.) It
+ is assumed that the router will cache the MAC address of the
+ requesting device. The ARP request should be sent 5 seconds
+ before the test frame stream starts in each trial. Trial
+ lengths of longer than 50 seconds may require that the router
+ be configured for an extended ARP timeout.
+
+ +--------+ +------------+
+ | | | phantom |------ P LAN
+ A
+ IN A------| DUT |------------| |------ P LAN
+ B
+ | | OUT A | router |------ P LAN
+ C
+ +--------+ +------------+
+
+ Figure 2
+
+ In the case where full routing is being used
+
+ C.2.4.2 Routing Update Frame
+ If the test does not involve adjacent net routing the tester
+ must supply proper routing information using a routing update.
+ A single routing update is used before each trial on each
+ "destination" port (see section C.24). This update includes
+ the network addresses that are reachable through a phantom
+ router on the network attached to the port. For a full mesh
+ test, one destination network address is present in the routing
+ update for each of the "input" ports. The test stream on each
+
+
+
+Bradner & McQuaid Informational [Page 25]
+
+RFC 1944 Benchmarking Methodology May 1996
+
+
+ "input" port consists of a repeating sequence of frames, one to
+ each of the "output" ports.
+
+ C.2.4.3 Management Query Frame
+ The management overhead test uses SNMP to query a set of
+ variables that should be present in all DUTs that support SNMP.
+ The variables for a single interface only are read by an NMS
+ at the appropriate intervals. The list of variables to
+ retrieve follow:
+
+ sysUpTime
+ ifInOctets
+ ifOutOctets
+ ifInUcastPkts
+ ifOutUcastPkts
+
+ C.2.4.4 Test Frames
+ The test frame is an UDP Echo Request with enough data to fill
+ out the required frame size. The data should not be all bits
+ off or all bits on since these patters can cause a "bit
+ stuffing" process to be used to maintain clock synchronization
+ on WAN links. This process will result in a longer frame than
+ was intended.
+
+ C.2.4.5 Frame Formats - TCP/IP on Ethernet
+ Each of the frames below are described for the 1st pair of DUT
+ ports, i.e. "input" port #1 and "output" port #1. Addresses
+ must be changed if the frame is to be used for other ports.
+
+ C.2.6.1 Learning Frame
+
+ ARP Request on Ethernet
+
+ -- DATAGRAM HEADER
+ offset data (hex) description
+ 00 FF FF FF FF FF FF dest MAC address send to
+ broadcast address
+ 06 xx xx xx xx xx xx set to source MAC address
+ 12 08 06 ARP type
+ 14 00 01 hardware type Ethernet = 1
+ 16 08 00 protocol type IP = 800
+ 18 06 hardware address length 48 bits
+ on Ethernet
+ 19 04 protocol address length 4 octets
+ for IP
+ 20 00 01 opcode request = 1
+ 22 xx xx xx xx xx xx source MAC address
+ 28 xx xx xx xx source IP address
+
+
+
+Bradner & McQuaid Informational [Page 26]
+
+RFC 1944 Benchmarking Methodology May 1996
+
+
+ 32 FF FF FF FF FF FF requesting DUT's MAC address
+ 38 xx xx xx xx DUT's IP address
+
+ C.2.6.2 Routing Update Frame
+
+ -- DATAGRAM HEADER
+ offset data (hex) description
+ 00 FF FF FF FF FF FF dest MAC address is broadcast
+ 06 xx xx xx xx xx xx source hardware address
+ 12 08 00 type
+
+ -- IP HEADER
+ 14 45 IP version - 4, header length (4
+ byte units) - 5
+ 15 00 service field
+ 16 00 EE total length
+ 18 00 00 ID
+ 20 40 00 flags (3 bits) 4 (do not
+ fragment),
+ fragment offset-0
+ 22 0A TTL
+ 23 11 protocol - 17 (UDP)
+ 24 C4 8D header checksum
+ 26 xx xx xx xx source IP address
+ 30 xx xx xx destination IP address
+ 33 FF host part = FF for broadcast
+
+ -- UDP HEADER
+ 34 02 08 source port 208 = RIP
+ 36 02 08 destination port 208 = RIP
+ 38 00 DA UDP message length
+ 40 00 00 UDP checksum
+
+ -- RIP packet
+ 42 02 command = response
+ 43 01 version = 1
+ 44 00 00 0
+
+ -- net 1
+ 46 00 02 family = IP
+ 48 00 00 0
+ 50 xx xx xx net 1 IP address
+ 53 00 net not node
+ 54 00 00 00 00 0
+ 58 00 00 00 00 0
+ 62 00 00 00 07 metric 7
+
+ -- net 2
+
+
+
+Bradner & McQuaid Informational [Page 27]
+
+RFC 1944 Benchmarking Methodology May 1996
+
+
+ 66 00 02 family = IP
+ 68 00 00 0
+ 70 xx xx xx net 2 IP address
+ 73 00 net not node
+ 74 00 00 00 00 0
+ 78 00 00 00 00 0
+ 82 00 00 00 07 metric 7
+
+ -- net 3
+ 86 00 02 family = IP
+ 88 00 00 0
+ 90 xx xx xx net 3 IP address
+ 93 00 net not node
+ 94 00 00 00 00 0
+ 98 00 00 00 00 0
+ 102 00 00 00 07 metric 7
+
+ -- net 4
+ 106 00 02 family = IP
+ 108 00 00 0
+ 110 xx xx xx net 4 IP address
+ 113 00 net not node
+ 114 00 00 00 00 0
+ 118 00 00 00 00 0
+ 122 00 00 00 07 metric 7
+
+ -- net 5
+ 126 00 02 family = IP
+ 128 00 00 0
+ 130 00 net 5 IP address
+ 133 00 net not node
+ 134 00 00 00 00 0
+ 138 00 00 00 00 0
+ 142 00 00 00 07 metric 7
+
+ -- net 6
+ 146 00 02 family = IP
+ 148 00 00 0
+ 150 xx xx xx net 6 IP address
+ 153 00 net not node
+ 154 00 00 00 00 0
+ 158 00 00 00 00 0
+ 162 00 00 00 07 metric 7
+
+ C.2.4.6 Management Query Frame
+
+ To be defined.
+
+
+
+
+Bradner & McQuaid Informational [Page 28]
+
+RFC 1944 Benchmarking Methodology May 1996
+
+
+ C.2.6.4 Test Frames
+ UDP echo request on Ethernet
+
+ -- DATAGRAM HEADER
+ offset data (hex) description
+ 00 xx xx xx xx xx xx set to dest MAC address
+ 06 xx xx xx xx xx xx set to source MAC address
+ 12 08 00 type
+
+ -- IP HEADER
+ 14 45 IP version - 4 header length 5 4
+ byte units
+ 15 00 TOS
+ 16 00 2E total length*
+ 18 00 00 ID
+ 20 00 00 flags (3 bits) - 0 fragment
+ offset-0
+ 22 0A TTL
+ 23 11 protocol - 17 (UDP)
+ 24 C4 8D header checksum*
+ 26 xx xx xx xx set to source IP address**
+ 30 xx xx xx xx set to destination IP address**
+
+ -- UDP HEADER
+ 34 C0 20 source port
+ 36 00 07 destination port 07 = Echo
+ 38 00 1A UDP message length*
+ 40 00 00 UDP checksum
+
+ -- UDP DATA
+ 42 00 01 02 03 04 05 06 07 some data***
+ 50 08 09 0A 0B 0C 0D 0E 0F
+
+ * - change for different length frames
+
+ ** - change for different logical streams
+
+ *** - fill remainder of frame with incrementing octets,
+ repeated if required by frame length
+
+
+
+
+
+
+
+
+
+
+
+
+Bradner & McQuaid Informational [Page 29]
+
+RFC 1944 Benchmarking Methodology May 1996
+
+
+ Values to be used in Total Length and UDP message length fields:
+
+ frame size total length UDP message length
+ 64 00 2E 00 1A
+ 128 00 6E 00 5A
+ 256 00 EE 00 9A
+ 512 01 EE 01 9A
+ 768 02 EE 02 9A
+ 1024 03 EE 03 9A
+ 1280 04 EE 04 9A
+ 1518 05 DC 05 C8
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+Bradner & McQuaid Informational [Page 30]
+