diff options
Diffstat (limited to 'doc/rfc/rfc3116.txt')
-rw-r--r-- | doc/rfc/rfc3116.txt | 7115 |
1 files changed, 7115 insertions, 0 deletions
diff --git a/doc/rfc/rfc3116.txt b/doc/rfc/rfc3116.txt new file mode 100644 index 0000000..2bd77df --- /dev/null +++ b/doc/rfc/rfc3116.txt @@ -0,0 +1,7115 @@ + + + + + + +Network Working Group J. Dunn +Request for Comments: 3116 C. Martin +Category: Informational ANC, Inc. + June 2001 + + + Methodology for ATM Benchmarking + +Status of this Memo + + This memo provides information for the Internet community. It does + not specify an Internet standard of any kind. Distribution of this + memo is unlimited. + +Copyright Notice + + Copyright (C) The Internet Society (2001). All Rights Reserved. + +Abstract + + This document discusses and defines a number of tests that may be + used to describe the performance characteristics of ATM (Asynchronous + Transfer Mode) based switching devices. In addition to defining the + tests this document also describes specific formats for reporting the + results of the tests. + + This memo is a product of the Benchmarking Methodology Working Group + (BMWG) of the Internet Engineering Task Force (IETF). + +Table of Contents + + 1. Introduction . . . . . . . . . . . . . . . . . . . . . . . . . 4 + 2. Background . . . . . . . . . . . . . . . . . . . . . . . . . . 5 + 2.1. Test Device Requirements . . . . . . . . . . . . . . . . . . 5 + 2.2. Systems Under Test (SUTs). . . . . . . . . . . . . . . . . . 5 + 2.3. Test Result Evaluation . . . . . . . . . . . . . . . . . . . 5 + 2.4. Requirements . . . . . . . . . . . . . . . . . . . . . . . . 5 + 2.5. Test Configurations for SONET. . . . . . . . . . . . . . . . 6 + 2.6. SUT Configuration. . . . . . . . . . . . . . . . . . . . . . 7 + 2.7. Frame Formats. . . . . . . . . . . . . . . . . . . . . . . . 8 + 2.8. Frame Sizes. . . . . . . . . . . . . . . . . . . . . . . . . 8 + 2.9. Verifying Received IP PDU's. . . . . . . . . . . . . . . . . 9 + 2.10. Modifiers . . . . . . . . . . . . . . . . . . . . . . . . . 9 + 2.10.1. Management IP PDU's . . . . . . . . . . . . . . . . . . . 9 + 2.10.2. Routing Update IP PDU's . . . . . . . . . . . . . . . . . 10 + 2.11. Filters . . . . . . . . . . . . . . . . . . . . . . . . . . 10 + 2.11.1. Filter Addresses. . . . . . . . . . . . . . . . . . . . . 11 + 2.12. Protocol Addresses. . . . . . . . . . . . . . . . . . . . . 12 + + + +Dunn & Martin Informational [Page 1] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + 2.13. Route Set Up. . . . . . . . . . . . . . . . . . . . . . . . 12 + 2.14. Bidirectional Traffic . . . . . . . . . . . . . . . . . . . 12 + 2.15. Single Stream Path. . . . . . . . . . . . . . . . . . . . . 12 + 2.16. Multi-port. . . . . . . . . . . . . . . . . . . . . . . . . 13 + 2.17. Multiple Protocols. . . . . . . . . . . . . . . . . . . . . 14 + 2.18. Multiple IP PDU Sizes . . . . . . . . . . . . . . . . . . . 14 + 2.19. Testing Beyond a Single SUT . . . . . . . . . . . . . . . . 14 + 2.20. Maximum IP PDU Rate . . . . . . . . . . . . . . . . . . . . 15 + 2.21. Busty Traffic . . . . . . . . . . . . . . . . . . . . . . . 15 + 2.22. Trial Description . . . . . . . . . . . . . . . . . . . . . 16 + 2.23. Trial Duration. . . . . . . . . . . . . . . . . . . . . . . 16 + 2.24. Address Resolution. . . . . . . . . . . . . . . . . . . . . 16 + 2.25. Synchronized Payload Bit Pattern. . . . . . . . . . . . . . 16 + 2.26. Burst Traffic Descriptors . . . . . . . . . . . . . . . . . 17 + 3. Performance Metrics. . . . . . . . . . . . . . . . . . . . . . 17 + 3.1. Physical Layer-SONET . . . . . . . . . . . . . . . . . . . . 17 + 3.1.1. Pointer Movements. . . . . . . . . . . . . . . . . . . . . 17 + 3.1.1.1. Pointer Movement Propagation . . . . . . . . . . . . . . 17 + 3.1.1.2. Cell Loss due to Pointer Movement. . . . . . . . . . . . 19 + 3.1.1.3. IP Packet Loss due to Pointer Movement . . . . . . . . . 20 + 3.1.2. Transport Overhead (TOH) Error Count . . . . . . . . . . . 21 + 3.1.2.1. TOH Error Propagation. . . . . . . . . . . . . . . . . . 21 + 3.1.2.2. Cell Loss due to TOH Error . . . . . . . . . . . . . . . 22 + 3.1.2.3. IP Packet Loss due to TOH Error. . . . . . . . . . . . . 23 + 3.1.3. Path Overhead (POH) Error Count. . . . . . . . . . . . . . 24 + 3.1.3.1. POH Error Propagation. . . . . . . . . . . . . . . . . . 24 + 3.1.3.2. Cell Loss due to POH Error . . . . . . . . . . . . . . . 25 + 3.1.3.3. IP Packet Loss due to POH Error. . . . . . . . . . . . . 26 + 3.2. ATM Layer. . . . . . . . . . . . . . . . . . . . . . . . . . 27 + 3.2.1. Two-Point Cell Delay Variation (CDV) . . . . . . . . . . . 27 + 3.2.1.1. Test Setup . . . . . . . . . . . . . . . . . . . . . . . 27 + 3.2.1.2. Two-point CDV/Steady Load/One VCC. . . . . . . . . . . . 27 + 3.2.1.3. Two-point CDV/Steady Load/Twelve VCCs. . . . . . . . . . 28 + 3.2.1.4. Two-point CDV/Steady Load/Maximum VCCs . . . . . . . . . 30 + 3.2.1.5. Two-point CDV/Bursty VBR Load/One VCC. . . . . . . . . . 31 + 3.2.1.6. Two-point CDV/Bursty VBR Load/Twelve VCCs. . . . . . . . 32 + 3.2.1.7. Two-point CDV/Bursty VBR Load/Maximum VCCs . . . . . . . 34 + 3.2.1.8. Two-point CDV/Mixed Load/Three VCC's . . . . . . . . . . 35 + 3.2.1.9. Two-point CDV/Mixed Load/Twelve VCCs . . . . . . . . . . 36 + 3.2.1.10. Two-point CDV/Mixed Load/Maximum VCCs . . . . . . . . . 38 + 3.2.2. Cell Error Ratio (CER) . . . . . . . . . . . . . . . . . . 39 + 3.2.2.1. Test Setup . . . . . . . . . . . . . . . . . . . . . . . 39 + 3.2.2.2. CER/Steady Load/One VCC. . . . . . . . . . . . . . . . . 40 + 3.2.2.3. CER/Steady Load/Twelve VCCs. . . . . . . . . . . . . . . 41 + 3.2.2.4. CER/Steady Load/Maximum VCCs . . . . . . . . . . . . . . 42 + 3.2.2.5. CER/Bursty VBR Load/One VCC. . . . . . . . . . . . . . . 43 + 3.2.2.6. CER/Bursty VBR Load/Twelve VCCs. . . . . . . . . . . . . 44 + 3.2.2.7. CER/Bursty VBR Load/Maximum VCCs . . . . . . . . . . . . 46 + + + +Dunn & Martin Informational [Page 2] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + 3.2.3. Cell Loss Ratio (CLR). . . . . . . . . . . . . . . . . . . 47 + 3.2.3.1. CLR/Steady Load/One VCC. . . . . . . . . . . . . . . . . 47 + 3.2.3.2. CLR/Steady Load/Twelve VCCs. . . . . . . . . . . . . . . 48 + 3.2.3.3. CLR/Steady Load/Maximum VCCs . . . . . . . . . . . . . . 49 + 3.2.3.4. CLR/Bursty VBR Load/One VCC. . . . . . . . . . . . . . . 51 + 3.2.3.5. CLR/Bursty VBR Load/Twelve VCCs. . . . . . . . . . . . . 52 + 3.2.3.6. CLR/Bursty VBR Load/Maximum VCCs . . . . . . . . . . . . 53 + 3.2.4. Cell Misinsertion Rate (CMR) . . . . . . . . . . . . . . . 54 + 3.2.4.1. CMR/Steady Load/One VCC. . . . . . . . . . . . . . . . . 54 + 3.2.4.2. CMR/Steady Load/Twelve VCCs. . . . . . . . . . . . . . . 55 + 3.2.4.3. CMR/Steady Load/Maximum VCCs . . . . . . . . . . . . . . 57 + 3.2.4.4. CMR/Bursty VBR Load/One VCC. . . . . . . . . . . . . . . 58 + 3.2.4.5. CMR/Bursty VBR Load/Twelve VCCs. . . . . . . . . . . . . 59 + 3.2.4.6. CMR/Bursty VBR Load/Maximum VCCs . . . . . . . . . . . . 60 + 3.2.5. CRC Error Ratio (CRC-ER) . . . . . . . . . . . . . . . . . 62 + 3.2.5.1. CRC-ER/Steady Load/One VCC . . . . . . . . . . . . . . . 62 + 3.2.5.2. CRC-ER/Steady Load/Twelve VCCs . . . . . . . . . . . . . 63 + 3.2.5.3. CRC-ER/Steady Load/Maximum VCCs. . . . . . . . . . . . . 64 + 3.2.5.4. CRC-ER/Bursty VBR Load/One VCC . . . . . . . . . . . . . 65 + 3.2.5.5. CRC-ER/Bursty VBR Load/Twelve VCCs . . . . . . . . . . . 66 + 3.2.5.6. CRC-ER/Bursty VBR Load/Maximum VCCs. . . . . . . . . . . 68 + 3.2.5.7. CRC-ER/Bursty UBR Load/One VCC . . . . . . . . . . . . . 69 + 3.2.5.8. CRC-ER/Bursty UBR Load/Twelve VCCs . . . . . . . . . . . 70 + 3.2.5.9. CRC-ER/Bursty UBR Load/Maximum VCCs. . . . . . . . . . . 71 + 3.2.5.10. CRC-ER/Bursty Mixed Load/Three VCC. . . . . . . . . . . 73 + 3.2.5.11. CRC-ER/Bursty Mixed Load/Twelve VCCs. . . . . . . . . . 74 + 3.2.5.12. CRC-ER/Bursty Mixed Load/Maximum VCCs . . . . . . . . . 75 + 3.2.6. Cell Transfer Delay (CTD). . . . . . . . . . . . . . . . . 76 + 3.2.6.1. Test Setup . . . . . . . . . . . . . . . . . . . . . . . 76 + 3.2.6.2. CTD/Steady Load/One VCC. . . . . . . . . . . . . . . . . 77 + 3.2.6.3. CTD/Steady Load/Twelve VCCs. . . . . . . . . . . . . . . 78 + 3.2.6.4. CTD/Steady Load/Maximum VCCs . . . . . . . . . . . . . . 79 + 3.2.6.5. CTD/Bursty VBR Load/One VCC. . . . . . . . . . . . . . . 81 + 3.2.6.6. CTD/Bursty VBR Load/Twelve VCCs. . . . . . . . . . . . . 82 + 3.2.6.7. CTD/Bursty VBR Load/Maximum VCCs . . . . . . . . . . . . 83 + 3.2.6.8. CTD/Bursty UBR Load/One VCC. . . . . . . . . . . . . . . 85 + 3.2.6.9. CTD/Bursty UBR Load/Twelve VCCs. . . . . . . . . . . . . 86 + 3.2.6.10. CTD/Bursty UBR Load/Maximum VCCs. . . . . . . . . . . . 87 + 3.2.6.11. CTD/Mixed Load/Three VCC's. . . . . . . . . . . . . . . 88 + 3.2.6.12. CTD/Mixed Load/Twelve VCCs. . . . . . . . . . . . . . . 90 + 3.2.6.13. CTD/Mixed Load/Maximum VCCs . . . . . . . . . . . . . . 91 + 3.3. ATM Adaptation Layer (AAL) Type 5 (AAL5) . . . . . . . . . . 93 + 3.3.1. IP Packet Loss due to AAL5 Re-assembly Errors. . . . . . . 93 + 3.3.2. AAL5 Re-assembly Time. . . . . . . . . . . . . . . . . . . 94 + 3.3.3. AAL5 CRC Error Ratio . . . . . . . . . . . . . . . . . . . 95 + 3.3.3.1. Test Setup . . . . . . . . . . . . . . . . . . . . . . . 95 + 3.3.3.2. AAL5-CRC-ER/Steady Load/One VCC. . . . . . . . . . . . . 95 + 3.3.3.3. AAL5-CRC-ER/Steady Load/Twelve VCCs. . . . . . . . . . . 96 + + + +Dunn & Martin Informational [Page 3] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + 3.3.3.4. AAL5-CRC-ER/Steady Load/Maximum VCCs . . . . . . . . . . 97 + 3.3.3.5. AAL5-CRC-ER/Bursty VBR Load/One VCC. . . . . . . . . . . 99 + 3.3.3.6. AAL5-CRC-ER/Bursty VBR Load/Twelve VCCs. . . . . . . . .100 + 3.3.3.7. AAL5-CRC-ER/Bursty VBR Load/Maximum VCCs . . . . . . . .101 + 3.3.3.8. AAL5-CRC-ER/Mixed Load/Three VCC's . . . . . . . . . . .102 + 3.3.3.9. AAL5-CRC-ER/Mixed Load/Twelve VCCs . . . . . . . . . . .104 + 3.3.3.10. AAL5-CRC-ER/Mixed Load/Maximum VCCs . . . . . . . . . .105 + 3.4. ATM Service: Signaling . . . . . . . . . . . . . . . . . . .106 + 3.4.1. CAC Denial Time and Connection Establishment Time. . . . .106 + 3.4.2. Connection Teardown Time . . . . . . . . . . . . . . . . .107 + 3.4.3. Crankback Time . . . . . . . . . . . . . . . . . . . . . .108 + 3.4.4. Route Update Response Time . . . . . . . . . . . . . . . .109 + 3.5. ATM Service: ILMI. . . . . . . . . . . . . . . . . . . . . .110 + 3.5.1. MIB Alignment Time . . . . . . . . . . . . . . . . . . . .110 + 3.5.2. Address Registration Time. . . . . . . . . . . . . . . . .111 + 4. Security Considerations . . . . . . . . . . . . . . . . . . .112 + 5. Notices. . . . . . . . . . . . . . . . . . . . . . . . . . . .112 + 6. References . . . . . . . . . . . . . . . . . . . . . . . . . .113 + 7. Authors' Addresses . . . . . . . . . . . . . . . . . . . . . .113 + APPENDIX A . . . . . . . . . . . . . . . . . . . . . . . . . . .114 + APPENDIX B . . . . . . . . . . . . . . . . . . . . . . . . . . .114 + APPENDIX C . . . . . . . . . . . . . . . . . . . . . . . . . . .116 + Full Copyright Statement . . . . . . . . . . . . . . . . . . . .127 + +1. Introduction + + This document defines a specific set of tests that vendors can use to + measure and report the performance characteristics of ATM network + devices. The results of these tests will provide the user comparable + data from different vendors with which to evaluate these devices. + The methods defined in this memo are based on RFC 2544 "Benchmarking + Methodology for Network Interconnect Devices". + + The document "Terminology for ATM Benchmarking" (RFC 2761), defines + many of the terms that are used in this document. The terminology + document should be consulted before attempting to make use of this + document. + + The BMWG produces two major classes of documents: Benchmarking + Terminology documents and Benchmarking Methodology documents. The + Terminology documents present the benchmarks and other related terms. + The Methodology documents define the procedures required to collect + the benchmarks cited in the corresponding Terminology documents. + + + + + + + + +Dunn & Martin Informational [Page 4] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + +2. Background + +2.1. Test Device Requirements + + This document is based on the requirement that a test device is + available. The test device can either be off the shelf or can be + easily built with current technologies. The test device must have a + transmitting and receiving port for the interface type under test. + The test device must be configured to transmit test PDUs and to + analyze received PDUs. The test device should be able to transmit + and analyze received data at the same time. + +2.2. Systems Under Test (SUTs) + + There are a number of tests described in this document that do not + apply to each SUT. Vendors should perform all of the tests that can + be supported by a specific product type. It will take some time to + perform all of the recommended tests under all of the recommended + conditions. + +2.3. Test Result Evaluation + + Performing all of the tests in this document will result in a great + deal of data. The applicability of this data to the evaluation of a + particular SUT will depend on its expected use and the configuration + of the network in which it will be used. For example, the time + required by a switch to provide ILMI services will not be a pertinent + measurement in a network that does not use the ILMI protocol, such as + an ATM WAN. Evaluating data relevant to a particular network + installation may require considerable experience, which may not be + readily available. Finally, test selection and evaluation of test + results must be done with an understanding of generally accepted + testing practices regarding repeatability, variance and the + statistical significance of a small numbers of trials. + +2.4. Requirements + + In this document, the words that are used to define the significance + of each particular requirement are capitalized. These words are: + + * "MUST" This word, or the words "REQUIRED" and "SHALL" mean that + the item is an absolute requirement of the specification. + + * "SHOULD" This word or the adjective "RECOMMENDED" means that there + may exist valid reasons in particular circumstances to ignore this + item, but the full implications should be understood and the case + carefully weighed before choosing a different course. + + + + +Dunn & Martin Informational [Page 5] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + * "MAY" This word or the adjective "OPTIONAL" means that this item + is truly optional. One vendor may choose to include the item + because a particular marketplace requires it or because it + enhances the product, for example; another vendor may omit the + same item. + + An implementation is not compliant if it fails to satisfy one or more + of the MUST requirements for the protocols it implements. An + implementation that satisfies all the MUST and all the SHOULD + requirements for its protocols is said to be "unconditionally + compliant"; one that satisfies all the MUST requirements but not all + the SHOULD requirements for its protocols is said to be + "conditionally compliant". + +2.5. Test Configurations for SONET + + The test device can be connected to the SUT in a variety of + configurations depending on the test point. The following + configurations will be used for the tests described in this document. + + 1) Uni-directional connection: The test devices transmit port + (labeled Tx) is connected to the SUT receive port (labeled Rx). + The SUTs transmit port is connected to the test device receive + port (see Figure 1). In this configuration, the test device can + verify that all transmitted packets are acknowledged correctly. + Note that this configuration does not verify internal system + functions, but verifies one port on the SUT. + + +-------------+ +-------------+ + | Tx|-------------->|Rx | + | Test Rx|<--------------|Tx SUT | + | Device | | | + +-------------+ +-------------+ + + Figure 1 + + 2) Bi-directional connection: The test devices first transmit port is + connected to the SUTs first receive port. The SUTs first transmit + port is connected to the test devices first receive port. The + test devices second transmit port is connected to the SUTs second + receive port. The SUTs second transmit port is connected to the + test devices second receive port (see Figure 2). In this + configuration, the test device can determine if all of the + transmitted packets were received and forwarded correctly. Note + that this configuration does verify internal system functions, + since it verifies two ports on the SUT. + + + + + +Dunn & Martin Informational [Page 6] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + +-------------+ +-------------+ + | Test Tx|-------------->|Rx | + | Device Rx|<--------------|Tx SUT | + | Tx Rx | | Tx Rx | + +-------------+ +-------------+ + | ^ | ^ + | | | | + | +------------------------+ | + | | + |---------------------------------| + + Figure 2 + + 3) Uni-directional passthrough connection: The test devices first + transmit port is connected to the SUT1 receive port. The SUT1 + transmit port is connected to the test devices first receive port. + The test devices second transmit port is connected to the SUT2 + receive port. The SUT2 transmit port is connected to the test + devices second receive port (see Figure 3). In this + configuration, the test device can determine if all of the packets + transmitted by SUT1 were correctly acknowledged by SUT2. Note + that this configuration does not verify internal system functions, + but verifies one port on each SUT. + + +-------------+ +-------------+ +-------------+ + | Tx|---------->|Rx Tx|---------->|Rx | + | SUT1 Rx|<----------|Tx Test Rx|<----------|Tx SUT2 | + | | | Device | | | + +-------------+ +-------------+ +-------------+ + + Figure 3 + +2.6. SUT Configuration + + The SUT MUST be configured as described in the SUT users guide. + Specifically, it is expected that all of the supported protocols will + be configured and enabled. It is expected that all of the tests will + be run without changing the configuration or setup of the SUT in any + way other than that required to do the specific test. For example, + it is not acceptable to disable all but one transport protocol when + testing the throughput of that protocol. If PNNI or BISUP is used to + initiate switched virtual connections (SVCs), the SUT configuration + SHOULD include the normally recommended routing update intervals and + keep alive frequency. The specific version of the software and the + exact SUT configuration, including what functions are disabled and + used during the tests MUST be included as part of the report of the + results. + + + + +Dunn & Martin Informational [Page 7] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + +2.7. Frame formats + + The formats of the test IP PDUs to use for TCP/IP and UPC/IP over ATM + are shown in Appendix C: Test Frame Formats. Note that these IP PDUs + are in accordance with RFC 2225. These exact IP PDU formats SHOULD + be used in the tests described in this document for this + protocol/media combination. These IP PDUs will be used as a template + for testing other protocol/media combinations. The specific formats + that are used to define the test IP PDUs for a particular test series + MUST be included in the report of the results. + +2.8. Frame sizes + + All of the described tests SHOULD be performed using a number of IP + PDU sizes. Specifically, the sizes SHOULD include the maximum and + minimum legitimate sizes for the protocol under test on the media + under test and enough sizes in between to be able to get a full + characterization of the SUT performance. Except where noted, at + least five IP PDU sizes SHOULD be tested for each test condition. + + Theoretically the minimum size UDP Echo request IP PDU would consist + of an IP header (minimum length 20 octets), a UDP header (8 octets), + AAL5 trailer (8 octets) and an LLC/SNAP code point header (8 octets); + therefore, the minimum size PDU will fit into one ATM cell. The + theoretical maximum IP PDU size is determined by the size of the + length field in the IP header. In almost all cases the actual + maximum and minimum sizes are determined by the limitations of the + media. In the case of ATM, the maximum IP PDU size SHOULD be the ATM + MTU size, which is 9180 octets. + + In theory it would be ideal to distribute the IP PDU sizes in a way + that would evenly distribute the theoretical IP PDU rates. These + recommendations incorporate this theory but specify IP PDU sizes, + which are easy to understand and remember. In addition, many of the + same IP PDU sizes are specified on each of the media types to allow + for easy performance comparisons. + + Note: The inclusion of an unrealistically small IP PDU size on some + of the media types (i.e., with little or no space for data) is to + help characterize the per-IP PDU processing overhead of the SUT. + + The IP PDU sizes that will be used are: + + 44, 64, 128, 256, 1024, 1518, 2048, 4472, 9180 + + The minimum size IP PDU for UDP on ATM is 44 octets, the minimum size + of 44 is recommended to allow direct comparison to token ring + performance. The IP PDU size of 4472 is recommended instead of the + + + +Dunn & Martin Informational [Page 8] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + theoretical FDDI maximum size of 4500 octets in order to permit the + same type of comparison. An IP (i.e., not UDP) IP PDU may be used in + addition if a higher data rate is desired, in which case the minimum + IP PDU size is 28 octets. + +2.9. Verifying received IP PDUs + + The test equipment SHOULD discard any IP PDUs received during a test + run that are not actual forwarded test IP PDUs. For example, keep- + alive and routing update IP PDUs SHOULD NOT be included in the count + of received IP PDUs. In any case, the test equipment SHOULD verify + the length of the received IP PDUs and check that they match the + expected length. + + Preferably, the test equipment SHOULD include sequence numbers in the + transmitted IP PDUs and check for these numbers on the received IP + PDUs. If this is done, the reported results SHOULD include, in + addition to the number of IP PDUs dropped, the number of IP PDUs that + were received out of order, the number of duplicate IP PDUs received + and the number of gaps in the received IP PDU numbering sequence. + This functionality is required for some of the described tests. + +2.10. Modifiers + + It is useful to characterize the SUTs performance under a number of + conditions. Some of these conditions are noted below. The reported + results SHOULD include as many of these conditions as the test + equipment is able to generate. The suite of tests SHOULD be run + first without any modifying conditions, then repeated under each of + the modifying conditions separately. To preserve the ability to + compare the results of these tests, any IP PDUs that are required to + generate the modifying conditions (excluding management queries) will + be included in the same data stream as that of the normal test IP + PDUs and in place of one of the test IP PDUs. They MUST not be + supplied to the SUT on a separate network port. + +2.10.1. Management IP PDUs + + Most ATM data networks now make use of ILMI, signaling and OAM. In + many environments, there can be a number of management stations + sending queries to the same SUT at the same time. + + Management queries MUST be made in accordance with the applicable + specification, e.g., ILMI sysUpTime getNext requests will be made in + accordance with ILMI 4.0. The response to the query MUST be verified + by the test equipment. Note that, for each management protocol in + + + + + +Dunn & Martin Informational [Page 9] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + use, this requires that the test equipment implement the associated + protocol state machine. One example of the specific query IP PDU + (ICMP) that should be used is shown in Appendix C. + +2.10.2. Routing update IP PDUs + + The processing of PNNI updates could have a significant impact on the + ability of a switch to forward cells and complete calls. If PNNI is + configured on the SUT, one routing update MUST be transmitted before + the first test IP PDU is transmitted during the trial. The test + SHOULD verify that the SUT has properly processed the routing update. + + PNNI routing update IP PDUs SHOULD be sent at the rate specified in + Appendix B. Appendix C defines one routing update PDU for the TCP/IP + over ATM example. The routing updates are designed to change the + routing on a number of networks that are not involved in the + forwarding of the test data. The first IP PDU sets the routing table + state to "A", the second one changes the state to "B". The IP PDUs + MUST be alternated during the trial. The test SHOULD verify that the + SUT has properly processed the routing update. + +2.11. Filters + + Filters are added to switches to selectively inhibit the forwarding + of cells that would normally be forwarded. This is usually done to + implement security controls on the data that is accepted between one + area and another. Different products have different capabilities to + implement filters. Filters are applicable only if the SUT supports + the filtering feature. + + The SUT SHOULD be first configured to add one filter condition and + the tests performed. This filter SHOULD permit the forwarding of the + test data stream. This filter SHOULD be of the form as described in + the SUT Users Guide. + + The SUT SHOULD be then reconfigured to implement a total of 25 + filters. The first 24 of these filters SHOULD be based on 24 + separate ATM NSAP Network Prefix addresses. + + The 24 ATM NSAP Network Prefix addresses SHOULD not be any that are + represented in the test data stream. The last filter SHOULD permit + the forwarding of the test data stream. By "first" and "last" we + mean to ensure that in the second case, 25 conditions must be checked + before the data IP over ATM PDUs will match the conditions that + permit the forwarding of the IP PDU. Of course, if the SUT reorders + the filters or does not use a linear scan of the filter rules the + effect of the sequence in which the filters are input is properly + lost. + + + +Dunn & Martin Informational [Page 10] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + The exact filters configuration command lines used SHOULD be included + with the report of the results. + +2.11.1. Filter Addresses + + Two sets of filter addresses are required, one for the single filter + case and one for the 25 filter case. + + The single filter case should permit traffic from ATM address [Switch + Network Prefix] 00 00 00 00 00 01 00 to ATM address [Switch Network + Prefix] 00 00 00 00 00 02 00 and deny all other traffic. Note that + the 13 octet Switch Network Prefix MUST be configured before this + test can be run. + + The 25 filter case should follow the following sequence. + + deny [Switch Network Prefix] 00 00 00 00 00 01 00 + to [Switch Network Prefix] 00 00 00 00 00 03 00 + deny [Switch Network Prefix] 00 00 00 00 00 01 00 + to [Switch Network Prefix] 00 00 00 00 00 04 00 + deny [Switch Network Prefix] 00 00 00 00 00 01 00 + to [Switch Network Prefix] 00 00 00 00 00 05 00 + ... + deny [Switch Network Prefix] 00 00 00 00 00 01 00 + to [Switch Network Prefix] 00 00 00 00 00 0C 00 + deny [Switch Network Prefix] 00 00 00 00 00 01 00 + to [Switch Network Prefix] 00 00 00 00 00 0D 00 + allow [Switch Network Prefix] 00 00 00 00 00 01 00 + to [Switch Network Prefix] 00 00 00 00 00 02 00 + deny [Switch Network Prefix] 00 00 00 00 00 01 00 + to [Switch Network Prefix] 00 00 00 00 00 0E 00 + deny [Switch Network Prefix] 00 00 00 00 00 01 00 + to [Switch Network Prefix] 00 00 00 00 00 0F 00 + ... + deny [Switch Network Prefix] 00 00 00 00 00 01 00 + to [Switch Network Prefix] 00 00 00 00 00 18 00 + deny all else + + All previous filter conditions should be cleared from the switch + before this sequence is entered. The sequence is selected to test to + see if the switch sorts the filter conditions or accepts them in the + order that they were entered. Both of these procedures will result + in a greater impact on performance than will some form of hash + coding. + + + + + + + +Dunn & Martin Informational [Page 11] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + +2.12. Protocol addresses + + It is easier to implement these tests using a single logical stream + of data, with one source ATM address and one destination ATM address, + and for some conditions like the filters described above, a practical + requirement. Networks in the real world are not limited to single + streams of data. The test suite SHOULD be first run with a single + ATM source and destination address pair. The tests SHOULD then be + repeated with using a random destination address. In the case of + testing single switches, the addresses SHOULD be random and uniformly + distributed over a range of 256 seven octet user parts. In the case + of testing multiple interconnected switches, the addresses SHOULD be + random and uniformly distributed over the 256 network prefixes, each + of which should support 256 seven octet user parts. The specific + address ranges to use for ATM are shown in Appendix A. IP to ATM + address mapping MUST be accomplished as described in RFC 2225. + +2.13. Route Set Up + + It is not reasonable that all of the routing information necessary to + forward the test stream, especially in the multiple address case, + will be manually set up. If PNNI and/or ILMI are running, at the + start of each trial a routing update MUST be sent to the SUT. This + routing update MUST include all of the ATM addresses that will be + required for the trial. This routing update will have to be repeated + at the interval required by PNNI or ILMI. An example of the format + and repetition interval of the update IP PDUs is given in Appendix B + (interval and size) and Appendix C (format). + +2.14. Bidirectional traffic + + Bidirectional performance tests SHOULD be run with the same data rate + being offered from each direction. The sum of the data rates should + not exceed the theoretical limit for the media. + +2.15. Single stream path + + The full suite of tests SHOULD be run with the appropriate modifiers + for a single receive and transmit port on the SUT. If the internal + design of the SUT has multiple distinct pathways, for example, + multiple interface cards each with multiple network ports, then all + possible permutations of pathways SHOULD be tested separately. If + multiple interconnected switches are tested, the test MUST specify + routes, which allow only one path between source and destination ATM + addresses. + + + + + + +Dunn & Martin Informational [Page 12] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + +2.16. Multi-port + + Many switch products provide several network ports on the same + interface module. Each port on an interface module SHOULD be + stimulated in an identical manner. Specifically, half of the ports + on each module SHOULD be receive ports and half SHOULD be transmit + ports. For example if a SUT has two interface module each of which + has four ports, two ports on each interface module be receive ports + and two will be transmit ports. Each receive port MUST be offered + the same data rate. The addresses in the input data streams SHOULD + be set so that an IP PDU will be directed to each of the transmit + ports in sequence. That is, all transmit ports will receive an + identical distribution of IP PDUs from a particular receive port. + + Consider the following 6 port SUT: + + -------------- + ---------| Rx A Tx X|-------- + ---------| Rx B Tx Y|-------- + ---------| Rx C Tx Z|-------- + -------------- + + The addressing of the data streams for each of the inputs SHOULD be: + + stream sent to Rx A: + IP PDU to Tx X, IP PDU to Tx Y, IP PDU to Tx Z + stream sent to Rx B: + IP PDU to Tx X, IP PDU to Tx Y, IP PDU to Tx Z + stream sent to Rx C + IP PDU to Tx X, IP PDU to Tx Y, IP PDU to Tx Z + + Note: Each stream contains the same sequence of IP destination + addresses; therefore, each transmit port will receive 3 IP PDUs + simultaneously. This procedure ensures that the SUT will have to + process multiple IP PDUs addressed to the same transmit port + simultaneously. + + The same configuration MAY be used to perform a bi-directional + multi-stream test. In this case all of the ports are considered both + receive and transmit ports. Each data stream MUST consist of IP PDUs + whose addresses correspond to the ATM addresses all of the other + ports. + + + + + + + + + +Dunn & Martin Informational [Page 13] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + +2.17. Multiple protocols + + This document does not address the issue of testing the effects of a + mixed protocol environment other than to suggest that if such tests + are wanted then PDUs SHOULD be distributed between all of the test + protocols. The distribution MAY approximate the conditions on the + network in which the SUT would be used. + +2.18. Multiple IP PDU sizes + + This document does not address the issue of testing the effects of a + mixed IP PDU size environment other than to suggest that, if such + tests are required, then IP PDU size SHOULD be evenly distributed + among all of the PDU sizes listed in this document. The distribution + MAY approximate the conditions on the network in which the SUT would + be used. + +2.19. Testing beyond a single SUT + + In the performance testing of a single SUT, the paradigm can be + described as applying some input to a SUT and monitoring the output. + The results of which can be used to form a basis of characterization + of that device under those test conditions. + + This model is useful when the test input and output are homogeneous + (e.g., 64-byte IP, AAL5 PDUs into the SUT; 64 byte IP, AAL5 PDUs + out). + + By extending the single SUT test model, reasonable benchmarks + regarding multiple SUTs or heterogeneous environments may be + collected. In this extension, the single SUT is replaced by a system + of interconnected network SUTs. This test methodology would support + the benchmarking of a variety of device/media/service/protocol + combinations. For example, a configuration for a LAN-to-WAN-to-LAN + test might be: + + (1) ATM UNI -> SUT 1 -> BISUP -> SUT 2 -> ATM UNI + + Or an extended LAN configuration might be: + + (2) ATM UNI -> SUT 1 -> PNNI Network -> SUT 2 -> ATM UNI + + In both examples 1 and 2, end-to-end benchmarks of each system could + be empirically ascertained. Other behavior may be characterized + through the use of intermediate devices. In example 2, the + configuration may be used to give an indication of the effect of PNNI + routing on IP throughput. + + + + +Dunn & Martin Informational [Page 14] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + Because multiple SUTs are treated as a single system, there are + limitations to this methodology. For instance, this methodology may + yield an aggregate benchmark for a tested system. That benchmark + alone, however, may not necessarily reflect asymmetries in behavior + between the SUTs, latencies introduced by other apparatus (e.g., + CSUs/DSUs, switches), etc. + + Further, care must be used when comparing benchmarks of different + systems by ensuring that the SUTs' features and configuration of the + tested systems have the appropriate common denominators to allow + comparison. + +2.20. Maximum IP PDU rate + + The maximum IP PDU rates that should be used when testing LAN + connections SHOULD be the listed theoretical maximum rate for the IP + PDU size on the media. + + The maximum IP PDU rate that should be used when testing WAN + connections SHOULD be greater than the listed theoretical maximum + rate for the IP PDU size on that speed connection. The higher rate + for WAN tests is to compensate for the fact that some vendors employ + various forms of header compression. + + A list of maximum IP PDU rates for LAN connections is included in + Appendix B. + +2.21. Bursty traffic + + It is convenient to measure the SUT performance under steady state + load; however, this is an unrealistic way to gauge the functioning of + a SUT. Actual network traffic normally consists of bursts of IP + PDUs. + + Some of the tests described below SHOULD be performed with both + constant bit rate, bursty Unspecified Bit Rate (UBR) Best Effort + [AF-TM4.1] and Variable Bit Rate Non-real Time (VBR-nrt) Best Effort + [AF-TM4.1]. The IP PDUs within a burst are transmitted with the + minimum legitimate inter-IP PDU gap. + + The objective of the test is to determine the minimum interval + between bursts that the SUT can process with no IP PDU loss. Tests + SHOULD be run with burst sizes of 10% of Maximum Burst Size (MBS), + 20% of MBS, 50% of MBS and 100% MBS. Note that the number of IP PDUs + in each burst will depend on the PDU size. For UBR, the MBS refers + to the associated VBR traffic parameters. + + + + + +Dunn & Martin Informational [Page 15] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + +2.22. Trial description + + A particular test consists of multiple trials. Each trial returns + one piece of information, for example the loss rate at a particular + input IP PDU rate. Each trial consists of five of phases: + + a) If the SUT is a switch supporting PNNI, send the routing update to + the SUT receive port and wait two seconds to be sure that the + routing has settled. + + b) Send an ATM ARP PDU to determine the ATM address corresponding to + the destination IP address. The formats of the ATM ARP PDU that + should be used are shown in the Test Frame Formats document and + MUST be in accordance with RFC 2225. + + c) Stimulate SUT with traffic load. + + d) Wait for two seconds for any residual IP PDUs to be received. + + e) Wait for at least five seconds for the SUT to restabilize. + +2.23. Trial duration + + The objective of the tests defined in this document is to accurately + characterize the behavior of a particular piece of network equipment + under varying traffic loads. The choice of test duration must be a + compromise between this objective and keeping the duration of the + benchmarking test suite within reasonable bounds. The SUT SHOULD be + stimulated for at least 60 seconds. If this time period results in a + high variance in the test results, the SUT SHOULD be stimulated for + at least 300 seconds. + +2.24. Address resolution + + The SUT MUST be able to respond to address resolution requests sent + by another SUT, an ATM ARP server or the test equipment in accordance + with RFC 2225. + +2.25. Synchronized Payload Bit Pattern. + + Some measurements assume that both the transmitter and receiver + payload information is synchronized. Synchronization MUST be + achieved by supplying a known bit pattern to both the transmitter and + receiver. This bit pattern MUST be one of the following: PRBS-15, + PRBS-23, 0xFF00, or 0xAA55. + + + + + + +Dunn & Martin Informational [Page 16] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + +2.26. Burst Traffic Descriptors. + + Some measurements require busty traffic patterns. These patterns + MUST conform to one of the following traffic descriptors: + +1) PCR=100% allotted line rate, SCR=50% allotted line rate, and MBS=8192 + +2) PCR=100% allotted line rate, SCR=50% allotted line rate, and MBS=4096 + +3) PCR=90% allotted line rate, SCR=50% allotted line rate, and MBS=8192 + +4) PCR=90% allotted line rate, SCR=50% allotted line rate, and MBS=4096 + +5) PCR=90% allotted line rate, SCR=45% allotted line rate, and MBS=8192 + +6) PCR=90% allotted line rate, SCR=45% allotted line rate, and MBS=4096 + +7) PCR=80% allotted line rate, SCR=40% allotted line rate, and MBS=65536 + +8) PCR=80% allotted line rate, SCR=40% allotted line rate, and MBS=32768 + + The allotted line rate refers to the total available line rate + divided by the number of VCCs in use. + +3. Performance Metrics + +3.1. Physical Layer-SONET + +3.1.1. Pointer Movements + +3.1.1.1. Pointer Movement Propagation. + + Objective: To determine that the SUT does not propagate pointer + movements as defined in RFC 2761 "Terminology for ATM Benchmarking". + + Procedure: + + 1) Set up the SUT and test device using the uni-directional + configuration. + + 2) Send a specific number of IP PDUs at a specific rate through the + SUT. Since this test is not a throughput test, the rate should + not be greater than 90% of line rate. The cell payload SHOULD + contain valid IP PDUs. The IP PDUs MUST be encapsulated in AAL5. + + + + + + + +Dunn & Martin Informational [Page 17] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + 3) Count the IP PDUs that are transmitted by the SUT to verify + connectivity and load. If the count on the test device is the + same on the SUT, continue the test, else lower the test device + traffic rate until the counts are the same. + + 4) Inject one forward payload pointer movement. Verify that the SUT + does not change the pointer. + + 5) Inject one forward payload pointer movement every 1 second. + Verify that the SUT does not change the pointer. + + 6) Discontinue the payload pointer movement. + + 7) Inject five forward payload pointer movements every 1 second. + Verify that the SUT does not change the pointer. + + 8) Discontinue the payload pointer movement. + + 9) Inject one backward payload pointer movement. Verify that the + SUT does not change the pointer. + + 10) Inject one backward payload pointer movement every 1 second. + Verify that the SUT does not change the pointer. + + 11) Discontinue the payload pointer movement. + + 12) Inject five backward payload pointer movements every 1 second. + Verify that the SUT does not change the pointer. + + 13) Discontinue the payload pointer movement. + + Reporting Format: + + The results of the pointer movement propagation test SHOULD be + reported in a form of a table. The rows SHOULD be labeled single + pointer movement, one pointer movement per second, and five + pointer movements per second. The columns SHOULD be labeled + pointer movement and loss of pointer. The elements of the table + SHOULD be either True or False, indicating whether the particular + condition was observed for each test. + + The table MUST also indicate the IP PDU size in octets and traffic + rate in IP PDUs per second as generated by the test device. + + + + + + + + +Dunn & Martin Informational [Page 18] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + +3.1.1.2. Cell Loss due to Pointer Movement. + + Objective: To determine if the SUT will drop cells due to pointer + movements as defined in RFC 2761 "Terminology for ATM Benchmarking". + + Procedure: + + 1) Set up the SUT and test device using the uni-directional + configuration. + + 2) Send a specific number of cells at a specific rate through the + SUT. Since this test is not a throughput test, the rate should + not be greater than 90% of line rate. The cell payload SHOULD + contain valid IP PDUs. The IP PDUs MUST be encapsulated in AAL5. + + 3) Count the cells that are transmitted by the SUT to verify + connectivity and load. If the count on the test device is the + same on the SUT, continue the test; else lower the test device + traffic rate until the counts are the same. + + 4) Inject one forward payload pointer movement. Verify that the SUT + does not drop any cells. + + 5) Inject one forward payload pointer movement every 1 second. + Verify that the SUT does not drop any cells. + + 6) Discontinue the payload pointer movement. + + 7) Inject five forward payload pointer movements every 1 second. + Verify that the SUT does not drop any cells. + + 8) Discontinue the payload pointer movement. + + 9) Inject one backward payload pointer movement. Verify that the + SUT does not drop any cells. + + 10) Inject one backward payload pointer movement every 1 second. + Verify that the SUT does not drop any cells. + + 11) Discontinue the payload pointer movement. + + 12) Inject five backward payload pointer movements every 1 second. + Verify that the SUT does not drop any cells. + + 13) Discontinue the payload pointer movement. + + + + + + +Dunn & Martin Informational [Page 19] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + Reporting Format: + + The results of the cell loss due to pointer movement test SHOULD + be reported in a form of a table. The rows SHOULD be labeled + single pointer movement, one pointer movement per second, and five + pointer movements per second. The columns SHOULD be labeled cell + loss and number of cells lost. The elements of column 1 SHOULD be + either True or False, indicating whether the particular condition + was observed for each test. The elements of column 2 SHOULD be + non-negative integers. + + The table MUST also indicate the traffic rate in IP PDUs per + second as generated by the test device. + +3.1.1.3. IP Packet Loss due to Pointer Movement. + + Objective: To determine if the SUT will drop IP packets due to + pointer movements as defined in RFC 2761 "Terminology for ATM + Benchmarking". + + Procedure: + + 1) Set up the SUT and test device using the uni-directional + configuration. + + 2) Send a specific number of IP packets at a specific rate through + the SUT. Since this test is not a throughput test, the rate + should not be greater than 90% of line rate. The IP PDUs MUST be + encapsulated in AAL5. + + 3) Count the IP packets that are transmitted by the SUT to verify + connectivity and load. If the count on the test device is the + same on the SUT, continue the test; else lower the test device + traffic rate until the counts are the same. + + 4) Inject one forward payload pointer movement. Verify that the SUT + does not drop any packets. + + 5) Inject one forward payload pointer movement every 1 second. + Verify that the SUT does not drop any packets. + + 6) Discontinue the payload pointer movement. + + 7) Inject five forward payload pointer movements every 1 second. + Verify that the SUT does not drop any packets. + + 8) Discontinue the payload pointer movement. + + + + +Dunn & Martin Informational [Page 20] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + 9) Inject one backward payload pointer movement. Verify that the + SUT does not drop any packets. + + 10) Inject one backward payload pointer movement every 1 second. + Verify that the SUT does not drop any packets. + + 11) Discontinue the payload pointer movement. + + 12) Inject five backward payload pointer movements every 1 second. + Verify that the SUT does not drop any packets. + + 13) Discontinue the payload pointer movement. + + Reporting Format: + + The results of the IP packet loss due to pointer movement test + SHOULD be reported in a form of a table. The rows SHOULD be + labeled single pointer movement, one pointer movement per second, + and five pointer movements per second. The columns SHOULD be + labeled packet loss and number of packets lost. The elements of + column 1 SHOULD be either True or False, indicating whether the + particular condition was observed for each test. The elements of + column 2 SHOULD be non-negative integers. + + The table MUST also indicate the packet size in octets and traffic + rate in packets per second as generated by the test device. + +3.1.2. Transport Overhead (TOH) Error Count + +3.1.2.1. TOH Error Propagation. + + Objective: To determine that the SUT does not propagate TOH errors as + defined in RFC 2761 "Terminology for ATM Benchmarking". + + Procedure: + + 1) Set up the SUT and test device using the uni-directional + configuration. + + 2) Send a specific number of IP PDUs at a specific rate through the + SUT. Since this test is not a throughput test, the rate should + not be greater than 90% of line rate. The cell payload SHOULD + contain valid IP PDUs. The IP PDUs MUST be encapsulated in AAL5. + + 3) Count the IP PDUs that are transmitted by the SUT to verify + connectivity and load. If the count on the test device is the + same on the SUT, continue the test, else lower the test device + traffic rate until the counts are the same. + + + +Dunn & Martin Informational [Page 21] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + 4) Inject one error in the first bit of the A1 and A2 Frameword. + Verify that the SUT does not propagate the error. + + 5) Inject one error in the first bit of the A1 and A2 Frameword + every 1 second. Verify that the SUT does not propagate the + error. + + 6) Discontinue the Frameword error. + + 7) Inject one error in the first bit of the A1 and A2 Frameword for + 4 consecutive IP PDUs in every 6 IP PDUs. Verify that the SUT + indicates Loss of Frame. + + 8) Discontinue the Frameword error. + + Reporting Format: + + The results of the TOH error propagation test SHOULD be reported + in a form of a table. The rows SHOULD be labeled single error, + one error per second, and four consecutive errors every 6 IP PDUs. + The columns SHOULD be labeled error propagated and loss of IP PDU. + The elements of the table SHOULD be either True or False, + indicating whether the particular condition was observed for each + test. + + The table MUST also indicate the IP PDU size in octets and traffic + rate in IP PDUs per second as generated by the test device. + +3.1.2.2. c TOH Error. + + Objective: To determine if the SUT will drop cells due TOH Errors as + defined in RFC 2761 "Terminology for ATM Benchmarking". + + Procedure: + + 1) Set up the SUT and test device using the uni-directional + configuration. + + 2) Send a specific number of cells at a specific rate through the + SUT. Since this test is not a throughput test, the rate should + not be greater than 90% of line rate. The cell payload SHOULD + contain valid IP PDUs. The IP PDUs MUST be encapsulated in AAL5. + + 3) Count the cells that are transmitted by the SUT to verify + connectivity and load. If the count on the test device is the + same on the SUT, continue the test; else lower the test device + traffic rate until the counts are the same. + + + + +Dunn & Martin Informational [Page 22] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + 4) Inject one error in the first bit of the A1 and A2 Frameword. + Verify that the SUT does not drop any cells. + + 5) Inject one error in the first bit of the A1 and A2 Frameword + every 1 second. Verify that the SUT does not drop any cells. + + 6) Discontinue the Frameword error. + + 7) Inject one error in the first bit of the A1 and A2 Frameword for + 4 consecutive IP PDUs in every 6 IP PDUs. Verify that the SUT + does drop cells. + + 8) Discontinue the Frameword error. + + Reporting Format: + + The results of the Cell Loss due to TOH errors test SHOULD be + reported in a form of a table. The rows SHOULD be labeled single + error, one error per second, and four consecutive errors every 6 + IP PDUs. The columns SHOULD be labeled cell loss and number of + cells lost. The elements of column 1 SHOULD be either True or + False, indicating whether the particular condition was observed + for each test. The elements of column 2 SHOULD be non-negative + integers. + + The table MUST also indicate the traffic rate in IP PDUs per + second as generated by the test device. + +3.1.2.3. IP Packet Loss due to TOH Error. + + Objective: To determine if the SUT will drop IP packets due to TOH + errors as defined in RFC 2761 "Terminology for ATM Benchmarking". + + Procedure: + + 1) Set up the SUT and test device using the uni-directional + configuration. + + 2) Send a specific number of IP packets at a specific rate through + the SUT. Since this test is not a throughput test, the rate + should not be greater than 90% of line rate. The IP PDUs MUST be + encapsulated in AAL5. + + 3) Count the IP packets that are transmitted by the SUT to verify + connectivity and load. If the count on the test device is the + same on the SUT, continue the test; else lower the test device + traffic rate until the counts are the same. + + + + +Dunn & Martin Informational [Page 23] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + 4) Inject one error in the first bit of the A1 and A2 Frameword. + Verify that the SUT does not drop any packets. + + 5) Inject one error in the first bit of the A1 and A2 Frameword + every 1 second. Verify that the SUT does not drop any packets. + + 6) Discontinue the Frameword error. + + 7) Inject one error in the first bit of the A1 and A2 Frameword for + 4 consecutive IP PDUs in every 6 IP PDUs. Verify that the SUT + does drop packets. + + 8) Discontinue the Frameword error. + + Reporting Format: + + The results of the IP packet loss due to TOH errors test SHOULD be + reported in a form of a table. The rows SHOULD be labeled single + error, one error per second, and four consecutive errors every 6 + IP PDUs. The columns SHOULD be labeled packet loss and number of + packets lost. The elements of column 1 SHOULD be either True or + False, indicating whether the particular condition was observed + for each test. The elements of column 2 SHOULD be non-negative + integers. + + The table MUST also indicate the packet size in octets and traffic + rate in packets per second as generated by the test device. + +3.1.3. Path Overhead (POH) Error Count + +3.1.3.1. POH Error Propagation. + + Objective: To determine that the SUT does not propagate POH errors as + defined in RFC 2761 "Terminology for ATM Benchmarking". + + Procedure: + + 1) Set up the SUT and test device using the uni-directional + configuration. + + 2) Send a specific number of IP PDUs at a specific rate through the + SUT. Since this test is not a throughput test, the rate should + not be greater than 90% of line rate. The cell payload SHOULD + contain valid IP PDUs. The IP PDUs MUST be encapsulated in AAL5. + + + + + + + +Dunn & Martin Informational [Page 24] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + 3) Count the IP PDUs that are transmitted by the SUT to verify + connectivity and load. If the count on the test device is the + same on the SUT, continue the test, else lower the test device + traffic rate until the counts are the same. + + 4) Inject one error in the B3 (Path BIP8) byte. Verify that the SUT + does not propagate the error. + + 5) Inject one error in the B3 byte every 1 second. Verify that the + SUT does not propagate the error. + + 6) Discontinue the POH error. + + Reporting Format: + + The results of the POH error propagation test SHOULD be reported + in a form of a table. The rows SHOULD be labeled single error + and one error per second. The columns SHOULD be labeled error + propagated and loss of IP PDU. The elements of the table SHOULD + be either True or False, indicating whether the particular + condition was observed for each test. + + The table MUST also indicate the IP PDU size in octets and + traffic rate in IP PDUs per second as generated by the test + device. + +3.1.3.2. Cell Loss due to POH Error. + + Objective: To determine if the SUT will drop cells due POH Errors as + defined in RFC 2761 "Terminology for ATM Benchmarking". + + Procedure: + + 1) Set up the SUT and test device using the uni-directional + configuration. + + 2) Send a specific number of cells at a specific rate through the + SUT. Since this test is not a throughput test, the rate should + not be greater than 90% of line rate. The cell payload SHOULD + contain valid IP PDUs. The IP PDUs MUST be encapsulated in AAL5. + + 3) Count the cells that are transmitted by the SUT to verify + connectivity and load. If the count on the test device is the + same on the SUT, continue the test; else lower the test device + traffic rate until the counts are the same. + + 4) Inject one error in the B3 (Path BIP8) byte. Verify that the SUT + does not drop any cells. + + + +Dunn & Martin Informational [Page 25] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + 5) Inject one error in the B3 byte every 1 second. Verify that the + SUT does not drop any cells. + + 6) Discontinue the POH error. + + Reporting Format: + + The results of the Cell Loss due to POH errors test SHOULD be + reported in a form of a table. The rows SHOULD be labeled single + error and one error per second. The columns SHOULD be labeled + cell loss and number of cells lost. The elements of column 1 + SHOULD be either True or False, indicating whether the particular + condition was observed for each test. The elements of column 2 + SHOULD be non-negative integers. + + The table MUST also indicate the traffic rate in IP PDUs per + second as generated by the test device. + +3.1.3.3. IP Packet Loss due to POH Error. + + Objective: To determine if the SUT will drop IP packets due to POH + errors as defined in RFC 2761 "Terminology for ATM Benchmarking". + + Procedure: + + 1) Set up the SUT and test device using the uni-directional + configuration. + + 2) Send a specific number of IP packets at a specific rate through + the SUT. Since this test is not a throughput test, the rate + should not be greater than 90% of line rate. The IP PDUs MUST be + encapsulated in AAL5. + + 3) Count the IP packets that are transmitted by the SUT to verify + connectivity and load. If the count on the test device is the + same on the SUT, continue the test; else lower the test device + traffic rate until the counts are the same. + + 4) Inject one error in the B3 (Path BIP8) byte. Verify that the SUT + does not drop any packets. + + 5) Inject one error in the B3 byte every 1 second. Verify that the + SUT does not drop any packets. + + 6) Discontinue the POH error. + + + + + + +Dunn & Martin Informational [Page 26] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + Reporting Format: + + The results of the IP packet loss due to POH errors test SHOULD be + reported in a form of a table. The rows SHOULD be labeled single + error and one error per second. The columns SHOULD be labeled + packet loss and number of packets lost. The elements of column 1 + SHOULD be either True or False, indicating whether the particular + condition was observed for each test. The elements of column 2 + SHOULD be non-negative integers. + + The table MUST also indicate the packet size in octets and traffic + rate in packets per second as generated by the test device. + +3.2. ATM Layer + +3.2.1. Two-Point Cell Delay Variation (CDV) + +3.2.1.1. Test Setup + + The cell delay measurements assume that both the transmitter and + receiver timestamp information is synchronized. Synchronization + SHOULD be achieved by supplying a common clock signal (minimum of 100 + Mhz or 10 ns resolution) to both the transmitter and receiver. The + maximum timestamp values MUST be recorded to ensure synchronization + in the case of counter rollover. The cell delay measurements SHOULD + utilize the O.191 cell (ITUT-O.191) encapsulated in a valid IP + packet. If the O.191 cell is not available, a test cell encapsulated + in a valid IP packet MAY be used. The test cell MUST contain a + transmit timestamp which can be correlated with a receive timestamp. + A description of the test cell MUST be included in the test results. + The description MUST include the timestamp length (in bits), counter + rollover value, and the timestamp accuracy (in ns). + +3.2.1.2. Two-point CDV/Steady Load/One VCC + + Objective: To determine the SUT variation in cell transfer delay with + one VCC as defined in RFC 2761 "Terminology for ATM Benchmarking". + + Procedure: + + 1) Set up the SUT and test device using the bi-directional + configuration. + + 2) Configure the SUT and test device with one VCC. The VCC SHOULD + contain one VPI/VCI. The VCC MUST be configured as either a CBR, + VBR, or UBR connection. The VPI/VCI MUST not be one of the + reserved ATM signaling channels (e.g., [0,5], [0,16]). + + + + +Dunn & Martin Informational [Page 27] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + 3) Send a specific number of IP packets containing timestamps at a + specific constant rate through the SUT via the defined test VCC. + Since this test is not a throughput test, the rate should not be + greater than 90% of line rate. The IP PDUs MUST be encapsulated + in AAL5. + + 4) Count the IP packets that are transmitted by the SUT to verify + connectivity and load. If the count on the test device is the + same on the SUT, continue the test; else lower the test device + traffic rate until the counts are the same. + + 5) Record the packets timestamps at the transmitter and receiver + ends of the test device. + + Reporting Format: + + The results of the Two-point CDV/Steady Load/One VCC test SHOULD + be reported in a form of text, graph, and histogram. + + The text results SHOULD display the numerical values of the CDV. + The values given SHOULD include: time period of test in s, test + VPI/VCI value, total number of cells transmitted and received on + the given VPI/VCI during the test in positive integers, maximum + and minimum CDV during the test in us, and peak-to-peak CDV in us. + + The graph results SHOULD display the cell delay values. The x- + coordinate SHOULD be the test run time in either seconds, minutes + or days depending on the total length of the test. The x- + coordinate time SHOULD be configurable. The y-coordinate SHOULD + be the cell delay in us. The integration time per point MUST be + indicated. + + The histogram results SHOULD display the peak-to-peak cell delay. + The x-coordinate SHOULD be the cell delay in us with at least 256 + bins. The y-coordinate SHOULD be the number of cells observed in + each bin. + + The results MUST also indicate the packet size in octets, traffic + rate in packets per second, and bearer class as generated by the + test device. The VCC and VPI/VCI values MUST be indicated. The + bearer class of the created VCC MUST also be indicated. + +3.2.1.3. Two-point CDV/Steady Load/Twelve VCCs + + Objective: To determine the SUT variation in cell transfer delay with + twelve VCCs as defined in RFC 2761 "Terminology for ATM + Benchmarking". + + + + +Dunn & Martin Informational [Page 28] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + Procedure: + + 1) Set up the SUT and test device using the bi-directional + configuration. + + 2) Configure the SUT and test device with twelve VCCs, using 1 VPI + and 12 VCIs. The VCC's MUST be configured as either a CBR, VBR, + or UBR connection. The VPI/VCIs MUST not be one of the reserved + ATM signaling channels (e.g., [0,5], [0,16]). + + 3) Send a specific number of IP packets containing timestamps at a + specific constant rate through the SUT via the defined test VCCs. + All of the VPI/VCI pairs will generate traffic at the same + traffic rate. Since this test is not a throughput test, the rate + should not be greater than 90% of line rate. The IP PDUs MUST be + encapsulated in AAL5. + + 4) Count the IP packets that are transmitted by the SUT on all VCCs + to verify connectivity and load. If the count on the test device + is the same on the SUT, continue the test; else lower the test + device traffic rate until the counts are the same. + + 5) Record the packets timestamps at the transmitter and receiver + ends of the test device for all VCCs. + + Reporting Format: + + The results of the Two-point CDV/Steady Load/Twelve VCCs test + SHOULD be reported in a form of text, graph, and histograms. + + The text results SHOULD display the numerical values of the CDV. + The values given SHOULD include: time period of test in s, test + VPI/VCI values, total number of cells transmitted and received on + each VCC during the test in positive integers, maximum and minimum + CDV on each VCC during the test in us, and peak-to-peak CDV on + each VCC in us. + + The graph results SHOULD display the cell delay values. The x- + coordinate SHOULD be the test run time in either seconds, minutes + or days depending on the total length of the test. The x- + coordinate time SHOULD be configurable. The y-coordinate SHOULD + be the cell delay for each VCC in ms. There SHOULD be 12 curves + on the graph, one curves indicated and labeled for each VCC. The + integration time per point MUST be indicated. + + + + + + + +Dunn & Martin Informational [Page 29] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + The histograms SHOULD display the peak-to-peak cell delay. There + will be one histogram for each VCC. The x-coordinate SHOULD be + the cell delay in us with at least 256 bins. The y-coordinate + SHOULD be the number of cells observed in each bin. + + The results MUST also indicate the packet size in octets, traffic + rate in packets per second, and bearer class as generated by the + test device. The VCC and VPI/VCI values MUST be indicated. The + bearer class of the created VCC MUST also be indicated. + +3.2.1.4. Two-point CDV/Steady Load/Maximum VCCs + + Objective: To determine the SUT variation in cell transfer delay with + the maximum number VCCs supported on the SUT as defined in RFC 2761 + "Terminology for ATM Benchmarking". + + Procedure: + + 1) Set up the SUT and test device using the bi-directional + configuration. + + 2) Configure the SUT and test device with the maximum number of VCCs + supported on the SUT. For example, if the maximum number of VCCs + supported on the SUT is 1024, define 256 VPIs with 4 VCIs per + VPI. The VCC's MUST be configured as either a CBR, VBR, or UBR + connection. The VPI/VCIs MUST not be one of the reserved ATM + signaling channels (e.g., [0,5], [0,16]). + + 3) Send a specific number of IP packets containing timestamps at a + specific constant rate through the SUT via the defined test VCCs. + All of the VPI/VCI pairs will generate traffic at the same + traffic rate. Since this test is not a throughput test, the rate + should not be greater than 90% of line rate. The IP PDUs MUST be + encapsulated in AAL5. + + 4) Count the IP packets that are transmitted by the SUT on all VCCs + to verify connectivity and load. If the count on the test device + is the same on the SUT, continue the test; else lower the test + device traffic rate until the counts are the same. + + 5) Record the packets timestamps at the transmitter and receiver + ends of the test device for all VCCs. + + Reporting Format: + + The results of the Two-point CDV/Steady Load/Maximum VCCs test + SHOULD be reported in a form of text, graphs, and histograms. + + + + +Dunn & Martin Informational [Page 30] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + The text results SHOULD display the numerical values of the CDV. + The values given SHOULD include: time period of test in s, test + VPI/VCI values, total number of cells transmitted and received on + each VCC during the test in positive integers, maximum and minimum + CDV on each VCC during the test in us, and peak-to-peak CDV on + each VCC in us. + + The graph results SHOULD display the cell delay values. There + will be (Max number of VCCs/10) graphs, with 10 VCCs indicated on + each graph. The x-coordinate SHOULD be the test run time in + either seconds, minutes or days depending on the total length of + the test. The x-coordinate time SHOULD be configurable. The y- + coordinate SHOULD be the cell delay for each VCC in us. There + SHOULD be no more than 10 curves on each graph, one curve + indicated and labeled for each VCC. The integration time per + point MUST be indicated. + + The histograms SHOULD display the peak-to-peak cell delay. There + will be one histogram for each VCC. The x-coordinate SHOULD be + the cell delay in us with at least 256 bins. The y-coordinate + SHOULD be the number of cells observed in each bin. + + The results MUST also indicate the packet size in octets, traffic + rate in packets per second, and bearer class as generated by the + test device. The VCC and VPI/VCI values MUST be indicated. The + bearer class of the created VCC MUST also be indicated. + +3.2.1.5. Two-point CDV/Bursty VBR Load/One VCC + + Objective: To determine the SUT variation in cell transfer delay with + one VCC as defined in RFC 2761 "Terminology for ATM Benchmarking". + + Procedure: + + 1) Set up the SUT and test device using the bi-directional + configuration. + + 2) Configure the SUT and test device with one VCC. The VCC SHOULD + contain one VPI/VCI. The VCC MUST be configured as either a CBR + or VBR connection. The VPI/VCI MUST not be one of the reserved + ATM signaling channels (e.g., [0,5], [0,16]). + + 3) Send a specific number of IP packets containing timestamps at a + specific VBR through the SUT via the defined test VCC. Since + this test is not a throughput test, the rate should not be + greater than 90% of line rate. The IP PDUs MUST be encapsulated + in AAL5. + + + + +Dunn & Martin Informational [Page 31] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + 4) Count the IP packets that are transmitted by the SUT to verify + connectivity and load. If the count on the test device is the + same on the SUT, continue the test; else lower the test device + traffic rate until the counts are the same. + + 5) Record the packets timestamps at the transmitter and receiver + ends of the test device. + + Reporting Format: + + The results of the Two-point CDV/Bursty VBR Load/One VCC test + SHOULD be reported in a form of text, graph, and histogram. + + The text results SHOULD display the numerical values of the CDV. + The values given SHOULD include: time period of test in s, test + VPI/VCI value, total number of cells transmitted and received on + the given VPI/VCI during the test in positive integers, maximum + and minimum CDV during the test in us, and peak-to-peak CDV in us. + + The graph results SHOULD display the cell delay values. The x- + coordinate SHOULD be the test run time in either seconds, minutes + or days depending on the total length of the test. The x- + coordinate time SHOULD be configurable. The y-coordinate SHOULD + be the cell delay in us. The integration time per point MUST be + indicated. + + The histogram results SHOULD display the peak-to-peak cell delay. + The x-coordinate SHOULD be the cell delay in us with at least 256 + bins. The y-coordinate SHOULD be the number of cells observed in + each bin. + + The results MUST also indicate the packet size in octets, traffic + rate in packets per second, and bearer class as generated by the + test device. The VCC and VPI/VCI values MUST be indicated. The + PCR, SCR, and MBS MUST be indicated. The bearer class of the + created VCC MUST also be indicated. + +3.2.1.6. Two-point CDV/Bursty VBR Load/Twelve VCCs + + Objective: To determine the SUT variation in cell transfer delay with + twelve VCCs as defined in RFC 2761 "Terminology for ATM + Benchmarking". + + Procedure: + + 1) Set up the SUT and test device using the bi-directional + configuration. + + + + +Dunn & Martin Informational [Page 32] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + 2) Configure the SUT and test device with twelve VCCs, using 1 VPI + and 12 VCIs. The VCC's MUST be configured as either a CBR or VBR + connection. The VPI/VCIs MUST not be one of the reserved ATM + signaling channels (e.g., [0,5], [0,16]). + + 3) Send a specific number of IP packets containing timestamps at a + specific VBR through the SUT via the defined test VCCs. All of + the VPI/VCI pairs will generate traffic at the same traffic rate. + Since this test is not a throughput test, the rate should not be + greater than 90% of line rate. The IP PDUs MUST be encapsulated + in AAL5. + + 4) Count the IP packets that are transmitted by the SUT on all VCCs + to verify connectivity and load. If the count on the test device + is the same on the SUT, continue the test; else lower the test + device traffic rate until the counts are the same. + + 5) Record the packets timestamps at the transmitter and receiver + ends of the test device for all VCCs. + + Reporting Format: + + The results of the Two-point CDV/Bursty VBR Load/Twelve VCCs test + SHOULD be reported in a form of text, graph, and histograms. + + The text results SHOULD display the numerical values of the CDV. + The values given SHOULD include: time period of test in s, test + VPI/VCI values, total number of cells transmitted and received on + each VCC during the test in positive integers, maximum and minimum + CDV on each VCC during the test in us, and peak-to-peak CDV on + each VCC in us. + + The graph results SHOULD display the cell delay values. The x- + coordinate SHOULD be the test run time in either seconds, minutes + or days depending on the total length of the test. The x- + coordinate time SHOULD be configurable. The y-coordinate SHOULD + be the cell delay for each VCC in ms. There SHOULD be 12 curves + on the graph, one curves indicated and labeled for each VCC. The + integration time per point MUST be indicated. + + The histograms SHOULD display the peak-to-peak cell delay. There + will be one histogram for each VCC. The x-coordinate SHOULD be + the cell delay in us with at least 256 bins. The y-coordinate + SHOULD be the number of cells observed in each bin. + + + + + + + +Dunn & Martin Informational [Page 33] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + The results MUST also indicate the packet size in octets, traffic + rate in packets per second, and bearer class as generated by the + test device. The VCC and VPI/VCI values MUST be indicated. The + PCR, SCR, and MBS MUST be indicated. The bearer class of the + created VCC MUST also be indicated. + +3.2.1.7. Two-point CDV/Bursty VBR Load/Maximum VCCs + + Objective: To determine the SUT variation in cell transfer delay with + the maximum number VCCs supported on the SUT as defined in RFC 2761 + "Terminology for ATM Benchmarking". + + Procedure: + + 1) Set up the SUT and test device using the bi-directional + configuration. + + 2) Configure the SUT and test device with the maximum number of VCCs + supported on the SUT. For example, if the maximum number of VCCs + supported on the SUT is 1024, define 256 VPIs with 4 VCIs per + VPI. The VCC's MUST be configured as either a CBR or VBR + connection. The VPI/VCIs MUST not be one of the reserved ATM + signaling channels (e.g., [0,5], [0,16]). + + 3) Send a specific number of IP packets containing timestamps at a + specific VBR through the SUT via the defined test VCCs. All of + the VPI/VCI pairs will generate traffic at the same traffic rate. + Since this test is not a throughput test, the rate should not be + greater than 90% of line rate. The IP PDUs MUST be encapsulated + in AAL5. + + 4) Count the IP packets that are transmitted by the SUT on all VCCs + to verify connectivity and load. If the count on the test device + is the same on the SUT, continue the test; else lower the test + device traffic rate until the counts are the same. + + 5) Record the packets timestamps at the transmitter and receiver + ends of the test device for all VCCs. + + Reporting Format: + + The results of the Two-point CDV/Bursty VBR Load/Maximum VCCs test + SHOULD be reported in a form of text, graphs, and histograms. + + The text results SHOULD display the numerical values of the CDV. + The values given SHOULD include: time period of test in s, test + VPI/VCI values, total number of cells transmitted and received on + + + + +Dunn & Martin Informational [Page 34] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + each VCC during the test in positive integers, maximum and minimum + CDV on each VCC during the test in us, and peak-to-peak CDV on + each VCC in us. + + The graph results SHOULD display the cell delay values. There + will be (Max number of VCCs/10) graphs, with 10 VCCs indicated on + each graph. The x-coordinate SHOULD be the test run time in + either seconds, minutes or days depending on the total length of + the test. The x-coordinate time SHOULD be configurable. The y- + coordinate SHOULD be the cell delay for each VCC in us. There + SHOULD be no more than 10 curves on each graph, one curve + indicated and labeled for each VCC. The integration time per + point MUST be indicated. + + The histograms SHOULD display the peak-to-peak cell delay. There + will be one histogram for each VCC. The x-coordinate SHOULD be + the cell delay in us with at least 256 bins. The y-coordinate + SHOULD be the number of cells observed in each bin. + + The results MUST also indicate the packet size in octets, traffic + rate in packets per second, and bearer class as generated by the + test device. The VCC and VPI/VCI values MUST be indicated. The + PCR, SCR, and MBS MUST be indicated. The bearer class of the + created VCC MUST also be indicated. + +3.2.1.8. Two-point CDV/Mixed Load/Three VCC's + + Objective: To determine the SUT variation in cell transfer delay with + three VCC's as defined in RFC 2761 "Terminology for ATM + Benchmarking". + + Procedure: + + 1) Set up the SUT and test device using the bi-directional + configuration. + + 2) Configure the SUT and test device with three VCC's. Each VCC + MUST be defined as a different Bearer class: one CBR, one UBR and + one VBR. Each VCC SHOULD contain one VPI/VCI. The VPI/VCI MUST + not be one of the reserved ATM signaling channels (e.g., [0,5], + [0,16]). + + 3) Send a specific number of IP packets containing timestamps + through the SUT via the defined test VCCs. Each generated VCC + stream MUST match the corresponding VCC Bearer class. All of the + VPI/VCI pairs will generate traffic at the same traffic rate. + + + + + +Dunn & Martin Informational [Page 35] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + Since this test is not a throughput test, the rate should not be + greater than 90% of line rate. The IP PDUs MUST be encapsulated + in AAL5. + + 4) Count the IP packets that are transmitted by the SUT to verify + connectivity and load. If the count on the test device is the + same on the SUT, continue the test; else lower the test device + traffic rate until the counts are the same. + + 5) Record the packets timestamps at the transmitter and receiver + ends of the test device for all VCC's. + + Reporting Format: + + The results of the Two-point CDV/Mixed Load/Three VCC test SHOULD + be reported in a form of text, graph, and histogram. + + The text results SHOULD display the numerical values of the CDV. + The values given SHOULD include: time period of test in s, test + VPI/VCI value, total number of cells transmitted and received on + the given VPI/VCI during the test in positive integers, maximum + and minimum CDV during the test in us, and peak-to-peak CDV in us. + + The graph results SHOULD display the cell delay values. The x- + coordinate SHOULD be the test run time in either seconds, minutes + or days depending on the total length of the test. The x- + coordinate time SHOULD be configurable. The y-coordinate SHOULD + be the cell delay in us. The integration time per point MUST be + indicated. + + The histogram results SHOULD display the peak-to-peak cell delay. + The x-coordinate SHOULD be the cell delay in us with at least 256 + bins. The y-coordinate SHOULD be the number of cells observed in + each bin. + + The results MUST also indicate the packet size in octets, traffic + rate in packets per second, and bearer class as generated by the + test device. The VCC and VPI/VCI values MUST be indicated. The + PCR, SCR, and MBS MUST be indicated. The bearer class of the + created VCC MUST also be indicated. + +3.2.1.9. Two-point CDV/Mixed Load/Twelve VCCs + + Objective: To determine the SUT variation in cell transfer delay with + twelve VCCs as defined in RFC 2761 "Terminology for ATM + Benchmarking". + + + + + +Dunn & Martin Informational [Page 36] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + Procedure: + + 1) Set up the SUT and test device using the bi-directional + configuration. + + 2) Configure the SUT and test device with twelve VCC's. Each VCC + MUST be defined as one of the Bearer classes for a total of four + CBR, four UBR and four VBR VCC's. Each VCC SHOULD contain one + VPI/VCI. The VPI/VCI MUST not be one of the reserved ATM + signaling channels (e.g., [0,5], [0,16]). + + 3) Send a specific number of IP packets containing timestamps + through the SUT via the defined test VCCs. Each generated VCC + stream MUST match the corresponding VCC Bearer class. All of the + VPI/VCI pairs will generate traffic at the same traffic rate. + Since this test is not a throughput test, the rate should not be + greater than 90% of line rate. The IP PDUs MUST be encapsulated + in AAL5. + + 4) Count the IP packets that are transmitted by the SUT on all VCCs + to verify connectivity and load. If the count on the test device + is the same on the SUT, continue the test; else lower the test + device traffic rate until the counts are the same. + + 5) Record the packets timestamps at the transmitter and receiver + ends of the test device for all VCCs. + + Reporting Format: + + The results of the Two-point CDV/Mixed Load/Twelve VCCs test + SHOULD be reported in a form of text, graph, and histograms. + + The text results SHOULD display the numerical values of the CDV. + The values given SHOULD include: time period of test in s, test + VPI/VCI values, total number of cells transmitted and received on + each VCC during the test in positive integers, maximum and minimum + CDV on each VCC during the test in us, and peak-to-peak CDV on + each VCC in us. + + The graph results SHOULD display the cell delay values. The x- + coordinate SHOULD be the test run time in either seconds, minutes + or days depending on the total length of the test. The x- + coordinate time SHOULD be configurable. The y-coordinate SHOULD + be the cell delay for each VCC in ms. There SHOULD be 12 curves + on the graph, one curves indicated and labeled for each VCC. The + integration time per point MUST be indicated. + + + + + +Dunn & Martin Informational [Page 37] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + The histograms SHOULD display the peak-to-peak cell delay. There + will be one histogram for each VCC. The x-coordinate SHOULD be + the cell delay in us with at least 256 bins. The y-coordinate + SHOULD be the number of cells observed in each bin. + + The results MUST also indicate the packet size in octets, traffic + rate in packets per second, and bearer class as generated by the + test device. The VCC and VPI/VCI values MUST be indicated. The + PCR, SCR, and MBS MUST be indicated. The bearer class of the + created VCC MUST also be indicated. + +3.2.1.10. Two-point CDV/Mixed Load/Maximum VCCs + + Objective: To determine the SUT variation in cell transfer delay with + the maximum number VCCs supported on the SUT as defined in RFC 2761 + "Terminology for ATM Benchmarking". + + Procedure: + + 1) Set up the SUT and test device using the bi-directional + configuration. + + 2) Configure the SUT and test device with maximum number of VCCs + supported on the SUT. For example, if the maximum number of VCCs + supported on the SUT is 1024, define 256 VPIs with 4 VCIs per + VPI. Each VCC MUST be defined as one of the Bearer classes for a + total of (max VCC/3) CBR, (max VCC/3) UBR and (max VCC/3) VBR + VCC's. If the maximum number of VCC's is not divisible by 3, the + total for each bearer class MUST be within 3 VCC's of each other. + The VPI/VCI MUST not be one of the reserved ATM signaling + channels (e.g., [0,5], [0,16]). + + 3) Send a specific number of IP packets containing timestamps + through the SUT via the defined test VCCs. Each generated VCC + stream MUST match the corresponding VCC Bearer class. All of the + VPI/VCI pairs will generate traffic at the same traffic rate. + Since this test is not a throughput test, the rate should not be + greater than 90% of line rate. The IP PDUs MUST be encapsulated + in AAL5. + + 4) Count the IP packets that are transmitted by the SUT on all VCCs + to verify connectivity and load. If the count on the test device + is the same on the SUT, continue the test; else lower the test + device traffic rate until the counts are the same. + + 5) Record the packets timestamps at the transmitter and receiver + ends of the test device for all VCCs. + + + + +Dunn & Martin Informational [Page 38] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + Reporting Format: + + The results of the Two-point CDV/Mixed Load/Maximum VCCs test + SHOULD be reported in a form of text, graphs, and histograms. + + The text results SHOULD display the numerical values of the CDV. + The values given SHOULD include: time period of test in s, test + VPI/VCI values, total number of cells transmitted and received on + each VCC during the test in positive integers, maximum and minimum + CDV on each VCC during the test in us, and peak-to-peak CDV on + each VCC in us. + + The graph results SHOULD display the cell delay values. There + will be (Max number of VCCs/10) graphs, with 10 VCCs indicated on + each graph. The x-coordinate SHOULD be the test run time in + either seconds, minutes or days depending on the total length of + the test. The x-coordinate time SHOULD be configurable. The y- + coordinate SHOULD be the cell delay for each VCC in us. There + SHOULD be no more than 10 curves on each graph, one curve + indicated and labeled for each VCC. The integration time per + point MUST be indicated. + + The histograms SHOULD display the peak-to-peak cell delay. There + will be one histogram for each VCC. The x-coordinate SHOULD be + the cell delay in us with at least 256 bins. The y-coordinate + SHOULD be the number of cells observed in each bin. + + The results MUST also indicate the packet size in octets, traffic + rate in packets per second, and bearer class as generated by the + test device. The VCC and VPI/VCI values MUST be indicated. The + PCR, SCR, and MBS MUST be indicated. The bearer class of the + created VCC MUST also be indicated. + +3.2.2. Cell Error Ratio (CER) + +3.2.2.1. Test Setup + + The cell error ratio measurements assume that both the transmitter + and receiver payload information is synchronized. Synchronization + MUST be achieved by supplying a known bit pattern to both the + transmitter and receiver. If this bit pattern is longer than the + packet size, the receiver MUST synchronize with the transmitter + before tests can be run. + + + + + + + + +Dunn & Martin Informational [Page 39] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + +3.2.2.2. CER/Steady Load/One VCC + + Objective: To determine the SUT ratio of errored cells on one VCC in + a transmission in relation to the total cells sent as defined in RFC + 2761 "Terminology for ATM Benchmarking". + + Procedure: + + 1) Set up the SUT and test device using the bi-directional + configuration. + + 2) Configure the SUT and test device with one VCC. The VCC SHOULD + contain one VPI/VCI. The VCC MUST be configured as either a CBR, + VBR, or UBR connection. The VPI/VCI MUST not be one of the + reserved ATM signaling channels (e.g., [0,5], [0,16]). + + 3) Send a specific number of IP packets containing one of the + specified bit patterns at a constant rate through the SUT via the + defined test VCC. Since this test is not a throughput test, the + rate should not be greater than 90% of line rate. The IP PDUs + MUST be encapsulated in AAL5. + + 4) Count the IP packets that are transmitted by the SUT to verify + connectivity and load. If the count on the test device is the + same on the SUT, continue the test; else lower the test device + traffic rate until the counts are the same. + + 5) Record the number of bit errors at the receiver end of the test + device. + + Reporting Format: + + The results of the CER/Steady Load/One VCC test SHOULD be reported + in a form of text and graph. + + The text results SHOULD display the numerical values of the CER. + The values given SHOULD include: time period of test in s, test + VPI/VCI value, total number of cells transmitted and received on + the given VPI/VCI during the test in positive integers, and the + CER for the entire test. + + The graph results SHOULD display the cell error ratio values. The + x-coordinate SHOULD be the test run time in either seconds, + minutes or days depending on the total length of the test. The + x-coordinate time SHOULD be configurable. The y-coordinate SHOULD + be the CER. The integration time per point MUST be indicated. + + + + + +Dunn & Martin Informational [Page 40] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + The results MUST also indicate the packet size in octets, traffic + rate in packets per second, and bearer class as generated by the + test device. The VCC and VPI/VCI values MUST be indicated. The + PCR, SCR, and MBS MUST be indicated. The bearer class of the + created VCC MUST be indicated. The generated bit pattern MUST + also be indicated. + +3.2.2.3. CER/Steady Load/Twelve VCCs + + Objective: To determine the SUT ratio of errored cells on twelve VCCs + in a transmission in relation to the total cells sent as defined in + RFC 2761 "Terminology for ATM Benchmarking". + + Procedure: + + 1) Set up the SUT and test device using the bi-directional + configuration. + + 2) Configure the SUT and test device with twelve VCCs, using 1 VPI + and 12 VCIs. The VCC's MUST be configured as either a CBR, VBR, + or UBR connection. The VPI/VCIs MUST not be one of the reserved + ATM signaling channels (e.g., [0,5], [0,16]). + + 3) Send a specific number of IP packets containing one of the + specified bit patterns at a constant rate through the SUT via the + defined test VCCs. All of the VPI/VCI pairs will generate + traffic at the same traffic rate. Since this test is not a + throughput test, the rate should not be greater than 90% of line + rate. The IP PDUs MUST be encapsulated in AAL5. + + 4) Count the IP packets that are transmitted by the SUT on all VCCs + to verify connectivity and load. If the count on the test device + is the same on the SUT, continue the test; else lower the test + device traffic rate until the counts are the same. + + 5) Record the number of bit errors at the receiver end of the test + device for all VCCs. + + Reporting Format: + + The results of the CER/Steady Load/Twelve VCCs test SHOULD be + reported in a form of text and graph. + + The text results SHOULD display the numerical values of the CER. + The values given SHOULD include: time period of test in s, test + VPI/VCI value, total number of cells transmitted and received on + the given VPI/VCI during the test in positive integers, and the + CER for the entire test. + + + +Dunn & Martin Informational [Page 41] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + The graph results SHOULD display the cell error ratio values. The + x-coordinate SHOULD be the test run time in either seconds, + minutes or days depending on the total length of the test. The + x-coordinate time SHOULD be configurable. The y-coordinate SHOULD + be the CER for each VCC. There should be 12 curves on the graph, + on curve indicated and labeled for each VCC. The integration time + per point MUST be indicated. + + The results MUST also indicate the packet size in octets, traffic + rate in packets per second, and bearer class as generated by the + test device. The VCC and VPI/VCI values MUST be indicated. The + PCR, SCR, and MBS MUST be indicated. The bearer class of the + created VCC MUST be indicated. The generated bit pattern MUST + also be indicated. + +3.2.2.4. CER/Steady Load/Maximum VCCs + + Objective: To determine the SUT ratio of errored cells with the + maximum number VCCs supported on the SUT in a transmission in + relation to the total cells sent as defined in RFC 2761 "Terminology + for ATM Benchmarking". + + Procedure: + + 1) Set up the SUT and test device using the bi-directional + configuration. + + 2) Configure the SUT and test device with the maximum number of VCCs + supported on the SUT. For example, if the maximum number of VCCs + supported on the SUT is 1024, define 256 VPIs with 4 VCIs per + VPI. The VCC's MUST be configured as either a CBR, VBR, or UBR + connection. The VPI/VCIs MUST not be one of the reserved ATM + signaling channels (e.g., [0,5], [0,16]). + + 3) Send a specific number of IP packets containing one of the + specified bit patterns at a constant rate through the SUT via the + defined test VCCs. All of the VPI/VCI pairs will generate + traffic at the same traffic rate. Since this test is not a + throughput test, the rate should not be greater than 90% of line + rate. The IP PDUs MUST be encapsulated in AAL5. + + 4) Count the IP packets that are transmitted by the SUT on all VCCs + to verify connectivity and load. If the count on the test device + is the same on the SUT, continue the test; else lower the test + device traffic rate until the counts are the same. + + 5) Record the number of bit errors at the receiver end of the test + device for all VCCs. + + + +Dunn & Martin Informational [Page 42] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + Reporting Format: + + The results of the CER/Steady Load/Maximum VCCs test SHOULD be + reported in a form of text and graph. + + The text results SHOULD display the numerical values of the CER. + The values given SHOULD include: time period of test in s, test + VPI/VCI value, total number of cells transmitted and received on + the given VPI/VCI during the test in positive integers, and the + CER for the entire test. + + The graph results SHOULD display the cell error ratio values. + There will be (Max number of VCCs/10) graphs, with 10 VCCs + indicated on each graph. The x-coordinate SHOULD be the test run + time in either seconds, minutes or days depending on the total + length of the test. The x-coordinate time SHOULD be configurable. + The y-coordinate SHOULD be the CER for each VCC. There SHOULD be + no more than 10 curves on each graph, one curve indicated and + labeled for each VCC. The integration time per point MUST be + indicated. + + The results MUST also indicate the packet size in octets, traffic + rate in packets per second, and bearer class as generated by the + test device. The VCC and VPI/VCI values MUST be indicated. The + PCR, SCR, and MBS MUST be indicated. The bearer class of the + created VCC MUST be indicated. The generated bit pattern MUST + also be indicated. + +3.2.2.5. CER/Bursty VBR Load/One VCC + + Objective: To determine the SUT ratio of errored cells on one VCC in + a transmission in relation to the total cells sent as defined in RFC + 2761 "Terminology for ATM Benchmarking". + + Procedure: + + 1) Set up the SUT and test device using the bi-directional + configuration. + + 2) Configure the SUT and test device with one VCC. The VCC SHOULD + contain one VPI/VCI. The VCC MUST be configured as either a CBR + or VBR connection. The VPI/VCI MUST not be one of the reserved + ATM signaling channels (e.g., [0,5], [0,16]). The PCR, SCR, and + MBS must be configured using one of the specified traffic + descriptors. + + + + + + +Dunn & Martin Informational [Page 43] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + 3) Send a specific number of IP packets containing one of the + specified bit patterns at a specific VBR rate through the SUT via + the defined test VCC. Since this test is not a throughput test, + the rate should not be greater than 90% of line rate. The IP + PDUs MUST be encapsulated in AAL5. + + 4) Count the IP packets that are transmitted by the SUT to verify + connectivity and load. If the count on the test device is the + same on the SUT, continue the test; else lower the test device + traffic rate until the counts are the same. + + 5) Record the number of bit errors at the receiver end of the test + device. + + Reporting Format: + + The results of the CER/Bursty VBR Load/One VCC test SHOULD be + reported in a form of text and graph. + + The text results SHOULD display the numerical values of the CER. + The values given SHOULD include: time period of test in s, test + VPI/VCI value, total number of cells transmitted and received on + the given VPI/VCI during the test in positive integers, and the + CER for the entire test. + + The graph results SHOULD display the cell error ratio values. The + x-coordinate SHOULD be the test run time in either seconds, + minutes or days depending on the total length of the test. The + x-coordinate time SHOULD be configurable. The y-coordinate SHOULD + be the CER. The integration time per point MUST be indicated. + + The results MUST also indicate the packet size in octets, traffic + rate in packets per second, and bearer class as generated by the + test device. The VCC and VPI/VCI values MUST be indicated. The + PCR, SCR, and MBS MUST be indicated. The bearer class of the + created VCC MUST be indicated. The generated bit pattern MUST + also be indicated. + +3.2.2.6. CER/Bursty VBR Load/Twelve VCCs + + Objective: To determine the SUT ratio of errored cells on twelve VCCs + in a transmission in relation to the total cells sent as defined in + RFC 2761 "Terminology for ATM Benchmarking". + + Procedure: + + 1) Set up the SUT and test device using the bi-directional + configuration. + + + +Dunn & Martin Informational [Page 44] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + 2) Configure the SUT and test device with twelve VCCs, using 1 VPI + and 12 VCIs. The VCC's MUST be configured as either a CBR or VBR + connection. The VPI/VCIs MUST not be one of the reserved ATM + signaling channels (e.g., [0,5], [0,16]). The PCR, SCR, and MBS + must be configured using one of the specified traffic + descriptors. + + 3) Send a specific number of IP packets containing one of the + specified bit patterns at a specific VBR rate through the SUT via + the defined test VCCs. All of the VPI/VCI pairs will generate + traffic at the same traffic rate. Since this test is not a + throughput test, the rate should not be greater than 90% of line + rate. The PCR, SCR, and MBS must be indicated. The IP PDUs MUST + be encapsulated in AAL5. + + 4) Count the IP packets that are transmitted by the SUT on all VCCs + to verify connectivity and load. If the count on the test device + is the same on the SUT, continue the test; else lower the test + device traffic rate until the counts are the same. + + 5) Record the number of bit errors at the receiver end of the test + device for all VCCs. + + Reporting Format: + + The results of the CER/Bursty VBR Load/Twelve VCCs test SHOULD be + reported in a form of text and graph. + + The text results SHOULD display the numerical values of the CER. + The values given SHOULD include: time period of test in s, test + VPI/VCI value, total number of cells transmitted and received on + the given VPI/VCI during the test in positive integers, and the + CER for the entire test. + + The graph results SHOULD display the cell error ratio values. The + x-coordinate SHOULD be the test run time in either seconds, + minutes or days depending on the total length of the test. The + x-coordinate time SHOULD be configurable. The y-coordinate SHOULD + be the CER for each VCC. There should be 12 curves on the graph, + on curve indicated and labeled for each VCC. The integration time + per point MUST be indicated. + + The results MUST also indicate the packet size in octets, traffic + rate in packets per second, and bearer class as generated by the + test device. The VCC and VPI/VCI values MUST be indicated. The + PCR, SCR, and MBS MUST be indicated. The bearer class of the + created VCC MUST be indicated. The generated bit pattern MUST + also be indicated. + + + +Dunn & Martin Informational [Page 45] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + +3.2.2.7. CER/Bursty VBR Load/Maximum VCCs + + Objective: To determine the SUT ratio of errored cells with the + maximum number VCCs supported on the SUT in a transmission in + relation to the total cells sent as defined in RFC 2761 "Terminology + for ATM Benchmarking". + + Procedure: + + 1) Set up the SUT and test device using the bi-directional + configuration. + + 2) Configure the SUT and test device with the maximum number of VCCs + supported on the SUT. For example, if the maximum number of VCCs + supported on the SUT is 1024, define 256 VPIs with 4 VCIs per + VPI. The VCC's MUST be configured as either a CBR or VBR + connection. The VPI/VCIs MUST not be one of the reserved ATM + signaling channels (e.g., [0,5], [0,16]). The PCR, SCR, and MBS + must be configured using one of the specified traffic + descriptors. + + 3) Send a specific number of IP packets containing one of the + specified bit patterns at a specific VBR rate through the SUT via + the defined test VCCs. All of the VPI/VCI pairs will generate + traffic at the same traffic rate. Since this test is not a + throughput test, the rate should not be greater than 90% of line + rate. The IP PDUs MUST be encapsulated in AAL5. + + 4) Count the IP packets that are transmitted by the SUT on all VCCs + to verify connectivity and load. If the count on the test device + is the same on the SUT, continue the test; else lower the test + device traffic rate until the counts are the same. + + 5) Record the number of bit errors at the receiver end of the test + device for all VCCs. + + Reporting Format: + + The results of the CER/Bursty VBR Load/Maximum VCCs test SHOULD be + reported in a form of text and graph. + + The text results SHOULD display the numerical values of the CER. + The values given SHOULD include: time period of test in s, test + VPI/VCI value, total number of cells transmitted and received on + the given VPI/VCI during the test in positive integers, and the + CER for the entire test. + + + + + +Dunn & Martin Informational [Page 46] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + The graph results SHOULD display the cell error ratio values. + There will be (Max number of VCCs/10) graphs, with 10 VCCs + indicated on each graph. The x-coordinate SHOULD be the test run + time in either seconds, minutes or days depending on the total + length of the test. The x-coordinate time SHOULD be configurable. + The y-coordinate SHOULD be the CER for each VCC. There SHOULD be + no more than 10 curves on each graph, one curve indicated and + labeled for each VCC. The integration time per point MUST be + indicated. + + The results MUST also indicate the packet size in octets, traffic + rate in packets per second, and bearer class as generated by the + test device. The VCC and VPI/VCI values MUST be indicated. The + PCR, SCR, and MBS MUST be indicated. The bearer class of the + created VCC MUST be indicated. The generated bit pattern MUST + also be indicated. + +3.2.3. Cell Loss Ratio (CLR) + +3.2.3.1. CLR/Steady Load/One VCC + + Objective: To determine the SUT ratio of lost cells on one VCC in a + transmission in relation to the total cells sent as defined in RFC + 2761 "Terminology for ATM Benchmarking". + + Procedure: + + 1) Set up the SUT and test device using the bi-directional + configuration. + + 2) Configure the SUT and test device with one VCC. The VCC SHOULD + contain one VPI/VCI. The VCC MUST be configured as either a CBR, + VBR, or UBR connection. The VPI/VCI MUST not be one of the + reserved ATM signaling channels (e.g., [0,5], [0,16]). + + 3) Send a specific number of IP packets at a specific constant rate + through the SUT via the defined test VCC. Since this test is not + a throughput test, the rate should not be greater than 90% of + line rate. The IP PDUs MUST be encapsulated in AAL5. + + 4) Count the IP packets that are transmitted by the SUT to verify + connectivity and load. If the count on the test device is the + same on the SUT, continue the test; else lower the test device + traffic rate until the counts are the same. + + 5) Record the number of cells transmitted and received on the test + device. + + + + +Dunn & Martin Informational [Page 47] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + Reporting Format: + + The results of the CLR/Steady Load/One VCC test SHOULD be reported + in a form of text and graph. + + The text results SHOULD display the numerical values of the CLR. + The values given SHOULD include: time period of test in s, test + VPI/VCI value, total number of cells transmitted and received on + the given VPI/VCI during the test in positive integers, and the + CLR for the entire test. + + The graph results SHOULD display the Cell Loss ratio values. The + x-coordinate SHOULD be the test run time in either seconds, + minutes or days depending on the total length of the test. The + x-coordinate time SHOULD be configurable. The y-coordinate SHOULD + be the CLR. The integration time per point MUST be indicated. + + The results MUST also indicate the packet size in octets, traffic + rate in packets per second, and bearer class as generated by the + test device. The VCC and VPI/VCI values MUST be indicated. The + PCR, SCR, and MBS MUST be indicated. The bearer class of the + created VCC MUST also be indicated. + +3.2.3.2. CLR/Steady Load/Twelve VCCs + + Objective: To determine the SUT ratio of lost cells on twelve VCCs in + a transmission in relation to the total cells sent as defined in RFC + 2761 "Terminology for ATM Benchmarking". + + Procedure: + + 1) Set up the SUT and test device using the bi-directional + configuration. + + 2) Configure the SUT and test device with twelve VCCs, using 1 VPI + and 12 VCIs. The VCC's MUST be configured as either a CBR, VBR, + or UBR connection. The VPI/VCIs MUST not be one of the reserved + ATM signaling channels (e.g., [0,5], [0,16]). + + 3) Send a specific number of IP packets at a specific constant rate + through the SUT via the defined test VCCs. All of the VPI/VCI + pairs will generate traffic at the same traffic rate. Since this + test is not a throughput test, the rate should not be greater + than 90% of line rate. The IP PDUs MUST be encapsulated in AAL5. + + + + + + + +Dunn & Martin Informational [Page 48] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + 4) Count the IP packets that are transmitted by the SUT on all VCCs + to verify connectivity and load. If the count on the test device + is the same on the SUT, continue the test; else lower the test + device traffic rate until the counts are the same. + + 5) Record the number of cells transmitted and received per VCC on + the test device. + + Reporting Format: + + The results of the CLR/Steady Load/Twelve VCCs test SHOULD be + reported in a form of text and graph. + + The text results SHOULD display the numerical values of the CLR. + The values given SHOULD include: time period of test in s, test + VPI/VCI value, total number of cells transmitted and received on + the given VPI/VCI during the test in positive integers, and the + CLR for the entire test. + + The graph results SHOULD display the Cell Loss ratio values. The + x-coordinate SHOULD be the test run time in either seconds, + minutes or days depending on the total length of the test. The + x-coordinate time SHOULD be configurable. The y-coordinate SHOULD + be the CLR for each VCC. There should be 12 curves on the graph, + on curve indicated and labeled for each VCC. The integration time + per point MUST be indicated. + + The results MUST also indicate the packet size in octets, traffic + rate in packets per second, and bearer class as generated by the + test device. The VCC and VPI/VCI values MUST be indicated. The + PCR, SCR, and MBS MUST be indicated. The bearer class of the + created VCC MUST also be indicated. + +3.2.3.3. CLR/Steady Load/Maximum VCCs + + Objective: To determine the SUT ratio of lost cells with the maximum + number VCCs supported on the SUT in a transmission in relation to the + total cells sent as defined in RFC 2761 "Terminology for ATM + Benchmarking". + + Procedure: + + 1) Set up the SUT and test device using the bi-directional + configuration. + + 2) Configure the SUT and test device with the maximum number of VCCs + supported on the SUT. For example, if the maximum number of VCCs + supported on the SUT is 1024, define 256 VPIs with 4 VCIs per + + + +Dunn & Martin Informational [Page 49] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + VPI. The VCC's MUST be configured as either a CBR, VBR, or UBR + connection. The VPI/VCIs MUST not be one of the reserved ATM + signaling channels (e.g., [0,5], [0,16]). + + 3) Send a specific number of IP packets at a specific constant rate + through the SUT via the defined test VCCs. All of the VPI/VCI + pairs will generate traffic at the same traffic rate. Since this + test is not a throughput test, the rate should not be greater + than 90% of line rate. The IP PDUs MUST be encapsulated in AAL5. + + 4) Count the IP packets that are transmitted by the SUT on all VCCs + to verify connectivity and load. If the count on the test device + is the same on the SUT, continue the test; else lower the test + device traffic rate until the counts are the same. + + 5) Record the number of cells transmitted and received per VCC on + the test device. + + Reporting Format: + + The results of the CLR/Steady Load/Maximum VCCs test SHOULD be + reported in a form of text and graph. + + The text results SHOULD display the numerical values of the CLR. + The values given SHOULD include: time period of test in s, test + VPI/VCI value, total number of cells transmitted and received on + the given VPI/VCI during the test in positive integers, and the + CLR for the entire test. + + The graph results SHOULD display the Cell Loss ratio values. + There will be (Max number of VCCs/10) graphs, with 10 VCCs + indicated on each graph. The x-coordinate SHOULD be the test run + time in either seconds, minutes or days depending on the total + length of the test. The x-coordinate time SHOULD be configurable. + The y-coordinate SHOULD be the CLR for each VCC. There SHOULD be + no more than 10 curves on each graph, one curve indicated and + labeled for each VCC. The integration time per point MUST be + indicated. + + The results MUST also indicate the packet size in octets, traffic + rate in packets per second, and bearer class as generated by the + test device. The VCC and VPI/VCI values MUST be indicated. The + PCR, SCR, and MBS MUST be indicated. The bearer class of the + created VCC MUST also be indicated. + + + + + + + +Dunn & Martin Informational [Page 50] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + +3.2.3.4. CLR/Bursty VBR Load/One VCC + + Objective: To determine the SUT ratio of lost cells on one VCC in a + transmission in relation to the total cells sent as defined in RFC + 2761 "Terminology for ATM Benchmarking". + + Procedure: + + 1) Set up the SUT and test device using the bi-directional + configuration. + + 2) Configure the SUT and test device with one VCC. The VCC SHOULD + contain one VPI/VCI. The VCC MUST be configured as either a CBR + or VBR connection. The VPI/VCI MUST not be one of the reserved + ATM signaling channels (e.g., [0,5], [0,16]). The PCR, SCR, and + MBS must be configured using one of the specified traffic + descriptors. + + 3) Send a specific number of IP packets containing one of the + specified bit patterns at a specific rate through the SUT via the + defined test VCC. Since this test is not a throughput test, the + rate should not be greater than 90% of line rate. The IP PDUs + MUST be encapsulated in AAL5. + + 4) Count the IP packets that are transmitted by the SUT to verify + connectivity and load. If the count on the test device is the + same on the SUT, continue the test; else lower the test device + traffic rate until the counts are the same. + + 5) Record the number of cells transmitted and received on the test + device. + + Reporting Format: + + The results of the CLR/Bursty VBR Load/One VCC test SHOULD be + reported in a form of text and graph. + + The text results SHOULD display the numerical values of the CLR. + The values given SHOULD include: time period of test in s, test + VPI/VCI value, total number of cells transmitted and received on + the given VPI/VCI during the test in positive integers, and the + CLR for the entire test. + + The graph results SHOULD display the Cell Loss ratio values. The + x-coordinate SHOULD be the test run time in either seconds, + minutes or days depending on the total length of the test. The + x-coordinate time SHOULD be configurable. The y-coordinate SHOULD + be the CLR. The integration time per point MUST be indicated. + + + +Dunn & Martin Informational [Page 51] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + The results MUST also indicate the packet size in octets, traffic + rate in packets per second, and bearer class as generated by the + test device. The VCC and VPI/VCI values MUST be indicated. The + PCR, SCR, and MBS MUST be indicated. The bearer class of the + created VCC MUST also be indicated. + +3.2.3.5. CLR/Bursty VBR Load/Twelve VCCs + + Objective: To determine the SUT ratio of lost cells on twelve VCCs in + a transmission in relation to the total cells sent as defined in RFC + 2761 "Terminology for ATM Benchmarking". + + Procedure: + + 1) Set up the SUT and test device using the bi-directional + configuration. + + 2) Configure the SUT and test device with twelve VCCs, using 1 VPI + and 12 VCIs. The VCC MUST be configured as either a CBR or VBR + connection. The VPI/VCIs MUST not be one of the reserved ATM + signaling channels (e.g., [0,5], [0,16]). The PCR, SCR, and MBS + must be configured using one of the specified traffic + descriptors. + + 3) Send a specific number of IP packets containing one of the + specified bit patterns at a specific rate through the SUT via the + defined test VCCs. All of the VPI/VCI pairs will generate + traffic at the same traffic rate. Since this test is not a + throughput test, the rate should not be greater than 90% of line + rate. The PCR, SCR, and MBS must be indicated. The IP PDUs MUST + be encapsulated in AAL5. + + 4) Count the IP packets that are transmitted by the SUT on all VCCs + to verify connectivity and load. If the count on the test device + is the same on the SUT, continue the test; else lower the test + device traffic rate until the counts are the same. + + 5) Record the number of cells transmitted and received per VCC on + the test device. + + Reporting Format: + + The results of the CLR/Bursty VBR Load/Twelve VCCs test SHOULD be + reported in a form of text and graph. + + The text results SHOULD display the numerical values of the CLR. + The values given SHOULD include: time period of test in s, test + VPI/VCI value, total number of cells transmitted and received on + + + +Dunn & Martin Informational [Page 52] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + the given VPI/VCI during the test in positive integers, and the + CLR for the entire test. + + The graph results SHOULD display the Cell Loss ratio values. The + x-coordinate SHOULD be the test run time in either seconds, + minutes or days depending on the total length of the test. The + x-coordinate time SHOULD be configurable. The y-coordinate SHOULD + be the CLR for each VCC. There should be 12 curves on the graph, + on curve indicated and labeled for each VCC. The integration time + per point MUST be indicated. + + The results MUST also indicate the packet size in octets, traffic + rate in packets per second, and bearer class as generated by the + test device. The VCC and VPI/VCI values MUST be indicated. The + PCR, SCR, and MBS MUST be indicated. The bearer class of the + created VCC MUST also be indicated. + +3.2.3.6. CLR/Bursty VBR Load/Maximum VCCs + + Objective: To determine the SUT ratio of lost cells with the maximum + number VCCs supported on the SUT in a transmission in relation to the + total cells sent as defined in RFC 2761 "Terminology for ATM + Benchmarking". + + Procedure: + + 1) Set up the SUT and test device using the bi-directional + configuration. + + 2) Configure the SUT and test device with the maximum number of VCCs + supported on the SUT. For example, if the maximum number of VCCs + supported on the SUT is 1024, define 256 VPIs with 4 VCIs per + VPI. The VCC MUST be configured as either a CBR or VBR + connection. The VPI/VCIs MUST not be one of the reserved ATM + signaling channels (e.g., [0,5], [0,16]). The PCR, SCR, and MBS + must be configured using one of the specified traffic + descriptors. + + 3) Send a specific number of IP packets containing one of the + specified bit patterns at a specific rate through the SUT via the + defined test VCCs. All of the VPI/VCI pairs will generate + traffic at the same traffic rate. Since this test is not a + throughput test, the rate should not be greater than 90% of line + rate. The IP PDUs MUST be encapsulated in AAL5. + + + + + + + +Dunn & Martin Informational [Page 53] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + 4) Count the IP packets that are transmitted by the SUT on all VCCs + to verify connectivity and load. If the count on the test device + is the same on the SUT, continue the test; else lower the test + device traffic rate until the counts are the same. + + 5) Record the number of cells transmitted and received per VCC on + the test device. + + Reporting Format: + + The results of the CLR/Bursty VBR Load/Maximum VCCs test SHOULD be + reported in a form of text and graph. + + The text results SHOULD display the numerical values of the CLR. + The values given SHOULD include: time period of test in s, test + VPI/VCI value, total number of cells transmitted and received on + the given VPI/VCI during the test in positive integers, and the + CLR for the entire test. + + The graph results SHOULD display the Cell Loss ratio values. + There will be (Max number of VCCs/10) graphs, with 10 VCCs + indicated on each graph. The x-coordinate SHOULD be the test run + time in either seconds, minutes or days depending on the total + length of the test. The x-coordinate time SHOULD be configurable. + The y-coordinate SHOULD be the CLR for each VCC. There SHOULD be + no more than 10 curves on each graph, one curve indicated and + labeled for each VCC. The integration time per point MUST be + indicated. + + The results MUST also indicate the packet size in octets, traffic + rate in packets per second, and bearer class as generated by the + test device. The VCC and VPI/VCI values MUST be indicated. The + PCR, SCR, and MBS MUST be indicated. The bearer class of the + created VCC MUST also be indicated. + +3.2.4. Cell Misinsertion Rate (CMR) + +3.2.4.1. CMR/Steady Load/One VCC + + Objective: To determine the SUT ratio of cell misinsertion on one VCC + in a transmission in relation to the total cells sent as defined in + RFC 2761 "Terminology for ATM Benchmarking". + + Procedure: + + 1) Set up the SUT and test device using the bi-directional + configuration. + + + + +Dunn & Martin Informational [Page 54] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + 2) Configure the SUT and test device with one VCC. The VCC MUST be + configured as either a CBR, VBR, or UBR connection. The VCC + SHOULD contain one VPI/VCI. The VPI/VCI MUST not be one of the + reserved ATM signaling channels (e.g., [0,5], [0,16]). + + 3) Send a specific number of IP packets at a specific constant rate + through the SUT via the defined test VCC. Since this test is not + a throughput test, the rate should not be greater than 90% of + line rate. The IP PDUs MUST be encapsulated in AAL5. + + 4) Count the IP packets that are transmitted by the SUT to verify + connectivity and load. If the count on the test device is the + same on the SUT, continue the test; else lower the test device + traffic rate until the counts are the same. + + 5) Record the number of cell misinsertion errors at the receiver end + of the test device. + + Reporting Format: + + The results of the CMR/Steady Load/One VCC test SHOULD be reported + in a form of text and graph. + + The text results SHOULD display the numerical values of the CMR. + The values given SHOULD include: time period of test in s, test + VPI/VCI value, total number of cells transmitted and received on + the given VPI/VCI during the test in positive integers, and the + CMR for the entire test. + + The graph results SHOULD display the Cell misinsertion rate + values. The x-coordinate SHOULD be the test run time in either + seconds, minutes or days depending on the total length of the + test. The x-coordinate time SHOULD be configurable. The y- + coordinate SHOULD be the CMR. The integration time per point MUST + be indicated. + + The results MUST also indicate the packet size in octets, traffic + rate in packets per second, and bearer class as generated by the + test device. The VCC and VPI/VCI values MUST be indicated. The + PCR, SCR, and MBS MUST be indicated. The bearer class of the + created VCC MUST also be indicated. + +3.2.4.2. CMR/Steady Load/Twelve VCCs + + Objective: To determine the SUT rate of misinserted cells on twelve + VCCs in a transmission in relation to the total cells sent as defined + in RFC 2761 "Terminology for ATM Benchmarking". + + + + +Dunn & Martin Informational [Page 55] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + Procedure: + + 1) Set up the SUT and test device using the bi-directional + configuration. + + 2) Configure the SUT and test device with twelve VCCs, using 1 VPI + and 12 VCIs. The VCC's MUST be configured as either a CBR, VBR, + or UBR connection. The VPI/VCIs MUST not be one of the reserved + ATM signaling channels (e.g., [0,5], [0,16]). + + 3) Send a specific number of IP packets at a specific constant rate + through the SUT via the defined test VCCs. All of the VPI/VCI + pairs will generate traffic at the same traffic rate. Since this + test is not a throughput test, the rate should not be greater + than 90% of line rate. The IP PDUs MUST be encapsulated in AAL5. + + 4) Count the IP packets that are transmitted by the SUT on all VCCs + to verify connectivity and load. If the count on the test device + is the same on the SUT, continue the test; else lower the test + device traffic rate until the counts are the same. + + 5) Record the number of cell misinsertion errors at the receiver end + of the test device per VCC. + + Reporting Format: + + The results of the CMR/Steady Load/Twelve VCCs test SHOULD be + reported in a form of text and graph. + + The text results SHOULD display the numerical values of the CMR. + The values given SHOULD include: time period of test in s, test + VPI/VCI value, total number of cells transmitted and received on + the given VPI/VCI during the test in positive integers, and the + CMR for the entire test. + + The graph results SHOULD display the Cell misinsertion rate + values. The x-coordinate SHOULD be the test run time in either + seconds, minutes or days depending on the total length of the + test. The x-coordinate time SHOULD be configurable. The y- + coordinate SHOULD be the CMR for each VCC. There should be 12 + curves on the graph, on curve indicated and labeled for each VCC. + The integration time per point MUST be indicated. + + The results MUST also indicate the packet size in octets, traffic + rate in packets per second, and bearer class as generated by the + test device. The VCC and VPI/VCI values MUST be indicated. The + PCR, SCR, and MBS MUST be indicated. The bearer class of the + created VCC MUST also be indicated. + + + +Dunn & Martin Informational [Page 56] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + +3.2.4.3. CMR/Steady Load/Maximum VCCs + + Objective: To determine the SUT rate of misinserted cells with the + maximum number VCCs supported on the SUT in a transmission in + relation to the total cells sent as defined in RFC 2761 "Terminology + for ATM Benchmarking". + + Procedure: + + 1) Set up the SUT and test device using the bi-directional + configuration. + + 2) Configure the SUT and test device with the maximum number of VCCs + supported on the SUT. For example, if the maximum number of VCCs + supported on the SUT is 1024, define 256 VPIs with 4 VCIs per + VPI. The VCC's MUST be configured as either a CBR, VBR, or UBR + connection. The VPI/VCIs MUST not be one of the reserved ATM + signaling channels (e.g., [0,5], [0,16]). + + 3) Send a specific number of IP packets at a specific constant rate + through the SUT via the defined test VCCs. All of the VPI/VCI + pairs will generate traffic at the same traffic rate. Since this + test is not a throughput test, the rate should not be greater + than 90% of line rate. The IP PDUs MUST be encapsulated in AAL5. + + 4) Count the IP packets that are transmitted by the SUT on all VCCs + to verify connectivity and load. If the count on the test device + is the same on the SUT, continue the test; else lower the test + device traffic rate until the counts are the same. + + 5) Record the number of cell misinsertion errors at the receiver end + of the test device per VCC. + + Reporting Format: + + The results of the CMR/Steady Load/Maximum VCCs test SHOULD be + reported in a form of text and graph. + + The text results SHOULD display the numerical values of the CMR. + The values given SHOULD include: time period of test in s, test + VPI/VCI value, total number of cells transmitted and received on + the given VPI/VCI during the test in positive integers, and the + CMR for the entire test. + + The graph results SHOULD display the Cell misinsertion rate + values. There will be (Max number of VCCs/10) graphs, with 10 + VCCs indicated on each graph. The x-coordinate SHOULD be the test + run time in either seconds, minutes or days depending on the total + + + +Dunn & Martin Informational [Page 57] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + length of the test. The x-coordinate time SHOULD be configurable. + The y-coordinate SHOULD be the CMR for each VCC. There SHOULD be + no more than 10 curves on each graph, one curve indicated and + labeled for each VCC. The integration time per point MUST be + indicated. + + The results MUST also indicate the packet size in octets, traffic + rate in packets per second, and bearer class as generated by the + test device. The VCC and VPI/VCI values MUST be indicated. The + PCR, SCR, and MBS MUST be indicated. The bearer class of the + created VCC MUST also be indicated. + +3.2.4.4. CMR/Bursty VBR Load/One VCC + + Objective: To determine the SUT rate of misinserted cells on one VCC + in a transmission in relation to the total cells sent as defined in + RFC 2761 "Terminology for ATM Benchmarking". + + Procedure: + + 1) Set up the SUT and test device using the bi-directional + configuration. + + 2) Configure the SUT and test device with one VCC. The VCC SHOULD + contain one VPI/VCI. The VCC MUST be configured as either a CBR + or VBR connection. The VPI/VCI MUST not be one of the reserved + ATM signaling channels (e.g., [0,5], [0,16]). The PCR, SCR, and + MBS must be configured using one of the specified traffic + descriptors. + + 3) Send a specific number of IP packets containing one of the + specified bit patterns at a specific rate through the SUT via the + defined test VCC. Since this test is not a throughput test, the + rate should not be greater than 90% of line rate. The IP PDUs + MUST be encapsulated in AAL5. + + 4) Count the IP packets that are transmitted by the SUT to verify + connectivity and load. If the count on the test device is the + same on the SUT, continue the test; else lower the test device + traffic rate until the counts are the same. + + 5) Record the number of cell misinsertion errors at the receiver end + of the test device. + + Reporting Format: + + The results of the CMR/Bursty VBR Load/One VCC test SHOULD be + reported in a form of text and graph. + + + +Dunn & Martin Informational [Page 58] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + The text results SHOULD display the numerical values of the CMR. + The values given SHOULD include: time period of test in s, test + VPI/VCI value, total number of cells transmitted and received on + the given VPI/VCI during the test in positive integers, and the + CMR for the entire test. + + The graph results SHOULD display the Cell misinsertion rate + values. The x-coordinate SHOULD be the test run time in either + seconds, minutes or days depending on the total length of the + test. The x-coordinate time SHOULD be configurable. The y- + coordinate SHOULD be the CMR. The integration time per point MUST + be indicated. + + The results MUST also indicate the packet size in octets, traffic + rate in packets per second, and bearer class as generated by the + test device. The VCC and VPI/VCI values MUST be indicated. The + PCR, SCR, and MBS MUST be indicated. The bearer class of the + created VCC MUST also be indicated. + +3.2.4.5. CMR/Bursty VBR Load/Twelve VCCs + + Objective: To determine the SUT rate of misinserted cells on twelve + VCCs in a transmission in relation to the total cells sent as defined + in RFC 2761 "Terminology for ATM Benchmarking". + + Procedure: + + 1) Set up the SUT and test device using the bi-directional + configuration. + + 2) Configure the SUT and test device with twelve VCCs, using 1 VPI + and 12 VCIs. The VCC's MUST be configured as either a CBR or VBR + connection. The VPI/VCIs MUST not be one of the reserved ATM + signaling channels (e.g., [0,5], [0,16]). The PCR, SCR, and MBS + must be configured using one of the specified traffic + descriptors. + + 3) Send a specific number of IP packets containing one of the + specified bit patterns at a specific rate through the SUT via the + defined test VCCs. All of the VPI/VCI pairs will generate + traffic at the same traffic rate. Since this test is not a + throughput test, the rate should not be greater than 90% of line + rate. The PCR, SCR, and MBS must be indicated. The IP PDUs MUST + be encapsulated in AAL5. + + + + + + + +Dunn & Martin Informational [Page 59] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + 4) Count the IP packets that are transmitted by the SUT on all VCCs + to verify connectivity and load. If the count on the test device + is the same on the SUT, continue the test; else lower the test + device traffic rate until the counts are the same. + + 5) Record the number of cell misinsertion errors at the receiver end + of the test device per VCC. + + Reporting Format: + + The results of the CMR/Bursty VBR Load/Twelve VCCs test SHOULD be + reported in a form of text and graph. + + The text results SHOULD display the numerical values of the CMR. + The values given SHOULD include: time period of test in s, test + VPI/VCI value, total number of cells transmitted and received on + the given VPI/VCI during the test in positive integers, and the + CMR for the entire test. + + The graph results SHOULD display the Cell misinsertion rate + values. The x-coordinate SHOULD be the test run time in either + seconds, minutes or days depending on the total length of the + test. The x-coordinate time SHOULD be configurable. The y- + coordinate SHOULD be the CMR for each VCC. There should be 12 + curves on the graph, on curve indicated and labeled for each VCC. + The integration time per point MUST be indicated. + + The results MUST also indicate the packet size in octets, traffic + rate in packets per second, and bearer class as generated by the + test device. The VCC and VPI/VCI values MUST be indicated. The + PCR, SCR, and MBS MUST be indicated. The bearer class of the + created VCC MUST also be indicated. + +3.2.4.6. CMR/Bursty VBR Load/Maximum VCCs + + Objective: To determine the SUT rate of misinserted cells with the + maximum number VCCs supported on the SUT in a transmission in + relation to the total cells sent as defined in RFC 2761 "Terminology + for ATM Benchmarking". + + Procedure: + + 1) Set up the SUT and test device using the bi-directional + configuration. + + 2) Configure the SUT and test device with the maximum number of VCCs + supported on the SUT. For example, if the maximum number of VCCs + supported on the SUT is 1024, define 256 VPIs with 4 VCIs per + + + +Dunn & Martin Informational [Page 60] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + VPI. The VCC's MUST be configured as either a CBR or VBR + connection. The VPI/VCIs MUST not be one of the reserved ATM + signaling channels (e.g., [0,5], [0,16]). The PCR, SCR, and MBS + must be configured using one of the specified traffic + descriptors. + + 3) Send a specific number of IP packets containing one of the + specified bit patterns at a specific rate through the SUT via the + defined test VCCs. All of the VPI/VCI pairs will generate + traffic at the same traffic rate. Since this test is not a + throughput test, the rate should not be greater than 90% of line + rate. The IP PDUs MUST be encapsulated in AAL5. + + 4) Count the IP packets that are transmitted by the SUT on all VCCs + to verify connectivity and load. If the count on the test device + is the same on the SUT, continue the test; else lower the test + device traffic rate until the counts are the same. + + 5) Record the number of cell misinsertion errors at the receiver end + of the test device per VCC. + + Reporting Format: + + The results of the CMR/Bursty VBR Load/Maximum VCCs test SHOULD be + reported in a form of text and graph. + + The text results SHOULD display the numerical values of the CMR. + The values given SHOULD include: time period of test in s, test + VPI/VCI value, total number of cells transmitted and received on + the given VPI/VCI during the test in positive integers, and the + CMR for the entire test. + + The graph results SHOULD display the Cell misinsertion rate + values. There will be (Max number of VCCs/10) graphs, with 10 + VCCs indicated on each graph. The x-coordinate SHOULD be the test + run time in either seconds, minutes or days depending on the total + length of the test. The x-coordinate time SHOULD be configurable. + The y-coordinate SHOULD be the CMR for each VCC. There SHOULD be + no more than 10 curves on each graph, one curve indicated and + labeled for each VCC. The integration time per point MUST be + indicated. + + The results MUST also indicate the packet size in octets, traffic + rate in packets per second, and bearer class as generated by the + test device. The VCC and VPI/VCI values MUST be indicated. The + PCR, SCR, and MBS MUST be indicated. The bearer class of the + created VCC MUST also be indicated. + + + + +Dunn & Martin Informational [Page 61] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + +3.2.5. CRC Error Ratio (CRC-ER) + +3.2.5.1. CRC-ER/Steady Load/One VCC + + Objective: To determine the SUT ratio of CRC errors on one VCC in a + transmission in relation to the total cells sent as defined in RFC + 2761 "Terminology for ATM Benchmarking". + + Procedure: + + 1) Set up the SUT and test device using the bi-directional + configuration. + + 2) Configure the SUT and test device with one VCC. The VCC SHOULD + contain one VPI/VCI. The VCC MUST be configured as either a CBR, + VBR, or UBR connection. The VPI/VCI MUST not be one of the + reserved ATM signaling channels (e.g., [0,5], [0,16]). + + 3) Send a specific number of IP packets at a specific constant rate + through the SUT via the defined test VCC. Since this test is not + a throughput test, the rate should not be greater than 90% of + line rate. The IP PDUs MUST be encapsulated in AAL5. + + 4) Count the IP packets that are transmitted by the SUT to verify + connectivity and load. If the count on the test device is the + same on the SUT, continue the test; else lower the test device + traffic rate until the counts are the same. + + 5) Record the number of CRC errored cells received on the test + device. + + Reporting Format: + + The results of the CRC-ER/Steady Load/One VCC test SHOULD be + reported in a form of text and graph. + + The text results SHOULD display the numerical values of the CRC- + ER. The values given SHOULD include: time period of test in s, + test VPI/VCI value, total number of cells transmitted and received + on the given VPI/VCI during the test in positive integers, and the + CRC-ER for the entire test. + + The graph results SHOULD display the CRC Error ratio values. The + x-coordinate SHOULD be the test run time in either seconds, + minutes or days depending on the total length of the test. The + x-coordinate time SHOULD be configurable. The y-coordinate SHOULD + be the CRC-ER. The integration time per point MUST be indicated. + + + + +Dunn & Martin Informational [Page 62] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + The results MUST also indicate the packet size in octets, traffic + rate in packets per second, and bearer class as generated by the + test device. The VCC and VPI/VCI values MUST be indicated. The + PCR, SCR, and MBS MUST be indicated. The bearer class of the + created VCC MUST also be indicated. + +3.2.5.2. CRC-ER/Steady Load/Twelve VCCs + + Objective: To determine the SUT ratio of lost cells on twelve VCCs in + a transmission in relation to the total cells sent as defined in RFC + 2761 "Terminology for ATM Benchmarking". + + Procedure: + + 1) Set up the SUT and test device using the bi-directional + configuration. + + 2) Configure the SUT and test device with twelve VCCs, using 1 VPI + and 12 VCIs. The VCC's MUST be configured as either a CBR, VBR, + or UBR connection. The VPI/VCIs MUST not be one of the reserved + ATM signaling channels (e.g., [0,5], [0,16]). + + 3) Send a specific number of IP packets at a specific constant rate + through the SUT via the defined test VCCs. All of the VPI/VCI + pairs will generate traffic at the same traffic rate. Since this + test is not a throughput test, the rate should not be greater + than 90% of line rate. The IP PDUs MUST be encapsulated in AAL5. + + 4) Count the IP packets that are transmitted by the SUT on all VCCs + to verify connectivity and load. If the count on the test device + is the same on the SUT, continue the test; else lower the test + device traffic rate until the counts are the same. + + 5) Record the number of CRC errored cells received per VCC on the + test device. + + Reporting Format: + + The results of the CRC-ER/Steady Load/Twelve VCCs test SHOULD be + reported in a form of text and graph. + + The text results SHOULD display the numerical values of the CRC- + ER. The values given SHOULD include: time period of test in s, + test VPI/VCI value, total number of cells transmitted and received + on the given VPI/VCI during the test in positive integers, and the + CRC-ER for the entire test. + + + + + +Dunn & Martin Informational [Page 63] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + The graph results SHOULD display the CRC Error ratio values. The + x-coordinate SHOULD be the test run time in either seconds, + minutes or days depending on the total length of the test. The + x-coordinate time SHOULD be configurable. The y-coordinate SHOULD + be the CRC-ER for each VCC. There should be 12 curves on the + graph, on curve indicated and labeled for each VCC. The + integration time per point MUST be indicated. + + The results MUST also indicate the packet size in octets, traffic + rate in packets per second, and bearer class as generated by the + test device. The VCC and VPI/VCI values MUST be indicated. The + PCR, SCR, and MBS MUST be indicated. The bearer class of the + created VCC MUST also be indicated. + +3.2.5.3. CRC-ER/Steady Load/Maximum VCCs + + Objective: To determine the SUT ratio of lost cells with the maximum + number VCCs supported on the SUT in a transmission in relation to the + total cells sent as defined in RFC 2761 "Terminology for ATM + Benchmarking". + + Procedure: + + 1) Set up the SUT and test device using the bi-directional + configuration. + + 2) Configure the SUT and test device with the maximum number of VCCs + supported on the SUT. For example, if the maximum number of VCCs + supported on the SUT is 1024, define 256 VPIs with 4 VCIs per + VPI. The VCC's MUST be configured as either a CBR, VBR, or UBR + connection. The VPI/VCIs MUST not be one of the reserved ATM + signaling channels (e.g., [0,5], [0,16]). + + 3) Send a specific number of IP packets at a specific constant rate + through the SUT via the defined test VCCs. All of the VPI/VCI + pairs will generate traffic at the same traffic rate. Since this + test is not a throughput test, the rate should not be greater + than 90% of line rate. The IP PDUs MUST be encapsulated in AAL5. + + 4) Count the IP packets that are transmitted by the SUT on all VCCs + to verify connectivity and load. If the count on the test device + is the same on the SUT, continue the test; else lower the test + device traffic rate until the counts are the same. + + 5) Record the number of CRC errored cells received per VCC on the + test device. + + + + + +Dunn & Martin Informational [Page 64] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + Reporting Format: + + The results of the CRC-ER/Steady Load/Maximum VCCs test SHOULD be + reported in a form of text and graph. + + The text results SHOULD display the numerical values of the CRC- + ER. The values given SHOULD include: time period of test in s, + test VPI/VCI value, total number of cells transmitted and received + on the given VPI/VCI during the test in positive integers, and the + CRC-ER for the entire test. + + The graph results SHOULD display the CRC Error ratio values. + There will be (Max number of VCCs/10) graphs, with 10 VCCs + indicated on each graph. The x-coordinate SHOULD be the test run + time in either seconds, minutes or days depending on the total + length of the test. The x-coordinate time SHOULD be configurable. + The y-coordinate SHOULD be the CRC-ER for each VCC. There SHOULD + be no more than 10 curves on each graph, one curve indicated and + labeled for each VCC. The integration time per point MUST be + indicated. + + The results MUST also indicate the packet size in octets, traffic + rate in packets per second, and bearer class as generated by the + test device. The VCC and VPI/VCI values MUST be indicated. The + PCR, SCR, and MBS MUST be indicated. The bearer class of the + created VCC MUST also be indicated. + +3.2.5.4. CRC-ER/Bursty VBR Load/One VCC + + Objective: To determine the SUT ratio of lost cells on one VCC in a + transmission in relation to the total cells sent as defined in RFC + 2761 "Terminology for ATM Benchmarking". + + Procedure: + + 1) Set up the SUT and test device using the bi-directional + configuration. + + 2) Configure the SUT and test device with one VCC. The VCC SHOULD + contain one VPI/VCI. The VCC MUST be configured as either a CBR + or VBR connection. The VPI/VCI MUST not be one of the reserved + ATM signaling channels (e.g., [0,5], [0,16]). The PCR, SCR, and + MBS must be configured using one of the specified traffic + descriptors. + + 3) Send a specific number of IP packets containing one of the + specified bit patterns at a specific rate through the SUT via the + defined test VCC. Since this test is not a throughput test, the + + + +Dunn & Martin Informational [Page 65] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + rate should not be greater than 90% of line rate. The IP PDUs + MUST be encapsulated in AAL5. + + 4) Count the IP packets that are transmitted by the SUT to verify + connectivity and load. If the count on the test device is the + same on the SUT, continue the test; else lower the test device + traffic rate until the counts are the same. + + 5) Record the number of CRC errored cells received per VCC on the + test device. + + Reporting Format: + + The results of the CRC-ER/Bursty VBR Load/One VCC test SHOULD be + reported in a form of text and graph. + + The text results SHOULD display the numerical values of the CRC- + ER. The values given SHOULD include: time period of test in s, + test VPI/VCI value, total number of cells transmitted and received + on the given VPI/VCI during the test in positive integers, and the + CRC-ER for the entire test. + + The graph results SHOULD display the CRC Error ratio values. The + x-coordinate SHOULD be the test run time in either seconds, + minutes or days depending on the total length of the test. The + x-coordinate time SHOULD be configurable. The y-coordinate SHOULD + be the CRC-ER. The integration time per point MUST be indicated. + + The results MUST also indicate the packet size in octets, traffic + rate in packets per second, and bearer class as generated by the + test device. The VCC and VPI/VCI values MUST be indicated. The + PCR, SCR, and MBS MUST be indicated. The bearer class of the + created VCC MUST also be indicated. + +3.2.5.5. CRC-ER/Bursty VBR Load/Twelve VCCs + + Objective: To determine the SUT ratio of lost cells on twelve VCCs in + a transmission in relation to the total cells sent as defined in RFC + 2761 "Terminology for ATM Benchmarking". + + Procedure: + + 1) Set up the SUT and test device using the bi-directional + configuration. + + 2) Configure the SUT and test device with twelve VCCs, using 1 VPI + and 12 VCIs. The VCC MUST be configured as either a CBR or VBR + connection. The VPI/VCIs MUST not be one of the reserved ATM + + + +Dunn & Martin Informational [Page 66] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + signaling channels (e.g., [0,5], [0,16]). The PCR, SCR, and MBS + must be configured using one of the specified traffic + descriptors. + + 3) Send a specific number of IP packets containing one of the + specified bit patterns at a specific rate through the SUT via the + defined test VCCs. All of the VPI/VCI pairs will generate + traffic at the same traffic rate. Since this test is not a + throughput test, the rate should not be greater than 90% of line + rate. The PCR, SCR, and MBS must be indicated. The IP PDUs MUST + be encapsulated in AAL5. + + 4) Count the IP packets that are transmitted by the SUT on all VCCs + to verify connectivity and load. If the count on the test device + is the same on the SUT, continue the test; else lower the test + device traffic rate until the counts are the same. + + 5) Record the number of CRC errored cells received per VCC on the + test device for all VCCs. + + Reporting Format: + + The results of the CRC-ER/Bursty VBR Load/Twelve VCCs test SHOULD + be reported in a form of text and graph. + + The text results SHOULD display the numerical values of the CRC- + ER. The values given SHOULD include: time period of test in s, + test VPI/VCI value, total number of cells transmitted and received + on the given VPI/VCI during the test in positive integers, and the + CRC-ER for the entire test. + + The graph results SHOULD display the CRC Error ratio values. The + x-coordinate SHOULD be the test run time in either seconds, + minutes or days depending on the total length of the test. The + x-coordinate time SHOULD be configurable. The y-coordinate SHOULD + be the CRC-ER for each VCC. There should be 12 curves on the + graph, on curve indicated and labeled for each VCC. The + integration time per point MUST be indicated. + + The results MUST also indicate the packet size in octets, traffic + rate in packets per second, and bearer class as generated by the + test device. The VCC and VPI/VCI values MUST be indicated. The + PCR, SCR, and MBS MUST be indicated. The bearer class of the + created VCC MUST also be indicated. + + + + + + + +Dunn & Martin Informational [Page 67] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + +3.2.5.6. CRC-ER/Bursty VBR Load/Maximum VCCs + + Objective: To determine the SUT ratio of lost cells with the maximum + number VCCs supported on the SUT in a transmission in relation to the + total cells sent as defined in RFC 2761 "Terminology for ATM + Benchmarking". + + Procedure: + + 1) Set up the SUT and test device using the bi-directional + configuration. + + 2) Configure the SUT and test device with the maximum number of VCCs + supported on the SUT. For example, if the maximum number of VCCs + supported on the SUT is 1024, define 256 VPIs with 4 VCIs per + VPI. The VCC MUST be configured as either a CBR or VBR + connection. The VPI/VCIs MUST not be one of the reserved ATM + signaling channels (e.g., [0,5], [0,16]). The PCR, SCR, and MBS + must be configured using one of the specified traffic + descriptors. + + 3) Send a specific number of IP packets containing one of the + specified bit patterns at a specific rate through the SUT via the + defined test VCCs. All of the VPI/VCI pairs will generate + traffic at the same traffic rate. Since this test is not a + throughput test, the rate should not be greater than 90% of line + rate. The IP PDUs MUST be encapsulated in AAL5. + + 4) Count the IP packets that are transmitted by the SUT on all VCCs + to verify connectivity and load. If the count on the test device + is the same on the SUT, continue the test; else lower the test + device traffic rate until the counts are the same. + + 5) Record the number of CRC errored cells received per VCC on the + test device for all VCCs. + + Reporting Format: + + The results of the CRC-ER/Bursty VBR Load/Maximum VCCs test SHOULD + be reported in a form of text and graph. + + The text results SHOULD display the numerical values of the CRC- + ER. The values given SHOULD include: time period of test in s, + test VPI/VCI value, total number of cells transmitted and received + on the given VPI/VCI during the test in positive integers, and the + CRC-ER for the entire test. + + The graph results SHOULD display the CRC Error ratio values. + + + +Dunn & Martin Informational [Page 68] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + There will be (Max number of VCCs/10) graphs, with 10 VCCs + indicated on each graph. The x-coordinate SHOULD be the test run + time in either seconds, minutes or days depending on the total + length of the test. The x-coordinate time SHOULD be configurable. + The y-coordinate SHOULD be the CRC-ER for each VCC. There SHOULD + be no more than 10 curves on each graph, one curve indicated and + labeled for each VCC. The integration time per point MUST be + indicated. + + The results MUST also indicate the packet size in octets, traffic + rate in packets per second, and bearer class as generated by the + test device. The VCC and VPI/VCI values MUST be indicated. The + PCR, SCR, and MBS MUST be indicated. The bearer class of the + created VCC MUST also be indicated. + +3.2.5.7. CRC-ER/Bursty UBR Load/One VCC + + Objective: To determine the SUT ratio of lost cells on one VCC in a + transmission in relation to the total cells sent as defined in RFC + 2761 "Terminology for ATM Benchmarking". + + Procedure: + + 1) Set up the SUT and test device using the bi-directional + configuration. + + 2) Configure the SUT and test device with one VCC. The VCC SHOULD + contain one VPI/VCI. The VCC MUST be configured as a UBR + connection. The VPI/VCI MUST not be one of the reserved ATM + signaling channels (e.g., [0,5], [0,16]). The PCR, SCR, and MBS + must be configured using one of the specified traffic + descriptors. + + 3) Send a specific number of IP packets containing one of the + specified bit patterns at a specific rate through the SUT via the + defined test VCC. Since this test is not a throughput test, the + rate should not be greater than 90% of line rate. The IP PDUs + MUST be encapsulated in AAL5. + + 4) Count the IP packets that are transmitted by the SUT to verify + connectivity and load. If the count on the test device is the + same on the SUT, continue the test; else lower the test device + traffic rate until the counts are the same. + + 5) Record the number of CRC errored cells received per VCC on the + test device. + + + + + +Dunn & Martin Informational [Page 69] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + Reporting Format: + + The results of the CRC-ER/Bursty UBR Load/One VCC test SHOULD be + reported in a form of text and graph. + + The text results SHOULD display the numerical values of the CRC- + ER. The values given SHOULD include: time period of test in s, + test VPI/VCI value, total number of cells transmitted and received + on the given VPI/VCI during the test in positive integers, and the + CRC-ER for the entire test. + + The graph results SHOULD display the CRC Error ratio values. The + x-coordinate SHOULD be the test run time in either seconds, + minutes or days depending on the total length of the test. The + x-coordinate time SHOULD be configurable. The y-coordinate SHOULD + be the CRC-ER. The integration time per point MUST be indicated. + + The results MUST also indicate the packet size in octets, traffic + rate in packets per second, and bearer class as generated by the + test device. The VCC and VPI/VCI values MUST be indicated. The + PCR, SCR, and MBS MUST be indicated. The bearer class of the + created VCC MUST also be indicated. + +3.2.5.8. CRC-ER/Bursty UBR Load/Twelve VCCs + + Objective: To determine the SUT ratio of lost cells on twelve VCCs in + a transmission in relation to the total cells sent as defined in RFC + 2761 "Terminology for ATM Benchmarking". + + Procedure: + + 1) Set up the SUT and test device using the bi-directional + configuration. + + 2) Configure the SUT and test device with twelve VCCs, using 1 VPI + and 12 VCIs. The VCC MUST be configured as a UBR connection. + The VPI/VCIs MUST not be one of the reserved ATM signaling + channels (e.g., [0,5], [0,16]). The PCR, SCR, and MBS must be + configured using one of the specified traffic descriptors. + + 3) Send a specific number of IP packets containing one of the + specified bit patterns at a specific rate through the SUT via the + defined test VCCs. All of the VPI/VCI pairs will generate + traffic at the same traffic rate. Since this test is not a + throughput test, the rate should not be greater than 90% of line + rate. The PCR, SCR, and MBS must be indicated. The IP PDUs MUST + be encapsulated in AAL5. + + + + +Dunn & Martin Informational [Page 70] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + 4) Count the IP packets that are transmitted by the SUT on all VCCs + to verify connectivity and load. If the count on the test device + is the same on the SUT, continue the test; else lower the test + device traffic rate until the counts are the same. + + 5) Record the number of CRC errored cells received per VCC on the + test device for all VCCs. + + Reporting Format: + + The results of the CRC-ER/Bursty UBR Load/Twelve VCCs test SHOULD + be reported in a form of text and graph. + + The text results SHOULD display the numerical values of the CRC- + ER. The values given SHOULD include: time period of test in s, + test VPI/VCI value, total number of cells transmitted and received + on the given VPI/VCI during the test in positive integers, and the + CRC-ER for the entire test. + + The graph results SHOULD display the CRC Error ratio values. The + x-coordinate SHOULD be the test run time in either seconds, + minutes or days depending on the total length of the test. The + x-coordinate time SHOULD be configurable. The y-coordinate SHOULD + be the CRC-ER for each VCC. There should be 12 curves on the + graph, on curve indicated and labeled for each VCC. The + integration time per point MUST be indicated. + + The results MUST also indicate the packet size in octets, traffic + rate in packets per second, and bearer class as generated by the + test device. The VCC and VPI/VCI values MUST be indicated. The + PCR, SCR, and MBS MUST be indicated. The bearer class of the + created VCC MUST also be indicated. + +3.2.5.9. CRC-ER/Bursty UBR Load/Maximum VCCs + + Objective: To determine the SUT ratio of lost cells with the maximum + number VCCs supported on the SUT in a transmission in relation to the + total cells sent as defined in RFC 2761 "Terminology for ATM + Benchmarking". + + Procedure: + + 1) Set up the SUT and test device using the bi-directional + configuration. + + 2) Configure the SUT and test device with the maximum number of VCCs + supported on the SUT. For example, if the maximum number of VCCs + supported on the SUT is 1024, define 256 VPIs with 4 VCIs per + + + +Dunn & Martin Informational [Page 71] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + VPI. The VCC MUST be configured as a UBR connection. The + VPI/VCIs MUST not be one of the reserved ATM signaling channels + (e.g., [0,5], [0,16]). The PCR, SCR, and MBS must be configured + using one of the specified traffic descriptors. + + 3) Send a specific number of IP packets containing one of the + specified bit patterns at a specific rate through the SUT via the + defined test VCCs. All of the VPI/VCI pairs will generate + traffic at the same traffic rate. Since this test is not a + throughput test, the rate should not be greater than 90% of line + rate. The IP PDUs MUST be encapsulated in AAL5. + + 4) Count the IP packets that are transmitted by the SUT on all VCCs + to verify connectivity and load. If the count on the test device + is the same on the SUT, continue the test; else lower the test + device traffic rate until the counts are the same. + + 5) Record the number of CRC errored cells received per VCC on the + test device for all VCCs. + + Reporting Format: + + The results of the CRC-ER/Bursty UBR Load/Maximum VCCs test SHOULD + be reported in a form of text and graph. + + The text results SHOULD display the numerical values of the CRC- + ER. The values given SHOULD include: time period of test in s, + test VPI/VCI value, total number of cells transmitted and received + on the given VPI/VCI during the test in positive integers, and the + CRC-ER for the entire test. + + The graph results SHOULD display the CRC Error ratio values. + There will be (Max number of VCCs/10) graphs, with 10 VCCs + indicated on each graph. The x-coordinate SHOULD be the test run + time in either seconds, minutes or days depending on the total + length of the test. The x-coordinate time SHOULD be configurable. + The y-coordinate SHOULD be the CRC-ER for each VCC. There SHOULD + be no more than 10 curves on each graph, one curve indicated and + labeled for each VCC. The integration time per point MUST be + indicated. + + The results MUST also indicate the packet size in octets, traffic + rate in packets per second, and bearer class as generated by the + test device. The VCC and VPI/VCI values MUST be indicated. The + PCR, SCR, and MBS MUST be indicated. The bearer class of the + created VCC MUST also be indicated. + + + + + +Dunn & Martin Informational [Page 72] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + +3.2.5.10. CRC-ER/Bursty Mixed Load/Three VCC + + Objective: To determine the SUT ratio of lost cells on three VCC's in + relation to the total cells sent as defined in RFC 2761 "Terminology + for ATM Benchmarking". + + Procedure: + + 1) Set up the SUT and test device using the bi-directional + configuration. + + 2) Configure the SUT and test device with three VCC's. Each VCC + MUST be defined as a different Bearer class; one CBR, one UBR and + one VBR. Each VCC SHOULD contain one VPI/VCI. The VPI/VCI MUST + not be one of the reserved ATM signaling channels (e.g., [0,5], + [0,16]). The PCR, SCR, and MBS must be configured using one of + the specified traffic descriptors. + + 3) Send a specific number of IP packets containing one of the + specified bit patterns through the SUT via the defined test VCCs. + Each generated VCC stream MUST match the corresponding VCC Bearer + class. All of the VPI/VCI pairs will generate traffic at the + same traffic rate. Since this test is not a throughput test, the + rate should not be greater than 90% of line rate. The IP PDUs + MUST be encapsulated in AAL5. + + 4) Count the IP packets that are transmitted by the SUT to verify + connectivity and load. If the count on the test device is the + same on the SUT, continue the test; else lower the test device + traffic rate until the counts are the same. + + 5) Record the number of CRC errored cells received per VCC on the + test device. + + Reporting Format: + + The results of the CRC-ER/Bursty Mixed Load/Three VCC test SHOULD + be reported in in a form of text and graph. + + The text results SHOULD display the numerical values of the CRC- + ER. The values given SHOULD include: time period of test in s, + test VPI/VCI value, total number of cells transmitted and received + on the given VPI/VCI during the test in positive integers, and the + CRC-ER for the entire test. + + The graph results SHOULD display the CRC Error ratio values. The + x-coordinate SHOULD be the test run time in either seconds, + minutes or days depending on the total length of the test. The + + + +Dunn & Martin Informational [Page 73] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + x-coordinate time SHOULD be configurable. The y-coordinate SHOULD + be the CRC-ER for each VCC. There should be 12 curves on the + graph, on curve indicated and labeled for each VCC. The + integration time per point MUST be indicated. + + The results MUST also indicate the packet size in octets, traffic + rate in packets per second, and bearer class as generated by the + test device. The VCC and VPI/VCI values MUST be indicated. The + PCR, SCR, and MBS MUST be indicated. The bearer class of the + created VCC MUST also be indicated. + +3.2.5.11. CRC-ER/Bursty Mixed Load/Twelve VCCs + + Objective: To determine the SUT ratio of lost cells on twelve VCCs in + a transmission in relation to the total cells sent as defined in RFC + 2761 "Terminology for ATM Benchmarking". + + Procedure: + + 1) Set up the SUT and test device using the bi-directional + configuration. + + 2) Configure the SUT and test device with twelve VCC's. Each VCC + MUST be defined as one of the Bearer classes for a total of four + CBR, four UBR and four VBR VCC's. Each VCC SHOULD contain one + VPI/VCI. The VPI/VCI MUST not be one of the reserved ATM + signaling channels (e.g., [0,5], [0,16]). + + 3) Send a specific number of IP packets containing one of the + specified bit patterns through the SUT via the defined test VCCs. + Each generated VCC stream MUST match the corresponding VCC Bearer + class. All of the VPI/VCI pairs will generate traffic at the + same traffic rate. Since this test is not a throughput test, the + rate should not be greater than 90% of line rate. The IP PDUs + MUST be encapsulated in AAL5. + + 4) Count the IP packets that are transmitted by the SUT on all VCCs + to verify connectivity and load. If the count on the test device + is the same on the SUT, continue the test; else lower the test + device traffic rate until the counts are the same. + + 5) Record the number of CRC errored cells received per VCC on the + test device for all VCCs. + + Reporting Format: + + The results of the CRC-ER/Bursty Mixed Load/Twelve VCCs test + SHOULD be reported in a form of text and graph. + + + +Dunn & Martin Informational [Page 74] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + The text results SHOULD display the numerical values of the CRC- + ER. The values given SHOULD include: time period of test in s, + test VPI/VCI value, total number of cells transmitted and received + on the given VPI/VCI during the test in positive integers, and the + CRC-ER for the entire test. + + The graph results SHOULD display the CRC Error ratio values. The + x-coordinate SHOULD be the test run time in either seconds, + minutes or days depending on the total length of the test. The + x-coordinate time SHOULD be configurable. The y-coordinate SHOULD + be the CRC-ER for each VCC. There should be 12 curves on the + graph, on curve indicated and labeled for each VCC. The + integration time per point MUST be indicated. + + The results MUST also indicate the packet size in octets, traffic + rate in packets per second, and bearer class as generated by the + test device. The VCC and VPI/VCI values MUST be indicated. The + PCR, SCR, and MBS MUST be indicated. The bearer class of the + created VCC MUST also be indicated. + +3.2.5.12. CRC-ER/Bursty Mixed Load/Maximum VCCs + + Objective: To determine the SUT ratio of lost cells with the maximum + number VCCs supported on the SUT in a transmission in relation to the + total cells sent as defined in RFC 2761 "Terminology for ATM + Benchmarking". + + Procedure: + + 1) Set up the SUT and test device using the bi-directional + configuration. + + 2) Configure the SUT and test device with maximum number of VCCs + supported on the SUT. For example, if the maximum number of VCCs + supported on the SUT is 1024, define 256 VPIs with 4 VCIs per + VPI. Each VCC MUST be defined as one of the Bearer classes for a + total of (max VCC/3) CBR, (max VCC/3) UBR and (max VCC/3) VBR + VCC's. The VPI/VCI MUST not be one of the reserved ATM signaling + channels (e.g., [0,5], [0,16]). + + 3) Send a specific number of IP packets containing one of the + specified bit patterns through the SUT via the defined test VCCs. + Each generated VCC stream MUST match the corresponding VCC Bearer + class. All of the VPI/VCI pairs will generate traffic at the + same traffic rate. Since this test is not a throughput test, the + rate should not be greater than 90% of line rate. The IP PDUs + MUST be encapsulated in AAL5. + + + + +Dunn & Martin Informational [Page 75] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + 4) Count the IP packets that are transmitted by the SUT on all VCCs + to verify connectivity and load. If the count on the test device + is the same on the SUT, continue the test; else lower the test + device traffic rate until the counts are the same. + + 5) Record the number of CRC errored cells received per VCC on the + test device for all VCCs. + + Reporting Format: + + The results of the CRC-ER/Bursty Mixed Load/Maximum VCCs test + SHOULD be reported in a form of text and graph. + + The text results SHOULD display the numerical values of the CRC- + ER. The values given SHOULD include: time period of test in s, + test VPI/VCI value, total number of cells transmitted and received + on the given VPI/VCI during the test in positive integers, and the + CRC-ER for the entire test. + + The graph results SHOULD display the CRC Error ratio values. + There will be (Max number of VCCs/10) graphs, with 10 VCCs + indicated on each graph. The x-coordinate SHOULD be the test run + time in either seconds, minutes or days depending on the total + length of the test. The x-coordinate time SHOULD be configurable. + The y-coordinate SHOULD be the CRC-ER for each VCC. There SHOULD + be no more than 10 curves on each graph, one curve indicated and + labeled for each VCC. The integration time per point MUST be + indicated. + + The results MUST also indicate the packet size in octets, traffic + rate in packets per second, and bearer class as generated by the + test device. The VCC and VPI/VCI values MUST be indicated. The + PCR, SCR, and MBS MUST be indicated. The bearer class of the + created VCC MUST also be indicated. + +3.2.6. Cell Transfer Delay (CTD) + +3.2.6.1. Test Setup + + The cell transfer delay measurements assume that both the transmitter + and receiver timestamp information is synchronized. Synchronization + SHOULD be achieved by supplying a common clock signal (minimum of 100 + Mhz or 10 ns resolution) to both the transmitter and receiver. The + maximum timestamp values MUST be recorded to ensure synchronization + in the case of counter rollover. The cell transfer delay + measurements SHOULD utilize the O.191 cell (ITUT-O.191) encapsulated + in a valid IP packet. If the O.191 cell is not available, a test + cell encapsulated in a valid IP packet MAY be used. The test cell + + + +Dunn & Martin Informational [Page 76] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + MUST contain a transmit timestamp which can be correlated with a + receive timestamp. A description of the test cell MUST be included + in the test results. The description MUST include the timestamp + length (in bits), counter rollover value, and the timestamp accuracy + (in ns). + +3.2.6.2. CTD/Steady Load/One VCC + + Objective: To determine the SUT variation in cell transfer delay with + one VCC as defined in RFC 2761 "Terminology for ATM Benchmarking". + + Procedure: + + 1) Set up the SUT and test device using the bi-directional + configuration. + + 2) Configure the SUT and test device with one VCC. The VCC SHOULD + contain one VPI/VCI. The VCC MUST be configured as either a CBR, + VBR, or UBR connection. The VPI/VCI MUST not be one of the + reserved ATM signaling channels (e.g., [0,5], [0,16]). + + 3) Send a specific number of IP packets containing timestamps at a + specific constant rate through the SUT via the defined test VCC. + Since this test is not a throughput test, the rate should not be + greater than 90% of line rate. The IP PDUs MUST be encapsulated + in AAL5. + + 4) Count the IP packets that are transmitted by the SUT to verify + connectivity and load. If the count on the test device is the + same on the SUT, continue the test; else lower the test device + traffic rate until the counts are the same. + + 5) Record the packets timestamps at the transmitter and receiver + ends of the test device. + + Reporting Format: + + The results of the CTD/Steady Load/One VCC test SHOULD be reported + in a form of text, graph, and histogram. + + The text results SHOULD display the numerical values of the CTD. + The values given SHOULD include: time period of test in s, test + VPI/VCI value, total number of cells transmitted and received on + the given VPI/VCI during the test in positive integers, minimum, + maximum, and mean CTD during the test in us. + + The graph results SHOULD display the cell transfer delay values. + The x-coordinate SHOULD be the test run time in either seconds, + + + +Dunn & Martin Informational [Page 77] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + minutes or days depending on the total length of the test. The + x-coordinate time SHOULD be configurable. The y-coordinate SHOULD + be the cell transfer delay in us. The integration time per point + MUST be indicated. + + The histogram results SHOULD display the cell transfer delay. The + x-coordinate SHOULD be the cell transfer delay in us with at least + 256 bins. The y-coordinate SHOULD be the number of cells observed + in each bin. + + The results MUST also indicate the packet size in octets, traffic + rate in packets per second, and bearer class as generated by the + test device. The VCC and VPI/VCI values MUST be indicated. The + bearer class of the created VCC MUST also be indicated. + +3.2.6.3. CTD/Steady Load/Twelve VCCs + + Objective: To determine the SUT variation in cell transfer delay with + twelve VCCs as defined in RFC 2761 "Terminology for ATM + Benchmarking". + + Procedure: + + 1) Set up the SUT and test device using the bi-directional + configuration. + + 2) Configure the SUT and test device with twelve VCCs, using 1 VPI + and 12 VCIs. The VCC's MUST be configured as either a CBR, VBR, + or UBR connection. The VPI/VCIs MUST not be one of the reserved + ATM signaling channels (e.g., [0,5], [0,16]). + + 3) Send a specific number of IP packets containing timestamps at a + specific constant rate through the SUT via the defined test VCCs. + All of the VPI/VCI pairs will generate traffic at the same + traffic rate. Since this test is not a throughput test, the rate + should not be greater than 90% of line rate. The IP PDUs MUST be + encapsulated in AAL5. + + 4) Count the IP packets that are transmitted by the SUT on all VCCs + to verify connectivity and load. If the count on the test device + is the same on the SUT, continue the test; else lower the test + device traffic rate until the counts are the same. + + 5) Record the packets timestamps at the transmitter and receiver + ends of the test device for all VCCs. + + + + + + +Dunn & Martin Informational [Page 78] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + Reporting Format: + + The results of the CTD/Steady Load/Twelve VCCs test SHOULD be + reported in a form of text, graph, and histograms. + + The text results SHOULD display the numerical values of the CTD. + The values given SHOULD include: time period of test in s, test + VPI/VCI values, total number of cells transmitted and received on + each VCC during the test in positive integers, maximum and minimum + CTD on each VCC during the test in us, and mean CTD on each VCC in + us. + + The graph results SHOULD display the cell transfer delay values. + The x-coordinate SHOULD be the test run time in either seconds, + minutes or days depending on the total length of the test. The + x-coordinate time SHOULD be configurable. The y-coordinate SHOULD + be the cell transfer delay for each VCC in ms. There SHOULD be 12 + curves on the graph, one curves indicated and labeled for each + VCC. The integration time per point MUST be indicated. + + The histograms SHOULD display the cell transfer delay. There will + be one histogram for each VCC. The x-coordinate SHOULD be the + cell transfer delay in us with at least 256 bins. The y- + coordinate SHOULD be the number of cells observed in each bin. + + The results MUST also indicate the packet size in octets, traffic + rate in packets per second, and bearer class as generated by the + test device. The VCC and VPI/VCI values MUST be indicated. The + bearer class of the created VCC MUST also be indicated. + +3.2.6.4. CTD/Steady Load/Maximum VCCs + + Objective: To determine the SUT variation in cell transfer delay with + the maximum number VCCs supported on the SUT as defined in RFC 2761 + "Terminology for ATM Benchmarking". + + Procedure: + + 1) Set up the SUT and test device using the bi-directional + configuration. + + 2) Configure the SUT and test device with the maximum number of VCCs + supported on the SUT. For example, if the maximum number of VCCs + supported on the SUT is 1024, define 256 VPIs with 4 VCIs per + VPI. The VCC's MUST be configured as either a CBR, VBR, or UBR + connection. The VPI/VCIs MUST not be one of the reserved ATM + signaling channels (e.g., [0,5], [0,16]). + + + + +Dunn & Martin Informational [Page 79] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + 3) Send a specific number of IP packets containing timestamps at a + specific constant rate through the SUT via the defined test VCCs. + All of the VPI/VCI pairs will generate traffic at the same + traffic rate. Since this test is not a throughput test, the rate + should not be greater than 90% of line rate. The IP PDUs MUST be + encapsulated in AAL5. + + 4) Count the IP packets that are transmitted by the SUT on all VCCs + to verify connectivity and load. If the count on the test device + is the same on the SUT, continue the test; else lower the test + device traffic rate until the counts are the same. + + 5) Record the packets timestamps at the transmitter and receiver + ends of the test device for all VCCs. + + Reporting Format: + + The results of the CTD/Steady Load/Maximum VCCs test SHOULD be + reported in a form of text, graphs, and histograms. + + The text results SHOULD display the numerical values of the CTD. + The values given SHOULD include: time period of test in s, test + VPI/VCI values, total number of cells transmitted and received on + each VCC during the test in positive integers, maximum and minimum + CTD on each VCC during the test in us, and mean CTD on each VCC in + us. + + The graph results SHOULD display the cell transfer delay values. + There will be (Max number of VCCs/10) graphs, with 10 VCCs + indicated on each graph. The x-coordinate SHOULD be the test run + time in either seconds, minutes or days depending on the total + length of the test. The x-coordinate time SHOULD be configurable. + The y-coordinate SHOULD be the cell transfer delay for each VCC in + us. There SHOULD be no more than 10 curves on each graph, one + curve indicated and labeled for each VCC. The integration time + per point MUST be indicated. + + The histograms SHOULD display the cell transfer delay. There will + be one histogram for each VCC. The x-coordinate SHOULD be the + cell transfer delay in us with at least 256 bins. The y- + coordinate SHOULD be the number of cells observed in each bin. + + The results MUST also indicate the packet size in octets, traffic + rate in packets per second, and bearer class as generated by the + test device. The VCC and VPI/VCI values MUST be indicated. The + bearer class of the created VCC MUST also be indicated. + + + + + +Dunn & Martin Informational [Page 80] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + +3.2.6.5. CTD/Bursty VBR Load/One VCC + + Objective: To determine the SUT variation in cell transfer delay with + one VCC as defined in RFC 2761 "Terminology for ATM Benchmarking". + + Procedure: + + 1) Set up the SUT and test device using the bi-directional + configuration. + + 2) Configure the SUT and test device with one VCC. The VCC SHOULD + contain one VPI/VCI. The VCC MUST be configured as either a CBR + or VBR connection. The VPI/VCI MUST not be one of the reserved + ATM signaling channels (e.g., [0,5], [0,16]). + + 3) Send a specific number of IP packets containing timestamps at a + specific VBR through the SUT via the defined test VCC. Since + this test is not a throughput test, the rate should not be + greater than 90% of line rate. The IP PDUs MUST be encapsulated + in AAL5. + + 4) Count the IP packets that are transmitted by the SUT to verify + connectivity and load. If the count on the test device is the + same on the SUT, continue the test; else lower the test device + traffic rate until the counts are the same. + + 5) Record the packets timestamps at the transmitter and receiver + ends of the test device. + + Reporting Format: + + The results of the CTD/Bursty VBR Load/One VCC test SHOULD be + reported in a form of text, graph, and histogram. + + The text results SHOULD display the numerical values of the CTD. + The values given SHOULD include: time period of test in s, test + VPI/VCI value, total number of cells transmitted and received on + the given VPI/VCI during the test in positive integers, minimum, + maximum, and mean CTD during the test in us. + + The graph results SHOULD display the cell transfer delay values. + The x-coordinate SHOULD be the test run time in either seconds, + minutes or days depending on the total length of the test. The + x-coordinate time SHOULD be configurable. The y-coordinate SHOULD + be the cell transfer delay in us. The integration time per point + MUST be indicated. + + + + + +Dunn & Martin Informational [Page 81] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + The histogram results SHOULD display the cell transfer delay. The + x-coordinate SHOULD be the cell transfer delay in us with at least + 256 bins. The y-coordinate SHOULD be the number of cells observed + in each bin. + + The results MUST also indicate the packet size in octets, traffic + rate in packets per second, and bearer class as generated by the + test device. The VCC and VPI/VCI values MUST be indicated. The + PCR, SCR, and MBS MUST be indicated. The bearer class of the + created VCC MUST also be indicated. + +3.2.6.6. CTD/Bursty VBR Load/Twelve VCCs + + Objective: To determine the SUT variation in cell transfer delay with + twelve VCCs as defined in RFC 2761 "Terminology for ATM + Benchmarking". + + Procedure: + + 1) Set up the SUT and test device using the bi-directional + configuration. + + 2) Configure the SUT and test device with twelve VCCs, using 1 VPI + and 12 VCIs. The VCC's MUST be configured as either a CBR or VBR + connection. The VPI/VCIs MUST not be one of the reserved ATM + signaling channels (e.g., [0,5], [0,16]). + + 3) Send a specific number of IP packets containing timestamps at a + specific VBR through the SUT via the defined test VCCs. All of + the VPI/VCI pairs will generate traffic at the same traffic rate. + Since this test is not a throughput test, the rate should not be + greater than 90% of line rate. The IP PDUs MUST be encapsulated + in AAL5. + + 4) Count the IP packets that are transmitted by the SUT on all VCCs + to verify connectivity and load. If the count on the test device + is the same on the SUT, continue the test; else lower the test + device traffic rate until the counts are the same. + + 5) Record the packets timestamps at the transmitter and receiver + ends of the test device for all VCCs. + + Reporting Format: + + The results of the CTD/Bursty VBR Load/Twelve VCCs test SHOULD be + reported in a form of text, graph, and histograms. + + + + + +Dunn & Martin Informational [Page 82] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + The text results SHOULD display the numerical values of the CTD. + The values given SHOULD include: time period of test in s, test + VPI/VCI values, total number of cells transmitted and received on + each VCC during the test in positive integers, maximum and minimum + CTD on each VCC during the test in us, and mean CTD on each VCC in + us. + + The graph results SHOULD display the cell transfer delay values. + The x-coordinate SHOULD be the test run time in either seconds, + minutes or days depending on the total length of the test. The + x-coordinate time SHOULD be configurable. The y-coordinate SHOULD + be the cell transfer delay for each VCC in ms. There SHOULD be 12 + curves on the graph, one curves indicated and labeled for each + VCC. The integration time per point MUST be indicated. + + The histograms SHOULD display the cell transfer delay. There will + be one histogram for each VCC. The x-coordinate SHOULD be the + cell transfer delay in us with at least 256 bins. The y- + coordinate SHOULD be the number of cells observed in each bin. + + The results MUST also indicate the packet size in octets, traffic + rate in packets per second, and bearer class as generated by the + test device. The VCC and VPI/VCI values MUST be indicated. The + PCR, SCR, and MBS MUST be indicated. The bearer class of the + created VCC MUST also be indicated. + +3.2.6.7. CTD/Bursty VBR Load/Maximum VCCs + + Objective: To determine the SUT variation in cell transfer delay with + the maximum number VCCs supported on the SUT as defined in RFC 2761 + "Terminology for ATM Benchmarking". + + Procedure: + + 1) Set up the SUT and test device using the bi-directional + configuration. + + 2) Configure the SUT and test device with the maximum number of VCCs + supported on the SUT. For example, if the maximum number of VCCs + supported on the SUT is 1024, define 256 VPIs with 4 VCIs per + VPI. The VCC's MUST be configured as either a CBR or VBR + connection. The VPI/VCIs MUST not be one of the reserved ATM + signaling channels (e.g., [0,5], [0,16]). + + 3) Send a specific number of IP packets containing timestamps at a + specific VBR through the SUT via the defined test VCCs. All of + the VPI/VCI pairs will generate traffic at the same traffic rate. + Since this test is not a throughput test, the rate should not be + + + +Dunn & Martin Informational [Page 83] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + greater than 90% of line rate. The IP PDUs MUST be encapsulated + in AAL5. + + 4) Count the IP packets that are transmitted by the SUT on all VCCs + to verify connectivity and load. If the count on the test device + is the same on the SUT, continue the test; else lower the test + device traffic rate until the counts are the same. + + 5) Record the packets timestamps at the transmitter and receiver + ends of the test device for all VCCs. + + Reporting Format: + + The results of the CTD/Bursty VBR Load/Maximum VCCs test SHOULD be + reported in a form of text, graphs, and histograms. + + The text results SHOULD display the numerical values of the CTD. + The values given SHOULD include: time period of test in s, test + VPI/VCI values, total number of cells transmitted and received on + each VCC during the test in positive integers, maximum and minimum + CTD on each VCC during the test in us, and mean CTD on each VCC in + us. + + The graph results SHOULD display the cell transfer delay values. + There will be (Max number of VCCs/10) graphs, with 10 VCCs + indicated on each graph. The x-coordinate SHOULD be the test run + time in either seconds, minutes or days depending on the total + length of the test. The x-coordinate time SHOULD be configurable. + The y-coordinate SHOULD be the cell transfer delay for each VCC in + us. There SHOULD be no more than 10 curves on each graph, one + curve indicated and labeled for each VCC. The integration time + per point MUST be indicated. + + The histograms SHOULD display the cell transfer delay. There will + be one histogram for each VCC. The x-coordinate SHOULD be the + cell transfer delay in us with at least 256 bins. The y- + coordinate SHOULD be the number of cells observed in each bin. + + The results MUST also indicate the packet size in octets, traffic + rate in packets per second, and bearer class as generated by the + test device. The VCC and VPI/VCI values MUST be indicated. The + PCR, SCR, and MBS MUST be indicated. The bearer class of the + created VCC MUST also be indicated. + + + + + + + + +Dunn & Martin Informational [Page 84] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + +3.2.6.8. CTD/Bursty UBR Load/One VCC + + Objective: To determine the SUT variation in cell transfer delay with + one VCC as defined in RFC 2761 "Terminology for ATM Benchmarking". + + Procedure: + + 1) Set up the SUT and test device using the bi-directional + configuration. + + 2) Configure the SUT and test device with one VCC. The VCC SHOULD + contain one VPI/VCI. The VCC MUST be configured as a UBR + connection. The VPI/VCI MUST not be one of the reserved ATM + signaling channels (e.g., [0,5], [0,16]). + + 3) Send a specific number of IP packets containing timestamps at a + specific UBR through the SUT via the defined test VCC. Since + this test is not a throughput test, the rate should not be + greater than 90% of line rate. The IP PDUs MUST be encapsulated + in AAL5. + + 4) Count the IP packets that are transmitted by the SUT to verify + connectivity and load. If the count on the test device is the + same on the SUT, continue the test; else lower the test device + traffic rate until the counts are the same. + + 5) Record the packets timestamps at the transmitter and receiver + ends of the test device. + + Reporting Format: + + The results of the CTD/Bursty UBR Load/One VCC test SHOULD be + reported in a form of text, graph, and histogram. + + The text results SHOULD display the numerical values of the CTD. + The values given SHOULD include: time period of test in s, test + VPI/VCI value, total number of cells transmitted and received on + the given VPI/VCI during the test in positive integers, minimum, + maximum, and mean CTD during the test in us. + + The graph results SHOULD display the cell transfer delay values. + The x-coordinate SHOULD be the test run time in either seconds, + minutes or days depending on the total length of the test. The + x-coordinate time SHOULD be configurable. The y-coordinate SHOULD + be the cell transfer delay in us. The integration time per point + MUST be indicated. + + + + + +Dunn & Martin Informational [Page 85] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + The histogram results SHOULD display the cell transfer delay. The + x-coordinate SHOULD be the cell transfer delay in us with at least + 256 bins. The y-coordinate SHOULD be the number of cells observed + in each bin. + + The results MUST also indicate the packet size in octets, traffic + rate in packets per second, and bearer class as generated by the + test device. The VCC and VPI/VCI values MUST be indicated. The + bearer class of the created VCC MUST also be indicated. + +3.2.6.9. CTD/Bursty UBR Load/Twelve VCCs + + Objective: To determine the SUT variation in cell transfer delay with + twelve VCCs as defined in RFC 2761 "Terminology for ATM + Benchmarking". + + Procedure: + + 1) Set up the SUT and test device using the bi-directional + configuration. + + 2) Configure the SUT and test device with twelve VCCs, using 1 VPI + and 12 VCIs. The VCC's MUST be configured as a UBR connection. + The VPI/VCIs MUST not be one of the reserved ATM signaling + channels (e.g., [0,5], [0,16]). + + 3) Send a specific number of IP packets containing timestamps at a + specific UBR through the SUT via the defined test VCCs. All of + the VPI/VCI pairs will generate traffic at the same traffic rate. + Since this test is not a throughput test, the rate should not be + greater than 90% of line rate. The IP PDUs MUST be encapsulated + in AAL5. + + 4) Count the IP packets that are transmitted by the SUT on all VCCs + to verify connectivity and load. If the count on the test device + is the same on the SUT, continue the test; else lower the test + device traffic rate until the counts are the same. + + 5) Record the packets timestamps at the transmitter and receiver + ends of the test device for all VCCs. + + Reporting Format: + + The results of the CTD/Bursty UBR Load/Twelve VCCs test SHOULD be + reported in a form of text, graph, and histograms. + + The text results SHOULD display the numerical values of the CTD. + The values given SHOULD include: time period of test in s, test + + + +Dunn & Martin Informational [Page 86] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + VPI/VCI values, total number of cells transmitted and received on + each VCC during the test in positive integers, maximum and minimum + CTD on each VCC during the test in us, and mean CTD on each VCC in + us. + + The graph results SHOULD display the cell transfer delay values. + The x-coordinate SHOULD be the test run time in either seconds, + minutes or days depending on the total length of the test. The + x-coordinate time SHOULD be configurable. The y-coordinate SHOULD + be the cell transfer delay for each VCC in ms. There SHOULD be 12 + curves on the graph, one curves indicated and labeled for each + VCC. The integration time per point MUST be indicated. + + The histograms SHOULD display the cell transfer delay. There will + be one histogram for each VCC. The x-coordinate SHOULD be the + cell transfer delay in us with at least 256 bins. The y- + coordinate SHOULD be the number of cells observed in each bin. + + The results MUST also indicate the packet size in octets, traffic + rate in packets per second, and bearer class as generated by the + test device. The VCC and VPI/VCI values MUST be indicated. The + bearer class of the created VCC MUST also be indicated. + +3.2.6.10. CTD/Bursty UBR Load/Maximum VCCs + + Objective: To determine the SUT variation in cell transfer delay with + the maximum number VCCs supported on the SUT as defined in RFC 2761 + "Terminology for ATM Benchmarking". + + Procedure: + + 1) Set up the SUT and test device using the bi-directional + configuration. + + 2) Configure the SUT and test device with the maximum number of VCCs + supported on the SUT. For example, if the maximum number of VCCs + supported on the SUT is 1024, define 256 VPIs with 4 VCIs per + VPI. The VCC MUST be configured as a UBR connection. The + VPI/VCIs MUST not be one of the reserved ATM signaling channels + (e.g., [0,5], [0,16]). + + 3) Send a specific number of IP packets containing timestamps at a + specific UBR through the SUT via the defined test VCCs. All of + the VPI/VCI pairs will generate traffic at the same traffic rate. + Since this test is not a throughput test, the rate should not be + greater than 90% of line rate. The IP PDUs MUST be encapsulated + in AAL5. + + + + +Dunn & Martin Informational [Page 87] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + 4) Count the IP packets that are transmitted by the SUT on all VCCs + to verify connectivity and load. If the count on the test device + is the same on the SUT, continue the test; else lower the test + device traffic rate until the counts are the same. + + 5) Record the packets timestamps at the transmitter and receiver + ends of the test device for all VCCs. + + Reporting Format: + + The results of the CTD/Bursty UBR Load/Maximum VCCs test SHOULD be + reported in a form of text, graphs, and histograms. + + The text results SHOULD display the numerical values of the CTD. + The values given SHOULD include: time period of test in s, test + VPI/VCI values, total number of cells transmitted and received on + each VCC during the test in positive integers, maximum and minimum + CTD on each VCC during the test in us, and mean CTD on each VCC in + us. + + The graph results SHOULD display the cell transfer delay values. + There will be (Max number of VCCs/10) graphs, with 10 VCCs + indicated on each graph. The x-coordinate SHOULD be the test run + time in either seconds, minutes or days depending on the total + length of the test. The x-coordinate time SHOULD be configurable. + The y-coordinate SHOULD be the cell transfer delay for each VCC in + us. There SHOULD be no more than 10 curves on each graph, one + curve indicated and labeled for each VCC. The integration time + per point MUST be indicated. + + The histograms SHOULD display the cell transfer delay. There will + be one histogram for each VCC. The x-coordinate SHOULD be the + cell transfer delay in us with at least 256 bins. The y- + coordinate SHOULD be the number of cells observed in each bin. + + The results MUST also indicate the packet size in octets, traffic + rate in packets per second, and bearer class as generated by the + test device. The VCC and VPI/VCI values MUST be indicated. The + bearer class of the created VCC MUST also be indicated. + +3.2.6.11. CTD/Mixed Load/Three VCC's + + Objective: To determine the SUT variation in cell transfer delay with + three VCC's as defined in RFC 2761 "Terminology for ATM + Benchmarking". + + + + + + +Dunn & Martin Informational [Page 88] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + Procedure: + + 1) Set up the SUT and test device using the bi-directional + configuration. + + 2) Configure the SUT and test device with three VCC's. Each VCC + MUST be defined as a different Bearer class: one CBR, one UBR and + one VBR. Each VCC SHOULD contain one VPI/VCI. The VPI/VCI MUST + not be one of the reserved ATM signaling channels (e.g., [0,5], + [0,16]). + + 3) Send a specific number of IP packets containing timestamps + through the SUT via the defined test VCCs. Each generated VCC + stream MUST match the corresponding VCC Bearer class. All of the + VPI/VCI pairs will generate traffic at the same traffic rate. + Since this test is not a throughput test, the rate should not be + greater than 90% of line rate. The IP PDUs MUST be encapsulated + in AAL5. + + 4) Count the IP packets that are transmitted by the SUT to verify + connectivity and load. If the count on the test device is the + same on the SUT, continue the test; else lower the test device + traffic rate until the counts are the same. + + 5) Record the packets timestamps at the transmitter and receiver + ends of the test device for all VCC's. + + Reporting Format: + + The results of the CTD/Mixed Load/Three VCC test SHOULD be + reported in a form of text, graph, and histogram. + + The text results SHOULD display the numerical values of the CTD. + The values given SHOULD include: time period of test in s, test + VPI/VCI value, total number of cells transmitted and received on + the given VPI/VCI during the test in positive integers, minimum, + maximum, and mean CTD during the test in us. + + The graph results SHOULD display the cell transfer delay values. + The x-coordinate SHOULD be the test run time in either seconds, + minutes or days depending on the total length of the test. The + x-coordinate time SHOULD be configurable. The y-coordinate SHOULD + be the cell transfer delay in us. The integration time per point + MUST be indicated. + + + + + + + +Dunn & Martin Informational [Page 89] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + The histogram results SHOULD display the cell transfer delay. The + x-coordinate SHOULD be the cell transfer delay in us with at least + 256 bins. The y-coordinate SHOULD be the number of cells observed + in each bin. + + The results MUST also indicate the packet size in octets, traffic + rate in packets per second, and bearer class as generated by the + test device. The VCC and VPI/VCI values MUST be indicated. The + PCR, SCR, and MBS MUST be indicated. The bearer class of the + created VCC MUST also be indicated. + +3.2.6.12. CTD/Mixed Load/Twelve VCCs + + Objective: To determine the SUT variation in cell transfer delay with + twelve VCCs as defined in RFC 2761 "Terminology for ATM + Benchmarking". + + Procedure: + + 1) Set up the SUT and test device using the bi-directional + configuration. + + 2) Configure the SUT and test device with twelve VCC's. Each VCC + MUST be defined as one of the Bearer classes for a total of four + CBR, four UBR and four VBR VCC's. Each VCC SHOULD contain one + VPI/VCI. The VPI/VCI MUST not be one of the reserved ATM + signaling channels (e.g., [0,5], [0,16]). + + 3) Send a specific number of IP packets containing timestamps + through the SUT via the defined test VCCs. Each generated VCC + stream MUST match the corresponding VCC Bearer class. All of the + VPI/VCI pairs will generate traffic at the same traffic rate. + Since this test is not a throughput test, the rate should not be + greater than 90% of line rate. The IP PDUs MUST be encapsulated + in AAL5. + + 4) Count the IP packets that are transmitted by the SUT on all VCCs + to verify connectivity and load. If the count on the test device + is the same on the SUT, continue the test; else lower the test + device traffic rate until the counts are the same. + + 5) Record the packets timestamps at the transmitter and receiver + ends of the test device for all VCCs. + + Reporting Format: + + The results of the CTD/Mixed Load/Twelve VCCs test SHOULD be + reported in a form of text, graph, and histograms. + + + +Dunn & Martin Informational [Page 90] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + The text results SHOULD display the numerical values of the CTD. + The values given SHOULD include: time period of test in s, test + VPI/VCI values, total number of cells transmitted and received on + each VCC during the test in positive integers, maximum and minimum + CTD on each VCC during the test in us, and mean CTD on each VCC in + us. + + The graph results SHOULD display the cell transfer delay values. + The x-coordinate SHOULD be the test run time in either seconds, + minutes or days depending on the total length of the test. The + x-coordinate time SHOULD be configurable. The y-coordinate SHOULD + be the cell transfer delay for each VCC in ms. There SHOULD be 12 + curves on the graph, one curves indicated and labeled for each + VCC. The integration time per point MUST be indicated. + + The histograms SHOULD display the cell transfer delay. There will + be one histogram for each VCC. The x-coordinate SHOULD be the + cell transfer delay in us with at least 256 bins. The y- + coordinate SHOULD be the number of cells observed in each bin. + + The results MUST also indicate the packet size in octets, traffic + rate in packets per second, and bearer class as generated by the + test device. The VCC and VPI/VCI values MUST be indicated. The + PCR, SCR, and MBS MUST be indicated. The bearer class of the + created VCC MUST also be indicated. + +3.2.6.13. CTD/Mixed Load/Maximum VCCs + + Objective: To determine the SUT variation in cell transfer delay with + the maximum number VCCs supported on the SUT as defined in RFC 2761 + "Terminology for ATM Benchmarking". + + Procedure: + + 1) Set up the SUT and test device using the bi-directional + configuration. + + 2) Configure the SUT and test device with maximum number of VCCs + supported on the SUT. For example, if the maximum number of VCCs + supported on the SUT is 1024, define 256 VPIs with 4 VCIs per + VPI. Each VCC MUST be defined as one of the Bearer classes for a + total of (max VCC/3) CBR, (max VCC/3) UBR and (max VCC/3) VBR + VCC's. If the maximum number of VCC's is not divisible by 3, the + total for each bearer class MUST be within 3 VCC's of each other. + The VPI/VCI MUST not be one of the reserved ATM signaling + channels (e.g., [0,5], [0,16]). + + + + + +Dunn & Martin Informational [Page 91] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + 3) Send a specific number of IP packets containing timestamps + through the SUT via the defined test VCCs. Each generated VCC + stream MUST match the corresponding VCC Bearer class. All of the + VPI/VCI pairs will generate traffic at the same traffic rate. + Since this test is not a throughput test, the rate should not be + greater than 90% of line rate. The IP PDUs MUST be encapsulated + in AAL5. + + 4) Count the IP packets that are transmitted by the SUT on all VCCs + to verify connectivity and load. If the count on the test device + is the same on the SUT, continue the test; else lower the test + device traffic rate until the counts are the same. + + 5) Record the packets timestamps at the transmitter and receiver + ends of the test device for all VCCs. + + Reporting Format: + + The results of the CTD/Mixed Load/Maximum VCCs test SHOULD be + reported in a form of text, graphs, and histograms. + + The text results SHOULD display the numerical values of the CTD. + The values given SHOULD include: time period of test in s, test + VPI/VCI values, total number of cells transmitted and received on + each VCC during the test in positive integers, maximum and minimum + CTD on each VCC during the test in us, and mean CTD on each VCC in + us. + + The graph results SHOULD display the cell transfer delay values. + There will be (Max number of VCCs/10) graphs, with 10 VCCs + indicated on each graph. The x-coordinate SHOULD be the test run + time in either seconds, minutes or days depending on the total + length of the test. The x-coordinate time SHOULD be configurable. + The y-coordinate SHOULD be the cell transfer delay for each VCC in + us. There SHOULD be no more than 10 curves on each graph, one + curve indicated and labeled for each VCC. The integration time + per point MUST be indicated. + + The histograms SHOULD display the cell transfer delay. There will + be one histogram for each VCC. The x-coordinate SHOULD be the + cell transfer delay in us with at least 256 bins. The y- + coordinate SHOULD be the number of cells observed in each bin. + + The results MUST also indicate the packet size in octets, traffic + rate in packets per second, and bearer class as generated by the + test device. The VCC and VPI/VCI values MUST be indicated. The + PCR, SCR, and MBS MUST be indicated. The bearer class of the + created VCC MUST also be indicated. + + + +Dunn & Martin Informational [Page 92] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + +3.3. ATM Adaptation Layer (AAL) Type 5 (AAL5) + +3.3.1. IP Packet Loss due to AAL5 Re-assembly Errors + + Objective: To determine if the SUT will drop IP packets due AAL5 Re- + assembly Errors as defined in RFC 2761 "Terminology for ATM + Benchmarking". + + Procedure: + + 1) Set up the SUT and test device using the uni-directional + configuration. + + 2) Send a specific number of cells at a specific rate through the + SUT. Since this test is not a throughput test, the rate should + not be greater than 90% of line rate. The cell payload SHOULD + contain valid IP PDUs. The IP PDUs MUST be encapsulated in AAL5. + + 3) Count the cells that are transmitted by the SUT to verify + connectivity and load. If the count on the test device is the + same on the SUT, continue the test; else lower the test device + traffic rate until the counts are the same. + + 4) Inject one error in the first bit of the AAL5 payload. Verify + that the SUT does not drop any AAL5 PDU's. + + 5) Discontinue the AAL5 payload error. + + 6) Inject one error in the first bit of the AAL5 header for 4 + consecutive IP PDUs in every 6 IP PDUs. Verify that the SUT does + drop the AAL5 PDU's. + + 7) Discontinue the AAL5 payload error. + + Reporting Format: + + The results of the AAL5 PDU Loss due to AAL5 PDU errors test + SHOULD be reported in a form of a table. The rows SHOULD be + labeled single error, one error per second, and four consecutive + errors every 6 IP PDUs. The columns SHOULD be labeled AAL5 PDU + loss and number of PDU's lost. The elements of column 1 SHOULD be + either True or False, indicating whether the particular condition + was observed for each test. The elements of column 2 SHOULD be + non-negative integers. + + The table MUST also indicate the traffic rate in IP PDUs per + second as generated by the test device. + + + + +Dunn & Martin Informational [Page 93] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + +3.3.2. AAL5 Reassembly Time. + + Objective: To determine the SUT AAL5 Reassembly Time as defined in + RFC 2761 "Terminology for ATM Benchmarking". + + Procedure: + + 1) Set up the SUT and test device using the uni-directional + configuration. + + 2) Send a specific number of IP packets at a specific rate through + the SUT. Since this test is not a throughput test, the rate + should not be greater than 90% of line rate. The IP PDUs MUST be + encapsulated in AAL5. The AAL5 PDU size is 65535 octets or 1365 + ATM cells. + + 3) Count the IP packets that are transmitted by the SUT to verify + connectivity and load. If the count on the test device is the + same on the SUT, continue the test; else lower the test device + traffic rate until the counts are the same. + + 4) Given an AAL5 reassembly timer of 'x' seconds, where 'x' is the + actual value of the AAL5 reassembly timer on the SUT, sent + traffic at 1365 cells per 'x' seconds. The expected results are + that no AAL5 PDU's will be dropped. + + 5) Send traffic at 1360 cells per 'x' seconds. The expected results + are that all AAL5 PDU's will be dropped. + + Reporting Format: + + The results of the IP packet loss due to AAL5 reassembly timeout + test SHOULD be reported in a form of a table. The rows SHOULD be + labeled 1365 cells per 'x' seconds and 1360 cells per 'x' seconds. + The columns SHOULD be labeled packet loss and number of packets + lost. The elements of column 1 SHOULD be either True or False, + indicating whether the particular condition was observed for each + test. The elements of column 2 SHOULD be non-negative integers. + + The table MUST also indicate the packet size in octets and traffic + rate in packets per second as generated by the test device, + including the value of + + + + + + + + + +Dunn & Martin Informational [Page 94] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + +3.3.3. AAL5 CRC Error Ratio. + +3.3.3.1. Test Setup + + The AAL5 CRC error ratio measurements assume that both the + transmitter and receiver payload information is synchronized. + Synchronization MUST be achieved by supplying a known bit pattern to + both the transmitter and receiver. If this bit pattern is longer + than the packet size, the receiver MUST synchronize with the + transmitter before tests can be run. + +3.3.3.2. AAL5-CRC-ER/Steady Load/One VCC + + Objective: To determine the SUT ratio of AAL5 CRC PDU errors on one + VCC in a transmission in relation to the total AAL5 PDU's sent as + defined in RFC 2761 "Terminology for ATM Benchmarking". + + Procedure: + + 1) Set up the SUT and test device using the bi-directional + configuration. + + 2) Configure the SUT and test device with one VCC. The VCC SHOULD + contain one VPI/VCI. The VCC MUST be configured as either a CBR, + VBR, or UBR connection. The VPI/VCI MUST not be one of the + reserved ATM signaling channels (e.g., [0,5], [0,16]). + + 3) Send a specific number of IP packets containing one of the + specified bit patterns at a constant rate through the SUT via the + defined test VCC. Since this test is not a throughput test, the + rate should not be greater than 90% of line rate. The IP PDUs + MUST be encapsulated in AAL5. + + 4) Count the IP packets that are transmitted by the SUT to verify + connectivity and load. If the count on the test device is the + same on the SUT, continue the test; else lower the test device + traffic rate until the counts are the same. + + 5) Record the number of AAL5 CRC errors at the receiver end of the + test device. + + Reporting Format: + + The results of the AAL5-CRC-ER/Steady Load/One VCC test SHOULD be + reported in a form of text and graph. + + + + + + +Dunn & Martin Informational [Page 95] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + The text results SHOULD display the numerical values of the AAL5- + CRC-ER. The values given SHOULD include: time period of test in + s, test VPI/VCI value, total number of AAL5 PDU's transmitted and + received on the given VPI/VCI during the test in positive + integers, and the AAL5-CRC-ER for the entire test. + + The graph results SHOULD display the AAL5 CRC error ratio values. + The x-coordinate SHOULD be the test run time in either seconds, + minutes or days depending on the total length of the test. The + x-coordinate time SHOULD be configurable. The y-coordinate SHOULD + be the AAL5-CRC-ER. The integration time per point MUST be + indicated. + + The results MUST also indicate the packet size in octets, traffic + rate in packets per second, and bearer class as generated by the + test device. The VCC and VPI/VCI values MUST be indicated. The + PCR, SCR, and MBS MUST be indicated. The bearer class of the + created VCC MUST be indicated. The generated bit pattern MUST + also be indicated. + +3.3.3.3. AAL5-CRC-ER/Steady Load/Twelve VCCs + + Objective: To determine the SUT ratio of AAL5 CRC PDU errors on + twelve VCC's in a transmission in relation to the total AAL5 PDU's + sent as defined in RFC 2761 "Terminology for ATM Benchmarking". + + Procedure: + + 1) Set up the SUT and test device using the bi-directional + configuration. + + 2) Configure the SUT and test device with twelve VCCs, using 1 VPI + and 12 VCIs. The VCC's MUST be configured as either a CBR, VBR, + or UBR connection. The VPI/VCIs MUST not be one of the reserved + ATM signaling channels (e.g., [0,5], [0,16]). + + 3) Send a specific number of IP packets containing one of the + specified bit patterns at a constant rate through the SUT via the + defined test VCCs. All of the VPI/VCI pairs will generate + traffic at the same traffic rate. + + Since this test is not a throughput test, the rate should not be + greater than 90% of line rate. The IP PDUs MUST be encapsulated + in AAL5. + + + + + + + +Dunn & Martin Informational [Page 96] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + 4) Count the IP packets that are transmitted by the SUT on all VCCs + to verify connectivity and load. If the count on the test device + is the same on the SUT, continue the test; else lower the test + device traffic rate until the counts are the same. + + 5) Record the number of AAL5 CRC errors at the receiver end of the + test device for all VCCs. + + Reporting Format: + + The results of the AAL5-CRC-ER/Steady Load/Twelve VCCs test SHOULD + be reported in a form of text and graph. + + The text results SHOULD display the numerical values of the AAL5- + CRC-ER. The values given SHOULD include: time period of test in + s, test VPI/VCI value, total number of AAL5 PDU's transmitted and + received on the given VPI/VCI during the test in positive + integers, and the AAL5-CRC-ER for the entire test. + + The graph results SHOULD display the AAL5 CRC error ratio values. + The x-coordinate SHOULD be the test run time in either seconds, + minutes or days depending on the total length of the test. The + x-coordinate time SHOULD be configurable. The y-coordinate SHOULD + be the AAL5-CRC-ER for each VCC. There should be 12 curves on the + graph, on curve indicated and labeled for each VCC. The + integration time per point MUST be indicated. + + The results MUST also indicate the packet size in octets, traffic + rate in packets per second, and bearer class as generated by the + test device. The VCC and VPI/VCI values MUST be indicated. The + PCR, SCR, and MBS MUST be indicated. The bearer class of the + created VCC MUST be indicated. The generated bit pattern MUST + also be indicated. + +3.3.3.4. AAL5-CRC-ER/Steady Load/Maximum VCCs + + Objective: To determine the SUT ratio of AAL5 CRC PDU errors with the + maximum number VCCs supported on the SUT in a transmission in + relation to the total AAL5 PDU's sent as defined in RFC 2761 + "Terminology for ATM Benchmarking". + + Procedure: + + 1) Set up the SUT and test device using the bi-directional + configuration. + + 2) Configure the SUT and test device with the maximum number of VCCs + supported on the SUT. For example, if the maximum number of VCCs + + + +Dunn & Martin Informational [Page 97] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + supported on the SUT is 1024, define 256 VPIs with 4 VCIs per + VPI. The VCC's MUST be configured as either a CBR, VBR, or UBR + connection. The VPI/VCIs MUST not be one of the reserved ATM + signaling channels (e.g., [0,5], [0,16]). + + 3) Send a specific number of IP packets containing one of the + specified bit patterns at a constant rate through the SUT via the + defined test VCCs. All of the VPI/VCI pairs will generate + traffic at the same traffic rate. Since this test is not a + throughput test, the rate should not be greater than 90% of line + rate. The IP PDUs MUST be encapsulated in AAL5. + + 4) Count the IP packets that are transmitted by the SUT on all VCCs + to verify connectivity and load. If the count on the test device + is the same on the SUT, continue the test; else lower the test + device traffic rate until the counts are the same. + + 5) Record the number of AAL5 CRC errors at the receiver end of the + test device for all VCCs. + + Reporting Format: + + The results of the AAL5-CRC-ER/Steady Load/Maximum VCCs test + SHOULD be reported in a form of text and graph. + + The text results SHOULD display the numerical values of the AAL5- + CRC-ER. The values given SHOULD include: time period of test in + s, test VPI/VCI value, total number of AAL5 PDU's transmitted and + received on the given VPI/VCI during the test in positive + integers, and the AAL5-CRC-ER for the entire test. + + The graph results SHOULD display the AAL5 CRC error ratio values. + There will be (Max number of VCCs/10) graphs, with 10 VCCs + indicated on each graph. The x-coordinate SHOULD be the test run + time in either seconds, minutes or days depending on the total + length of the test. The x-coordinate time SHOULD be configurable. + The y-coordinate SHOULD be the AAL5-CRC-ER for each VCC. There + SHOULD be no more than 10 curves on each graph, one curve + indicated and labeled for each VCC. The integration time per + point MUST be indicated. + + The results MUST also indicate the packet size in octets, traffic + rate in packets per second, and bearer class as generated by the + test device. The VCC and VPI/VCI values MUST be indicated. The + PCR, SCR, and MBS MUST be indicated. The bearer class of the + created VCC MUST be indicated. The generated bit pattern MUST + also be indicated. + + + + +Dunn & Martin Informational [Page 98] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + +3.3.3.5. AAL5-CRC-ER/Bursty VBR Load/One VCC + + Objective: To determine the SUT ratio of AAL5 CRC PDU errors on one + VCC in a transmission in relation to the total AAL5 PDU's sent as + defined in RFC 2761 "Terminology for ATM Benchmarking". + + Procedure: + + 1) Set up the SUT and test device using the bi-directional + configuration. + + 2) Configure the SUT and test device with one VCC. The VCC SHOULD + contain one VPI/VCI. The VCC MUST be configured as either a CBR + or VBR connection. The VPI/VCI MUST not be one of the reserved + ATM signaling channels (e.g., [0,5], [0,16]). The PCR, SCR, and + MBS must be configured using one of the specified traffic + descriptors. + + 3) Send a specific number of IP packets containing one of the + specified bit patterns at a specific VBR rate through the SUT via + the defined test VCC. Since this test is not a throughput test, + the rate should not be greater than 90% of line rate. The IP + PDUs MUST be encapsulated in AAL5. + + 4) Count the IP packets that are transmitted by the SUT to verify + connectivity and load. If the count on the test device is the + same on the SUT, continue the test; else lower the test device + traffic rate until the counts are the same. + + 5) Record the number of AAL5 CRC errors at the receiver end of the + test device. + + Reporting Format: + + The results of the AAL5-CRC-ER/Bursty VBR Load/One VCC test SHOULD + be reported in a form of text and graph. + + The text results SHOULD display the numerical values of the AAL5- + CRC-ER. The values given SHOULD include: time period of test in + s, test VPI/VCI value, total number of AAL5 PDU's transmitted and + received on the given VPI/VCI during the test in positive + integers, and the AAL5-CRC-ER for the entire test. + + The graph results SHOULD display the AAL5 CRC error ratio values. + The x-coordinate SHOULD be the test run time in either seconds, + minutes or days depending on the total length of the test. The + + + + + +Dunn & Martin Informational [Page 99] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + x-coordinate time SHOULD be configurable. The y-coordinate SHOULD + be the AAL5-CRC-ER. The integration time per point MUST be + indicated. + + The results MUST also indicate the packet size in octets, traffic + rate in packets per second, and bearer class as generated by the + test device. The VCC and VPI/VCI values MUST be indicated. The + PCR, SCR, and MBS MUST be indicated. The bearer class of the + created VCC MUST be indicated. The generated bit pattern MUST + also be indicated. + +3.3.3.6. AAL5-CRC-ER/Bursty VBR Load/Twelve VCCs + + Objective: To determine the SUT ratio of AAL5 CRC PDU errors on + twelve VCC's in a transmission in relation to the total AAL5 PDU's + sent as defined in RFC 2761 "Terminology for ATM Benchmarking". + + Procedure: + + 1) Set up the SUT and test device using the bi-directional + configuration. + + 2) Configure the SUT and test device with twelve VCCs, using 1 VPI + and 12 VCIs. The VCC's MUST be configured as either a CBR or VBR + connection. The VPI/VCIs MUST not be one of the reserved ATM + signaling channels (e.g., [0,5], [0,16]). The PCR, SCR, and MBS + must be configured using one of the specified traffic + descriptors. + + 3) Send a specific number of IP packets containing one of the + specified bit patterns at a specific VBR rate through the SUT via + the defined test VCCs. All of the VPI/VCI pairs will generate + traffic at the same traffic rate. Since this test is not a + throughput test, the rate should not be greater than 90% of line + rate. The PCR, SCR, and MBS must be indicated. The IP PDUs MUST + be encapsulated in AAL5. + + 4) Count the IP packets that are transmitted by the SUT on all VCCs + to verify connectivity and load. If the count on the test device + is the same on the SUT, continue the test; else lower the test + device traffic rate until the counts are the same. + + 5) Record the number of AAL5 CRC errors at the receiver end of the + test device for all VCCs. + + + + + + + +Dunn & Martin Informational [Page 100] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + Reporting Format: + + The results of the AAL5-CRC-ER/Bursty VBR Load/Twelve VCCs test + SHOULD be reported in a form of text and graph. + + The text results SHOULD display the numerical values of the AAL5- + CRC-ER. The values given SHOULD include: time period of test in + s, test VPI/VCI value, total number of AAL5 PDU's transmitted and + received on the given VPI/VCI during the test in positive + integers, and the AAL5-CRC-ER for the entire test. + + The graph results SHOULD display the AAL5 CRC error ratio values. + The x-coordinate SHOULD be the test run time in either seconds, + minutes or days depending on the total length of the test. The + x-coordinate time SHOULD be configurable. The y-coordinate SHOULD + be the AAL5-CRC-ER for each VCC. There should be 12 curves on the + graph, on curve indicated and labeled for each VCC. The + integration time per point MUST be indicated. + + The results MUST also indicate the packet size in octets, traffic + rate in packets per second, and bearer class as generated by the + test device. The VCC and VPI/VCI values MUST be indicated. The + PCR, SCR, and MBS MUST be indicated. The bearer class of the + created VCC MUST be indicated. The generated bit pattern MUST + also be indicated. + +3.3.3.7. AAL5-CRC-ER/Bursty VBR Load/Maximum VCCs + + Objective: To determine the SUT ratio of AAL5 CRC PDU errors with the + maximum number VCCs supported on the SUT in a transmission in + relation to the total AAL5 PDU's sent as defined in RFC 2761 + "Terminology for ATM Benchmarking". + + Procedure: + + 1) Set up the SUT and test device using the bi-directional + configuration. + + 2) Configure the SUT and test device with the maximum number of VCCs + supported on the SUT. For example, if the maximum number of VCCs + supported on the SUT is 1024, define 256 VPIs with 4 VCIs per + VPI. The VCC's MUST be configured as either a CBR or VBR + connection. The VPI/VCIs MUST not be one of the reserved ATM + signaling channels (e.g., [0,5], [0,16]). The PCR, SCR, and MBS + must be configured using one of the specified traffic + descriptors. + + + + + +Dunn & Martin Informational [Page 101] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + 3) Send a specific number of IP packets containing one of the + specified bit patterns at a specific VBR rate through the SUT via + the defined test VCCs. All of the VPI/VCI pairs will generate + traffic at the same traffic rate. Since this test is not a + throughput test, the rate should not be greater than 90% of line + rate. The IP PDUs MUST be encapsulated in AAL5. + + 4) Count the IP packets that are transmitted by the SUT on all VCCs + to verify connectivity and load. If the count on the test device + is the same on the SUT, continue the test; else lower the test + device traffic rate until the counts are the same. + + 5) Record the number of AAL5 CRC errors at the receiver end of the + test device for all VCCs. + + Reporting Format: + + The results of the AAL5-CRC-ER/Bursty VBR Load/Maximum VCCs test + SHOULD be reported in a form of text and graph. + + The text results SHOULD display the numerical values of the AAL5- + CRC-ER. The values given SHOULD include: time period of test in + s, test VPI/VCI value, total number of AAL5 PDU's transmitted and + received on the given VPI/VCI during the test in positive + integers, and the AAL5-CRC-ER for the entire test. + + The graph results SHOULD display the AAL5 CRC error ratio values. + There will be (Max number of VCCs/10) graphs, with 10 VCCs + indicated on each graph. The x-coordinate SHOULD be the test run + time in either seconds, minutes or days depending on the total + length of the test. The x-coordinate time SHOULD be configurable. + The y-coordinate SHOULD be the AAL5-CRC-ER for each VCC. There + SHOULD be no more than 10 curves on each graph, one curve + indicated and labeled for each VCC. The integration time per + point MUST be indicated. + + The results MUST also indicate the packet size in octets, traffic + rate in packets per second, and bearer class as generated by the + test device. The VCC and VPI/VCI values MUST be indicated. The + PCR, SCR, and MBS MUST be indicated. The bearer class of the + created VCC MUST be indicated. The generated bit pattern MUST + also be indicated. + +3.3.3.8. AAL5-CRC-ER/Mixed Load/Three VCC's + + Objective: To determine the SUT ratio of AAL5 CRC PDU errors on three + VCC's in a transmission in relation to the total AAL5 PDU's sent as + defined in RFC 2761 "Terminology for ATM Benchmarking". + + + +Dunn & Martin Informational [Page 102] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + Procedure: + + 1) Set up the SUT and test device using the bi-directional + configuration. + + 2) Configure the SUT and test device with three VCC's. Each VCC + MUST be defined as a different Bearer class; one CBR, one UBR and + one VBR. Each VCC SHOULD contain one VPI/VCI. The VPI/VCI MUST + not be one of the reserved ATM signaling channels (e.g., [0,5], + [0,16]). + + 3) Send a specific number of IP packets containing one of the + specified bit patterns through the SUT via the defined test VCCs. + Each generated VCC stream MUST match the corresponding VCC Bearer + class. All of the VPI/VCI pairs will generate traffic at the + same traffic rate. Since this test is not a throughput test, the + rate should not be greater than 90% of line rate. The IP PDUs + MUST be encapsulated in AAL5. + + 4) Count the IP packets that are transmitted by the SUT to verify + connectivity and load. If the count on the test device is the + same on the SUT, continue the test; else lower the test device + traffic rate until the counts are the same. + + 5) Record the number of AAL5 CRC errors at the receiver end of the + test device for all VCCs. + + Reporting Format: + + The results of the AAL5-CRC-ER/Bursty Mixed Load/Three VCCs test + SHOULD be reported in a form of text and graph. + + The text results SHOULD display the numerical values of the AAL5- + CRC-ER. The values given SHOULD include: time period of test in + s, test VPI/VCI value, total number of AAL5 PDU's transmitted and + received on the given VPI/VCI during the test in positive + integers, and the AAL5-CRC-ER for the entire test. + + The graph results SHOULD display the AAL5 CRC error ratio values. + The x-coordinate SHOULD be the test run time in either seconds, + minutes or days depending on the total length of the test. The + x-coordinate time SHOULD be configurable. The y-coordinate SHOULD + be the AAL5-CRC-ER for each VCC. There should be 12 curves on the + graph, on curve indicated and labeled for each VCC. The + integration time per point MUST be indicated. + + The results MUST also indicate the packet size in octets, traffic + rate in packets per second, and bearer class as generated by the + + + +Dunn & Martin Informational [Page 103] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + test device. The VCC and VPI/VCI values MUST be indicated. The + PCR, SCR, and MBS MUST be indicated. The bearer class of the + created VCC MUST be indicated. The generated bit pattern MUST + also be indicated. + +3.3.3.9. AAL5-CRC-ER/Mixed Load/Twelve VCCs + + Objective: To determine the SUT ratio of AAL5 CRC PDU errors on + twelve VCC's in a transmission in relation to the total AAL5 PDU's + sent as defined in RFC 2761 "Terminology for ATM Benchmarking". + + Procedure: + + 1) Set up the SUT and test device using the bi-directional + configuration. + + 2) Configure the SUT and test device with twelve VCC's. Each VCC + MUST be defined as one of the Bearer classes for a total of four + CBR, four UBR and four VBR VCC's. Each VCC SHOULD contain one + VPI/VCI. The VPI/VCI MUST not be one of the reserved ATM + signaling channels (e.g., [0,5], [0,16]). + + 3) Send a specific number of IP packets containing one of the + specified bit patterns through the SUT via the defined test VCCs. + Each generated VCC stream MUST match the corresponding VCC Bearer + class. All of the VPI/VCI pairs will generate traffic at the + same traffic rate. Since this test is not a throughput test, the + rate should not be greater than 90% of line rate. The IP PDUs + MUST be encapsulated in AAL5. + + 4) Count the IP packets that are transmitted by the SUT on all VCCs + to verify connectivity and load. If the count on the test device + is the same on the SUT, continue the test; else lower the test + device traffic rate until the counts are the same. + + 5) Record the number of AAL5 CRC errors at the receiver end of the + test device for all VCCs. + + Reporting Format: + + The results of the AAL5-CRC-ER/Bursty Mixed Load/Twelve VCCs test + SHOULD be reported in a form of text and graph. + + The text results SHOULD display the numerical values of the AAL5- + CRC-ER. The values given SHOULD include: time period of test in + s, test VPI/VCI value, total number of AAL5 PDU's transmitted and + received on the given VPI/VCI during the test in positive + integers, and the AAL5-CRC-ER for the entire test. + + + +Dunn & Martin Informational [Page 104] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + The graph results SHOULD display the AAL5 CRC error ratio values. + The x-coordinate SHOULD be the test run time in either seconds, + minutes or days depending on the total length of the test. The + x-coordinate time SHOULD be configurable. The y-coordinate SHOULD + be the AAL5-CRC-ER for each VCC. There should be 12 curves on the + graph, on curve indicated and labeled for each VCC. The + integration time per point MUST be indicated. + + The results MUST also indicate the packet size in octets, traffic + rate in packets per second, and bearer class as generated by the + test device. The VCC and VPI/VCI values MUST be indicated. The + PCR, SCR, and MBS MUST be indicated. The bearer class of the + created VCC MUST be indicated. The generated bit pattern MUST + also be indicated. + +3.3.3.10. AAL5-CRC-ER/Mixed Load/Maximum VCCs + + Objective: To determine the SUT ratio of AAL5 CRC PDU errors with the + maximum number VCCs supported on the SUT in a transmission in + relation to the total AAL5 PDU's sent as defined in RFC 2761 + "Terminology for ATM Benchmarking". + + Procedure: + + 1) Set up the SUT and test device using the bi-directional + configuration. + + 2) Configure the SUT and test device with maximum number of VCCs + supported on the SUT. For example, if the maximum number of VCCs + supported on the SUT is 1024, define 256 VPIs with 4 VCIs per + VPI. Each VCC MUST be defined as one of the Bearer classes for a + total of (max VCC/3) CBR, (max VCC/3) UBR and (max VCC/3) VBR + VCC's. The VPI/VCI MUST not be one of the reserved ATM signaling + channels (e.g., [0,5], [0,16]). + + 3) Send a specific number of IP packets containing one of the + specified bit patterns through the SUT via the defined test VCCs. + Each generated VCC stream MUST match the corresponding VCC Bearer + class. All of the VPI/VCI pairs will generate traffic at the + same traffic rate. Since this test is not a throughput test, the + rate should not be greater than 90% of line rate. The IP PDUs + MUST be encapsulated in AAL5. + + 4) Count the IP packets that are transmitted by the SUT on all VCCs + to verify connectivity and load. If the count on the test device + is the same on the SUT, continue the test; else lower the test + device traffic rate until the counts are the same. + + + + +Dunn & Martin Informational [Page 105] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + 5) Record the number of AAL5 CRC errors at the receiver end of the + test device for all VCCs. + + Reporting Format: + + The results of the AAL5-CRC-ER/Bursty Mixed Load/Maximum VCCs test + SHOULD be reported in a form of text and graph. + + The text results SHOULD display the numerical values of the AAL5- + CRC-ER. The values given SHOULD include: time period of test in + s, test VPI/VCI value, total number of AAL5 PDU's transmitted and + received on the given VPI/VCI during the test in positive + integers, and the AAL5-CRC-ER for the entire test. + + The graph results SHOULD display the AAL5 CRC error ratio values. + There will be (Max number of VCCs/10) graphs, with 10 VCCs + indicated on each graph. The x-coordinate SHOULD be the test run + time in either seconds, minutes or days depending on the total + length of the test. The x-coordinate time SHOULD be configurable. + The y-coordinate SHOULD be the AAL5-CRC-ER for each VCC. There + SHOULD be no more than 10 curves on each graph, one curve + indicated and labeled for each VCC. The integration time per + point MUST be indicated. + + The results MUST also indicate the packet size in octets, traffic + rate in packets per second, and bearer class as generated by the + test device. The VCC and VPI/VCI values MUST be indicated. The + PCR, SCR, and MBS MUST be indicated. The bearer class of the + created VCC MUST be indicated. The generated bit pattern MUST + also be indicated. + +3.4. ATM Service: Signaling + +3.4.1. CAC Denial Time and Connection Establishment Time + + Objective: To determine the CAC rejection time and Connection + Establishment Time on the SUT as defined in RFC 2761 "Terminology for + ATM Benchmarking". + + Procedure: + + 1) Set up the SUT and test device using the bi-directional + configuration. + + 2) Create a UNI signaling setup message, as described in Appendix C, + specifying a PCR which will not allow CAC to reject the call. + + + + + +Dunn & Martin Informational [Page 106] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + 3) Send the UNI signaling setup message. Note the time the setup + message was sent. Verify that the SVC has been setup with the + correct parameters. Note the time the connect message was + received + + 4) Create a UNI signaling setup message, as described in Appendix C, + specifying a PCR which will allow CAC to reject the call. + + 5) Send the UNI signaling setup message. Note the time the setup + message was sent. Verify that the SVC has been rejected with the + correct cause code. Note the time the release complete message + was received. + + 6) Compute the rejection time as the difference between the time the + release complete message was received and the time setup message + was send. + + Reporting Format: + + The results of the CAC Denial Time and Connection Establishment + Time tests SHOULD be reported in a form of a table. The rows + SHOULD be labeled call accepted and call rejected. The columns + SHOULD be labeled time setup sent, time response received, and + correct response. The elements of the columns 1 and 2 SHOULD be + in seconds. The elements of column 3 SHOULD be be either True or + False, indicating whether the particular condition was observed + for each test. + + The table MUST also indicate the packet size in octets and traffic + rate in packets per second as generated by the test device. + +3.4.2. Connection Teardown Time + + Objective: To determine the Connection Teardown Time on the SUT as + defined in RFC 2761 "Terminology for ATM Benchmarking". + + Procedure: + + 1) Set up the SUT and test device using the bi-directional + configuration. + + 2) Create a UNI signaling setup message, as described in Appendix C, + specifying a PCR which will not allow CAC to reject the call. + + 3) Send the UNI signaling setup message. Note the time the setup + message was sent. Verify that the SVC has been setup with the + correct parameters. Note the time the connect message was + received + + + +Dunn & Martin Informational [Page 107] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + 4) Create a UNI signaling release message, as described in Appendix + C, specifying a cause code of normal call clearing. + + 5) Send the UNI signaling release message. Note the time the + release message was sent. Verify that the SVC has been + terminated with the correct cause code. Note the time the + release complete message was received. + + 6) Compute the release time as the difference between the time the + release complete message was received and the time release + message was send. + + Reporting Format: + + The results of the Connection Teardown Time tests SHOULD be + reported in a form of a table. The rows SHOULD be labeled call + accepted and call released. The columns SHOULD be labeled time + message sent, time response received, and correct response. The + elements of the columns 1 and 2 SHOULD be in seconds. The + elements of column 3 SHOULD be be either True or False, indicating + whether the particular condition was observed for each test. + + The table MUST also indicate the packet size in octets and traffic + rate in packets per second as generated by the test device. + +3.4.3. Crankback Time + + Objective: To determine the Crankback Time on the SUT as defined in + RFC 2761 "Terminology for ATM Benchmarking". + + Procedure: + + 1) Set up the SUT and test device using the uni-directional + passthrough configuration. + + 2) Create a PNNI signaling setup message, as described in Appendix + C, specifying a DTL which is not blocked by the far end SUT. + + 3) Send the PNNI signaling setup message. Note the time the setup + message was sent. Verify that the connect message has been + received by the near-end switch. Note the time the connect + message was received + + 4) Create a PNNI signaling setup message, as described in Appendix + C, specifying a DTL which is blocked by the far end SUT. + + 5) Send the PNNI signaling release message. Note the time the + release message was sent. Note the time the release complete + + + +Dunn & Martin Informational [Page 108] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + message was received. Note the time the near-end switch sends + it's own PNNI setup message (referred to as the near-end setup + message) specifying the non- blocked DTL. + + 6) Compute the crankback time as the difference between the time the + near-end setup message was received and the time release message + was send. + + Reporting Format: + + The results of the Crankback Time tests SHOULD be reported in a + form of a table. The rows SHOULD be labeled DTL call accepted and + call released. The columns SHOULD be labeled time message sent, + time response received, and correct response. The elements of the + columns 1 and 2 SHOULD be in seconds. The elements of column 3 + SHOULD be be either True or False, indicating whether the + particular condition was observed for each test. + + The table MUST also indicate the packet size in octets and traffic + rate in packets per second as generated by the test device. + +3.4.4. Route Update Response Time + + Objective: To determine the Route Update Response Time on the SUT as + defined in RFC 2761 "Terminology for ATM Benchmarking". + + Procedure: + + 1) Set up the SUT and test device using the uni-directional + passthrough configuration. + + 2) Create a PNNI PTSE as described in Appendix C, specifying a + routing topology. Verify that the routing tables on the far-end + and near-end switches are empty. + + 3) Send the PTSE message to the far-end switch. Note the time the + PTSE message was sent. Verify that the PTSE message has been + received by the far-end switch. Note the time the PTSE message + was received. + + 4) Create another PNNI PTSE as described in Appendix C, specifying a + change in the routing topology. Verify that the routing tables + on the far-end and near-end switches contain the previous PTSE + routes. + + 5) Send the PTSE message to the far-end switch. Note the time the + PTSE message was sent. Verify that the PTSE message has been + received by the far-end switch. Note the time the PTSE message + + + +Dunn & Martin Informational [Page 109] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + was received. Note the time the PTSE was sent to the near-end + switch. Note the time the PTSE message was received on the + near-end switch. + + 6) Compute the Route Update Response time as the difference between + the time the far-end PTSE message was sent and the time far-end + PTSE message was received by the near-end. + + Reporting Format: + + The results of the Route Update Response Time tests SHOULD be + reported in a form of a table. The rows SHOULD be labeled PTSE + call accepted, far-end PTSE message send, and near-end message + received. The columns SHOULD be labeled time message sent, time + response received, and correct response. The elements of the + columns 1 and 2 SHOULD be in seconds. The elements of column 3 + SHOULD be be either True or False, indicating whether the + particular condition was observed for each test. + + The table MUST also indicate the packet size in octets and traffic + rate in packets per second as generated by the test device. + +3.5. ATM Service: ILMI + +3.5.1. MIB Alignment Time + + Objective: To determine the MIB Alignment Time on the SUT as defined + in RFC 2761 "Terminology for ATM Benchmarking". + + Procedure: + + 1) Set up the SUT and test device using the bi-directional + configuration. + + 2) Send a Cold Start message to the SUT. Note the time the message + was sent to the SUT. Verify that the Cold Start message has been + received by the SUT. Note the time the message was received. + + 3) Send a Get Request message to the SUT. Note the time the message + was sent to the SUT. Verify that the Get Request message has + been received by the SUT. Note the time the message was + received. + + 4) After all MIB elements are exchanged, verify that the final Get + Request message has been received by the SUT. Note the time the + message was send and received by the SUT. + + + + + +Dunn & Martin Informational [Page 110] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + 5) Compute the MIB Alignment Time as the difference between the time + the Cold Start message was sent and the time the final Get + Request was received by the SUT. + + Reporting Format: + + The results of the MIB Alignment Time tests SHOULD be reported in + a form of a table. The rows SHOULD be labeled Cold Start Send, + Cold Start accepted, Final Get Request send, and Final Get Request + received. The columns SHOULD be labeled time message sent, time + response received, and correct response. The elements of the + columns 1 and 2 SHOULD be in seconds. The elements of column 3 + SHOULD be be either True or False, indicating whether the + particular condition was observed for each test. + + The table MUST also indicate the packet size in octets and traffic + rate in packets per second as generated by the test device. + +3.5.2. Address Registration Time + + Objective: To determine the Address Registration Time on the SUT as + defined in RFC 2761 "Terminology for ATM Benchmarking". + + Procedure: + + 1) Set up the SUT and test device using the bi-directional + configuration. + + 2) Send a Set Request message to the SUT. Note the time the message + was sent to the SUT. Verify that the Set Request message has + been received by the SUT. Note the time the message was + received. + + 3) Send a Get Request message to the SUT. Note the time the message + was sent to the SUT. Verify that the Get Request message has + been received by the SUT. Note the time the message was + received. + + 4) After all MIB elements are exchanged, verify that the final Get + Request message has been received by the SUT. Note the time the + message was send and received by the SUT. + + 5) Compute the Address Registration Time as the difference between + the time the Set Request message was sent and the time the final + Get Request was received by the SUT. + + + + + + +Dunn & Martin Informational [Page 111] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + Reporting Format: + + The results of the Address Registration Time tests SHOULD be + reported in a form of a table. The rows SHOULD be labeled Set + Request Send, Set Request accepted, Final Get Request send, and + Final Get Request received. The columns SHOULD be labeled time + message sent, time response received, and correct response. The + elements of the columns 1 and 2 SHOULD be in seconds. The + elements of column 3 SHOULD be be either True or False, indicating + whether the particular condition was observed for each test. + + The table MUST also indicate the packet size in octets and traffic + rate in packets per second as generated by the test device. + +4. Security Considerations + + As this document is solely for the purpose of providing methodology + and describes neither a protocol nor an implementation, there are no + security considerations associated with this document. + +5. Notices + + The IETF takes no position regarding the validity or scope of any + intellectual property or other rights that might be claimed to + pertain to the implementation or use of the technology described in + this document or the extent to which any license under such rights + might or might not be available; neither does it represent that it + has made any effort to identify any such rights. Information on the + IETFs procedures with respect to rights in standards-track and + standards-related documentation can be found in BCP-11. Copies of + claims of rights made available for publication and any assurances of + licenses to be made available, or the result of an attempt made to + obtain a general license or permission for the use of such + proprietary rights by implementors or users of this specification can + be obtained from the IETF Secretariat. + + The IETF invites any interested party to bring to its attention any + copyrights, patents or patent applications, or other proprietary + rights which may cover technology that may be required to practice + this standard. Please address the information to the IETF Executive + Director. + + + + + + + + + + +Dunn & Martin Informational [Page 112] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + +6. References + + [RFC2544] Bradner, S. and J. McQuaid, "Benchmarking Methodology + for Network Interconnect Devices", RFC 2544, March + 1999. + + [RFC2225] Laubach, M. and J. Halpern, "Classical IP and ARP over + ATM", RFC 2225, April 1998. + + [RFC2761] Dunn, J. and C. Martin, "Terminology for ATM + Benchmarking", RFC 2761, February 2000. + + [AF-ILMI4.0] ATM Forum Integrated Local Management Interface + Version 4.0, af-ilmi-0065.000, September 1996. + + [AF-TEST-0022] Introduction to ATM Forum Test Specifications, af- + test-0022.00, December 1994. + + [AF-TM4.1] ATM Forum, Traffic Management Specification Version + 4.1, af-tm-0121.00, April 1996. + + [AF-UNI3.1] ATM Forum, User Network Interface Specification + Version 3.1, September 1994. + + [AF-UNI4.0] ATM Forum, User Network Interface Specification + Version 4.0, July 1996. + +7. Authors' Addresses + + Jeffrey Dunn + Advanced Network Consultants, Inc. + 4214 Crest Place + Ellicott City, MD 21043, USA + + Phone: +1 (410) 750-1700 + EMail: Jeffrey.Dunn@worldnet.att.net + + + Cynthia Martin + Advanced Network Consultants, Inc. + 4214 Crest Place + Ellicott City, MD 21043, USA + + Phone: +1 (410) 750-1700 + EMail: Cynthia.E.Martin@worldnet.att.net + + + + + + +Dunn & Martin Informational [Page 113] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + +Appendix A: Ranges + + ATM NSAP Network Prefix. + 39 0000 0000 0000 0000 0000 0000-39 0000 0000 0000 0000 0000 00FF + 39 0000 0000 0000 0000 0001 0000-39 0000 0000 0000 0000 0001 00FF + 39 0000 0000 0000 0001 0000 0000 + 39 0000 0000 0000 0002 0020 0000 + 39 0000 0000 0300 0002 0030 0000 + 39 0000 0000 4000 0002 0060 0000 + 39 0000 0006 0060 0002 0030 0000 + 39 0000 0006 0050 0002 0030 0000 + 39 0000 0009 0300 0002 0030 0000 + 39 0000 00A0 0300 0002 0030 0000 + 39 0000 0B00 0300 0002 0030 0000 + 39 0000 C000 0300 0002 0030 0000 + + ATM NSAP End System Identifier. + 1111 1111 1111 00-1111 1111 11FF 00 + 2222 2222 2000 00-2222 2222 2222 00 + 9999 999A 0000 00-9999 999C 0000 00 + +Appendix B: Rates + + PNNI Routing Update Size. + + 1) 1 PNNI routing entry update on non-aggregated addresses + + 2) 2 PNNI routing entry updates on non-aggregated addresses + + 3) 5 PNNI routing entry updates on non-aggregated addresses + + 4) 1 % of total available bandwidth or 1 Mb/s, whichever is less on + non- aggregated addresses + + 5) 1 % of total available bandwidth or 1 Mb/s, whichever is less on + of non-aggregated addresses and of aggregated addresses + + 6) 1 % of total available bandwidth or 1 Mb/s, whichever is less on + aggregated addresses + + 7) 2 % of total available bandwidth or 2 Mb/s, whichever is less on + non- aggregated addresses + + 8) 2 % of total available bandwidth or 2 Mb/s, whichever is less on + of non-aggregated addresses and of aggregated addresses + + 9) 2 % of total available bandwidth or 2 Mb/s, whichever is less on + aggregated addresses + + + +Dunn & Martin Informational [Page 114] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + PNNI Routing Update Repetition Interval. + + Repetition Interval begins after initial PNNI routing table + stabilizes. + + 1) 1 update every 1 hour, for 24 hours + + 2) 1 update every 30 minutes, for 24 hours + + 3) 1 update every 5 minutes, for 1 hour + + 4) 1 update every 1 minute, for 15 minutes + + 5) 1 update every 30 seconds, for 5 minutes + + 6) 1 update every 30 seconds, for 1 minute + + 7) 1 update every 1 second, for 30 seconds + + Maximum WAN Connection rates in packets per second (pps): + + 25.6 OC-3c OC-12c + IP Packet Size + octets/cells + 44/2 30188 176603 706412 + 64/2 30188 176603 706412 + 128/3 20125 117735 470940 + 256/6 10062 58867 235468 + 1024/22 2744 16054 64216 + 1518/32 1886 11037 44148 + 2048/43 1404 8214 32856 + 4472/94 642 3757 15028 + 9180/192 314 1839 7356 + + Maximum LAN Connection rates in packets per second (pps): + + DS-1 DS-3 E1 E3 + IP Packet Size + octets/cells + 44/2 1811 52133 2340 40000 + 64/2 1811 52133 2340 40000 + 128/3 1207 34755 1560 26666 + 256/6 603 17377 780 13333 + 1024/22 164 4739 212 3636 + 1518/32 113 3258 146 2500 + 2048/43 84 2424 108 1860 + 4472/94 38 1109 49 851 + 9180/192 18 543 24 416 + + + +Dunn & Martin Informational [Page 115] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + Notes: 1. PDU size in cells is computed based on ceiling( ( PDU size + in octets + 16) / 48). This assumes an 8 octet LLC/SNAP header and + an 8 octet AAL/5 trailer. + + 2. Due to the number of possible configurations, IMA pps rates are + not listed, but may be derived from the following formula: floor + (IDCR/cells per packet), where cells per packet is computed as in + note 1. + + 3. The following cell rates were used: DS-1 = 3622 cps (using ATM TC) + E1 = 4681 cps 25.6 Mb/s = 60377 cps E3 = 80000 cps (using ATM TC) + DS-3 = 104266 cps (using ATM TC) OC-3c = 353207 cps OC-12c = 1412828 + cps + +Appendix C: PDU's + + TCP/IP over ATM Example 1. + LLC: DSAP 0xAA (SNAP-SAP) + SSAP 0xAA (SNAP-SAP) + Control 0x03 (Unnumbered Information) + SNAP: OUI 0x00-00-00 (Ethertype) + PID 0x0800 (Internet Protocol) + IP: Version = 4 + Header length = 20 + Type of service = 0 + 000. .... Precedence = Routine(0) + ...0 .... Delay = Normal (0) + .... 0... Throughput = Normal (0) + .... .0.. Reliability = Normal (0) + Packet length = 40 + Id = 0 + Fragmentation Info = 0x0000 + .0.. .... .... .... Don't Fragment Bit = FALSE + ..0. .... .... .... More Fragments Bit = FALSE + ...0 0000 0000 0000 Fragment offset = 0 + Time to live = 255 + Protocol = TCP (6) + Header checksum = F9CF + Source address = 15.19.209.236 + Destination address = 15.19.209.237 + TCP: Source port = smtp (25) + Destination port = smtp (25) + Sequence number = 1 + Ack number = 0 + Data offset = 20 + Flags = 0x02 + ..0. .... URGENT Flag = FALSE + ...0 .... ACK Flag = FALSE + + + +Dunn & Martin Informational [Page 116] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + .... 0... PUSH Flag = FALSE + .... .0.. RST Flag = FALSE + .... ..1. SYN Flag = TRUE + .... ...0 FIN Flag = FALSE + Window = 0 + Checksum = EDAF + Urgent pointer = 00000000 + + TCP/IP over ATM Example 2. +LLC: DSAP 0xAA (SNAP-SAP) + SSAP 0xAA (SNAP-SAP) + Control 0x03 (Unnumbered Information) + SNAP: OUI 0x00-00-00 (Ethertype) + PID 0x0800 (Internet Protocol) + IP: Version = 4 + Header length = 20 + Type of service = 0 + 000. .... Precedence = Routine(0) + ...0 .... Delay = Normal (0) + .... 0... Throughput = Normal (0) + .... .0.. Reliability = Normal (0) + Packet length = 40 + Id = 0 + Fragmentation Info = 0x0000 + .0.. .... .... .... Don't Fragment Bit = FALSE + ..0. .... .... .... More Fragments Bit = FALSE + ...0 0000 0000 0000 Fragment offset = 0 + Time to live = 255 + Protocol = TCP (6) + Header checksum = F9CF + Source address = 15.19.209.236 + Destination address = 15.19.209.237 + TCP: Source port = ftp-data (20) + Destination port = 2000 + Sequence number = 1 + Ack number = 0 + Data offset = 20 + Flags = 0x02 + ..0. .... URGENT Flag = FALSE + ...0 .... ACK Flag = FALSE + .... 0... PUSH Flag = FALSE + .... .0.. RST Flag = FALSE + .... ..1. SYN Flag = TRUE + .... ...0 FIN Flag = FALSE + Window = 0 + Checksum = E5FD + Urgent pointer = 00000000 + + + + +Dunn & Martin Informational [Page 117] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + UDP/IP over ATM Example. + LLC: DSAP 0xAA (SNAP-SAP) + SSAP 0xAA (SNAP-SAP) + Control 0x03 (Unnumbered Information) + SNAP: OUI 0x00-00-00 (Ethertype) + PID 0x0800 (Internet Protocol) + IP: Version = 4 + Header length = 20 + Type of service = 0 + 000. .... Precedence = Routine(0) + ...0 .... Delay = Normal (0) + .... 0... Throughput = Normal (0) + .... .0.. Reliability = Normal (0) + Packet length = 28 + Id = 0 + Fragmentation Info = 0x0000 + .0.. .... .... .... Don't Fragment Bit = FALSE + ..0. .... .... .... More Fragments Bit = FALSE + ...0 0000 0000 0000 Fragment offset = 0 + Time to live = 255 + Protocol = ICMP (1) + Header checksum = F9E0 + Source address = 15.19.209.236 + Destination address = 15.19.209.237 + ICMP: Type = Echo request (8) + Code = 0 + Checksum = F7FF + Identifier = 0 (0x0) + Sequence Number = 0 (0x0) + + RIP Routing Update over ATM. + + -- DATAGRAM HEADER + offset data (hex) description + 00 FF FF FF FF FF FF dest MAC address is broadcast + 06 xx xx xx xx xx xx source hardware address + 12 08 00 type + + -- IP HEADER + 14 45 IP version - 4, header length (4 + byte units) - 5 + 15 00 service field + 16 00 EE total length + 18 00 00 ID + 20 40 00 flags (3 bits) 4 (do not + fragment), + fragment offset-0 + 22 0A TTL + + + +Dunn & Martin Informational [Page 118] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + 23 11 protocol - 17 (UDP) + 24 C4 8D header checksum + 26 xx xx xx xx source IP address + 30 xx xx xx destination IP address + 33 FF host part = FF for broadcast + + -- UDP HEADER + 34 02 08 source port 208 = RIP + 36 02 08 destination port 208 = RIP + 38 00 DA UDP message length + 40 00 00 UDP checksum + + -- RIP packet + 42 02 command = response + 43 01 version = 1 + 44 00 00 0 + + -- net 1 + 46 00 02 family = IP + 48 00 00 0 + 50 xx xx xx net 1 IP address + 53 00 net not node + 54 00 00 00 00 0 + 58 00 00 00 00 0 + 62 00 00 00 07 metric 7 + + -- net 2 + + 66 00 02 family = IP + 68 00 00 0 + 70 xx xx xx net 2 IP address + 73 00 net not node + 74 00 00 00 00 0 + 78 00 00 00 00 0 + 82 00 00 00 07 metric 7 + + -- net 3 + 86 00 02 family = IP + 88 00 00 0 + 90 xx xx xx net 3 IP address + 93 00 net not node + 94 00 00 00 00 0 + 98 00 00 00 00 0 + 102 00 00 00 07 metric 7 + + -- net 4 + 106 00 02 family = IP + 108 00 00 0 + + + +Dunn & Martin Informational [Page 119] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + 110 xx xx xx net 4 IP address + 113 00 net not node + 114 00 00 00 00 0 + 118 00 00 00 00 0 + 122 00 00 00 07 metric 7 + + -- net 5 + 126 00 02 family = IP + 128 00 00 0 + 130 00 net 5 IP address + 133 00 net not node + 134 00 00 00 00 0 + 138 00 00 00 00 0 + 142 00 00 00 07 metric 7 + + -- net 6 + 146 00 02 family = IP + 148 00 00 0 + 150 xx xx xx net 6 IP address + 153 00 net not node + 154 00 00 00 00 0 + 158 00 00 00 00 0 + 162 00 00 00 07 metric 7 + + UNI 3.1 Signaling Setup Message Example. PCR will not allow CAC to + reject the call. + + Protocol Discriminator : Q.93B UNI call control + Call Reference Length : 3 + Call Reference Flag : orig + Call Reference Value : 0 + Message Type : SETUP + Ext : last octet + Action Indicator : clear call + Message Length : 50 + Information Element ID : ATM Traffic Descriptor + Ext : last octet + Coding Standard : ITU-T standard + Action Indicator : clear call + IE Length : 9 + Cell Rate Subfield ID : forward peak CR(CLP=0+1) + Forward Peak Cell Rate : 1 + Cell Rate Subfield ID : backward peak CR(CLP=0+1) + Backward Peak Cell Rate : 1 + Cell Rate Subfield ID : best effort indicator + Information Element ID : Broadband Bearer Capability + Ext : last octet + Coding Standard : ITU-T standard + + + +Dunn & Martin Informational [Page 120] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + Action Indicator : clear call + IE Length : 2 + Ext : last octet + Bearer Class : BCOB-X + Ext : last octet + Clipping Susceptibility : not susceptible to clipping + User Plane Connection CFG : point-to-point + Information Element ID : Called Party Number + Ext : last octet + Coding Standard : ITU-T standard + Action Indicator : clear call + IE Length : 21 + Ext : last octet + Addressing/Numbering Plan : ISO NSAP addressing + ISO NSAP Address Octets : 3900000000000000000000000011111111111100 + Information Element ID : Quality of Service Parameter + Ext : last octet + Coding Standard : ITU-T standard + Action Indicator : clear call + IE Length : 2 + QoS Class Forward : QoS class 0 - unspecified + QoS Class Backward : QoS class 0 - unspecified + + UNI 3.1 Signaling Setup Message Reject Example. PCR will allow + CAC to reject the call. + + Protocol Discriminator : Q.93B UNI call control + Call Reference Length : 3 + Call Reference Flag : orig + Call Reference Value : 0 + Message Type : SETUP + Ext : last octet + Action Indicator : clear call + Message Length : 50 + Information Element ID : ATM Traffic Descriptor + Ext : last octet + Coding Standard : ITU-T standard + Action Indicator : clear call + IE Length : 8 + Cell Rate Subfield ID : forward peak CR(CLP=0+1) + Forward Peak Cell Rate : 300000 + Cell Rate Subfield ID : backward peak CR(CLP=0+1) + Backward Peak Cell Rate : 300000 + Information Element ID : Broadband Bearer Capability + Ext : last octet + Coding Standard : ITU-T standard + Flag : not significant + Action Indicator : clear call + + + +Dunn & Martin Informational [Page 121] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + IE Length : 3 + Ext : another octet + Bearer Class : BCOB-X + Ext : last octet + Traffic Type : constant bit rate + Timing Requirements : end-to-end timing required + Ext : last octet + Clipping Susceptibility : not susceptible to clipping + User Plane Connection CFG : point-to-point + Information Element ID : Called Party Number + Ext : last octet + Coding Standard : ITU-T standard + Action Indicator : clear call + IE Length : 21 + Ext : last octet + Addressing/Numbering Plan : ISO NSAP addressing + ISO NSAP Address Octets : 3900000000000000000000000011111111111100 + Information Element ID : Quality of Service Parameter + Ext : last octet + Coding Standard : ITU-T standard + Action Indicator : clear call + IE Length : 2 + QoS Class Forward : QoS class 0 - unspecified + QoS Class Backward : QoS class 0 - unspecified + + UNI 3.1 Signaling Release Message, specifying a cause code of normal + call clearing. + + Protocol Discriminator : Q.93B UNI call control + Call Reference Length : 3 + Call Reference Flag : orig + Call Reference Value : 0 + Message Type : RELEASE + Ext : last octet + Action Indicator : clear call + Message Length : 6 + Information Element ID : Cause + Ext : last octet + Coding Standard : ITU-T standard + Action Indicator : clear call + IE Length : 2 + Ext : last octet + Location : user + Ext : last octet + Cause Value : NE:normal call clearing + + PNNI Signaling Setup Message, specifying a DTL which is not blocked + by the far end SUT. + + + +Dunn & Martin Informational [Page 122] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + Protocol Discriminator : PNNI signalling + Call Reference Length : 3 + Call Reference Flag : from + Message Type : SETUP + Ext : last octet + Pass Along Request : no pass along request + Action Indicator : clear call + Message Length : 56 + Information Element ID : ATM Traffic Descriptor + Ext : last octet + Coding Standard : ITU-T standardized + Pass Along Request : no pass along request + Action Indicator : clear call + IE Length : 0 + Information Element ID : Broadband Bearer Capability + Ext : last octet + Coding Standard : ITU-T standardized + Pass Along Request : no pass along request + Action Indicator : clear call + IE Length : 3 + Ext : another octet + Bearer Class : BCOB-X + Ext : last octet + ATM Transfer Capability : reserved for bwd compatibility + Ext : last octet + Clipping Susceptibility : not susceptible to clipping + User Plane Connection cfg : point-to-point + Information Element ID : Called Party Number + Ext : last octet + Coding Standard : ITU-T standardized + Pass Along Request : no pass along request + Action Indicator : clear call + IE Length : 8 + Ext : last octet + Type of Number : unknown + Addressing/Numbering Plan : ATM endsystem address + ATM Endsystem Address Oct : 11111111111101 + Information Element ID : Designated Transit List + Ext : last octet + Coding Standard : ATM Forum specific + Pass Along Request : no pass along request + Action Indicator : clear call + IE Length : 29 + Current Transit Pointer : 0 + Logical Node/Port Indicat : Logical Node/Port Indicator + Logical Node Identifier : 3900000000000000000000000011111111111100 + + + + + +Dunn & Martin Informational [Page 123] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + PNNI Signaling Setup Message Reject, specifying a DTL which is + blocked by the far end SUT. + +Protocol Discriminator : PNNI signalling + Call Reference Length : 3 + Call Reference Flag : from + Call Reference Value : 0 + Message Type : SETUP + Ext : last octet + Pass Along Request : no pass along request + Action Indicator : clear call + Message Length : 56 + Information Element ID : ATM Traffic Descriptor + Ext : last octet + Coding Standard : ITU-T standardized + Pass Along Request : no pass along request + Action Indicator : clear call + IE Length : 0 + Information Element ID : Broadband Bearer Capability + Ext : last octet + Coding Standard : ITU-T standardized + Pass Along Request : no pass along request + Action Indicator : clear call + IE Length : 3 + Bearer Class : BCOB-X + Ext : last octet + ATM Transfer Capability : reserved for bwd compatibility + Ext : last octet + Clipping Susceptibility : not susceptible to clipping + User Plane Connection cfg : point-to-point + Information Element ID : Called Party Number + Ext : last octet + Coding Standard : ITU-T standardized + Pass Along Request : no pass along request + Action Indicator : clear call + IE Length : 8 + Ext : last octet + Addressing/Numbering Plan : ATM endsystem address + ATM Endsystem Address Oct : 11111111111101 + Information Element ID : Designated Transit List + Ext : last octet + Coding Standard : ATM Forum specific + Pass Along Request : no pass along request + Action Indicator : clear call + IE Length : 29 + Current Transit Pointer : 0 + Logical Node/Port Indicat : Logical Node/Port Indicator + Logical Node Identifier : 3900000000000000000000000011111111111100 + + + +Dunn & Martin Informational [Page 124] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + PNNI Far End Request Message. + +Header: Packet Type 5 (PTSE REQUEST) + Packet Length 40 + Protocol Version 1 + Newest Version Supported 1 + Oldest Version Supported 0 + Reserved 0 + IG: Information Group Type 513 (Requested PTSE Header) + Information Group Length 32 + Originating Node ID + 00013900-00000000-00000000-00000011-11111111-1100 + PTSE Request Count 1 + PTSE Identifier 0 + + PNNI PTSE, specifying a routing topology. + +Header: Packet Type 4 (DATABASE SUMMARY) + Packet Length 76 + Protocol Version 1 + Newest Version Supported 1 + Oldest Version Supported 0 + Reserved 0 + Initialize (I)Bit 1 (during init. of DB syn + process) + More (M)Bit 1 (PTSEs to summarize) + Master (MS)Bit 1 (both nodes) + Reserved 0 + Reserved 0 + DS Sequence Number 0 + IG: Information Group Type 512 (Nodal PTSE Summaries) + Information Group Length 60 + Originating Node ID + 00013900-00000000-00000000-00000011-11111111-1100 + Originating Node's Peer Group 00000000-00000000-00000000- + 0001 + Reserved 0 + PTSE Summary Count 1 + PTSE Type 0 + Reserved 0 + PTSE Identifier 0 + PTSE Sequence Number 0 + PTSE Checksum 0 + PTSE Remaining Lifetime 0 + + + + + + + +Dunn & Martin Informational [Page 125] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + + PNNI PTSE Update, specifying a change in the routing topology. + +Header: Packet Type 2 (PTSP) + Packet Length 96 + Protocol Version 1 + Newest Version Supported 1 + Oldest Version Supported 0 + Reserved 0 + Originating Node ID + 00013900-00000000-00000000-00000011-11111111-1100 + Originating Node's Peer Group 00000000-00000000-00000000- + 0001 + IG: Information Group Type 64 (PTSE) + Information Group Length 52 + PTSE Type 0 + Reserved 0 + PTSE Identifier 0 + PTSE Sequence Number 0 + PTSE Checksum 42252 + PTSE Remaining Lifetime 3600 + IG: Information Group Type 224 (Internal Reachable ATM + Addresses) + Information Group Length 32 + VP Capability Flag 1 (VPCs supported) + Reserved 0 + Reserved 0 + Port ID 0 + Scope of Advertisement 96 + Address Information Length 14 + Address Information Count 1 + Prefix Length 13 + Reachable Address Prefix 39000000-00000000-00000000-01 + + + + + + + + + + + + + + + + + + + +Dunn & Martin Informational [Page 126] + +RFC 3116 Methodology for ATM Benchmarking June 2001 + + +Full Copyright Statement + + Copyright (C) The Internet Society (2001). All Rights Reserved. + + This document and translations of it may be copied and furnished to + others, and derivative works that comment on or otherwise explain it + or assist in its implementation may be prepared, copied, published + and distributed, in whole or in part, without restriction of any + kind, provided that the above copyright notice and this paragraph are + included on all such copies and derivative works. However, this + document itself may not be modified in any way, such as by removing + the copyright notice or references to the Internet Society or other + Internet organizations, except as needed for the purpose of + developing Internet standards in which case the procedures for + copyrights defined in the Internet Standards process must be + followed, or as required to translate it into languages other than + English. + + The limited permissions granted above are perpetual and will not be + revoked by the Internet Society or its successors or assigns. + + This document and the information contained herein is provided on an + "AS IS" basis and THE INTERNET SOCIETY AND THE INTERNET ENGINEERING + TASK FORCE DISCLAIMS ALL WARRANTIES, EXPRESS OR IMPLIED, INCLUDING + BUT NOT LIMITED TO ANY WARRANTY THAT THE USE OF THE INFORMATION + HEREIN WILL NOT INFRINGE ANY RIGHTS OR ANY IMPLIED WARRANTIES OF + MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. + +Acknowledgement + + Funding for the RFC Editor function is currently provided by the + Internet Society. + + + + + + + + + + + + + + + + + + + +Dunn & Martin Informational [Page 127] + |