diff options
Diffstat (limited to 'doc/rfc/rfc8911.txt')
-rw-r--r-- | doc/rfc/rfc8911.txt | 1922 |
1 files changed, 1922 insertions, 0 deletions
diff --git a/doc/rfc/rfc8911.txt b/doc/rfc/rfc8911.txt new file mode 100644 index 0000000..dbdcd3c --- /dev/null +++ b/doc/rfc/rfc8911.txt @@ -0,0 +1,1922 @@ + + + + +Internet Engineering Task Force (IETF) M. Bagnulo +Request for Comments: 8911 UC3M +Category: Standards Track B. Claise +ISSN: 2070-1721 Huawei + P. Eardley + BT + A. Morton + AT&T Labs + A. Akhter + Consultant + November 2021 + + + Registry for Performance Metrics + +Abstract + + This document defines the format for the IANA Registry of Performance + Metrics. This document also gives a set of guidelines for Registered + Performance Metric requesters and reviewers. + +Status of This Memo + + This is an Internet Standards Track document. + + This document is a product of the Internet Engineering Task Force + (IETF). It represents the consensus of the IETF community. It has + received public review and has been approved for publication by the + Internet Engineering Steering Group (IESG). Further information on + Internet Standards is available in Section 2 of RFC 7841. + + Information about the current status of this document, any errata, + and how to provide feedback on it may be obtained at + https://www.rfc-editor.org/info/rfc8911. + +Copyright Notice + + Copyright (c) 2021 IETF Trust and the persons identified as the + document authors. All rights reserved. + + This document is subject to BCP 78 and the IETF Trust's Legal + Provisions Relating to IETF Documents + (https://trustee.ietf.org/license-info) in effect on the date of + publication of this document. Please review these documents + carefully, as they describe your rights and restrictions with respect + to this document. Code Components extracted from this document must + include Revised BSD License text as described in Section 4.e of the + Trust Legal Provisions and are provided without warranty as described + in the Revised BSD License. + +Table of Contents + + 1. Introduction + 2. Terminology + 3. Scope + 4. Motivations for the Performance Metrics Registry + 4.1. Interoperability + 4.2. Single Point of Reference for Performance Metrics + 4.3. Side Benefits + 5. Criteria for Performance Metrics Registration + 6. Performance Metrics Registry: Prior Attempt + 6.1. Why This Attempt Should Succeed + 7. Definition of the Performance Metrics Registry + 7.1. Summary Category + 7.1.1. Identifier + 7.1.2. Name + 7.1.3. URI + 7.1.4. Description + 7.1.5. Reference + 7.1.6. Change Controller + 7.1.7. Version (of Registry Format) + 7.2. Metric Definition Category + 7.2.1. Reference Definition + 7.2.2. Fixed Parameters + 7.3. Method of Measurement Category + 7.3.1. Reference Method + 7.3.2. Packet Stream Generation + 7.3.3. Traffic Filter + 7.3.4. Sampling Distribution + 7.3.5. Runtime Parameters + 7.3.6. Role + 7.4. Output Category + 7.4.1. Type + 7.4.2. Reference Definition + 7.4.3. Metric Units + 7.4.4. Calibration + 7.5. Administrative Information + 7.5.1. Status + 7.5.2. Requester + 7.5.3. Revision + 7.5.4. Revision Date + 7.6. Comments and Remarks + 8. Processes for Managing the Performance Metrics Registry Group + 8.1. Adding New Performance Metrics to the Performance Metrics + Registry + 8.2. Backward-Compatible Revision of Registered Performance + Metrics + 8.3. Non-Backward-Compatible Deprecation of Registered + Performance Metrics + 8.4. Obsolete Registry Entries + 8.5. Registry Format Version and Future Changes/Extensions + 9. Security Considerations + 10. IANA Considerations + 10.1. Registry Group + 10.2. Performance Metrics Name Elements + 10.3. New Performance Metrics Registry + 11. Blank Registry Template + 11.1. Summary + 11.1.1. ID (Identifier) + 11.1.2. Name + 11.1.3. URI + 11.1.4. Description + 11.1.5. Reference + 11.1.6. Change Controller + 11.1.7. Version (of Registry Format) + 11.2. Metric Definition + 11.2.1. Reference Definition + 11.2.2. Fixed Parameters + 11.3. Method of Measurement + 11.3.1. Reference Method + 11.3.2. Packet Stream Generation + 11.3.3. Traffic Filtering (Observation) Details + 11.3.4. Sampling Distribution + 11.3.5. Runtime Parameters and Data Format + 11.3.6. Roles + 11.4. Output + 11.4.1. Type + 11.4.2. Reference Definition + 11.4.3. Metric Units + 11.4.4. Calibration + 11.5. Administrative Items + 11.5.1. Status + 11.5.2. Requester + 11.5.3. Revision + 11.5.4. Revision Date + 11.6. Comments and Remarks + 12. References + 12.1. Normative References + 12.2. Informative References + Acknowledgments + Authors' Addresses + +1. Introduction + + The IETF specifies and uses Performance Metrics of protocols and + applications transported over its protocols. Performance Metrics are + an important part of network operations using IETF protocols, and + [RFC6390] specifies guidelines for their development. + + The definition and use of Performance Metrics in the IETF have been + fostered in various working groups (WGs). Most notably: + + * The "IP Performance Metrics" (IPPM) WG is the WG primarily + focusing on Performance Metrics definition at the IETF. + + * The "Benchmarking Methodology" WG (BMWG) defines many Performance + Metrics for use in laboratory benchmarking of internetworking + technologies. + + * The "Metric Blocks for use with RTCP's Extended Report Framework" + (XRBLOCK) WG (concluded) specified many Performance Metrics + related to "RTP Control Protocol Extended Reports (RTCP XR)" + [RFC3611], which establishes a framework to allow new information + to be conveyed in RTCP, supplementing the original report blocks + defined in "RTP: A Transport Protocol for Real-Time Applications" + [RFC3550]. + + * The "IP Flow Information eXport" (IPFIX) WG (concluded) specified + an Internet Assigned Numbers Authority (IANA) process for new + Information Elements. Some Information Elements related to + Performance Metrics are proposed on a regular basis. + + * The "Performance Metrics for Other Layers" (PMOL) WG (concluded) + defined some Performance Metrics related to Session Initiation + Protocol (SIP) voice quality [RFC6035]. + + It is expected that more Performance Metrics will be defined in the + future -- not only IP-based metrics but also metrics that are + protocol specific and application specific. + + Despite the importance of Performance Metrics, there are two related + problems for the industry: + + * First, ensuring that when one party requests that another party + measure (or report or in some way act on) a particular Performance + Metric, both parties have exactly the same understanding of what + Performance Metric is being referred to. + + * Second, discovering which Performance Metrics have been specified, + to avoid developing a new Performance Metric that is very similar + but not quite interoperable. + + These problems can be addressed by creating a Registry for + Performance Metrics with the Internet Assigned Numbers Authority + (IANA). As such, this document defines the new IANA Registry for + Performance Metrics. + + Per this document, IANA has created and now maintains the Performance + Metrics Registry, according to the maintenance procedures and the + format defined in the sections below. The resulting Performance + Metrics Registry is for use by the IETF and others. Although the + Registry formatting specifications herein are primarily for Registry + creation by IANA, any other organization that wishes to create a + Performance Metrics Registry may use the same formatting + specifications for their purposes. The authors make no guarantee of + the Registry format's applicability to any possible set of + Performance Metrics envisaged by other organizations, but we + encourage others to apply it. In the remainder of this document, + unless we explicitly say otherwise, we will refer to the IANA- + maintained Performance Metrics Registry as simply the Performance + Metrics Registry. + +2. Terminology + + The key words "MUST", "MUST NOT", "REQUIRED", "SHALL", "SHALL NOT", + "SHOULD", "SHOULD NOT", "RECOMMENDED", "NOT RECOMMENDED", "MAY", and + "OPTIONAL" in this document are to be interpreted as described in + BCP 14 [RFC2119] [RFC8174] when, and only when, they appear in all + capitals, as shown here. + + Performance Metric: A quantitative measure of performance, targeted + to an IETF-specified protocol or targeted to an application + transported over an IETF-specified protocol. Examples of + Performance Metrics are the FTP response time for a complete file + download, the DNS Response time to resolve the IP address(es), a + database logging time, etc. This definition is consistent with + the definition of a metric in [RFC2330] and broader than the + definition of a Performance Metric in [RFC6390]. + + Registered Performance Metric: A Performance Metric expressed as an + entry in the Performance Metrics Registry, administered by IANA. + Such a Performance Metric has met all of the Registry review + criteria defined in this document in order to be included in the + Registry. + + Performance Metrics Registry: The IANA Registry containing + Registered Performance Metrics. + + Proprietary Registry: A set of metrics that are registered in a + proprietary Registry, as opposed to the Performance Metrics + Registry. + + Performance Metrics Experts: A group of designated experts [RFC8126] + selected by the IESG to validate the Performance Metrics before + updating the Performance Metrics Registry. The Performance + Metrics Experts work closely with IANA. + + Parameter: An input factor defined as a variable in the definition + of a Performance Metric. A Parameter is a numerical or other + specified factor forming one of a set that defines a metric or + sets the conditions of its operation. All Parameters must be + known in order to make a measurement using a metric and interpret + the results. There are two types of Parameters: Fixed and + Runtime. For the Fixed Parameters, the value of the variable is + specified in the Performance Metrics Registry Entry and different + Fixed Parameter values results in different Registered Performance + Metrics. For the Runtime Parameters, the value of the variable is + defined when the Metric Measurement Method is executed and a given + Registered Performance Metric supports multiple values for the + Parameter. Although Runtime Parameters do not change the + fundamental nature of the Performance Metric's definition, some + have substantial influence on the network property being assessed + and interpretation of the results. + + | Note: Consider the case of packet loss in the following two + | Active Measurement Method cases. The first case is packet loss + | as background loss where the Runtime Parameter set includes a + | very sparse Poisson stream and only characterizes the times + | when packets were lost. Actual user streams likely see much + | higher loss at these times, due to tail drop or radio errors. + | The second case is packet loss ratio as the complimentary + | probability of delivery ratio where the Runtime Parameter set + | includes a very dense, bursty stream, and characterizes the + | loss experienced by a stream that approximates a user stream. + | These are both "Loss metrics", but the difference in + | interpretation of the results is highly dependent on the + | Runtime Parameters (at least), to the extreme where we are + | actually using loss ratio to infer its complimentary + | probability: delivery ratio. + + Active Measurement Methods: Methods of Measurement conducted on + traffic that serves only the purpose of measurement and is + generated for that reason alone, and whose traffic characteristics + are known a priori. The complete definition of Active Methods is + specified in Section 3.4 of [RFC7799]. Examples of Active + Measurement Methods are the Measurement Methods for the one-way + delay metric defined in [RFC7679] and the round-trip delay metric + defined in [RFC2681]. + + Passive Measurement Methods: Methods of Measurement conducted on + network traffic, generated by either (1) the end users or + (2) network elements that would exist regardless of whether the + measurement was being conducted or not. The complete definition + of Passive Methods is specified in Section 3.6 of [RFC7799]. One + characteristic of Passive Measurement Methods is that sensitive + information may be observed and, as a consequence, stored in the + measurement system. + + Hybrid Measurement Methods: Methods of Measurement that use a + combination of Active Methods and Passive Methods, to assess + Active Metrics, Passive Metrics, or new metrics derived from the + a priori knowledge and observations of the stream of interest. + The complete definition of Hybrid Methods is specified in + Section 3.8 of [RFC7799]. + +3. Scope + + This document is intended for two different audiences: + + 1. For those preparing a candidate Performance Metric, it provides + criteria that the proposal SHOULD meet (see Section 5). It also + provides instructions for writing the text for each column of the + candidate Performance Metric and the references required for the + new Performance Metrics Registry Entry (up to and including the + publication of one or more immutable documents such as an RFC) + (see Section 7). + + 2. For the appointed Performance Metrics Experts and for IANA + personnel administering the new IANA Performance Metrics + Registry, it defines a set of acceptance criteria against which a + candidate Registered Performance Metric should be evaluated, and + requirements for the composition of a candidate Performance + Metric Registry Entry. + + Other organizations that standardize performance metrics are + encouraged to use the process defined in this memo to propose a + candidate Registered Performance Metric. In addition, this document + may be useful for other organizations who are defining a Performance + Metrics Registry of their own and may reuse the features of the + Performance Metrics Registry defined in this document. + + This Performance Metrics Registry is applicable to Performance + Metrics derived from Active Measurement, Passive Measurement, and any + other form of Performance Metric. This Registry is designed to + encompass Performance Metrics developed throughout the IETF and + especially for the technologies specified in the following working + groups: IPPM, XRBLOCK, IPFIX, and BMWG. This document analyzes a + prior attempt to set up a Performance Metrics Registry and the + reasons why this design was inadequate [RFC6248]. + + [RFC8912] populates the new Registry with the initial set of entries. + +4. Motivations for the Performance Metrics Registry + + In this section, we detail several motivations for the Performance + Metrics Registry. + +4.1. Interoperability + + As with any IETF Registry, the primary intention is to manage the + registration of Identifiers for use within one or more protocols. In + the particular case of the Performance Metrics Registry, there are + two types of protocols that will use the Performance Metrics in the + Performance Metrics Registry during their operation (by referring to + the index values): + + Control Protocol: This type of protocol is used to allow one entity + to request that another entity perform a measurement using a + specific metric defined by the Performance Metrics Registry. One + particular example is the Large-scale Measurement of Broadband + Performance (LMAP) framework [RFC7594]. Using the LMAP + terminology, the Performance Metrics Registry is used in the LMAP + Control Protocol to allow a Controller to schedule a Measurement + Task for one or more Measurement Agents. In order to enable this + use case, the entries in the Performance Metrics Registry must be + sufficiently defined to allow a Measurement Agent implementation + to trigger a specific Measurement Task upon the reception of a + Control Protocol message. This requirement heavily constrains the + types of entries that are acceptable for the Performance Metrics + Registry. + + Report Protocol: This type of protocol is used to allow an entity to + report Measurement Results to another entity. By referencing to a + specific Registered Performance Metric, it is possible to properly + characterize the Measurement Result data being reported. Using + the LMAP terminology, the Performance Metrics Registry is used in + the LMAP Report Protocol to allow a Measurement Agent to report + Measurement Results to a Collector. + + It should be noted that the LMAP framework explicitly allows for + using not only the IANA-maintained Performance Metrics Registry but + also other registries containing Performance Metrics, i.e., either + (1) registries defined by other organizations or (2) private + registries. However, others who are creating registries to be used + in the context of an LMAP framework are encouraged to use the + Registry format defined in this document, because this makes it + easier for developers of LMAP Measurement Agents to programmatically + use information found in those other registries' entries. + +4.2. Single Point of Reference for Performance Metrics + + A Performance Metrics Registry serves as a single point of reference + for Performance Metrics defined in different working groups in the + IETF. As we mentioned earlier, there are several working groups that + define Performance Metrics in the IETF, and it is hard to keep track + of all of them. This results in multiple definitions of similar + Performance Metrics that attempt to measure the same phenomena but in + slightly different (and incompatible) ways. Having a Registry would + allow the IETF community and others to have a single list of relevant + Performance Metrics defined by the IETF (and others, where + appropriate). The single list is also an essential aspect of + communication about Performance Metrics, where different entities + that request measurements, execute measurements, and report the + results can benefit from a common understanding of the referenced + Performance Metric. + +4.3. Side Benefits + + There are a couple of side benefits of having such a Registry. + First, the Performance Metrics Registry could serve as an inventory + of useful and used Performance Metrics that are normally supported by + different implementations of Measurement Agents. Second, the results + of measurements using the Performance Metrics should be comparable + even if they are performed by different implementations and in + different networks, as the Performance Metric is properly defined. + BCP 176 [RFC6576] examines whether the results produced by + independent implementations are equivalent in the context of + evaluating the completeness and clarity of metric specifications. + [RFC6576] is a BCP [RFC2026] that defines the Standards Track + advancement testing for (Active) IPPM Metrics, and the same process + will likely suffice to determine whether Registered Performance + Metrics are sufficiently well specified to result in comparable (or + equivalent) results. If a Registered Performance Metric has + undergone such testing, this SHOULD be noted in "Comments and + Remarks" (see Section 7.6), with a reference to the test results. + +5. Criteria for Performance Metrics Registration + + It is neither possible nor desirable to populate the Performance + Metrics Registry with all combinations of Parameters of all + Performance Metrics. A Registered Performance Metric SHOULD be: + + 1. Interpretable by the human user. + + 2. Implementable by the software or hardware designer. + + 3. Deployable by network operators. + + 4. Accurate in terms of producing equivalent results, and for + interoperability and deployment across vendors. + + 5. Operationally useful, so that it has significant industry + interest and/or has seen deployment. + + 6. Sufficiently tightly defined, so that different values for the + Runtime Parameters do not change the fundamental nature of the + measurement or change the practicality of its implementation. + + In essence, there needs to be evidence that (1) a candidate + Registered Performance Metric has significant industry interest or + has seen deployment and (2) there is agreement that the candidate + Registered Performance Metric serves its intended purpose. + +6. Performance Metrics Registry: Prior Attempt + + There was a previous attempt to define a Metrics Registry [RFC4148]. + However, it was obsoleted by [RFC6248] because it was "found to be + insufficiently detailed to uniquely identify IPPM metrics... [there + was too much] variability possible when characterizing a metric + exactly", which led to the IPPM Metrics Registry defined in [RFC4148] + having "very few users, if any." + + Three interesting additional quotes from [RFC6248] might help to + understand the issues related to that registry. + + 1. "It is not believed to be feasible or even useful to register + every possible combination of Type P, metric parameters, and + Stream parameters using the current structure of the IPPM Metrics + Registry." + + 2. "The current registry structure has been found to be + insufficiently detailed to uniquely identify IPPM metrics." + + 3. "Despite apparent efforts to find current or even future users, + no one responded to the call for interest in the RFC 4148 + registry during the second half of 2010." + + The current approach learns from this by tightly defining each + Registered Performance Metric with only a few variable (Runtime) + Parameters to be specified by the measurement designer, if any. The + idea is that entries in the Performance Metrics Registry stem from + different Measurement Methods that require input (Runtime) Parameters + to set factors like Source and Destination addresses (which do not + change the fundamental nature of the measurement). The downside of + this approach is that it could result in a large number of entries in + the Performance Metrics Registry. There is agreement that less is + more in this context -- it is better to have a reduced set of useful + metrics rather than a large set of metrics, some with questionable + usefulness. + +6.1. Why This Attempt Should Succeed + + As mentioned in the previous section, one of the main issues with the + previous Registry was that the metrics contained in the Registry were + too generic to be useful. This document specifies stricter criteria + for Performance Metric registration (see Section 5) and imposes a + group of Performance Metrics Experts that will provide guidelines to + assess if a Performance Metric is properly specified. + + Another key difference between this attempt and the previous one is + that in this case there is at least one clear user for the + Performance Metrics Registry: the LMAP framework and protocol. + Because the LMAP protocol will use the Performance Metrics Registry + values in its operation, this actually helps to determine if a metric + is properly defined -- in particular, since we expect that the LMAP + Control Protocol will enable a Controller to request that a + Measurement Agent perform a measurement using a given metric by + embedding the Performance Metrics Registry Identifier in the + protocol. Such a metric and method are properly specified if they + are defined well enough so that it is possible (and practical) to + implement them in the Measurement Agent. This was the failure of the + previous attempt: a Registry Entry with an undefined Type-P + (Section 13 of [RFC2330]) allows measurement results to vary + significantly. + +7. Definition of the Performance Metrics Registry + + This Performance Metrics Registry is applicable to Performance + Metrics used for Active Measurement, Passive Measurement, and any + other form of Performance Measurement. Each category of measurement + has unique properties, so some of the columns defined below are not + applicable for a given metric category. In this case, the column(s) + SHOULD be populated with the "N/A" value (Not Applicable). However, + the "N/A" value MUST NOT be used by any metric in the following + columns: Identifier, Name, URI, Status, Requester, Revision, Revision + Date, Description. In the future, a new category of metrics could + require additional columns, and adding new columns is a recognized + form of Registry extension. The specification defining the new + column(s) MUST give general guidelines for populating the new + column(s) for existing entries. + + The columns of the Performance Metrics Registry are defined below. + The columns are grouped into "Categories" to facilitate the use of + the Registry. Categories are described at the "Section 7.x" heading + level, and columns are described at the "Section 7.x.y" heading + level. The figure below illustrates this organization. An entry + (row) therefore gives a complete description of a Registered + Performance Metric. + + Each column serves as a checklist item and helps to avoid omissions + during registration and Expert Review [RFC8126]. + + Registry Categories and Columns are shown below in this format: + + Category + ------------------... + Column | Column |... + + Summary + ---------------------------------------------------------------- + Identifier | Name | URI | Desc. | Reference | Change | Ver | + | | | | | Controller | + + Metric Definition + ----------------------------------------- + Reference Definition | Fixed Parameters | + + Method of Measurement + --------------------------------------------------------------------- + Reference | Packet | Traffic | Sampling | Runtime | Role | + Method | Stream | Filter | Distribution | Parameters | | + | Generation | + Output + ----------------------------------------- + Type | Reference | Units | Calibration | + | Definition | | | + + Administrative Information + ------------------------------------- + Status |Requester | Rev | Rev. Date | + + Comments and Remarks + -------------------- + + There is a blank template of the Registry template provided in + Section 11 of this memo. + +7.1. Summary Category + +7.1.1. Identifier + + This column provides a numeric Identifier for the Registered + Performance Metric. The Identifier of each Registered Performance + Metric MUST be unique. Note that revising a Metric according to the + process in Section 8.2 creates a new entry in the Performance Metrics + Registry with the same identifier. + + The Registered Performance Metric unique Identifier is an unbounded + integer (range 0 to infinity). + + The Identifier 0 should be Reserved. The Identifier values from + 64512 to 65535 are reserved for private or experimental use, and the + user may encounter overlapping uses. + + When adding new Registered Performance Metrics to the Performance + Metrics Registry, IANA SHOULD assign the lowest available Identifier + to the new Registered Performance Metric. + + If a Performance Metrics Expert providing review determines that + there is a reason to assign a specific numeric Identifier, possibly + leaving a temporary gap in the numbering, then the Performance + Metrics Expert SHALL inform IANA of this decision. + +7.1.2. Name + + As the Name of a Registered Performance Metric is the first thing a + potential human implementer will use when determining whether it is + suitable for their measurement study, it is important to be as + precise and descriptive as possible. In the future, users will + review the Names to determine if the metric they want to measure has + already been registered, or if a similar entry is available, as a + basis for creating a new entry. + + Names are composed of the following elements, separated by an + underscore character "_": + + MetricType_Method_SubTypeMethod_... Spec_Units_Output + + MetricType: A combination of the directional properties and the + metric measured, such as and not limited to: + + +-----------+--------------------------------------+ + | RTDelay | Round-Trip Delay | + +-----------+--------------------------------------+ + | RTDNS | Response Time Domain Name Service | + +-----------+--------------------------------------+ + | RLDNS | Response Loss Domain Name Service | + +-----------+--------------------------------------+ + | OWDelay | One-Way Delay | + +-----------+--------------------------------------+ + | RTLoss | Round-Trip Loss | + +-----------+--------------------------------------+ + | OWLoss | One-Way Loss | + +-----------+--------------------------------------+ + | OWPDV | One-Way Packet Delay Variation | + +-----------+--------------------------------------+ + | OWIPDV | One-Way Inter-Packet Delay Variation | + +-----------+--------------------------------------+ + | OWReorder | One-Way Packet Reordering | + +-----------+--------------------------------------+ + | OWDuplic | One-Way Packet Duplication | + +-----------+--------------------------------------+ + | OWBTC | One-Way Bulk Transport Capacity | + +-----------+--------------------------------------+ + | OWMBM | One-Way Model-Based Metric | + +-----------+--------------------------------------+ + | SPMonitor | Single-Point Monitor | + +-----------+--------------------------------------+ + | MPMonitor | Multi-Point Monitor | + +-----------+--------------------------------------+ + + Table 1 + + Method: One of the methods defined in [RFC7799], such as and not + limited to: + + +-------------+----------------------------------------------+ + | Active | depends on a dedicated measurement packet | + | | stream and observations of the stream as | + | | described in [RFC7799] | + +-------------+----------------------------------------------+ + | Passive | depends *solely* on observation of one or | + | | more existing packet streams as described in | + | | [RFC7799] | + +-------------+----------------------------------------------+ + | HybridType1 | Hybrid Type I observations on one stream | + | | that combine both Active Methods and Passive | + | | Methods as described in [RFC7799] | + +-------------+----------------------------------------------+ + | HybridType2 | Hybrid Type II observations on two or more | + | | streams that combine both Active Methods and | + | | Passive Methods as described in [RFC7799] | + +-------------+----------------------------------------------+ + | Spatial | spatial metric as described in [RFC5644] | + +-------------+----------------------------------------------+ + + Table 2 + + SubTypeMethod: One or more subtypes to further describe the features + of the entry, such as and not limited to: + + +----------------+------------------------------------------------+ + | ICMP | Internet Control Message Protocol | + +----------------+------------------------------------------------+ + | IP | Internet Protocol | + +----------------+------------------------------------------------+ + | DSCPxx | where xx is replaced by a Diffserv code point | + +----------------+------------------------------------------------+ + | UDP | User Datagram Protocol | + +----------------+------------------------------------------------+ + | TCP | Transport Control Protocol | + +----------------+------------------------------------------------+ + | QUIC | QUIC transport protocol | + +----------------+------------------------------------------------+ + | HS | Hand-Shake, such as TCP's 3-way HS | + +----------------+------------------------------------------------+ + | Poisson | packet generation using Poisson distribution | + +----------------+------------------------------------------------+ + | Periodic | periodic packet generation | + +----------------+------------------------------------------------+ + | SendOnRcv | sender keeps one packet in transit by sending | + | | when previous packet arrives | + +----------------+------------------------------------------------+ + | PayloadxxxxB | where xxxx is replaced by an integer, the | + | | number of octets or 8-bit Bytes in the Payload | + +----------------+------------------------------------------------+ + | SustainedBurst | capacity test, worst case | + +----------------+------------------------------------------------+ + | StandingQueue | test of bottleneck queue behavior | + +----------------+------------------------------------------------+ + + Table 3 + + SubTypeMethod values are separated by a hyphen "-" character, + which indicates that they belong to this element and that their + order is unimportant when considering Name uniqueness. + + Spec: An immutable document Identifier combined with a document + section Identifier. For RFCs, this consists of the RFC number and + major section number that specifies this Registry Entry in the + form "RFCXXXXsecY", e.g., RFC7799sec3. Note: The RFC number is + not the primary reference specification for the metric definition + (e.g., [RFC7679] as the primary reference specification for one- + way delay metrics); it will contain the placeholder "RFCXXXXsecY" + until the RFC number is assigned to the specifying document and + would remain blank in Private Registry Entries without a + corresponding RFC. Anticipating the "RFC10K" problem, the number + of the RFC continues to replace "RFCXXXX", regardless of the + number of digits in the RFC number. Anticipating Registry Entries + from other standards bodies, the form of this Name Element MUST be + proposed and reviewed for consistency and uniqueness by the Expert + Reviewer. + + Units: The units of measurement for the output, such as and not + limited to: + + +------------+----------------------------+ + | Seconds | | + +------------+----------------------------+ + | Ratio | unitless | + +------------+----------------------------+ + | Percent | value multiplied by 100% | + +------------+----------------------------+ + | Logical | 1 or 0 | + +------------+----------------------------+ + | Packets | | + +------------+----------------------------+ + | BPS | bits per second | + +------------+----------------------------+ + | PPS | packets per second | + +------------+----------------------------+ + | EventTotal | for unitless counts | + +------------+----------------------------+ + | Multiple | more than one type of unit | + +------------+----------------------------+ + | Enumerated | a list of outcomes | + +------------+----------------------------+ + | Unitless | | + +------------+----------------------------+ + + Table 4 + + Output: The type of output resulting from measurement, such as and + not limited to: + + +--------------+------------------------------------+ + | Singleton | | + +--------------+------------------------------------+ + | Raw | multiple singletons | + +--------------+------------------------------------+ + | Count | | + +--------------+------------------------------------+ + | Minimum | | + +--------------+------------------------------------+ + | Maximum | | + +--------------+------------------------------------+ + | Median | | + +--------------+------------------------------------+ + | Mean | | + +--------------+------------------------------------+ + | 95Percentile | 95th percentile | + +--------------+------------------------------------+ + | 99Percentile | 99th percentile | + +--------------+------------------------------------+ + | StdDev | standard deviation | + +--------------+------------------------------------+ + | Variance | | + +--------------+------------------------------------+ + | PFI | pass, fail, inconclusive | + +--------------+------------------------------------+ + | FlowRecords | descriptions of flows observed | + +--------------+------------------------------------+ + | LossRatio | lost packets to total packets, <=1 | + +--------------+------------------------------------+ + + Table 5 + + An example, as described in Section 4 of [RFC8912], is + + RTDelay_Active_IP-UDP-Periodic_RFC8912sec4_Seconds_95Percentile + + Note that private registries following the format described here + SHOULD use the prefix "Priv_" on any Name to avoid unintended + conflicts (further considerations are described in Section 10). + Private Registry Entries usually have no specifying RFC; thus, the + Spec: element has no clear interpretation. + +7.1.3. URI + + The URI column MUST contain a URL [RFC3986] that uniquely identifies + and locates the Metric Entry so it is accessible through the + Internet. The URL points to a file containing all of the human- + readable information for one Registry Entry. The URL SHALL reference + a target file that is preferably HTML-formatted and contains URLs to + referenced sections of HTMLized RFCs, or other reference + specifications. These target files for different entries can be more + easily edited and reused when preparing new entries. The exact form + of the URL for each target file, and the target file itself, will be + determined by IANA and reside on <https://www.iana.org/>. Section 4 + of [RFC8912], as well as subsequent major sections of that document, + provide an example of a target file in HTML form. + +7.1.4. Description + + A Registered Performance Metric description is a written + representation of a particular Performance Metrics Registry Entry. + It supplements the Registered Performance Metric Name to help + Performance Metrics Registry users select relevant Registered + Performance Metrics. + +7.1.5. Reference + + This entry gives the specification containing the candidate Registry + Entry that was reviewed and agreed upon, if such an RFC or other + specification exists. + +7.1.6. Change Controller + + This entry names the entity responsible for approving revisions to + the Registry Entry and SHALL provide contact information (for an + individual, where appropriate). + +7.1.7. Version (of Registry Format) + + This column gives the version number for the Registry format used, at + the time the Performance Metric is registered. The format complying + with this memo MUST use 1.0. A new RFC that changes the Registry + format will designate a new version number corresponding to that + format. The version number of Registry Entries SHALL NOT change + unless the Registry Entry is updated to reflect the Registry format + (following the procedures in Section 8). + +7.2. Metric Definition Category + + This category includes columns to prompt all necessary details + related to the metric definition, including the immutable document + reference and values of input factors, called "Fixed Parameters", + which are left open in the immutable document but have a particular + value defined by the Performance Metric. + +7.2.1. Reference Definition + + This entry provides a reference (or references) to the relevant + sections of the document or documents that define the metric, as well + as any supplemental information needed to ensure an unambiguous + definition for implementations. A given reference needs to be an + immutable document, such as an RFC; for other standards bodies, it is + likely to be necessary to reference a specific, dated version of a + specification. + +7.2.2. Fixed Parameters + + Fixed Parameters are Parameters whose values must be specified in the + Performance Metrics Registry. The measurement system uses these + values. + + Where referenced metrics supply a list of Parameters as part of their + descriptive template, a subset of the Parameters will be designated + as Fixed Parameters. As an example for Active Metrics, Fixed + Parameters determine most or all of the IPPM framework convention + "packets of Type-P" as described in [RFC2330], such as transport + protocol, payload length, TTL, etc. An example for Passive Metrics + is for an RTP packet loss calculation that relies on the validation + of a packet as RTP, which is a multi-packet validation controlled by + the MIN_SEQUENTIAL variable as defined by [RFC3550]. Varying + MIN_SEQUENTIAL values can alter the loss report, and this variable + could be set as a Fixed Parameter. + + Parameters MUST have well-defined names. For human readers, the + hanging-indent style is preferred, and any Parameter names and + definitions that do not appear in the Reference Method Specification + MUST appear in this column (or the Runtime Parameters column). + + Parameters MUST have a well-specified data format. + + A Parameter that is a Fixed Parameter for one Performance Metrics + Registry Entry may be designated as a Runtime Parameter for another + Performance Metrics Registry Entry. + +7.3. Method of Measurement Category + + This category includes columns for references to relevant sections of + the immutable document(s) and any supplemental information needed to + ensure an unambiguous method for implementations. + +7.3.1. Reference Method + + This entry provides references to relevant sections of immutable + documents, such as RFC(s) (for other standards bodies, it is likely + to be necessary to reference a specific, dated version of a + specification) describing the Method of Measurement, as well as any + supplemental information needed to ensure unambiguous interpretation + for implementations referring to the immutable document text. + + Specifically, this section should include pointers to pseudocode or + actual code that could be used for an unambiguous implementation. + +7.3.2. Packet Stream Generation + + This column applies to Performance Metrics that generate traffic as + part of their Measurement Method, including, but not necessarily + limited to, Active Metrics. The generated traffic is referred to as + a "stream", and this column describes its characteristics. + + Each entry for this column contains the following information: + + Value: The name of the packet stream scheduling discipline + + Reference: The specification where the Parameters of the stream are + defined + + The packet generation stream may require Parameters such as the + average packet rate and distribution truncation value for streams + with Poisson-distributed inter-packet sending times. If such + Parameters are needed, they should be included in either the Fixed + Parameters column or the Runtime Parameters column, depending on + whether they will be fixed or will be an input for the metric. + + The simplest example of stream specification is singleton scheduling + (see [RFC2330]), where a single atomic measurement is conducted. + Each atomic measurement could consist of sending a single packet + (such as a DNS request) or sending several packets (for example, to + request a web page). Other streams support a series of atomic + measurements using pairs of packets, where the packet stream follows + a schedule defining the timing between transmitted packets, and an + atomic measurement assesses the reception time between successive + packets (e.g., a measurement of Inter-Packet Delay Variation). More + complex streams and measurement relationships are possible. + Principally, two different streams are used in IPPM Metrics: + (1) Poisson, distributed as described in [RFC2330] and (2) periodic, + as described in [RFC3432]. Both Poisson and periodic have their own + unique Parameters, and the relevant set of Parameter names and values + should be included in either the Fixed Parameters column or the + Runtime Parameters column. + +7.3.3. Traffic Filter + + This column applies to Performance Metrics that observe packets + flowing through (the device with) the Measurement Agent, i.e., + packets that are not necessarily addressed to the Measurement Agent. + This includes, but is not limited to, Passive Metrics. The filter + specifies the traffic that is measured. This includes protocol field + values/ranges, such as address ranges, and flow or session + Identifiers. + + The Traffic Filter itself depends on the needs of the metric itself + and a balance of an operator's measurement needs and a user's need + for privacy. Mechanics for conveying the filter criteria might be + the BPF (Berkeley Packet Filter) or PSAMP (Packet Sampling) [RFC5475] + Property Match Filtering, which reuses IPFIX [RFC7012]. An example + BPF string for matching TCP/80 traffic to remote Destination net + 192.0.2.0/24 would be "dst net 192.0.2.0/24 and tcp dst port 80". + More complex filter engines may allow for matching using Deep Packet + Inspection (DPI) technology. + + The Traffic Filter includes the following information: + + Type: The type of Traffic Filter used, e.g., BPF, PSAMP, OpenFlow + rule, etc., as defined by a normative reference + + Value: The actual set of rules expressed + +7.3.4. Sampling Distribution + + The sampling distribution defines, out of all of the packets that + match the Traffic Filter, which one or more of those packets are + actually used for the measurement. One possibility is "all", which + implies that all packets matching the Traffic Filter are considered, + but there may be other sampling strategies. It includes the + following information: + + Value: The name of the sampling distribution + + Reference definition: Pointer to the specification where the + sampling distribution is properly defined + + The sampling distribution may require Parameters. If such Parameters + are needed, they should be included in either the Fixed Parameters + column or the Runtime Parameters column, depending on whether they + will be fixed or will be an input for the metric. + + PSAMP is documented in "Sampling and Filtering Techniques for IP + Packet Selection" [RFC5475], while "A Framework for Packet Selection + and Reporting" [RFC5474] provides more background information. The + sampling distribution Parameters might be expressed in terms of the + model described in "Information Model for Packet Sampling Exports" + [RFC5477] and the process provided in "Flow Selection Techniques" + [RFC7014]. + +7.3.5. Runtime Parameters + + In contrast to the Fixed Parameters, Runtime Parameters are + Parameters that do not change the fundamental nature of the + measurement and their values are not specified in the Performance + Metrics Registry. They are left as variables in the Registry, as an + aid to the measurement system implementer or user. Their values are + supplied on execution, configured into the measurement system, and + reported with the Measurement Results (so that the context is + complete). + + Where metrics supply a list of Parameters as part of their + descriptive template, a subset of the Parameters will be designated + as Runtime Parameters. + + Parameters MUST have well-defined names. For human readers, the + hanging-indent style is preferred, and the names and definitions that + do not appear in the Reference Method Specification MUST appear in + this column. + + A data format for each Runtime Parameter MUST be specified in this + column, to simplify the control and implementation of measurement + devices. For example, Parameters that include an IPv4 address can be + encoded as a 32-bit integer (i.e., a binary base64-encoded value) or + "ip-address" as defined in [RFC6991]. The actual encoding(s) used + must be explicitly defined for each Runtime Parameter. IPv6 + addresses and options MUST be accommodated, allowing Registered + Performance Metrics to be used in that address family. Other address + families are permissible. + + Examples of Runtime Parameters include IP addresses, measurement + point designations, start times and end times for measurement, and + other information essential to the Method of Measurement. + +7.3.6. Role + + In some Methods of Measurement, there may be several Roles defined, + e.g., for a one-way packet delay Active Measurement, there is one + Measurement Agent that generates the packets and another Agent that + receives the packets. This column contains the name of the Role(s) + for this particular entry. In the one-way delay example above, there + should be two entries in the Registry's Role column, one for each + Role (Source and Destination). When a Measurement Agent is + instructed to perform the "Source" Role for the one-way delay metric, + the Agent knows that it is required to generate packets. The values + for this field are defined in the Reference Method of Measurement + (and this frequently results in abbreviated Role names such as + "Src"). + + When the Role column of a Registry Entry defines more than one Role, + the Role SHALL be treated as a Runtime Parameter and supplied for + execution. It should be noted that the LMAP framework [RFC7594] + distinguishes the Role from other Runtime Parameters. + +7.4. Output Category + + For entries that involve a stream and many singleton measurements, a + statistic may be specified in this column to summarize the results to + a single value. If the complete set of measured singletons is + output, this will be specified here. + + Some metrics embed one specific statistic in the reference metric + definition, while others allow several output types or statistics. + +7.4.1. Type + + This column contains the name of the output type. The output type + defines a single type of result that the metric produces. It can be + the raw results (packet send times and singleton metrics), or it can + be a summary statistic. The specification of the output type MUST + define the format of the output. In some systems, format + specifications will simplify both measurement implementation and + collection/storage tasks. Note that if two different statistics are + required from a single measurement (for example, both "Xth percentile + mean" and "Raw"), then a new output type must be defined ("Xth + percentile mean AND Raw"). See Section 7.1.2 above for a list of + output types. + +7.4.2. Reference Definition + + This column contains a pointer to the specification(s) where the + output type and format are defined. + +7.4.3. Metric Units + + The measured results must be expressed using some standard dimension + or units of measure. This column provides the units. + + When a sample of singletons (see Section 11 of [RFC2330] for + definitions of these terms) is collected, this entry will specify the + units for each measured value. + +7.4.4. Calibration + + Some specifications for Methods of Measurement include the ability to + perform an error calibration. Section 3.7.3 of [RFC7679] is one + example. In the Registry Entry, this field will identify a method of + calibration for the metric, and, when available, the measurement + system SHOULD perform the calibration when requested and produce the + output with an indication that it is the result of a calibration + method. In-situ calibration could be enabled with an internal + loopback that includes as much of the measurement system as possible, + performs address manipulation as needed, and provides some form of + isolation (e.g., deterministic delay) to avoid send-receive interface + contention. Some portion of the random and systematic error can be + characterized in this way. + + For one-way delay measurements, the error calibration must include an + assessment of the internal clock synchronization with its external + reference (this internal clock is supplying timestamps for + measurement). In practice, the time offsets of clocks at both the + Source and Destination are needed to estimate the systematic error + due to imperfect clock synchronization (the time offsets are + smoothed; thus, the random variation is not usually represented in + the results). + + Both internal loopback calibration and clock synchronization can be + used to estimate the *available accuracy* of the Output Metric Units. + For example, repeated loopback delay measurements will reveal the + portion of the output result resolution that is the result of system + noise and is thus inaccurate. + +7.5. Administrative Information + +7.5.1. Status + + This entry indicates the status of the specification of this + Registered Performance Metric. Allowed values are 'Current', + 'Deprecated', and 'Obsolete'. All newly defined Registered + Performance Metrics have 'Current' Status. + +7.5.2. Requester + + This entry indicates the requester for the Registered Performance + Metric. The requester MAY be a document (such as an RFC) or a + person. + +7.5.3. Revision + + This entry indicates the revision number of a Registered Performance + Metric, starting at 0 for Registered Performance Metrics at the time + of definition and incremented by one for each revision. However, in + the case of a non-backward-compatible revision, see Section 8.3. + +7.5.4. Revision Date + + This entry indicates the date of acceptance of the most recent + revision for the Registered Performance Metric. The date SHALL be + determined by IANA and the reviewing Performance Metrics Expert. + +7.6. Comments and Remarks + + Besides providing additional details that do not appear in other + categories, this open category (single column) allows unforeseen + issues to be addressed by simply updating this informational entry. + +8. Processes for Managing the Performance Metrics Registry Group + + Once a Performance Metric or set of Performance Metrics has been + identified for a given application, candidate Performance Metrics + Registry Entry specifications prepared in accordance with Section 7 + should be submitted to IANA to follow the process for review by the + Performance Metrics Experts, as defined below. This process is also + used for other changes to a Performance Metrics Registry Entry, such + as deprecation or revision, as described later in this section. + + It is desirable that the author(s) of a candidate Performance Metrics + Registry Entry seek review in the relevant IETF working group or + offer the opportunity for review on the working group mailing list. + +8.1. Adding New Performance Metrics to the Performance Metrics Registry + + Requests to add Registered Performance Metrics in the Performance + Metrics Registry SHALL be submitted to IANA, which forwards the + request to a designated group of experts (Performance Metrics + Experts) appointed by the IESG; these are the reviewers called for by + the Specification Required policy [RFC8126] defined for the + Performance Metrics Registry. The Performance Metrics Experts review + the request for such things as compliance with this document, + compliance with other applicable Performance Metrics-related RFCs, + and consistency with the currently defined set of Registered + Performance Metrics. The most efficient path for submission begins + with preparation of an Internet-Draft containing the proposed + Performance Metrics Registry Entry using the template in Section 11, + so that the submission formatting will benefit from the normal IETF + Internet-Draft submission processing (including HTMLization). + + Submission to IANA may be during IESG review (leading to IETF + Standards Action), where an Internet-Draft proposes one or more + Registered Performance Metrics to be added to the Performance Metrics + Registry, including the text of the proposed Registered Performance + Metric(s). + + If an RFC-to-be includes a Performance Metric and a proposed + Performance Metrics Registry Entry but the Performance Metrics + Expert's review determines that one or more of the criteria listed in + Section 5 have not been met, then the proposed Performance Metrics + Registry Entry MUST be removed from the text. Once evidence exists + that the Performance Metric meets the criteria in Section 5, the + proposed Performance Metrics Registry Entry SHOULD be submitted to + IANA to be evaluated in consultation with the Performance Metrics + Experts for registration at that time. + + Authors of proposed Registered Performance Metrics SHOULD review + compliance with the specifications in this document to check their + submissions before sending them to IANA. + + At least one Performance Metrics Expert should endeavor to complete + referred reviews in a timely manner. If the request is acceptable, + the Performance Metrics Experts signify their approval to IANA, and + IANA updates the Performance Metrics Registry. If the request is not + acceptable, the Performance Metrics Experts MAY coordinate with the + requester to change the request so that it is compliant; otherwise, + IANA SHALL coordinate resolution of issues on behalf of the expert. + The Performance Metrics Experts MAY choose to reject clearly + frivolous or inappropriate change requests outright, but such + exceptional circumstances should be rare. + + If the proposed Metric is unique in a significant way, in order to + properly describe the Metric, it may be necessary to propose a new + Name Element Registry, or (more likely) a new Entry in an existing + Name Element Registry. This proposal is part of the request for the + new Metric, so that it undergoes the same IANA review and approval + process. + + Decisions by the Performance Metrics Experts may be appealed per + Section 10 of [RFC8126]. + +8.2. Backward-Compatible Revision of Registered Performance Metrics + + A request for revision is only permitted when the requested changes + maintain backward compatibility with implementations of the prior + Performance Metrics Registry Entry describing a Registered + Performance Metric (entries with lower revision numbers but having + the same Identifier and Name). + + The purpose of the Status field in the Performance Metrics Registry + is to indicate whether the entry for a Registered Performance Metric + is 'Current', 'Deprecated', or 'Obsolete'. The term 'deprecated' is + used when an entry is replaced, either with a backwards-compatible + revision (this sub-section) or with a non-backwards-compatible + revision (in Section 8.3). + + In addition, no policy is defined for revising the Performance Metric + Entries in the IANA Registry or addressing errors therein. To be + clear, changes and deprecations within the Performance Metrics + Registry are not encouraged and should be avoided to the extent + possible. However, in recognition that change is inevitable, the + provisions of this section address the need for revisions. + + Revisions are initiated by sending a candidate Registered Performance + Metric definition to IANA, per Section 8.1, identifying the existing + Performance Metrics Registry Entry, and explaining how and why the + existing entry should be revised. + + The primary requirement in the definition of procedures for managing + changes to existing Registered Performance Metrics is avoidance of + measurement interoperability problems; the Performance Metrics + Experts must work to maintain interoperability above all else. + Changes to Registered Performance Metrics may only be done in an + interoperable way; necessary changes that cannot be done in a way + that allows interoperability with unchanged implementations MUST + result in the creation of a new Registered Performance Metric (with a + new Name, replacing the RFCXXXXsecY portion of the Name) and possibly + the deprecation of the earlier metric. + + A change to a Registered Performance Metric SHALL be determined to be + backward compatible when: + + 1. it involves the correction of an error that is obviously only + editorial, or + + 2. it corrects an ambiguity in the Registered Performance Metric's + definition, which itself leads to issues severe enough to prevent + the Registered Performance Metric's usage as originally defined, + or + + 3. it corrects missing information in the metric definition without + changing its meaning (e.g., the explicit definition of 'quantity' + semantics for numeric fields without a Data Type Semantics + value), or + + 4. it harmonizes with an external reference that was itself + corrected, or + + 5. if the current Registry format has been revised by adding a new + column that is not relevant to an existing Registered Performance + Metric (i.e., the new column can be safely filled in with "Not + Applicable"). + + If a Performance Metric revision is deemed permissible and backward + compatible by the Performance Metrics Experts, according to the rules + in this document, IANA SHOULD execute the change(s) in the + Performance Metrics Registry. The requester of the change is + appended to the original requester in the Performance Metrics + Registry. The Name of the revised Registered Performance Metric, + including the RFCXXXXsecY portion of the Name, SHALL remain unchanged + even when the change is the result of IETF Standards Action. The + revised Registry Entry SHOULD reference the new immutable document, + such as an RFC. For other standards bodies, it is likely to be + necessary to reference a specific, dated version of a specification, + in an appropriate category and column. + + Each Registered Performance Metric in the Performance Metrics + Registry has a revision number, starting at zero. Each change to a + Registered Performance Metric following this process increments the + revision number by one. + + When a revised Registered Performance Metric is accepted into the + Performance Metrics Registry, the date of acceptance of the most + recent revision is placed into the Revision Date column of the + Registry for that Registered Performance Metric. + + Where applicable, additions to Registered Performance Metrics in the + form of text in the Comments or Remarks column should include the + date, but such additions may not constitute a revision according to + this process. + + Older versions of the updated Metric Entries are kept in the Registry + for archival purposes. The older entries are kept with all fields + unmodified (including Revision Date) except for the Status field, + which SHALL be changed to 'Deprecated'. + + This process should not in any way be construed as allowing the + Performance Metrics Experts to overrule IETF consensus. + Specifically, any Registered Performance Metrics that were added to + the Performance Metrics Registry with IETF consensus require IETF + consensus for revision or deprecation. + +8.3. Non-Backward-Compatible Deprecation of Registered Performance + Metrics + + This section describes how to make a non-backward-compatible update + to a Registered Performance Metric. A Registered Performance Metric + MAY be deprecated and replaced when: + + 1. the Registered Performance Metric definition has an error or + shortcoming that cannot be permissibly changed per Section 8.2 + ("Revising Registered Performance Metrics"), or + + 2. the deprecation harmonizes with an external reference that was + itself deprecated through that reference's accepted deprecation + method. + + A request for deprecation is sent to IANA, which passes it to the + Performance Metrics Experts for review. When deprecating a + Performance Metric, the Performance Metric Description in the + Performance Metrics Registry MUST be updated to explain the + deprecation, as well as to refer to the new Performance Metric + created to replace the deprecated Performance Metric. + + When a new, non-backward-compatible Performance Metric replaces a + (now) deprecated metric, the revision number of the new Registered + Performance Metric is incremented over the value in the deprecated + version, and the current date is entered as the Revision Date of the + new Registered Performance Metric. + + The intentional use of deprecated Registered Performance Metrics + should result in a log entry or human-readable warning by the + respective application. + + Names and Metric IDs of deprecated Registered Performance Metrics + must not be reused. + + The deprecated entries are kept with all Administrative columns + unmodified, except the Status field (which is changed to + 'Deprecated'). + +8.4. Obsolete Registry Entries + + Existing Registry Entries may become obsolete over time due to: + + 1. the Registered Performance Metric is found to contain + considerable errors (and no one sees the value in the effort to + fix it), or + + 2. one or more critical References (or sections thereof) have been + designated obsolete by the SDO, or + + 3. other reasons brought to the attention of IANA and the Registry + Experts. + + When a Performance Metric Registry Entry is declared obsolete, the + Performance Metric Description in the Performance Metrics Registry is + updated to explain the reasons the Entry is now obsolete and has not + been replaced (Deprecation always involves replacement). + + Obsolete entries are kept with all Administrative columns unmodified, + except the Status field (which is changed to 'Obsolete'). + +8.5. Registry Format Version and Future Changes/Extensions + + The Registry Format Version defined in this memo is 1.0, and + candidate Registry Entries complying with this memo MUST use 1.0. + + The Registry Format can only be updated by publishing a new RFC with + the new format (Standards Action). + + When a Registered Performance Metric is created or revised, then it + uses the most recent Registry Format Version. + + Only one form of Registry extension is envisaged: + + Adding columns, or both categories and columns, to accommodate + unanticipated aspects of new measurements and metric categories. + + If the Performance Metrics Registry is extended in this way, the + version number of future entries complying with the extension SHALL + be incremented (in either the unit or the tenths digit, depending on + the degree of extension). + +9. Security Considerations + + This document defines a Registry structure and does not itself + introduce any new security considerations for the Internet. The + definition of Performance Metrics for this Registry may introduce + some security concerns, but the mandatory references should have + their own considerations for security, and such definitions should be + reviewed with security in mind if the security considerations are not + covered by one or more reference standards. + + The aggregated results of the Performance Metrics described in this + Registry might reveal network topology information that may be + considered sensitive. If such cases are found, then access control + mechanisms should be applied. + +10. IANA Considerations + + With the background and processes described in earlier sections, IANA + has taken the actions described below. + +10.1. Registry Group + + The new Registry group is named Performance Metrics. This document + refers to it as the "Performance Metrics Group" or "Registry Group", + meaning all registrations appearing on + <https://www.iana.org/assignments/performance-metrics> + (https://www.iana.org/assignments/performance-metrics). + + For clarity, note that this document and [RFC8912] use the following + conventions to refer to the various IANA registries related to + Performance Metrics. + + +===============+===========================+=====================+ + | | RFC 8911 and RFC 8912 | IANA Web page | + +===============+===========================+=====================+ + | Page Title | Performance Metrics Group | Performance Metrics | + +---------------+---------------------------+---------------------+ + | Main Registry | Performance Metrics | Performance Metrics | + | | Registry | Registry | + +---------------+---------------------------+---------------------+ + | Registry Row | Performance Metrics | registration (also | + | | Registry Entry | template) | + +---------------+---------------------------+---------------------+ + + Table 6 + + Registration Procedure: Specification Required + + Reference: RFC 8911 + + Experts: Performance Metrics Experts + +10.2. Performance Metrics Name Elements + + This memo specifies and populates the Registries for the Performance + Metric Name Elements. The Name assigned to a Performance Metric + Registry Entry consists of multiple Elements separated by an "_" + (underscore), in the order defined in Section 7.1.2. IANA has + created the following registries, which contain the current set of + possibilities for each Element in the Performance Metric Name. + + MetricType + + Method + + SubTypeMethod + + Spec + + Units + + Output + + At creation, IANA has populated the Registered Performance Metrics + Name Elements using the lists of values for each Name Element listed + in Section 7.1.2. The Name Elements in each Registry are case + sensitive. + + When preparing a Metric Entry for registration, the developer SHOULD + choose Name Elements from among the registered elements. However, if + the proposed metric is unique in a significant way, it may be + necessary to propose a new Name Element to properly describe the + metric, as described below. + + A candidate Metric Entry proposes a set of values for its Name + Elements. These are reviewed by IANA and an Expert Reviewer. It is + possible that a candidate Metric Entry proposes a new value for a + Name Element (that is, one that is not in the existing list of + possibilities), or even that it proposes a new Name Element. Such + new assignments are administered by IANA through the Specification + Required policy [RFC8126], which includes Expert Review (i.e., review + by one of a group of Performance Metrics Experts, who are appointed + by the IESG upon recommendation of the Transport Area Directors). + +10.3. New Performance Metrics Registry + + This document specifies the Performance Metrics Registry. The + Registry contains the following columns in the Summary category: + + Identifier + + Name + + URI + + Description + + Reference + + Change Controller + + Version + + Descriptions of these columns and additional information found in the + template for Registry Entries (categories and columns) are further + defined in Section 7. + + The Identifier 0 should be Reserved. The Registered Performance + Metric unique Identifier is an unbounded integer (range 0 to + infinity). The Identifier values from 64512 to 65535 are reserved + for private or experimental use, and the user may encounter + overlapping uses. When adding new Registered Performance Metrics to + the Performance Metrics Registry, IANA SHOULD assign the lowest + available Identifier to the new Registered Performance Metric. If a + Performance Metrics Expert providing review determines that there is + a reason to assign a specific numeric Identifier, possibly leaving a + temporary gap in the numbering, then the Performance Metrics Expert + SHALL inform IANA of this decision. + + Names starting with the prefix "Priv_" are reserved for private use + and are not considered for registration. The Name column entries are + further defined in Section 7. + + The URI column will have a URL to each completed Registry Entry. The + Registry Entry text SHALL be HTMLized to aid the reader (similar to + the way that Internet-Drafts are HTMLized, the same tool can perform + the function), with links to referenced section(s) of an RFC or + another immutable document. + + The Reference column will include an RFC number, an approved + specification designator from another standards body, or some other + immutable document. + + New assignments for the Performance Metrics Registry will be + administered by IANA through the Specification Required policy + [RFC8126] (which includes Expert Review, i.e., review by one of a + group of experts -- in the case of this document, the Performance + Metrics Experts, who are appointed by the IESG upon recommendation of + the Transport Area Directors) or by Standards Action. The experts + can be initially drawn from the Working Group Chairs, document + editors, and members of the Performance Metrics Directorate, among + other sources of experts. + + Extensions to the Performance Metrics Registry require IETF Standards + Action. Only one form of Registry extension is envisaged: + + * Adding columns, or both categories and columns, to accommodate + unanticipated aspects of new measurements and metric categories. + + If the Performance Metrics Registry is extended in this way, the + version number of future entries complying with the extension SHALL + be incremented (in either the unit or the tenths digit, depending on + the degree of extension). + +11. Blank Registry Template + + This section provides a blank template to help IANA and Registry + Entry writers. + +11.1. Summary + + This category includes multiple indexes to the Registry Entry: the + element ID and Metric Name. + +11.1.1. ID (Identifier) + + <insert a numeric Identifier, an integer, TBD> + +11.1.2. Name + + <insert the Name, according to the metric naming convention> + +11.1.3. URI + + URL: https://www.iana.org/assignments/performance-metrics/ ... <Name> + +11.1.4. Description + + <provide a description> + +11.1.5. Reference + + <provide the RFC or other specification that contains the approved + candidate Registry Entry> + +11.1.6. Change Controller + + <provide information regarding the entity responsible for approving + revisions to the Registry Entry (including contact information for an + individual, where appropriate)> + +11.1.7. Version (of Registry Format) + +11.2. Metric Definition + + This category includes columns to prompt the entry of all necessary + details related to the metric definition, including the immutable + document reference and values of input factors, called "Fixed + Parameters". + +11.2.1. Reference Definition + + <provide a full bibliographic reference to an immutable document> + + <provide a specific section reference and additional clarifications, + if needed> + +11.2.2. Fixed Parameters + + <list and specify Fixed Parameters, input factors that must be + determined and embedded in the measurement system for use when + needed> + +11.3. Method of Measurement + + This category includes columns for references to relevant sections of + the immutable document(s) and any supplemental information needed to + ensure an unambiguous method for implementations. + +11.3.1. Reference Method + + <for the metric, insert relevant section references and supplemental + info> + +11.3.2. Packet Stream Generation + + <provide a list of generation Parameters and section/spec references + if needed> + +11.3.3. Traffic Filtering (Observation) Details + + This category provides the filter details (when present), which + qualify the set of packets that contribute to the measured results + from among all packets observed. + + <provide a section reference> + +11.3.4. Sampling Distribution + + <insert time distribution details, or how this is different from the + filter> + +11.3.5. Runtime Parameters and Data Format + + Runtime Parameters are input factors that must be determined, + configured into the measurement system, and reported with the results + for the context to be complete. + + <provide a list of Runtime Parameters and their data formats> + +11.3.6. Roles + + <list the names of the different Roles from the Measurement Method> + +11.4. Output + + This category specifies all details of the output of measurements + using the metric. + +11.4.1. Type + + <insert the name of the output type -- raw results or a selected + summary statistic> + +11.4.2. Reference Definition + + <describe the reference data format for each type of result> + +11.4.3. Metric Units + + <insert units for the measured results, and provide the reference + specification> + +11.4.4. Calibration + + <insert information on calibration> + +11.5. Administrative Items + + This category provides administrative information. + +11.5.1. Status + + <provide status: 'Current' or 'Deprecated'> + +11.5.2. Requester + + <provide a person's name, an RFC number, etc.> + +11.5.3. Revision + + <provide the revision number: starts at 0> + +11.5.4. Revision Date + + <provide the date, in YYYY-MM-DD format> + +11.6. Comments and Remarks + + <list any additional (informational) details for this entry> + +12. References + +12.1. Normative References + + [RFC2026] Bradner, S., "The Internet Standards Process -- Revision + 3", BCP 9, RFC 2026, DOI 10.17487/RFC2026, October 1996, + <https://www.rfc-editor.org/info/rfc2026>. + + [RFC2119] Bradner, S., "Key words for use in RFCs to Indicate + Requirement Levels", BCP 14, RFC 2119, + DOI 10.17487/RFC2119, March 1997, + <https://www.rfc-editor.org/info/rfc2119>. + + [RFC2330] Paxson, V., Almes, G., Mahdavi, J., and M. Mathis, + "Framework for IP Performance Metrics", RFC 2330, + DOI 10.17487/RFC2330, May 1998, + <https://www.rfc-editor.org/info/rfc2330>. + + [RFC3986] Berners-Lee, T., Fielding, R., and L. Masinter, "Uniform + Resource Identifier (URI): Generic Syntax", STD 66, + RFC 3986, DOI 10.17487/RFC3986, January 2005, + <https://www.rfc-editor.org/info/rfc3986>. + + [RFC5644] Stephan, E., Liang, L., and A. Morton, "IP Performance + Metrics (IPPM): Spatial and Multicast", RFC 5644, + DOI 10.17487/RFC5644, October 2009, + <https://www.rfc-editor.org/info/rfc5644>. + + [RFC6390] Clark, A. and B. Claise, "Guidelines for Considering New + Performance Metric Development", BCP 170, RFC 6390, + DOI 10.17487/RFC6390, October 2011, + <https://www.rfc-editor.org/info/rfc6390>. + + [RFC6576] Geib, R., Ed., Morton, A., Fardid, R., and A. Steinmitz, + "IP Performance Metrics (IPPM) Standard Advancement + Testing", BCP 176, RFC 6576, DOI 10.17487/RFC6576, March + 2012, <https://www.rfc-editor.org/info/rfc6576>. + + [RFC7799] Morton, A., "Active and Passive Metrics and Methods (with + Hybrid Types In-Between)", RFC 7799, DOI 10.17487/RFC7799, + May 2016, <https://www.rfc-editor.org/info/rfc7799>. + + [RFC8126] Cotton, M., Leiba, B., and T. Narten, "Guidelines for + Writing an IANA Considerations Section in RFCs", BCP 26, + RFC 8126, DOI 10.17487/RFC8126, June 2017, + <https://www.rfc-editor.org/info/rfc8126>. + + [RFC8174] Leiba, B., "Ambiguity of Uppercase vs Lowercase in RFC + 2119 Key Words", BCP 14, RFC 8174, DOI 10.17487/RFC8174, + May 2017, <https://www.rfc-editor.org/info/rfc8174>. + +12.2. Informative References + + [RFC2681] Almes, G., Kalidindi, S., and M. Zekauskas, "A Round-trip + Delay Metric for IPPM", RFC 2681, DOI 10.17487/RFC2681, + September 1999, <https://www.rfc-editor.org/info/rfc2681>. + + [RFC3432] Raisanen, V., Grotefeld, G., and A. Morton, "Network + performance measurement with periodic streams", RFC 3432, + DOI 10.17487/RFC3432, November 2002, + <https://www.rfc-editor.org/info/rfc3432>. + + [RFC3550] Schulzrinne, H., Casner, S., Frederick, R., and V. + Jacobson, "RTP: A Transport Protocol for Real-Time + Applications", STD 64, RFC 3550, DOI 10.17487/RFC3550, + July 2003, <https://www.rfc-editor.org/info/rfc3550>. + + [RFC3611] Friedman, T., Ed., Caceres, R., Ed., and A. Clark, Ed., + "RTP Control Protocol Extended Reports (RTCP XR)", + RFC 3611, DOI 10.17487/RFC3611, November 2003, + <https://www.rfc-editor.org/info/rfc3611>. + + [RFC4148] Stephan, E., "IP Performance Metrics (IPPM) Metrics + Registry", BCP 108, RFC 4148, DOI 10.17487/RFC4148, August + 2005, <https://www.rfc-editor.org/info/rfc4148>. + + [RFC5474] Duffield, N., Ed., Chiou, D., Claise, B., Greenberg, A., + Grossglauser, M., and J. Rexford, "A Framework for Packet + Selection and Reporting", RFC 5474, DOI 10.17487/RFC5474, + March 2009, <https://www.rfc-editor.org/info/rfc5474>. + + [RFC5475] Zseby, T., Molina, M., Duffield, N., Niccolini, S., and F. + Raspall, "Sampling and Filtering Techniques for IP Packet + Selection", RFC 5475, DOI 10.17487/RFC5475, March 2009, + <https://www.rfc-editor.org/info/rfc5475>. + + [RFC5477] Dietz, T., Claise, B., Aitken, P., Dressler, F., and G. + Carle, "Information Model for Packet Sampling Exports", + RFC 5477, DOI 10.17487/RFC5477, March 2009, + <https://www.rfc-editor.org/info/rfc5477>. + + [RFC6035] Pendleton, A., Clark, A., Johnston, A., and H. Sinnreich, + "Session Initiation Protocol Event Package for Voice + Quality Reporting", RFC 6035, DOI 10.17487/RFC6035, + November 2010, <https://www.rfc-editor.org/info/rfc6035>. + + [RFC6248] Morton, A., "RFC 4148 and the IP Performance Metrics + (IPPM) Registry of Metrics Are Obsolete", RFC 6248, + DOI 10.17487/RFC6248, April 2011, + <https://www.rfc-editor.org/info/rfc6248>. + + [RFC6991] Schoenwaelder, J., Ed., "Common YANG Data Types", + RFC 6991, DOI 10.17487/RFC6991, July 2013, + <https://www.rfc-editor.org/info/rfc6991>. + + [RFC7012] Claise, B., Ed. and B. Trammell, Ed., "Information Model + for IP Flow Information Export (IPFIX)", RFC 7012, + DOI 10.17487/RFC7012, September 2013, + <https://www.rfc-editor.org/info/rfc7012>. + + [RFC7014] D'Antonio, S., Zseby, T., Henke, C., and L. Peluso, "Flow + Selection Techniques", RFC 7014, DOI 10.17487/RFC7014, + September 2013, <https://www.rfc-editor.org/info/rfc7014>. + + [RFC7594] Eardley, P., Morton, A., Bagnulo, M., Burbridge, T., + Aitken, P., and A. Akhter, "A Framework for Large-Scale + Measurement of Broadband Performance (LMAP)", RFC 7594, + DOI 10.17487/RFC7594, September 2015, + <https://www.rfc-editor.org/info/rfc7594>. + + [RFC7679] Almes, G., Kalidindi, S., Zekauskas, M., and A. Morton, + Ed., "A One-Way Delay Metric for IP Performance Metrics + (IPPM)", STD 81, RFC 7679, DOI 10.17487/RFC7679, January + 2016, <https://www.rfc-editor.org/info/rfc7679>. + + [RFC8912] Morton, A., Bagnulo, M., Eardley, P., and K. D'Souza, + "Initial Performance Metrics Registry Entries", RFC 8912, + DOI 10.17487/RFC8912, November 2021, + <https://www.rfc-editor.org/info/rfc8912>. + +Acknowledgments + + Thanks to Brian Trammell and Bill Cerveny, IPPM co-chairs during the + development of this memo, for leading several brainstorming sessions + on this topic. Thanks to Barbara Stark and Juergen Schoenwaelder for + the detailed feedback and suggestions. Thanks to Andrew McGregor for + suggestions on metric naming. Thanks to Michelle Cotton for her + early IANA review, and to Amanda Baber for answering questions + related to the presentation of the Registry and accessibility of the + complete template via URL. Thanks to Roni Even for his review and + suggestions to generalize the procedures. Thanks to all of the Area + Directors for their reviews. + +Authors' Addresses + + Marcelo Bagnulo + Universidad Carlos III de Madrid + Av. Universidad 30 + 28911 Leganes Madrid + Spain + + Phone: 34 91 6249500 + Email: marcelo@it.uc3m.es + URI: http://www.it.uc3m.es + + + Benoit Claise + Huawei + + Email: benoit.claise@huawei.com + + + Philip Eardley + BT + Adastral Park, Martlesham Heath + Ipswich + United Kingdom + + Email: philip.eardley@bt.com + + + Al Morton + AT&T Labs + 200 Laurel Avenue South + Middletown, NJ 07748 + United States of America + + Email: acmorton@att.com + + + Aamer Akhter + Consultant + 118 Timber Hitch + Cary, NC + United States of America + + Email: aakhter@gmail.com |