summaryrefslogtreecommitdiff
path: root/doc/rfc/rfc8172.txt
diff options
context:
space:
mode:
authorThomas Voss <mail@thomasvoss.com> 2024-11-27 20:54:24 +0100
committerThomas Voss <mail@thomasvoss.com> 2024-11-27 20:54:24 +0100
commit4bfd864f10b68b71482b35c818559068ef8d5797 (patch)
treee3989f47a7994642eb325063d46e8f08ffa681dc /doc/rfc/rfc8172.txt
parentea76e11061bda059ae9f9ad130a9895cc85607db (diff)
doc: Add RFC documents
Diffstat (limited to 'doc/rfc/rfc8172.txt')
-rw-r--r--doc/rfc/rfc8172.txt843
1 files changed, 843 insertions, 0 deletions
diff --git a/doc/rfc/rfc8172.txt b/doc/rfc/rfc8172.txt
new file mode 100644
index 0000000..d5bafa9
--- /dev/null
+++ b/doc/rfc/rfc8172.txt
@@ -0,0 +1,843 @@
+
+
+
+
+
+
+Internet Engineering Task Force (IETF) A. Morton
+Request for Comments: 8172 AT&T Labs
+Category: Informational July 2017
+ISSN: 2070-1721
+
+
+ Considerations for Benchmarking Virtual Network Functions
+ and Their Infrastructure
+
+Abstract
+
+ The Benchmarking Methodology Working Group has traditionally
+ conducted laboratory characterization of dedicated physical
+ implementations of internetworking functions. This memo investigates
+ additional considerations when network functions are virtualized and
+ performed in general-purpose hardware.
+
+Status of This Memo
+
+ This document is not an Internet Standards Track specification; it is
+ published for informational purposes.
+
+ This document is a product of the Internet Engineering Task Force
+ (IETF). It represents the consensus of the IETF community. It has
+ received public review and has been approved for publication by the
+ Internet Engineering Steering Group (IESG). Not all documents
+ approved by the IESG are a candidate for any level of Internet
+ Standard; see Section 2 of RFC 7841.
+
+ Information about the current status of this document, any errata,
+ and how to provide feedback on it may be obtained at
+ http://www.rfc-editor.org/info/rfc8172.
+
+Copyright Notice
+
+ Copyright (c) 2017 IETF Trust and the persons identified as the
+ document authors. All rights reserved.
+
+ This document is subject to BCP 78 and the IETF Trust's Legal
+ Provisions Relating to IETF Documents
+ (http://trustee.ietf.org/license-info) in effect on the date of
+ publication of this document. Please review these documents
+ carefully, as they describe your rights and restrictions with respect
+ to this document. Code Components extracted from this document must
+ include Simplified BSD License text as described in Section 4.e of
+ the Trust Legal Provisions and are provided without warranty as
+ described in the Simplified BSD License.
+
+
+
+
+Morton Informational [Page 1]
+
+RFC 8172 Benchmarking VNFs and Related Infrastructure July 2017
+
+
+Table of Contents
+
+ 1. Introduction . . . . . . . . . . . . . . . . . . . . . . . . 2
+ 1.1. Requirements Language . . . . . . . . . . . . . . . . . . 3
+ 2. Scope . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4
+ 3. Considerations for Hardware and Testing . . . . . . . . . . . 4
+ 3.1. Hardware Components . . . . . . . . . . . . . . . . . . . 4
+ 3.2. Configuration Parameters . . . . . . . . . . . . . . . . 5
+ 3.3. Testing Strategies . . . . . . . . . . . . . . . . . . . 6
+ 3.4. Attention to Shared Resources . . . . . . . . . . . . . . 7
+ 4. Benchmarking Considerations . . . . . . . . . . . . . . . . . 8
+ 4.1. Comparison with Physical Network Functions . . . . . . . 8
+ 4.2. Continued Emphasis on Black-Box Benchmarks . . . . . . . 8
+ 4.3. New Benchmarks and Related Metrics . . . . . . . . . . . 9
+ 4.4. Assessment of Benchmark Coverage . . . . . . . . . . . . 10
+ 4.5. Power Consumption . . . . . . . . . . . . . . . . . . . . 12
+ 5. Security Considerations . . . . . . . . . . . . . . . . . . . 12
+ 6. IANA Considerations . . . . . . . . . . . . . . . . . . . . . 13
+ 7. References . . . . . . . . . . . . . . . . . . . . . . . . . 13
+ 7.1. Normative References . . . . . . . . . . . . . . . . . . 13
+ 7.2. Informative References . . . . . . . . . . . . . . . . . 14
+ Acknowledgements . . . . . . . . . . . . . . . . . . . . . . . . 15
+ Author's Address . . . . . . . . . . . . . . . . . . . . . . . . 15
+
+1. Introduction
+
+ The Benchmarking Methodology Working Group (BMWG) has traditionally
+ conducted laboratory characterization of dedicated physical
+ implementations of internetworking functions (or physical network
+ functions (PNFs)). The black-box benchmarks of throughput, latency,
+ forwarding rates, and others have served our industry for many years.
+ [RFC1242] and [RFC2544] are the cornerstones of the work.
+
+ A set of service provider and vendor development goals has emerged:
+ reduce costs while increasing flexibility of network devices and
+ drastically reduce deployment time. Network Function Virtualization
+ (NFV) has the promise to achieve these goals and therefore has
+ garnered much attention. It now seems certain that some network
+ functions will be virtualized following the success of cloud
+ computing and virtual desktops supported by sufficient network path
+ capacity, performance, and widespread deployment; many of the same
+ techniques will help achieve NFV.
+
+ In the context of Virtual Network Functions (VNFs), the supporting
+ Infrastructure requires general-purpose computing systems, storage
+ systems, networking systems, virtualization support systems (such as
+ hypervisors), and management systems for the virtual and physical
+ resources. There will be many potential suppliers of Infrastructure
+
+
+
+Morton Informational [Page 2]
+
+RFC 8172 Benchmarking VNFs and Related Infrastructure July 2017
+
+
+ systems and significant flexibility in configuring the systems for
+ best performance. There are also many potential suppliers of VNFs,
+ adding to the combinations possible in this environment. The
+ separation of hardware and software suppliers has a profound
+ implication on benchmarking activities: much more of the internal
+ configuration of the black-box Device Under Test (DUT) must now be
+ specified and reported with the results, to foster both repeatability
+ and comparison testing at a later time.
+
+ Consider the following user story as further background and
+ motivation:
+
+ I'm designing and building my NFV Infrastructure platform. The
+ first steps were easy because I had a small number of categories
+ of VNFs to support and the VNF vendor gave hardware
+ recommendations that I followed. Now I need to deploy more VNFs
+ from new vendors, and there are different hardware
+ recommendations. How well will the new VNFs perform on my
+ existing hardware? Which among several new VNFs in a given
+ category are most efficient in terms of capacity they deliver?
+ And, when I operate multiple categories of VNFs (and PNFs)
+ *concurrently* on a hardware platform such that they share
+ resources, what are the new performance limits, and what are the
+ software design choices I can make to optimize my chosen hardware
+ platform? Conversely, what hardware platform upgrades should I
+ pursue to increase the capacity of these concurrently operating
+ VNFs?
+
+ See <http://www.etsi.org/technologies-clusters/technologies/nfv> for
+ more background; the white papers there may be a useful starting
+ place. The "NFV Performance & Portability Best Practices" document
+ [NFV.PER001] is particularly relevant to BMWG. There are also
+ documents available among the Approved ETSI NFV Specifications
+ [Approved_ETSI_NFV], including documents describing Infrastructure
+ performance aspects and service quality metrics, and drafts in the
+ ETSI NFV Open Area [Draft_ETSI_NFV], which may also have relevance to
+ benchmarking.
+
+1.1. Requirements Language
+
+ The key words "MUST", "MUST NOT", "REQUIRED", "SHALL", "SHALL NOT",
+ "SHOULD", "SHOULD NOT", "RECOMMENDED", "NOT RECOMMENDED", "MAY", and
+ "OPTIONAL" in this document are to be interpreted as described in BCP
+ 14 [RFC2119] [RFC8174] when, and only when, they appear in all
+ capitals, as shown here.
+
+
+
+
+
+
+Morton Informational [Page 3]
+
+RFC 8172 Benchmarking VNFs and Related Infrastructure July 2017
+
+
+2. Scope
+
+ At the time of this writing, BMWG is considering the new topic of
+ Virtual Network Functions and related Infrastructure to ensure that
+ common issues are recognized from the start; background materials
+ from respective standards development organizations and Open Source
+ development projects (e.g., IETF, ETSI NFV, and the Open Platform for
+ Network Function Virtualization (OPNFV)) are being used.
+
+ This memo investigates additional methodological considerations
+ necessary when benchmarking VNFs instantiated and hosted in general-
+ purpose hardware, using bare metal hypervisors [BareMetal] or other
+ isolation environments such as Linux containers. An essential
+ consideration is benchmarking physical and Virtual Network Functions
+ in the same way when possible, thereby allowing direct comparison.
+ Benchmarking combinations of physical and virtual devices and
+ functions in a System Under Test (SUT) is another topic of keen
+ interest.
+
+ A clearly related goal is investigating benchmarks for the capacity
+ of a general-purpose platform to host a plurality of VNF instances.
+ Existing networking technology benchmarks will also be considered for
+ adaptation to NFV and closely associated technologies.
+
+ A non-goal is any overlap with traditional computer benchmark
+ development and their specific metrics (e.g., SPECmark suites such as
+ SPEC CPU).
+
+ A continued non-goal is any form of architecture development related
+ to NFV and associated technologies in BMWG, consistent with all
+ chartered work since BMWG began in 1989.
+
+3. Considerations for Hardware and Testing
+
+ This section lists the new considerations that must be addressed to
+ benchmark VNF(s) and their supporting Infrastructure. The SUT is
+ composed of the hardware platform components, the VNFs installed, and
+ many other supporting systems. It is critical to document all
+ aspects of the SUT to foster repeatability.
+
+3.1. Hardware Components
+
+ The following new hardware components will become part of the test
+ setup:
+
+ 1. High-volume server platforms (general-purpose, possibly with
+ virtual technology enhancements)
+
+
+
+
+Morton Informational [Page 4]
+
+RFC 8172 Benchmarking VNFs and Related Infrastructure July 2017
+
+
+ 2. Storage systems with large capacity, high speed, and high
+ reliability
+
+ 3. Network interface ports specially designed for efficient service
+ of many virtual Network Interface Cards (NICs)
+
+ 4. High-capacity Ethernet switches
+
+ The components above are subjects for development of specialized
+ benchmarks that focus on the special demands of network function
+ deployment.
+
+ Labs conducting comparisons of different VNFs may be able to use the
+ same hardware platform over many studies, until the steady march of
+ innovations overtakes their capabilities (as happens with the lab's
+ traffic generation and testing devices today).
+
+3.2. Configuration Parameters
+
+ It will be necessary to configure and document the settings for the
+ entire general-purpose platform to ensure repeatability and foster
+ future comparisons, including, but clearly not limited to, the
+ following:
+
+ o number of server blades (shelf occupation)
+
+ o CPUs
+
+ o caches
+
+ o memory
+
+ o storage system
+
+ o I/O
+
+ as well as configurations that support the devices that host the VNF
+ itself:
+
+ o Hypervisor (or other forms of virtual function hosting)
+
+ o Virtual Machine (VM)
+
+ o Infrastructure virtual network (which interconnects virtual
+ machines with physical network interfaces or with each other
+ through virtual switches, for example)
+
+
+
+
+
+Morton Informational [Page 5]
+
+RFC 8172 Benchmarking VNFs and Related Infrastructure July 2017
+
+
+ and finally, the VNF itself, with items such as:
+
+ o specific function being implemented in VNF
+
+ o reserved resources for each function (e.g., CPU pinning and Non-
+ Uniform Memory Access (NUMA) node assignment)
+
+ o number of VNFs (or sub-VNF components, each with its own VM) in
+ the service function chain (see Section 1.1 of [RFC7498] for a
+ definition of service function chain)
+
+ o number of physical interfaces and links transited in the service
+ function chain
+
+ In the physical device benchmarking context, most of the
+ corresponding Infrastructure configuration choices were determined by
+ the vendor. Although the platform itself is now one of the
+ configuration variables, it is important to maintain emphasis on the
+ networking benchmarks and capture the platform variables as input
+ factors.
+
+3.3. Testing Strategies
+
+ The concept of characterizing performance at capacity limits may
+ change. For example:
+
+ 1. It may be more representative of system capacity to characterize
+ the case where the VMs hosting the VNFs are operating at 50%
+ utilization and therefore sharing the "real" processing power
+ across many VMs.
+
+ 2. Another important test case stems from the need to partition (or
+ isolate) network functions. A noisy neighbor (VM hosting a VNF
+ in an infinite loop) would ideally be isolated; the performance
+ of other VMs would continue according to their specifications,
+ and tests would evaluate the degree of isolation.
+
+ 3. System errors will likely occur as transients, implying a
+ distribution of performance characteristics with a long tail
+ (like latency) and leading to the need for longer-term tests of
+ each set of configuration and test parameters.
+
+ 4. The desire for elasticity and flexibility among network functions
+ will include tests where there is constant flux in the number of
+ VM instances, the resources the VMs require, and the setup/
+ teardown of network paths that support VM connectivity. Requests
+ for and instantiation of new VMs, along with releases for VMs
+ hosting VNFs that are no longer needed, would be a normal
+
+
+
+Morton Informational [Page 6]
+
+RFC 8172 Benchmarking VNFs and Related Infrastructure July 2017
+
+
+ operational condition. In other words, benchmarking should
+ include scenarios with production life-cycle management of VMs
+ and their VNFs and network connectivity in progress, including
+ VNF scaling up/down operations, as well as static configurations.
+
+ 5. All physical things can fail, and benchmarking efforts can also
+ examine recovery aided by the virtual architecture with different
+ approaches to resiliency.
+
+ 6. The sheer number of test conditions and configuration
+ combinations encourage increased efficiency, including automated
+ testing arrangements, combination sub-sampling through an
+ understanding of inter-relationships, and machine-readable test
+ results.
+
+3.4. Attention to Shared Resources
+
+ Since many components of the new NFV Infrastructure are virtual, test
+ setup design must have prior knowledge of interactions/dependencies
+ within the various resource domains in the SUT. For example, a
+ virtual machine performing the role of a traditional tester function,
+ such as generating and/or receiving traffic, should avoid sharing any
+ SUT resources with the DUT. Otherwise, the results will have
+ unexpected dependencies not encountered in physical device
+ benchmarking.
+
+ Note that the term "tester" has traditionally referred to devices
+ dedicated to testing in BMWG literature. In this new context,
+ "tester" additionally refers to functions dedicated to testing, which
+ may be either virtual or physical. "Tester" has never referred to
+ the individuals performing the tests.
+
+ The possibility to use shared resources in test design while
+ producing useful results remains one of the critical challenges to
+ overcome. Benchmarking setups may designate isolated resources for
+ the DUT and other critical support components (such as the host/
+ kernel) as the first baseline step and add other loading processes.
+ The added complexity of each setup leads to shared-resource testing
+ scenarios, where the characteristics of the competing load (in terms
+ of memory, storage, and CPU utilization) will directly affect the
+ benchmarking results (and variability of the results), but the
+ results should reconcile with the baseline.
+
+ The physical test device remains a solid foundation to compare with
+ results using combinations of physical and virtual test functions or
+ results using only virtual testers when necessary to assess virtual
+ interfaces and other virtual functions.
+
+
+
+
+Morton Informational [Page 7]
+
+RFC 8172 Benchmarking VNFs and Related Infrastructure July 2017
+
+
+4. Benchmarking Considerations
+
+ This section discusses considerations related to benchmarks
+ applicable to VNFs and their associated technologies.
+
+4.1. Comparison with Physical Network Functions
+
+ In order to compare the performance of VNFs and system
+ implementations with their physical counterparts, identical
+ benchmarks must be used. Since BMWG has already developed
+ specifications for many network functions, there will be re-use of
+ existing benchmarks through references, while allowing for the
+ possibility of benchmark curation during development of new
+ methodologies. Consideration should be given to quantifying the
+ number of parallel VNFs required to achieve comparable scale/capacity
+ with a given physical device or whether some limit of scale was
+ reached before the VNFs could achieve the comparable level. Again,
+ implementation based on different hypervisors or other virtual
+ function hosting remain as critical factors in performance
+ assessment.
+
+4.2. Continued Emphasis on Black-Box Benchmarks
+
+ When the network functions under test are based on open-source code,
+ there may be a tendency to rely on internal measurements to some
+ extent, especially when the externally observable phenomena only
+ support an inference of internal events (such as routing protocol
+ convergence observed in the data plane). Examples include CPU/Core
+ utilization, network utilization, storage utilization, and memory
+ committed/used. These "white-box" metrics provide one view of the
+ resource footprint of a VNF. Note that the resource utilization
+ metrics do not easily match the 3x4 Matrix, described in Section 4.4.
+
+ However, external observations remain essential as the basis for
+ benchmarks. Internal observations with fixed specification and
+ interpretation may be provided in parallel (as auxiliary metrics), to
+ assist the development of operations procedures when the technology
+ is deployed, for example. Internal metrics and measurements from
+ open-source implementations may be the only direct source of
+ performance results in a desired dimension, but corroborating
+ external observations are still required to assure the integrity of
+ measurement discipline was maintained for all reported results.
+
+ A related aspect of benchmark development is where the scope includes
+ multiple approaches to a common function under the same benchmark.
+ For example, there are many ways to arrange for activation of a
+ network path between interface points, and the activation times can
+ be compared if the start-to-stop activation interval has a generic
+
+
+
+Morton Informational [Page 8]
+
+RFC 8172 Benchmarking VNFs and Related Infrastructure July 2017
+
+
+ and unambiguous definition. Thus, generic benchmark definitions are
+ preferred over technology/protocol-specific definitions where
+ possible.
+
+4.3. New Benchmarks and Related Metrics
+
+ There will be new classes of benchmarks needed for network design and
+ assistance when developing operational practices (possibly automated
+ management and orchestration of deployment scale). Examples follow
+ in the paragraphs below, many of which are prompted by the goals of
+ increased elasticity and flexibility of the network functions, along
+ with reduced deployment times.
+
+ o Time to deploy VNFs: In cases where the general-purpose hardware
+ is already deployed and ready for service, it is valuable to know
+ the response time when a management system is tasked with
+ "standing up" 100s of virtual machines and the VNFs they will
+ host.
+
+ o Time to migrate VNFs: In cases where a rack or shelf of hardware
+ must be removed from active service, it is valuable to know the
+ response time when a management system is tasked with "migrating"
+ some number of virtual machines and the VNFs they currently host
+ to alternate hardware that will remain in service.
+
+ o Time to create a virtual network in the general-purpose
+ Infrastructure: This is a somewhat simplified version of existing
+ benchmarks for convergence time, in that the process is initiated
+ by a request from (centralized or distributed) control, rather
+ than inferred from network events (link failure). The successful
+ response time would remain dependent on data-plane observations to
+ confirm that the network is ready to perform.
+
+ o Effect of verification measurements on performance: A complete
+ VNF, or something as simple as a new policy to implement in a VNF,
+ is implemented. The action to verify instantiation of the VNF or
+ policy could affect performance during normal operation.
+
+ Also, it appears to be valuable to measure traditional packet
+ transfer performance metrics during the assessment of traditional and
+ new benchmarks, including metrics that may be used to support service
+ engineering such as the spatial composition metrics found in
+ [RFC6049]. Examples include mean one-way delay in Section 4.1 of
+ [RFC6049], Packet Delay Variation (PDV) in [RFC5481], and Packet
+ Reordering [RFC4737] [RFC4689].
+
+
+
+
+
+
+Morton Informational [Page 9]
+
+RFC 8172 Benchmarking VNFs and Related Infrastructure July 2017
+
+
+4.4. Assessment of Benchmark Coverage
+
+ It can be useful to organize benchmarks according to their applicable
+ life-cycle stage and the performance criteria they were designed to
+ assess. The table below (derived from [X3.102]) provides a way to
+ organize benchmarks such that there is a clear indication of coverage
+ for the intersection of life-cycle stages and performance criteria.
+
+ |----------------------------------------------------------|
+ | | | | |
+ | | SPEED | ACCURACY | RELIABILITY |
+ | | | | |
+ |----------------------------------------------------------|
+ | | | | |
+ | Activation | | | |
+ | | | | |
+ |----------------------------------------------------------|
+ | | | | |
+ | Operation | | | |
+ | | | | |
+ |----------------------------------------------------------|
+ | | | | |
+ | De-activation | | | |
+ | | | | |
+ |----------------------------------------------------------|
+
+ For example, the "Time to deploy VNFs" benchmark described above
+ would be placed in the intersection of Activation and Speed, making
+ it clear that there are other potential performance criteria to
+ benchmark, such as the "percentage of unsuccessful VM/VNF stand-ups"
+ in a set of 100 attempts. This example emphasizes that the
+ Activation and De-activation life-cycle stages are key areas for NFV
+ and related Infrastructure and encourages expansion beyond
+ traditional benchmarks for normal operation. Thus, reviewing the
+ benchmark coverage using this table (sometimes called the 3x3 Matrix)
+ can be a worthwhile exercise in BMWG.
+
+ In one of the first applications of the 3x3 Matrix in BMWG
+ [SDN-BENCHMARK], we discovered that metrics on measured size,
+ capacity, or scale do not easily match one of the three columns
+ above. Following discussion, this was resolved in two ways:
+
+ o Add a column, Scale, for use when categorizing and assessing the
+ coverage of benchmarks (without measured results). An example of
+ this use is found in [OPNFV-BENCHMARK] (and a variation may be
+ found in [SDN-BENCHMARK]). This is the 3x4 Matrix.
+
+
+
+
+
+Morton Informational [Page 10]
+
+RFC 8172 Benchmarking VNFs and Related Infrastructure July 2017
+
+
+ o If using the matrix to report results in an organized way, keep
+ size, capacity, and scale metrics separate from the 3x3 Matrix and
+ incorporate them in the report with other qualifications of the
+ results.
+
+ Note that the resource utilization (e.g., CPU) metrics do not fit in
+ the matrix. They are not benchmarks, and omitting them confirms
+ their status as auxiliary metrics. Resource assignments are
+ configuration parameters, and these are reported separately.
+
+ This approach encourages use of the 3x3 Matrix to organize reports of
+ results, where the capacity at which the various metrics were
+ measured could be included in the title of the matrix (and results
+ for multiple capacities would result in separate 3x3 Matrices, if
+ there were sufficient measurements/results to organize in that way).
+
+ For example, results for each VM and VNF could appear in the 3x3
+ Matrix, organized to illustrate resource occupation (CPU Cores) in a
+ particular physical computing system, as shown below.
+
+ VNF#1
+ .-----------.
+ |__|__|__|__|
+ Core 1 |__|__|__|__|
+ |__|__|__|__|
+ | | | | |
+ '-----------'
+ VNF#2
+ .-----------.
+ |__|__|__|__|
+ Cores 2-5 |__|__|__|__|
+ |__|__|__|__|
+ | | | | |
+ '-----------'
+ VNF#3 VNF#4 VNF#5
+ .-----------. .-----------. .-----------.
+ |__|__|__|__| |__|__|__|__| |__|__|__|__|
+ Core 6 |__|__|__|__| |__|__|__|__| |__|__|__|__|
+ |__|__|__|__| |__|__|__|__| |__|__|__|__|
+ | | | | | | | | | | | | | | |
+ '-----------' '-----------' '-----------'
+ VNF#6
+ .-----------.
+ |__|__|__|__|
+ Core 7 |__|__|__|__|
+ |__|__|__|__|
+ | | | | |
+ '-----------'
+
+
+
+Morton Informational [Page 11]
+
+RFC 8172 Benchmarking VNFs and Related Infrastructure July 2017
+
+
+ The combination of tables above could be built incrementally,
+ beginning with VNF#1 and one Core, then adding VNFs according to
+ their supporting Core assignments. X-Y plots of critical benchmarks
+ would also provide insight to the effect of increased hardware
+ utilization. All VNFs might be of the same type, or to match a
+ production environment, there could be VNFs of multiple types and
+ categories. In this figure, VNFs #3-#5 are assumed to require small
+ CPU resources, while VNF#2 requires four Cores to perform its
+ function.
+
+4.5. Power Consumption
+
+ Although there is incomplete work to benchmark physical network
+ function power consumption in a meaningful way, the desire to measure
+ the physical Infrastructure supporting the virtual functions only
+ adds to the need. Both maximum power consumption and dynamic power
+ consumption (with varying load) would be useful. The Intelligent
+ Platform Management Interface (IPMI) standard [IPMI2.0] has been
+ implemented by many manufacturers and supports measurement of
+ instantaneous energy consumption.
+
+ To assess the instantaneous energy consumption of virtual resources,
+ it may be possible to estimate the value using an overall metric
+ based on utilization readings, according to [NFVIaas-FRAMEWORK].
+
+5. Security Considerations
+
+ Benchmarking activities as described in this memo are limited to
+ technology characterization of a DUT/SUT using controlled stimuli in
+ a laboratory environment, with dedicated address space and the
+ constraints specified in the sections above.
+
+ The benchmarking network topology will be an independent test setup
+ and MUST NOT be connected to devices that may forward the test
+ traffic into a production network or misroute traffic to the test
+ management network.
+
+ Further, benchmarking is performed on a "black-box" basis, relying
+ solely on measurements observable external to the DUT/SUT.
+
+ Special capabilities SHOULD NOT exist in the DUT/SUT specifically for
+ benchmarking purposes. Any implications for network security arising
+ from the DUT/SUT SHOULD be identical in the lab and in production
+ networks.
+
+
+
+
+
+
+
+Morton Informational [Page 12]
+
+RFC 8172 Benchmarking VNFs and Related Infrastructure July 2017
+
+
+6. IANA Considerations
+
+ This document does not require any IANA actions.
+
+7. References
+
+7.1. Normative References
+
+ [NFV.PER001]
+ ETSI, "Network Function Virtualization: Performance &
+ Portability Best Practices", ETSI GS NFV-PER 001, V1.1.2,
+ December 2014.
+
+ [RFC2119] Bradner, S., "Key words for use in RFCs to Indicate
+ Requirement Levels", BCP 14, RFC 2119,
+ DOI 10.17487/RFC2119, March 1997,
+ <http://www.rfc-editor.org/info/rfc2119>.
+
+ [RFC2544] Bradner, S. and J. McQuaid, "Benchmarking Methodology for
+ Network Interconnect Devices", RFC 2544,
+ DOI 10.17487/RFC2544, March 1999,
+ <http://www.rfc-editor.org/info/rfc2544>.
+
+ [RFC4689] Poretsky, S., Perser, J., Erramilli, S., and S. Khurana,
+ "Terminology for Benchmarking Network-layer Traffic
+ Control Mechanisms", RFC 4689, DOI 10.17487/RFC4689,
+ October 2006, <http://www.rfc-editor.org/info/rfc4689>.
+
+ [RFC4737] Morton, A., Ciavattone, L., Ramachandran, G., Shalunov,
+ S., and J. Perser, "Packet Reordering Metrics", RFC 4737,
+ DOI 10.17487/RFC4737, November 2006,
+ <http://www.rfc-editor.org/info/rfc4737>.
+
+ [RFC7498] Quinn, P., Ed. and T. Nadeau, Ed., "Problem Statement for
+ Service Function Chaining", RFC 7498,
+ DOI 10.17487/RFC7498, April 2015,
+ <http://www.rfc-editor.org/info/rfc7498>.
+
+ [RFC8174] Leiba, B., "Ambiguity of Uppercase vs Lowercase in RFC
+ 2119 Key Words", BCP 14, RFC 8174, DOI 10.17487/RFC8174,
+ May 2017, <http://www.rfc-editor.org/info/rfc8174>.
+
+
+
+
+
+
+
+
+
+
+Morton Informational [Page 13]
+
+RFC 8172 Benchmarking VNFs and Related Infrastructure July 2017
+
+
+7.2. Informative References
+
+ [Approved_ETSI_NFV]
+ ETSI, Network Functions Virtualisation Technical
+ Committee, "ETSI NFV",
+ <http://www.etsi.org/standards-search>.
+
+ [BareMetal]
+ Popek, G. and R. Goldberg, "Formal requirements for
+ virtualizable third generation architectures",
+ Communications of the ACM, Volume 17, Issue 7, Pages
+ 412-421, DOI 10.1145/361011.361073, July 1974.
+
+ [Draft_ETSI_NFV]
+ ETSI, "Network Functions Virtualisation: Specifications",
+ <http://www.etsi.org/technologies-clusters/technologies/
+ nfv>.
+
+ [IPMI2.0] Intel Corporation, Hewlett-Packard Company, NEC
+ Corporation, and Dell Inc., "Intelligent Platform
+ Management Interface Specification v2.0 (with latest
+ errata)", April 2015,
+ <http://www.intel.com/content/dam/www/public/us/en/
+ documents/specification-updates/ipmi-intelligent-platform-
+ mgt-interface-spec-2nd-gen-v2-0-spec-update.pdf>.
+
+ [NFVIaas-FRAMEWORK]
+ Krishnan, R., Figueira, N., Krishnaswamy, D., Lopez, D.,
+ Wright, S., Hinrichs, T., Krishnaswamy, R., and A. Yerra,
+ "NFVIaaS Architectural Framework for Policy Based Resource
+ Placement and Scheduling", Work in Progress,
+ draft-krishnan-nfvrg-policy-based-rm-nfviaas-06, March
+ 2016.
+
+ [OPNFV-BENCHMARK]
+ Tahhan, M., O'Mahony, B., and A. Morton, "Benchmarking
+ Virtual Switches in OPNFV", Work in Progress,
+ draft-ietf-bmwg-vswitch-opnfv-04, June 2017.
+
+ [RFC1242] Bradner, S., "Benchmarking Terminology for Network
+ Interconnection Devices", RFC 1242, DOI 10.17487/RFC1242,
+ July 1991, <http://www.rfc-editor.org/info/rfc1242>.
+
+ [RFC5481] Morton, A. and B. Claise, "Packet Delay Variation
+ Applicability Statement", RFC 5481, DOI 10.17487/RFC5481,
+ March 2009, <http://www.rfc-editor.org/info/rfc5481>.
+
+
+
+
+
+Morton Informational [Page 14]
+
+RFC 8172 Benchmarking VNFs and Related Infrastructure July 2017
+
+
+ [RFC6049] Morton, A. and E. Stephan, "Spatial Composition of
+ Metrics", RFC 6049, DOI 10.17487/RFC6049, January 2011,
+ <http://www.rfc-editor.org/info/rfc6049>.
+
+ [SDN-BENCHMARK]
+ Vengainathan, B., Basil, A., Tassinari, M., Manral, V.,
+ and S. Banks, "Terminology for Benchmarking SDN Controller
+ Performance", Work in Progress, draft-ietf-bmwg-sdn-
+ controller-benchmark-term-04, June 2017.
+
+ [X3.102] ANSI, "Information Systems - Data Communication Systems
+ and Services - User-Oriented Performance Parameters
+ Communications Framework", ANSI X3.102, 1983.
+
+Acknowledgements
+
+ The author acknowledges an encouraging conversation on this topic
+ with Mukhtiar Shaikh and Ramki Krishnan in November 2013. Bhavani
+ Parise and Ilya Varlashkin have provided useful suggestions to expand
+ these considerations. Bhuvaneswaran Vengainathan has already tried
+ the 3x3 Matrix with the SDN controller document and contributed to
+ many discussions. Scott Bradner quickly pointed out shared resource
+ dependencies in an early vSwitch measurement proposal, and the topic
+ was included here as a key consideration. Further development was
+ encouraged by Barry Constantine's comments following the BMWG session
+ at IETF 92: the session itself was an affirmation for this memo.
+ There have been many interesting contributions from Maryam Tahhan,
+ Marius Georgescu, Jacob Rapp, Saurabh Chattopadhyay, and others.
+
+Author's Address
+
+ Al Morton
+ AT&T Labs
+ 200 Laurel Avenue South
+ Middletown, NJ 07748
+ United States of America
+
+ Phone: +1 732 420 1571
+ Fax: +1 732 368 1192
+ Email: acmorton@att.com
+
+
+
+
+
+
+
+
+
+
+
+Morton Informational [Page 15]
+