summaryrefslogtreecommitdiff
path: root/doc/rfc/rfc8735.txt
diff options
context:
space:
mode:
authorThomas Voss <mail@thomasvoss.com> 2024-11-27 20:54:24 +0100
committerThomas Voss <mail@thomasvoss.com> 2024-11-27 20:54:24 +0100
commit4bfd864f10b68b71482b35c818559068ef8d5797 (patch)
treee3989f47a7994642eb325063d46e8f08ffa681dc /doc/rfc/rfc8735.txt
parentea76e11061bda059ae9f9ad130a9895cc85607db (diff)
doc: Add RFC documents
Diffstat (limited to 'doc/rfc/rfc8735.txt')
-rw-r--r--doc/rfc/rfc8735.txt766
1 files changed, 766 insertions, 0 deletions
diff --git a/doc/rfc/rfc8735.txt b/doc/rfc/rfc8735.txt
new file mode 100644
index 0000000..0afe78a
--- /dev/null
+++ b/doc/rfc/rfc8735.txt
@@ -0,0 +1,766 @@
+
+
+
+
+Internet Engineering Task Force (IETF) A. Wang
+Request for Comments: 8735 China Telecom
+Category: Informational X. Huang
+ISSN: 2070-1721 C. Kou
+ BUPT
+ Z. Li
+ China Mobile
+ P. Mi
+ Huawei Technologies
+ February 2020
+
+
+ Scenarios and Simulation Results of PCE in a Native IP Network
+
+Abstract
+
+ Requirements for providing the End-to-End (E2E) performance assurance
+ are emerging within the service provider networks. While there are
+ various technology solutions, there is no single solution that can
+ fulfill these requirements for a native IP network. In particular,
+ there is a need for a universal E2E solution that can cover both
+ intra- and inter-domain scenarios.
+
+ One feasible E2E traffic-engineering solution is the addition of
+ central control in a native IP network. This document describes
+ various complex scenarios and simulation results when applying the
+ Path Computation Element (PCE) in a native IP network. This
+ solution, referred to as Centralized Control Dynamic Routing (CCDR),
+ integrates the advantage of using distributed protocols and the power
+ of a centralized control technology, providing traffic engineering
+ for native IP networks in a manner that applies equally to intra- and
+ inter-domain scenarios.
+
+Status of This Memo
+
+ This document is not an Internet Standards Track specification; it is
+ published for informational purposes.
+
+ This document is a product of the Internet Engineering Task Force
+ (IETF). It represents the consensus of the IETF community. It has
+ received public review and has been approved for publication by the
+ Internet Engineering Steering Group (IESG). Not all documents
+ approved by the IESG are candidates for any level of Internet
+ Standard; see Section 2 of RFC 7841.
+
+ Information about the current status of this document, any errata,
+ and how to provide feedback on it may be obtained at
+ https://www.rfc-editor.org/info/rfc8735.
+
+Copyright Notice
+
+ Copyright (c) 2020 IETF Trust and the persons identified as the
+ document authors. All rights reserved.
+
+ This document is subject to BCP 78 and the IETF Trust's Legal
+ Provisions Relating to IETF Documents
+ (https://trustee.ietf.org/license-info) in effect on the date of
+ publication of this document. Please review these documents
+ carefully, as they describe your rights and restrictions with respect
+ to this document. Code Components extracted from this document must
+ include Simplified BSD License text as described in Section 4.e of
+ the Trust Legal Provisions and are provided without warranty as
+ described in the Simplified BSD License.
+
+Table of Contents
+
+ 1. Introduction
+ 2. Terminology
+ 3. CCDR Scenarios
+ 3.1. QoS Assurance for Hybrid Cloud-Based Application
+ 3.2. Link Utilization Maximization
+ 3.3. Traffic Engineering for Multi-domain
+ 3.4. Network Temporal Congestion Elimination
+ 4. CCDR Simulation
+ 4.1. Case Study for CCDR Algorithm
+ 4.2. Topology Simulation
+ 4.3. Traffic Matrix Simulation
+ 4.4. CCDR End-to-End Path Optimization
+ 4.5. Network Temporal Congestion Elimination
+ 5. CCDR Deployment Consideration
+ 6. Security Considerations
+ 7. IANA Considerations
+ 8. References
+ 8.1. Normative References
+ 8.2. Informative References
+ Acknowledgements
+ Contributors
+ Authors' Addresses
+
+1. Introduction
+
+ A service provider network is composed of thousands of routers that
+ run distributed protocols to exchange reachability information. The
+ path for the destination network is mainly calculated, and
+ controlled, by the distributed protocols. These distributed
+ protocols are robust enough to support most applications; however,
+ they have some difficulties supporting the complexities needed for
+ traffic-engineering applications, e.g., E2E performance assurance, or
+ maximizing the link utilization within an IP network.
+
+ Multiprotocol Label Switching (MPLS) using Traffic-Engineering (TE)
+ technology (MPLS-TE) [RFC3209] is one solution for TE networks, but
+ it introduces an MPLS network along with related technology, which
+ would be an overlay of the IP network. MPLS-TE technology is often
+ used for Label Switched Path (LSP) protection and setting up complex
+ paths within a domain. It has not been widely deployed for meeting
+ E2E (especially in inter-domain) dynamic performance assurance
+ requirements for an IP network.
+
+ Segment Routing [RFC8402] is another solution that integrates some
+ advantages of using a distributed protocol and central control
+ technology, but it requires the underlying network, especially the
+ provider edge router, to do an in-depth label push and pop action
+ while adding complexity when coexisting with the non-segment routing
+ network. Additionally, it can only maneuver the E2E paths for MPLS
+ and IPv6 traffic via different mechanisms.
+
+ Deterministic Networking (DetNet) [RFC8578] is another possible
+ solution. It is primarily focused on providing bounded latency for a
+ flow and introduces additional requirements on the domain edge
+ router. The current DetNet scope is within one domain. The use
+ cases defined in this document do not require the additional
+ complexity of deterministic properties and so differ from the DetNet
+ use cases.
+
+ This document describes several scenarios for a native IP network
+ where a Centralized Control Dynamic Routing (CCDR) framework can
+ produce qualitative improvement in efficiency without requiring a
+ change to the data-plane behavior on the router. Using knowledge of
+ the Border Gateway Protocol (BGP) session-specific prefixes
+ advertised by a router, the network topology and the near-real-time
+ link-utilization information from network management systems, a
+ central PCE is able to compute an optimal path and give the
+ underlying routers the destination address to use to reach the BGP
+ nexthop, such that the distributed routing protocol will use the
+ computed path via traditional recursive lookup procedure. Some
+ results from simulations of path optimization are also presented to
+ concretely illustrate a variety of scenarios where CCDR shows
+ significant improvement over traditional distributed routing
+ protocols.
+
+ This document is the base document of the following two documents:
+ the universal solution document, which is suitable for intra-domain
+ and inter-domain TE scenario, is described in [PCE-NATIVE-IP]; and
+ the related protocol extension contents is described in
+ [PCEP-NATIVE-IP-EXT].
+
+2. Terminology
+
+ In this document, PCE is used as defined in [RFC5440]. The following
+ terms are used as described here:
+
+ BRAS: Broadband Remote Access Server
+
+ CD: Congestion Degree
+
+ CR: Core Router
+
+ CCDR: Centralized Control Dynamic Routing
+
+ E2E: End to End
+
+ IDC: Internet Data Center
+
+ MAN: Metro Area Network
+
+ QoS: Quality of Service
+
+ SR: Service Router
+
+ TE: Traffic Engineering
+
+ UID: Utilization Increment Degree
+
+ WAN: Wide Area Network
+
+3. CCDR Scenarios
+
+ The following sections describe various deployment scenarios where
+ applying the CCDR framework is intuitively expected to produce
+ improvements based on the macro-scale properties of the framework and
+ the scenario.
+
+3.1. QoS Assurance for Hybrid Cloud-Based Application
+
+ With the emergence of cloud computing technologies, enterprises are
+ putting more and more services on a public-oriented cloud environment
+ while keeping core business within their private cloud. The
+ communication between the private and public cloud sites spans the
+ WAN. The bandwidth requirements between them are variable, and the
+ background traffic between these two sites varies over time.
+ Enterprise applications require assurance of the E2E QoS performance
+ on demand for variable bandwidth services.
+
+ CCDR, which integrates the merits of distributed protocols and the
+ power of centralized control, is suitable for this scenario. The
+ possible solution framework is illustrated below:
+
+ +------------------------+
+ | Cloud-Based Application|
+ +------------------------+
+ |
+ +-----------+
+ | PCE |
+ +-----------+
+ |
+ |
+ //--------------\\
+ ///// \\\\\
+ Private Cloud Site || Distributed |Public Cloud Site
+ | Control Network |
+ \\\\\ /////
+ \\--------------//
+
+ Figure 1: Hybrid Cloud Communication Scenario
+
+ As illustrated in Figure 1, the source and destination of the "Cloud-
+ Based Application" traffic are located at "Private Cloud Site" and
+ "Public Cloud Site", respectively.
+
+ By default, the traffic path between the private and public cloud
+ site is determined by the distributed control network. When an
+ application requires E2E QoS assurance, it can send these
+ requirements to the PCE and let the PCE compute one E2E path, which
+ is based on the underlying network topology and real traffic
+ information, in order to accommodate the application's QoS
+ requirements. Section 4.4 of this document describes the simulation
+ results for this use case.
+
+3.2. Link Utilization Maximization
+
+ Network topology within a Metro Area Network (MAN) is generally in a
+ star mode as illustrated in Figure 2, with different devices
+ connected to different customer types. The traffic from these
+ customers is often in a tidal pattern with the links between the Core
+ Router (CR) / Broadband Remote Access Server (BRAS) and CR/Service
+ Router (SR) experiencing congestion in different periods due to
+ subscribers under BRAS often using the network at night and the
+ leased line users under SR often using the network during the
+ daytime. The link between BRAS/SR and CR must satisfy the maximum
+ traffic volume between them, respectively, which causes these links
+ to often be underutilized.
+
+ +--------+
+ | CR |
+ +----|---+
+ |
+ |-------|--------|-------|
+ | | | |
+ +--|-+ +-|+ +--|-+ +-|+
+ |BRAS| |SR| |BRAS| |SR|
+ +----+ +--+ +----+ +--+
+
+ Figure 2: Star-Mode Network Topology within MAN
+
+ If we consider connecting the BRAS/SR with a local link loop (which
+ is usually lower cost) and control the overall MAN topology with the
+ CCDR framework, we can exploit the tidal phenomena between the BRAS/
+ CR and SR/CR links, maximizing the utilization of these central trunk
+ links (which are usually higher cost than the local loops).
+
+ +-------+
+ ----- PCE |
+ | +-------+
+ +----|---+
+ | CR |
+ +----|---+
+ |
+ |-------|--------|-------|
+ | | | |
+ +--|-+ +-|+ +--|-+ +-|+
+ |BRAS-----SR| |BRAS-----SR|
+ +----+ +--+ +----+ +--+
+
+ Figure 3: Link Utilization Maximization via CCDR
+
+3.3. Traffic Engineering for Multi-domain
+
+ Service provider networks are often comprised of different domains,
+ interconnected with each other, forming a very complex topology as
+ illustrated in Figure 4. Due to the traffic pattern to/from the MAN
+ and IDC, the utilization of the links between them are often
+ asymmetric. It is almost impossible to balance the utilization of
+ these links via a distributed protocol, but this unbalance can be
+ overcome utilizing the CCDR framework.
+
+ +---+ +---+
+ |MAN|----------------|IDC|
+ +---+ | +---+
+ | ---------- |
+ |-----|Backbone|-----|
+ | ----|----- |
+ | | |
+ +---+ | +---+
+ |IDC|----------------|MAN|
+ +---+ +---+
+
+ Figure 4: Traffic Engineering for Complex Multi-domain Topology
+
+ A solution for this scenario requires the gathering of NetFlow
+ information, analysis of the source/destination autonomous system
+ (AS), and determining what the main cause of the congested link(s)
+ is. After this, the operator can use the external Border Gateway
+ Protocol (eBGP) sessions to schedule the traffic among the different
+ domains according to the solution described in the CCDR framework.
+
+3.4. Network Temporal Congestion Elimination
+
+ In more general situations, there is often temporal congestion within
+ the service provider's network, for example, due to daily or weekly
+ periodic bursts or large events that are scheduled well in advance.
+ Such congestion phenomena often appear regularly, and if the service
+ provider has methods to mitigate it, it will certainly improve their
+ network operation capabilities and increase satisfaction for
+ customers. CCDR is also suitable for such scenarios, as the
+ controller can schedule traffic out of the congested links, lowering
+ their utilization during these times. Section 4.5 describes the
+ simulation results of this scenario.
+
+4. CCDR Simulation
+
+ The following sections describe a specific case study to illustrate
+ the workings of the CCDR algorithm with concrete paths/metrics, as
+ well as a procedure for generating topology and traffic matrices and
+ the results from simulations applying CCDR for E2E QoS (assured path
+ and congestion elimination) over the generated topologies and traffic
+ matrices. In all cases examined, the CCDR algorithm produces
+ qualitatively significant improvement over the reference (OSPF)
+ algorithm, suggesting that CCDR will have broad applicability.
+
+ The structure and scale of the simulated topology is similar to that
+ of the real networks. Multiple different traffic matrices were
+ generated to simulate different congestion conditions in the network.
+ Only one of them is illustrated since the others produce similar
+ results.
+
+4.1. Case Study for CCDR Algorithm
+
+ In this section, we consider a specific network topology for case
+ study: examining the path selected by OSPF and CCDR and evaluating
+ how and why the paths differ. Figure 5 depicts the topology of the
+ network in this case. There are eight forwarding devices in the
+ network. The original cost and utilization are marked on it as shown
+ in the figure. For example, the original cost and utilization for
+ the link (1, 2) are 3 and 50%, respectively. There are two flows: f1
+ and f2. Both of these two flows are from node 1 to node 8. For
+ simplicity, it is assumed that the bandwidth of the link in the
+ network is 10 Mb/s. The flow rate of f1 is 1 Mb/s and the flow rate
+ of f2 is 2 Mb/s. The threshold of the link in congestion is 90%.
+
+ If the OSPF protocol, which adopts Dijkstra's algorithm (IS-IS is
+ similar because it also uses Dijkstra's algorithm), is applied in the
+ network, the two flows from node 1 to node 8 can only use the OSPF
+ path (p1: 1->2->3->8). This is because Dijkstra's algorithm mainly
+ considers the original cost of the link. Since CCDR considers cost
+ and utilization simultaneously, the same path as OSPF will not be
+ selected due to the severe congestion of the link (2, 3). In this
+ case, f1 will select the path (p2: 1->5->6->7->8) since the new cost
+ of this path is better than that of the OSPF path. Moreover, the
+ path p2 is also better than the path (p3: 1->2->4->7->8) for flow f1.
+ However, f2 will not select the same path since it will cause new
+ congestion in the link (6, 7). As a result, f2 will select the path
+ (p3: 1->2->4->7->8).
+
+
+ +----+ f1 +-------> +-----+ ----> +-----+
+ |Edge|-----------+ |+--------| 3 |-------| 8 |
+ |Node|---------+ | ||+-----> +-----+ ----> +-----+
+ +----+ | | 4/95%||| 6/50% |
+ | | ||| 5/60%|
+ | v ||| |
+ +----+ +-----+ -----> +-----+ +-----+ +-----+
+ |Edge|-------| 1 |--------| 2 |------| 4 |------| 7 |
+ |Node|-----> +-----+ -----> +-----+7/60% +-----+5/45% +-----+
+ +----+ f2 | 3/50% |
+ | |
+ | 3/60% +-----+ 5/55%+-----+ 3/75% |
+ +-----------| 5 |------| 6 |---------+
+ +-----+ +-----+
+ (a) Dijkstra's Algorithm (OSPF/IS-IS)
+
+
+ +----+ f1 +-----+ ----> +-----+
+ |Edge|-----------+ +--------| 3 |-------| 8 |
+ |Node|---------+ | | +-----+ ----> +-----+
+ +----+ | | 4/95% | 6/50% ^|^
+ | | | 5/60%|||
+ | v | |||
+ +----+ +-----+ -----> +-----+ ---> +-----+ ---> +-----+
+ |Edge|-------| 1 |--------| 2 |------| 4 |------| 7 |
+ |Node|-----> +-----+ +-----+7/60% +-----+5/45% +-----+
+ +----+ f2 || 3/50% |^
+ || ||
+ || 3/60% +-----+5/55% +-----+ 3/75% ||
+ |+-----------| 5 |------| 6 |---------+|
+ +----------> +-----+ ---> +-----+ ---------+
+ (b) CCDR Algorithm
+
+ Figure 5: Case Study for CCDR's Algorithm
+
+4.2. Topology Simulation
+
+ Moving on from the specific case study, we now consider a class of
+ networks more representative of real deployments, with a fully linked
+ core network that serves to connect edge nodes, which themselves
+ connect to only a subset of the core. An example of such a topology
+ is shown in Figure 6 for the case of 4 core nodes and 5 edge nodes.
+ The CCDR simulations presented in this work use topologies involving
+ 100 core nodes and 400 edge nodes. While the resulting graph does
+ not fit on this page, this scale of network is similar to what is
+ deployed in production environments.
+
+ +----+
+ /|Edge|\
+ | +----+ |
+ | |
+ | |
+ +----+ +----+ +----+
+ |Edge|----|Core|-----|Core|---------+
+ +----+ +----+ +----+ |
+ / | \ / | |
+ +----+ | \ / | |
+ |Edge| | X | |
+ +----+ | / \ | |
+ \ | / \ | |
+ +----+ +----+ +----+ |
+ |Edge|----|Core|-----|Core| |
+ +----+ +----+ +----+ |
+ | | |
+ | +------\ +----+
+ | ---|Edge|
+ +-----------------/ +----+
+
+ Figure 6: Topology of Simulation
+
+ For the simulations, the number of links connecting one edge node to
+ the set of core nodes is randomly chosen between two and thirty, and
+ the total number of links is more than 20,000. Each link has a
+ congestion threshold, which can be arbitrarily set, for example, to
+ 90% of the nominal link capacity without affecting the simulation
+ results.
+
+4.3. Traffic Matrix Simulation
+
+ For each topology, a traffic matrix is generated based on the link
+ capacity of the topology. It can result in many kinds of situations
+ such as congestion, mild congestion, and non-congestion.
+
+ In the CCDR simulation, the dimension of the traffic matrix is
+ 500*500 (100 core nodes plus 400 edge nodes). About 20% of links are
+ overloaded when the Open Shortest Path First (OSPF) protocol is used
+ in the network.
+
+4.4. CCDR End-to-End Path Optimization
+
+ The CCDR E2E path optimization entails finding the best path, which
+ is the lowest in metric value, as well as having utilization far
+ below the congestion threshold for each link of the path. Based on
+ the current state of the network, the PCE within CCDR framework
+ combines the shortest path algorithm with a penalty theory of
+ classical optimization and graph theory.
+
+ Given a background traffic matrix, which is unscheduled, when a set
+ of new flows comes into the network, the E2E path optimization finds
+ the optimal paths for them. The selected paths bring the least
+ congestion degree to the network.
+
+ The link Utilization Increment Degree (UID), when the new flows are
+ added into the network, is shown in Figure 7. The first graph in
+ Figure 7 is the UID with OSPF, and the second graph is the UID with
+ CCDR E2E path optimization. The average UID of the first graph is
+ more than 30%. After path optimization, the average UID is less than
+ 5%. The results show that the CCDR E2E path optimization has an eye-
+ catching decrease in UID relative to the path chosen based on OSPF.
+
+ While real-world results invariably differ from simulations (for
+ example, real-world topologies are likely to exhibit correlation in
+ the attachment patterns for edge nodes to the core, which are not
+ reflected in these results), the dramatic nature of the improvement
+ in UID and the choice of simulated topology to resemble real-world
+ conditions suggest that real-world deployments will also experience
+ significant improvement in UID results.
+
+ +-----------------------------------------------------------+
+ | * * * *|
+ 60| * * * * * *|
+ |* * ** * * * * * ** * * * * **|
+ |* * ** * * ** *** ** * * ** * * * ** * * *** **|
+ |* * * ** * ** ** *** *** ** **** ** *** **** ** *** **|
+ 40|* * * ***** ** *** *** *** ** **** ** *** ***** ****** **|
+ UID(%)|* * ******* ** *** *** ******* **** ** *** ***** *********|
+ |*** ******* ** **** *********** *********** ***************|
+ |******************* *********** *********** ***************|
+ 20|******************* ***************************************|
+ |******************* ***************************************|
+ |***********************************************************|
+ |***********************************************************|
+ 0+-----------------------------------------------------------+
+ 0 100 200 300 400 500 600 700 800 900 1000
+ +-----------------------------------------------------------+
+ | |
+ 60| |
+ | |
+ | |
+ | |
+ 40| |
+ UID(%)| |
+ | |
+ | |
+ 20| |
+ | *|
+ | * *|
+ | * * * * * ** * *|
+ 0+-----------------------------------------------------------+
+ 0 100 200 300 400 500 600 700 800 900 1000
+ Flow Number
+
+ Figure 7: Simulation Results with Congestion Elimination
+
+4.5. Network Temporal Congestion Elimination
+
+ During the simulations, different degrees of network congestion were
+ considered. To examine the effect of CCDR on link congestion, we
+ consider the Congestion Degree (CD) of a link, defined as the link
+ utilization beyond its threshold.
+
+ The CCDR congestion elimination performance is shown in Figure 8.
+ The first graph is the CD distribution before the process of
+ congestion elimination. The average CD of all congested links is
+ about 20%. The second graph shown in Figure 8 is the CD distribution
+ after using the congestion elimination process. It shows that only
+ twelve links among the total 20,000 exceed the threshold, and all the
+ CD values are less than 3%. Thus, after scheduling the traffic away
+ from the congested paths, the degree of network congestion is greatly
+ eliminated and the network utilization is in balance.
+
+ Before congestion elimination
+ +-----------------------------------------------------------+
+ | * ** * ** ** *|
+ 20| * * **** * ** ** *|
+ |* * ** * ** ** **** * ***** *********|
+ |* * * * * **** ****** * ** *** **********************|
+ 15|* * * ** * ** **** ********* *****************************|
+ |* * ****** ******* ********* *****************************|
+ CD(%) |* ********* ******* ***************************************|
+ 10|* ********* ***********************************************|
+ |*********** ***********************************************|
+ |***********************************************************|
+ 5|***********************************************************|
+ |***********************************************************|
+ |***********************************************************|
+ 0+-----------------------------------------------------------+
+ 0 0.5 1 1.5 2
+
+ After congestion elimination
+ +-----------------------------------------------------------+
+ | |
+ 20| |
+ | |
+ | |
+ 15| |
+ | |
+ CD(%) | |
+ 10| |
+ | |
+ | |
+ 5 | |
+ | |
+ | * ** * * * ** * ** * |
+ 0 +-----------------------------------------------------------+
+ 0 0.5 1 1.5 2
+ Link Number(*10000)
+
+ Figure 8: Simulation Results with Congestion Elimination
+
+ It is clear that by using an active path-computation mechanism that
+ is able to take into account observed link traffic/congestion, the
+ occurrence of congestion events can be greatly reduced. Only when a
+ preponderance of links in the network are near their congestion
+ threshold will the central controller be unable to find a clear path
+ as opposed to when a static metric-based procedure is used, which
+ will produce congested paths once a single bottleneck approaches its
+ capacity. More detailed information about the algorithm can be found
+ in [PTCS].
+
+5. CCDR Deployment Consideration
+
+ The above CCDR scenarios and simulation results demonstrate that a
+ single general solution can be found that copes with multiple complex
+ situations. The specific situations considered are not known to have
+ any special properties, so it is expected that the benefits
+ demonstrated will have general applicability. Accordingly, the
+ integrated use of a centralized controller for the more complex
+ optimal path computations in a native IP network should result in
+ significant improvements without impacting the underlying network
+ infrastructure.
+
+ For intra-domain or inter-domain native IP TE scenarios, the
+ deployment of a CCDR solution is similar with the centralized
+ controller being able to compute paths along with no changes being
+ required to the underlying network infrastructure. This universal
+ deployment characteristic can facilitate a generic traffic-
+ engineering solution where operators do not need to differentiate
+ between intra-domain and inter-domain TE cases.
+
+ To deploy the CCDR solution, the PCE should collect the underlying
+ network topology dynamically, for example, via Border Gateway
+ Protocol - Link State (BGP-LS) [RFC7752]. It also needs to gather
+ the network traffic information periodically from the network
+ management platform. The simulation results show that the PCE can
+ compute the E2E optimal path within seconds; thus, it can cope with a
+ change to the underlying network in a matter of minutes. More agile
+ requirements would need to increase the sample rate of the underlying
+ network and decrease the detection and notification interval of the
+ underlying network. The methods of gathering this information as
+ well as decreasing its latency are out of the scope of this document.
+
+6. Security Considerations
+
+ This document considers mainly the integration of distributed
+ protocols and the central control capability of a PCE. While it can
+ certainly simplify the management of a network in various traffic-
+ engineering scenarios as described in this document, the centralized
+ control also brings a new point that may be easily attacked.
+ Solutions for CCDR scenarios need to consider protection of the PCE
+ and communication with the underlying devices.
+
+ [RFC5440] and [RFC8253] provide additional information.
+
+ The control priority and interaction process should also be carefully
+ designed for the combination of the distributed protocol and central
+ control. Generally, the central control instructions should have
+ higher priority than the forwarding actions determined by the
+ distributed protocol. When communication between PCE and the
+ underlying devices is disrupted, the distributed protocol should take
+ control of the underlying network. [PCE-NATIVE-IP] provides more
+ considerations corresponding to the solution.
+
+7. IANA Considerations
+
+ This document has no IANA actions.
+
+8. References
+
+8.1. Normative References
+
+ [RFC5440] Vasseur, JP., Ed. and JL. Le Roux, Ed., "Path Computation
+ Element (PCE) Communication Protocol (PCEP)", RFC 5440,
+ DOI 10.17487/RFC5440, March 2009,
+ <https://www.rfc-editor.org/info/rfc5440>.
+
+ [RFC7752] Gredler, H., Ed., Medved, J., Previdi, S., Farrel, A., and
+ S. Ray, "North-Bound Distribution of Link-State and
+ Traffic Engineering (TE) Information Using BGP", RFC 7752,
+ DOI 10.17487/RFC7752, March 2016,
+ <https://www.rfc-editor.org/info/rfc7752>.
+
+ [RFC8253] Lopez, D., Gonzalez de Dios, O., Wu, Q., and D. Dhody,
+ "PCEPS: Usage of TLS to Provide a Secure Transport for the
+ Path Computation Element Communication Protocol (PCEP)",
+ RFC 8253, DOI 10.17487/RFC8253, October 2017,
+ <https://www.rfc-editor.org/info/rfc8253>.
+
+8.2. Informative References
+
+ [PCE-NATIVE-IP]
+ Wang, A., Zhao, Q., Khasanov, B., and H. Chen, "PCE in
+ Native IP Network", Work in Progress, Internet-Draft,
+ draft-ietf-teas-pce-native-ip-05, 9 January 2020,
+ <https://tools.ietf.org/html/draft-ietf-teas-pce-native-
+ ip-05>.
+
+ [PCEP-NATIVE-IP-EXT]
+ Wang, A., Khasanov, B., Fang, S., and C. Zhu, "PCEP
+ Extension for Native IP Network", Work in Progress,
+ Internet-Draft, draft-ietf-pce-pcep-extension-native-ip-
+ 05, 17 February 2020, <https://tools.ietf.org/html/draft-
+ ietf-pce-pcep-extension-native-ip-05>.
+
+ [PTCS] Zhang, P., Xie, K., Kou, C., Huang, X., Wang, A., and Q.
+ Sun, "A Practical Traffic Control Scheme With Load
+ Balancing Based on PCE Architecture",
+ DOI 10.1109/ACCESS.2019.2902610, IEEE Access 18526773,
+ March 2019,
+ <https://ieeexplore.ieee.org/document/8657733>.
+
+ [RFC3209] Awduche, D., Berger, L., Gan, D., Li, T., Srinivasan, V.,
+ and G. Swallow, "RSVP-TE: Extensions to RSVP for LSP
+ Tunnels", RFC 3209, DOI 10.17487/RFC3209, December 2001,
+ <https://www.rfc-editor.org/info/rfc3209>.
+
+ [RFC8402] Filsfils, C., Ed., Previdi, S., Ed., Ginsberg, L.,
+ Decraene, B., Litkowski, S., and R. Shakir, "Segment
+ Routing Architecture", RFC 8402, DOI 10.17487/RFC8402,
+ July 2018, <https://www.rfc-editor.org/info/rfc8402>.
+
+ [RFC8578] Grossman, E., Ed., "Deterministic Networking Use Cases",
+ RFC 8578, DOI 10.17487/RFC8578, May 2019,
+ <https://www.rfc-editor.org/info/rfc8578>.
+
+Acknowledgements
+
+ The authors would like to thank Deborah Brungard, Adrian Farrel,
+ Huaimo Chen, Vishnu Beeram, and Lou Berger for their support and
+ comments on this document.
+
+ Thanks to Benjamin Kaduk for his careful review and valuable
+ suggestions on this document. Also, thanks to Roman Danyliw, Alvaro
+ Retana, and Éric Vyncke for their reviews and comments.
+
+Contributors
+
+ Lu Huang contributed to the content of this document.
+
+Authors' Addresses
+
+ Aijun Wang
+ China Telecom
+ Beiqijia Town, Changping District
+ Beijing
+ Beijing, 102209
+ China
+
+ Email: wangaj3@chinatelecom.cn
+
+
+ Xiaohong Huang
+ Beijing University of Posts and Telecommunications
+ No.10 Xitucheng Road, Haidian District
+ Beijing
+ China
+
+ Email: huangxh@bupt.edu.cn
+
+
+ Caixia Kou
+ Beijing University of Posts and Telecommunications
+ No.10 Xitucheng Road, Haidian District
+ Beijing
+ China
+
+ Email: koucx@lsec.cc.ac.cn
+
+
+ Zhenqiang Li
+ China Mobile
+ 32 Xuanwumen West Ave, Xicheng District
+ Beijing
+ 100053
+ China
+
+ Email: li_zhenqiang@hotmail.com
+
+
+ Penghui Mi
+ Huawei Technologies
+ Tower C of Bldg.2, Cloud Park, No.2013 of Xuegang Road
+ Shenzhen
+ Bantian,Longgang District, 518129
+ China
+
+ Email: mipenghui@huawei.com