diff options
author | Thomas Voss <mail@thomasvoss.com> | 2024-11-27 20:54:24 +0100 |
---|---|---|
committer | Thomas Voss <mail@thomasvoss.com> | 2024-11-27 20:54:24 +0100 |
commit | 4bfd864f10b68b71482b35c818559068ef8d5797 (patch) | |
tree | e3989f47a7994642eb325063d46e8f08ffa681dc /doc/rfc/rfc8735.txt | |
parent | ea76e11061bda059ae9f9ad130a9895cc85607db (diff) |
doc: Add RFC documents
Diffstat (limited to 'doc/rfc/rfc8735.txt')
-rw-r--r-- | doc/rfc/rfc8735.txt | 766 |
1 files changed, 766 insertions, 0 deletions
diff --git a/doc/rfc/rfc8735.txt b/doc/rfc/rfc8735.txt new file mode 100644 index 0000000..0afe78a --- /dev/null +++ b/doc/rfc/rfc8735.txt @@ -0,0 +1,766 @@ + + + + +Internet Engineering Task Force (IETF) A. Wang +Request for Comments: 8735 China Telecom +Category: Informational X. Huang +ISSN: 2070-1721 C. Kou + BUPT + Z. Li + China Mobile + P. Mi + Huawei Technologies + February 2020 + + + Scenarios and Simulation Results of PCE in a Native IP Network + +Abstract + + Requirements for providing the End-to-End (E2E) performance assurance + are emerging within the service provider networks. While there are + various technology solutions, there is no single solution that can + fulfill these requirements for a native IP network. In particular, + there is a need for a universal E2E solution that can cover both + intra- and inter-domain scenarios. + + One feasible E2E traffic-engineering solution is the addition of + central control in a native IP network. This document describes + various complex scenarios and simulation results when applying the + Path Computation Element (PCE) in a native IP network. This + solution, referred to as Centralized Control Dynamic Routing (CCDR), + integrates the advantage of using distributed protocols and the power + of a centralized control technology, providing traffic engineering + for native IP networks in a manner that applies equally to intra- and + inter-domain scenarios. + +Status of This Memo + + This document is not an Internet Standards Track specification; it is + published for informational purposes. + + This document is a product of the Internet Engineering Task Force + (IETF). It represents the consensus of the IETF community. It has + received public review and has been approved for publication by the + Internet Engineering Steering Group (IESG). Not all documents + approved by the IESG are candidates for any level of Internet + Standard; see Section 2 of RFC 7841. + + Information about the current status of this document, any errata, + and how to provide feedback on it may be obtained at + https://www.rfc-editor.org/info/rfc8735. + +Copyright Notice + + Copyright (c) 2020 IETF Trust and the persons identified as the + document authors. All rights reserved. + + This document is subject to BCP 78 and the IETF Trust's Legal + Provisions Relating to IETF Documents + (https://trustee.ietf.org/license-info) in effect on the date of + publication of this document. Please review these documents + carefully, as they describe your rights and restrictions with respect + to this document. Code Components extracted from this document must + include Simplified BSD License text as described in Section 4.e of + the Trust Legal Provisions and are provided without warranty as + described in the Simplified BSD License. + +Table of Contents + + 1. Introduction + 2. Terminology + 3. CCDR Scenarios + 3.1. QoS Assurance for Hybrid Cloud-Based Application + 3.2. Link Utilization Maximization + 3.3. Traffic Engineering for Multi-domain + 3.4. Network Temporal Congestion Elimination + 4. CCDR Simulation + 4.1. Case Study for CCDR Algorithm + 4.2. Topology Simulation + 4.3. Traffic Matrix Simulation + 4.4. CCDR End-to-End Path Optimization + 4.5. Network Temporal Congestion Elimination + 5. CCDR Deployment Consideration + 6. Security Considerations + 7. IANA Considerations + 8. References + 8.1. Normative References + 8.2. Informative References + Acknowledgements + Contributors + Authors' Addresses + +1. Introduction + + A service provider network is composed of thousands of routers that + run distributed protocols to exchange reachability information. The + path for the destination network is mainly calculated, and + controlled, by the distributed protocols. These distributed + protocols are robust enough to support most applications; however, + they have some difficulties supporting the complexities needed for + traffic-engineering applications, e.g., E2E performance assurance, or + maximizing the link utilization within an IP network. + + Multiprotocol Label Switching (MPLS) using Traffic-Engineering (TE) + technology (MPLS-TE) [RFC3209] is one solution for TE networks, but + it introduces an MPLS network along with related technology, which + would be an overlay of the IP network. MPLS-TE technology is often + used for Label Switched Path (LSP) protection and setting up complex + paths within a domain. It has not been widely deployed for meeting + E2E (especially in inter-domain) dynamic performance assurance + requirements for an IP network. + + Segment Routing [RFC8402] is another solution that integrates some + advantages of using a distributed protocol and central control + technology, but it requires the underlying network, especially the + provider edge router, to do an in-depth label push and pop action + while adding complexity when coexisting with the non-segment routing + network. Additionally, it can only maneuver the E2E paths for MPLS + and IPv6 traffic via different mechanisms. + + Deterministic Networking (DetNet) [RFC8578] is another possible + solution. It is primarily focused on providing bounded latency for a + flow and introduces additional requirements on the domain edge + router. The current DetNet scope is within one domain. The use + cases defined in this document do not require the additional + complexity of deterministic properties and so differ from the DetNet + use cases. + + This document describes several scenarios for a native IP network + where a Centralized Control Dynamic Routing (CCDR) framework can + produce qualitative improvement in efficiency without requiring a + change to the data-plane behavior on the router. Using knowledge of + the Border Gateway Protocol (BGP) session-specific prefixes + advertised by a router, the network topology and the near-real-time + link-utilization information from network management systems, a + central PCE is able to compute an optimal path and give the + underlying routers the destination address to use to reach the BGP + nexthop, such that the distributed routing protocol will use the + computed path via traditional recursive lookup procedure. Some + results from simulations of path optimization are also presented to + concretely illustrate a variety of scenarios where CCDR shows + significant improvement over traditional distributed routing + protocols. + + This document is the base document of the following two documents: + the universal solution document, which is suitable for intra-domain + and inter-domain TE scenario, is described in [PCE-NATIVE-IP]; and + the related protocol extension contents is described in + [PCEP-NATIVE-IP-EXT]. + +2. Terminology + + In this document, PCE is used as defined in [RFC5440]. The following + terms are used as described here: + + BRAS: Broadband Remote Access Server + + CD: Congestion Degree + + CR: Core Router + + CCDR: Centralized Control Dynamic Routing + + E2E: End to End + + IDC: Internet Data Center + + MAN: Metro Area Network + + QoS: Quality of Service + + SR: Service Router + + TE: Traffic Engineering + + UID: Utilization Increment Degree + + WAN: Wide Area Network + +3. CCDR Scenarios + + The following sections describe various deployment scenarios where + applying the CCDR framework is intuitively expected to produce + improvements based on the macro-scale properties of the framework and + the scenario. + +3.1. QoS Assurance for Hybrid Cloud-Based Application + + With the emergence of cloud computing technologies, enterprises are + putting more and more services on a public-oriented cloud environment + while keeping core business within their private cloud. The + communication between the private and public cloud sites spans the + WAN. The bandwidth requirements between them are variable, and the + background traffic between these two sites varies over time. + Enterprise applications require assurance of the E2E QoS performance + on demand for variable bandwidth services. + + CCDR, which integrates the merits of distributed protocols and the + power of centralized control, is suitable for this scenario. The + possible solution framework is illustrated below: + + +------------------------+ + | Cloud-Based Application| + +------------------------+ + | + +-----------+ + | PCE | + +-----------+ + | + | + //--------------\\ + ///// \\\\\ + Private Cloud Site || Distributed |Public Cloud Site + | Control Network | + \\\\\ ///// + \\--------------// + + Figure 1: Hybrid Cloud Communication Scenario + + As illustrated in Figure 1, the source and destination of the "Cloud- + Based Application" traffic are located at "Private Cloud Site" and + "Public Cloud Site", respectively. + + By default, the traffic path between the private and public cloud + site is determined by the distributed control network. When an + application requires E2E QoS assurance, it can send these + requirements to the PCE and let the PCE compute one E2E path, which + is based on the underlying network topology and real traffic + information, in order to accommodate the application's QoS + requirements. Section 4.4 of this document describes the simulation + results for this use case. + +3.2. Link Utilization Maximization + + Network topology within a Metro Area Network (MAN) is generally in a + star mode as illustrated in Figure 2, with different devices + connected to different customer types. The traffic from these + customers is often in a tidal pattern with the links between the Core + Router (CR) / Broadband Remote Access Server (BRAS) and CR/Service + Router (SR) experiencing congestion in different periods due to + subscribers under BRAS often using the network at night and the + leased line users under SR often using the network during the + daytime. The link between BRAS/SR and CR must satisfy the maximum + traffic volume between them, respectively, which causes these links + to often be underutilized. + + +--------+ + | CR | + +----|---+ + | + |-------|--------|-------| + | | | | + +--|-+ +-|+ +--|-+ +-|+ + |BRAS| |SR| |BRAS| |SR| + +----+ +--+ +----+ +--+ + + Figure 2: Star-Mode Network Topology within MAN + + If we consider connecting the BRAS/SR with a local link loop (which + is usually lower cost) and control the overall MAN topology with the + CCDR framework, we can exploit the tidal phenomena between the BRAS/ + CR and SR/CR links, maximizing the utilization of these central trunk + links (which are usually higher cost than the local loops). + + +-------+ + ----- PCE | + | +-------+ + +----|---+ + | CR | + +----|---+ + | + |-------|--------|-------| + | | | | + +--|-+ +-|+ +--|-+ +-|+ + |BRAS-----SR| |BRAS-----SR| + +----+ +--+ +----+ +--+ + + Figure 3: Link Utilization Maximization via CCDR + +3.3. Traffic Engineering for Multi-domain + + Service provider networks are often comprised of different domains, + interconnected with each other, forming a very complex topology as + illustrated in Figure 4. Due to the traffic pattern to/from the MAN + and IDC, the utilization of the links between them are often + asymmetric. It is almost impossible to balance the utilization of + these links via a distributed protocol, but this unbalance can be + overcome utilizing the CCDR framework. + + +---+ +---+ + |MAN|----------------|IDC| + +---+ | +---+ + | ---------- | + |-----|Backbone|-----| + | ----|----- | + | | | + +---+ | +---+ + |IDC|----------------|MAN| + +---+ +---+ + + Figure 4: Traffic Engineering for Complex Multi-domain Topology + + A solution for this scenario requires the gathering of NetFlow + information, analysis of the source/destination autonomous system + (AS), and determining what the main cause of the congested link(s) + is. After this, the operator can use the external Border Gateway + Protocol (eBGP) sessions to schedule the traffic among the different + domains according to the solution described in the CCDR framework. + +3.4. Network Temporal Congestion Elimination + + In more general situations, there is often temporal congestion within + the service provider's network, for example, due to daily or weekly + periodic bursts or large events that are scheduled well in advance. + Such congestion phenomena often appear regularly, and if the service + provider has methods to mitigate it, it will certainly improve their + network operation capabilities and increase satisfaction for + customers. CCDR is also suitable for such scenarios, as the + controller can schedule traffic out of the congested links, lowering + their utilization during these times. Section 4.5 describes the + simulation results of this scenario. + +4. CCDR Simulation + + The following sections describe a specific case study to illustrate + the workings of the CCDR algorithm with concrete paths/metrics, as + well as a procedure for generating topology and traffic matrices and + the results from simulations applying CCDR for E2E QoS (assured path + and congestion elimination) over the generated topologies and traffic + matrices. In all cases examined, the CCDR algorithm produces + qualitatively significant improvement over the reference (OSPF) + algorithm, suggesting that CCDR will have broad applicability. + + The structure and scale of the simulated topology is similar to that + of the real networks. Multiple different traffic matrices were + generated to simulate different congestion conditions in the network. + Only one of them is illustrated since the others produce similar + results. + +4.1. Case Study for CCDR Algorithm + + In this section, we consider a specific network topology for case + study: examining the path selected by OSPF and CCDR and evaluating + how and why the paths differ. Figure 5 depicts the topology of the + network in this case. There are eight forwarding devices in the + network. The original cost and utilization are marked on it as shown + in the figure. For example, the original cost and utilization for + the link (1, 2) are 3 and 50%, respectively. There are two flows: f1 + and f2. Both of these two flows are from node 1 to node 8. For + simplicity, it is assumed that the bandwidth of the link in the + network is 10 Mb/s. The flow rate of f1 is 1 Mb/s and the flow rate + of f2 is 2 Mb/s. The threshold of the link in congestion is 90%. + + If the OSPF protocol, which adopts Dijkstra's algorithm (IS-IS is + similar because it also uses Dijkstra's algorithm), is applied in the + network, the two flows from node 1 to node 8 can only use the OSPF + path (p1: 1->2->3->8). This is because Dijkstra's algorithm mainly + considers the original cost of the link. Since CCDR considers cost + and utilization simultaneously, the same path as OSPF will not be + selected due to the severe congestion of the link (2, 3). In this + case, f1 will select the path (p2: 1->5->6->7->8) since the new cost + of this path is better than that of the OSPF path. Moreover, the + path p2 is also better than the path (p3: 1->2->4->7->8) for flow f1. + However, f2 will not select the same path since it will cause new + congestion in the link (6, 7). As a result, f2 will select the path + (p3: 1->2->4->7->8). + + + +----+ f1 +-------> +-----+ ----> +-----+ + |Edge|-----------+ |+--------| 3 |-------| 8 | + |Node|---------+ | ||+-----> +-----+ ----> +-----+ + +----+ | | 4/95%||| 6/50% | + | | ||| 5/60%| + | v ||| | + +----+ +-----+ -----> +-----+ +-----+ +-----+ + |Edge|-------| 1 |--------| 2 |------| 4 |------| 7 | + |Node|-----> +-----+ -----> +-----+7/60% +-----+5/45% +-----+ + +----+ f2 | 3/50% | + | | + | 3/60% +-----+ 5/55%+-----+ 3/75% | + +-----------| 5 |------| 6 |---------+ + +-----+ +-----+ + (a) Dijkstra's Algorithm (OSPF/IS-IS) + + + +----+ f1 +-----+ ----> +-----+ + |Edge|-----------+ +--------| 3 |-------| 8 | + |Node|---------+ | | +-----+ ----> +-----+ + +----+ | | 4/95% | 6/50% ^|^ + | | | 5/60%||| + | v | ||| + +----+ +-----+ -----> +-----+ ---> +-----+ ---> +-----+ + |Edge|-------| 1 |--------| 2 |------| 4 |------| 7 | + |Node|-----> +-----+ +-----+7/60% +-----+5/45% +-----+ + +----+ f2 || 3/50% |^ + || || + || 3/60% +-----+5/55% +-----+ 3/75% || + |+-----------| 5 |------| 6 |---------+| + +----------> +-----+ ---> +-----+ ---------+ + (b) CCDR Algorithm + + Figure 5: Case Study for CCDR's Algorithm + +4.2. Topology Simulation + + Moving on from the specific case study, we now consider a class of + networks more representative of real deployments, with a fully linked + core network that serves to connect edge nodes, which themselves + connect to only a subset of the core. An example of such a topology + is shown in Figure 6 for the case of 4 core nodes and 5 edge nodes. + The CCDR simulations presented in this work use topologies involving + 100 core nodes and 400 edge nodes. While the resulting graph does + not fit on this page, this scale of network is similar to what is + deployed in production environments. + + +----+ + /|Edge|\ + | +----+ | + | | + | | + +----+ +----+ +----+ + |Edge|----|Core|-----|Core|---------+ + +----+ +----+ +----+ | + / | \ / | | + +----+ | \ / | | + |Edge| | X | | + +----+ | / \ | | + \ | / \ | | + +----+ +----+ +----+ | + |Edge|----|Core|-----|Core| | + +----+ +----+ +----+ | + | | | + | +------\ +----+ + | ---|Edge| + +-----------------/ +----+ + + Figure 6: Topology of Simulation + + For the simulations, the number of links connecting one edge node to + the set of core nodes is randomly chosen between two and thirty, and + the total number of links is more than 20,000. Each link has a + congestion threshold, which can be arbitrarily set, for example, to + 90% of the nominal link capacity without affecting the simulation + results. + +4.3. Traffic Matrix Simulation + + For each topology, a traffic matrix is generated based on the link + capacity of the topology. It can result in many kinds of situations + such as congestion, mild congestion, and non-congestion. + + In the CCDR simulation, the dimension of the traffic matrix is + 500*500 (100 core nodes plus 400 edge nodes). About 20% of links are + overloaded when the Open Shortest Path First (OSPF) protocol is used + in the network. + +4.4. CCDR End-to-End Path Optimization + + The CCDR E2E path optimization entails finding the best path, which + is the lowest in metric value, as well as having utilization far + below the congestion threshold for each link of the path. Based on + the current state of the network, the PCE within CCDR framework + combines the shortest path algorithm with a penalty theory of + classical optimization and graph theory. + + Given a background traffic matrix, which is unscheduled, when a set + of new flows comes into the network, the E2E path optimization finds + the optimal paths for them. The selected paths bring the least + congestion degree to the network. + + The link Utilization Increment Degree (UID), when the new flows are + added into the network, is shown in Figure 7. The first graph in + Figure 7 is the UID with OSPF, and the second graph is the UID with + CCDR E2E path optimization. The average UID of the first graph is + more than 30%. After path optimization, the average UID is less than + 5%. The results show that the CCDR E2E path optimization has an eye- + catching decrease in UID relative to the path chosen based on OSPF. + + While real-world results invariably differ from simulations (for + example, real-world topologies are likely to exhibit correlation in + the attachment patterns for edge nodes to the core, which are not + reflected in these results), the dramatic nature of the improvement + in UID and the choice of simulated topology to resemble real-world + conditions suggest that real-world deployments will also experience + significant improvement in UID results. + + +-----------------------------------------------------------+ + | * * * *| + 60| * * * * * *| + |* * ** * * * * * ** * * * * **| + |* * ** * * ** *** ** * * ** * * * ** * * *** **| + |* * * ** * ** ** *** *** ** **** ** *** **** ** *** **| + 40|* * * ***** ** *** *** *** ** **** ** *** ***** ****** **| + UID(%)|* * ******* ** *** *** ******* **** ** *** ***** *********| + |*** ******* ** **** *********** *********** ***************| + |******************* *********** *********** ***************| + 20|******************* ***************************************| + |******************* ***************************************| + |***********************************************************| + |***********************************************************| + 0+-----------------------------------------------------------+ + 0 100 200 300 400 500 600 700 800 900 1000 + +-----------------------------------------------------------+ + | | + 60| | + | | + | | + | | + 40| | + UID(%)| | + | | + | | + 20| | + | *| + | * *| + | * * * * * ** * *| + 0+-----------------------------------------------------------+ + 0 100 200 300 400 500 600 700 800 900 1000 + Flow Number + + Figure 7: Simulation Results with Congestion Elimination + +4.5. Network Temporal Congestion Elimination + + During the simulations, different degrees of network congestion were + considered. To examine the effect of CCDR on link congestion, we + consider the Congestion Degree (CD) of a link, defined as the link + utilization beyond its threshold. + + The CCDR congestion elimination performance is shown in Figure 8. + The first graph is the CD distribution before the process of + congestion elimination. The average CD of all congested links is + about 20%. The second graph shown in Figure 8 is the CD distribution + after using the congestion elimination process. It shows that only + twelve links among the total 20,000 exceed the threshold, and all the + CD values are less than 3%. Thus, after scheduling the traffic away + from the congested paths, the degree of network congestion is greatly + eliminated and the network utilization is in balance. + + Before congestion elimination + +-----------------------------------------------------------+ + | * ** * ** ** *| + 20| * * **** * ** ** *| + |* * ** * ** ** **** * ***** *********| + |* * * * * **** ****** * ** *** **********************| + 15|* * * ** * ** **** ********* *****************************| + |* * ****** ******* ********* *****************************| + CD(%) |* ********* ******* ***************************************| + 10|* ********* ***********************************************| + |*********** ***********************************************| + |***********************************************************| + 5|***********************************************************| + |***********************************************************| + |***********************************************************| + 0+-----------------------------------------------------------+ + 0 0.5 1 1.5 2 + + After congestion elimination + +-----------------------------------------------------------+ + | | + 20| | + | | + | | + 15| | + | | + CD(%) | | + 10| | + | | + | | + 5 | | + | | + | * ** * * * ** * ** * | + 0 +-----------------------------------------------------------+ + 0 0.5 1 1.5 2 + Link Number(*10000) + + Figure 8: Simulation Results with Congestion Elimination + + It is clear that by using an active path-computation mechanism that + is able to take into account observed link traffic/congestion, the + occurrence of congestion events can be greatly reduced. Only when a + preponderance of links in the network are near their congestion + threshold will the central controller be unable to find a clear path + as opposed to when a static metric-based procedure is used, which + will produce congested paths once a single bottleneck approaches its + capacity. More detailed information about the algorithm can be found + in [PTCS]. + +5. CCDR Deployment Consideration + + The above CCDR scenarios and simulation results demonstrate that a + single general solution can be found that copes with multiple complex + situations. The specific situations considered are not known to have + any special properties, so it is expected that the benefits + demonstrated will have general applicability. Accordingly, the + integrated use of a centralized controller for the more complex + optimal path computations in a native IP network should result in + significant improvements without impacting the underlying network + infrastructure. + + For intra-domain or inter-domain native IP TE scenarios, the + deployment of a CCDR solution is similar with the centralized + controller being able to compute paths along with no changes being + required to the underlying network infrastructure. This universal + deployment characteristic can facilitate a generic traffic- + engineering solution where operators do not need to differentiate + between intra-domain and inter-domain TE cases. + + To deploy the CCDR solution, the PCE should collect the underlying + network topology dynamically, for example, via Border Gateway + Protocol - Link State (BGP-LS) [RFC7752]. It also needs to gather + the network traffic information periodically from the network + management platform. The simulation results show that the PCE can + compute the E2E optimal path within seconds; thus, it can cope with a + change to the underlying network in a matter of minutes. More agile + requirements would need to increase the sample rate of the underlying + network and decrease the detection and notification interval of the + underlying network. The methods of gathering this information as + well as decreasing its latency are out of the scope of this document. + +6. Security Considerations + + This document considers mainly the integration of distributed + protocols and the central control capability of a PCE. While it can + certainly simplify the management of a network in various traffic- + engineering scenarios as described in this document, the centralized + control also brings a new point that may be easily attacked. + Solutions for CCDR scenarios need to consider protection of the PCE + and communication with the underlying devices. + + [RFC5440] and [RFC8253] provide additional information. + + The control priority and interaction process should also be carefully + designed for the combination of the distributed protocol and central + control. Generally, the central control instructions should have + higher priority than the forwarding actions determined by the + distributed protocol. When communication between PCE and the + underlying devices is disrupted, the distributed protocol should take + control of the underlying network. [PCE-NATIVE-IP] provides more + considerations corresponding to the solution. + +7. IANA Considerations + + This document has no IANA actions. + +8. References + +8.1. Normative References + + [RFC5440] Vasseur, JP., Ed. and JL. Le Roux, Ed., "Path Computation + Element (PCE) Communication Protocol (PCEP)", RFC 5440, + DOI 10.17487/RFC5440, March 2009, + <https://www.rfc-editor.org/info/rfc5440>. + + [RFC7752] Gredler, H., Ed., Medved, J., Previdi, S., Farrel, A., and + S. Ray, "North-Bound Distribution of Link-State and + Traffic Engineering (TE) Information Using BGP", RFC 7752, + DOI 10.17487/RFC7752, March 2016, + <https://www.rfc-editor.org/info/rfc7752>. + + [RFC8253] Lopez, D., Gonzalez de Dios, O., Wu, Q., and D. Dhody, + "PCEPS: Usage of TLS to Provide a Secure Transport for the + Path Computation Element Communication Protocol (PCEP)", + RFC 8253, DOI 10.17487/RFC8253, October 2017, + <https://www.rfc-editor.org/info/rfc8253>. + +8.2. Informative References + + [PCE-NATIVE-IP] + Wang, A., Zhao, Q., Khasanov, B., and H. Chen, "PCE in + Native IP Network", Work in Progress, Internet-Draft, + draft-ietf-teas-pce-native-ip-05, 9 January 2020, + <https://tools.ietf.org/html/draft-ietf-teas-pce-native- + ip-05>. + + [PCEP-NATIVE-IP-EXT] + Wang, A., Khasanov, B., Fang, S., and C. Zhu, "PCEP + Extension for Native IP Network", Work in Progress, + Internet-Draft, draft-ietf-pce-pcep-extension-native-ip- + 05, 17 February 2020, <https://tools.ietf.org/html/draft- + ietf-pce-pcep-extension-native-ip-05>. + + [PTCS] Zhang, P., Xie, K., Kou, C., Huang, X., Wang, A., and Q. + Sun, "A Practical Traffic Control Scheme With Load + Balancing Based on PCE Architecture", + DOI 10.1109/ACCESS.2019.2902610, IEEE Access 18526773, + March 2019, + <https://ieeexplore.ieee.org/document/8657733>. + + [RFC3209] Awduche, D., Berger, L., Gan, D., Li, T., Srinivasan, V., + and G. Swallow, "RSVP-TE: Extensions to RSVP for LSP + Tunnels", RFC 3209, DOI 10.17487/RFC3209, December 2001, + <https://www.rfc-editor.org/info/rfc3209>. + + [RFC8402] Filsfils, C., Ed., Previdi, S., Ed., Ginsberg, L., + Decraene, B., Litkowski, S., and R. Shakir, "Segment + Routing Architecture", RFC 8402, DOI 10.17487/RFC8402, + July 2018, <https://www.rfc-editor.org/info/rfc8402>. + + [RFC8578] Grossman, E., Ed., "Deterministic Networking Use Cases", + RFC 8578, DOI 10.17487/RFC8578, May 2019, + <https://www.rfc-editor.org/info/rfc8578>. + +Acknowledgements + + The authors would like to thank Deborah Brungard, Adrian Farrel, + Huaimo Chen, Vishnu Beeram, and Lou Berger for their support and + comments on this document. + + Thanks to Benjamin Kaduk for his careful review and valuable + suggestions on this document. Also, thanks to Roman Danyliw, Alvaro + Retana, and Éric Vyncke for their reviews and comments. + +Contributors + + Lu Huang contributed to the content of this document. + +Authors' Addresses + + Aijun Wang + China Telecom + Beiqijia Town, Changping District + Beijing + Beijing, 102209 + China + + Email: wangaj3@chinatelecom.cn + + + Xiaohong Huang + Beijing University of Posts and Telecommunications + No.10 Xitucheng Road, Haidian District + Beijing + China + + Email: huangxh@bupt.edu.cn + + + Caixia Kou + Beijing University of Posts and Telecommunications + No.10 Xitucheng Road, Haidian District + Beijing + China + + Email: koucx@lsec.cc.ac.cn + + + Zhenqiang Li + China Mobile + 32 Xuanwumen West Ave, Xicheng District + Beijing + 100053 + China + + Email: li_zhenqiang@hotmail.com + + + Penghui Mi + Huawei Technologies + Tower C of Bldg.2, Cloud Park, No.2013 of Xuegang Road + Shenzhen + Bantian,Longgang District, 518129 + China + + Email: mipenghui@huawei.com |