diff options
Diffstat (limited to 'doc/rfc/rfc4829.txt')
-rw-r--r-- | doc/rfc/rfc4829.txt | 1067 |
1 files changed, 1067 insertions, 0 deletions
diff --git a/doc/rfc/rfc4829.txt b/doc/rfc/rfc4829.txt new file mode 100644 index 0000000..5a89550 --- /dev/null +++ b/doc/rfc/rfc4829.txt @@ -0,0 +1,1067 @@ + + + + + + +Network Working Group J. de Oliveira, Ed. +Request for Comments: 4829 Drexel University +Category: Informational JP. Vasseur, Ed. + Cisco Systems, Inc. + L. Chen + Verizon Laboratories + C. Scoglio + Kansas State University + April 2007 + + + Label Switched Path (LSP) Preemption Policies for + MPLS Traffic Engineering + +Status of This Memo + + This memo provides information for the Internet community. It does + not specify an Internet standard of any kind. Distribution of this + memo is unlimited. + +Copyright Notice + + Copyright (C) The IETF Trust (2007). + +IESG Note + + This RFC is not a candidate for any level of Internet Standard. The + IETF disclaims any knowledge of the fitness of this RFC for any + purpose and, in particular, notes that the decision to publish is not + based on IETF review for such things as security, congestion control, + or inappropriate interaction with deployed protocols. The RFC Editor + has chosen to publish this document at its discretion. Readers of + this document should exercise caution in evaluating its value for + implementation and deployment. See RFC 3932 for more information. + +Abstract + + When the establishment of a higher priority (Traffic Engineering + Label Switched Path) TE LSP requires the preemption of a set of lower + priority TE LSPs, a node has to make a local decision to select which + TE LSPs will be preempted. The preempted LSPs are then rerouted by + their respective Head-end Label Switch Router (LSR). This document + presents a flexible policy that can be used to achieve different + objectives: preempt the lowest priority LSPs; preempt the minimum + number of LSPs; preempt the set of TE LSPs that provide the closest + amount of bandwidth to the required bandwidth for the preempting TE + LSPs (to minimize bandwidth wastage); preempt the LSPs that will have + the maximum chance to get rerouted. Simulation results are given and + + + +de Oliveira, et al. Informational [Page 1] + +RFC 4829 LSP Preemption Policies for MPLS-TE April 2007 + + + a comparison among several different policies, with respect to + preemption cascading, number of preempted LSPs, priority, wasted + bandwidth and blocking probability is also included. + +Table of Contents + + 1. Motivation . . . . . . . . . . . . . . . . . . . . . . . . . . 3 + 2. Introduction . . . . . . . . . . . . . . . . . . . . . . . . . 3 + 3. LSP Setup Procedure and Preemption . . . . . . . . . . . . . . 5 + 4. Preemption Cascading . . . . . . . . . . . . . . . . . . . . . 6 + 5. Preemption Heuristic . . . . . . . . . . . . . . . . . . . . . 7 + 5.1. Preempting Resources on a Path . . . . . . . . . . . . . . 7 + 5.2. Preemption Heuristic Algorithm . . . . . . . . . . . . . . 8 + 6. Examples . . . . . . . . . . . . . . . . . . . . . . . . . . . 10 + 6.1. Simple Case: Single Link . . . . . . . . . . . . . . . . . 10 + 6.2. Network Case . . . . . . . . . . . . . . . . . . . . . . . 12 + 7. Security Considerations . . . . . . . . . . . . . . . . . . . 16 + 8. Acknowledgements . . . . . . . . . . . . . . . . . . . . . . . 16 + 9. Informative References . . . . . . . . . . . . . . . . . . . . 17 + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +de Oliveira, et al. Informational [Page 2] + +RFC 4829 LSP Preemption Policies for MPLS-TE April 2007 + + +1. Motivation + + The IETF Traffic Engineering Working Group has defined the + requirements and protocol extensions for DiffServ-aware MPLS Traffic + Engineering (DS-TE) [RFC3564] [RFC4124]. Several Bandwidth + Constraint models for use with DS-TE have been proposed [RFC4127] + [RFC4128] [RFC4126] and their performance was analyzed with respect + to the use of preemption. + + Preemption can be used as a tool to help ensure that high priority + LSPs can always be routed through relatively favorable paths. + Preemption can also be used to implement various prioritized access + policies as well as restoration policies following fault events + [RFC2702]. + + Although not a mandatory attribute in the traditional IP world, + preemption becomes important in networks using online, distributed + Constrained Shortest Path First (CSPF) strategies for their Traffic + Engineering Label Switched Path (TE LSP) path computation to limit + the impact of bandwidth fragmentation. Moreover, preemption is an + attractive strategy in an MPLS network in which traffic is treated in + a differentiated manner and high-importance traffic may be given + special treatment over lower-importance traffic [DEC-PREP, ATM-PREP]. + Nevertheless, in the DS-TE approach, whose issues and requirements + are discussed in [RFC3564], the preemption policy is considered an + important piece on the bandwidth reservation and management puzzle, + but no preemption strategy is defined. Note that preemption also + plays an important role in regular MPLS Traffic Engineering + environments (with a single pool of bandwidth). + + This document proposes a flexible preemption policy that can be + adjusted in order to give different weight to various preemption + criteria: priority of LSPs to be preempted, number of LSPs to be + preempted, amount of bandwidth preempted, blocking probability. The + implications (cascading effect, bandwidth wastage, priority of + preempted LSPs) of selecting a certain order of importance for the + criteria are discussed for the examples given. + +2. Introduction + + In [RFC2702], issues and requirements for Traffic Engineering in an + MPLS network are highlighted. In order to address both traffic- + oriented and resource-oriented performance objectives, the authors + point out the need for priority and preemption parameters as Traffic + Engineering attributes of traffic trunks. The notion of preemption + and preemption priority is defined in [RFC3272], and preemption + attributes are defined in [RFC2702] and [RFC3209]. + + + + +de Oliveira, et al. Informational [Page 3] + +RFC 4829 LSP Preemption Policies for MPLS-TE April 2007 + + + A traffic trunk is defined as an aggregate of traffic flows belonging + to the same class that are placed inside an LSP [RFC3564]. In this + context, preemption is the act of selecting an LSP that will be + removed from a given path in order to give room to another LSP with a + higher priority (lower preemption number). More specifically, the + preemption attributes determine whether an LSP with a certain setup + preemption priority can preempt another LSP with a lower holding + preemption priority from a given path, when there is competition for + available resources. Note that competing for resources is one + situation in which preemption can be triggered, but other situations + may exist, themselves controlled by a policy. + + For readability, a number of definitions from [RFC3564] are repeated + here: + + Class-Type (CT): The set of Traffic Trunks crossing a link that is + governed by a specific set of Bandwidth constraints. CT is used for + the purposes of link bandwidth allocation, constraint-based routing, + and admission control. A given Traffic Trunk belongs to the same CT + on all links. + + TE-Class: A pair of: + + i. A Class-Type. + + ii. A preemption priority allowed for that Class-Type. This means + that an LSP transporting a Traffic Trunk from that Class-Type can use + that preemption priority as the set-up priority, as the holding + priority, or both. + + By definition, there may be more than one TE-Class using the same CT, + as long as each TE-Class uses a different preemption priority. Also, + there may be more than one TE-Class with the same preemption + priority, provided that each TE-Class uses a different CT. The + network administrator may define the TE-Classes in order to support + preemption across CTs, to avoid preemption within a certain CT, or to + avoid preemption completely, when so desired. To ensure coherent + operation, the same TE-Classes must be configured in every Label + Switched Router (LSR) in the DS-TE domain. + + As a consequence of a per-TE-Class treatment, the Interior Gateway + Protocol (IGP) needs to advertise separate Traffic Engineering + information for each TE-Class, which consists of the Unreserved + Bandwidth (UB) information [RFC4124]. The UB information will be + used by the routers, checking against the bandwidth constraint model + parameters, to decide whether preemption is needed. Details on how + to calculate the UB are given in [RFC4124]. + + + + +de Oliveira, et al. Informational [Page 4] + +RFC 4829 LSP Preemption Policies for MPLS-TE April 2007 + + +3. LSP Setup Procedure and Preemption + + A new LSP setup request has two important parameters: bandwidth and + preemption priority. The set of LSPs to be preempted can be selected + by optimizing an objective function that represents these two + parameters, and the number of LSPs to be preempted. More + specifically, the objective function could be any, or a combination, + of the following [DEC-PREP, ATM-PREP]: + + * Preempt the LSPs that have the least priority (preemption + priority). The Quality of Service (QoS) of high priority traffic + would be better satisfied, and the cascading effect described below + can be limited. + + * Preempt the least number of LSPs. The number of LSPs that need to + be rerouted would be lower. + + * Preempt the least amount of bandwidth that still satisfies the + request. Resource utilization could be improved. The preemption + of larger TE LSPs (more than requested) by the newly signaled TE + LSP implies a larger amount of bandwidth to be rerouted, which is + likely to increase the probability of blocking (inability to find a + path for some TE LSPs). + + * Preempt LSPs that minimize the blocking probability (risk that + preempted TE LSP cannot be rerouted). + + After the preemption selection phase is finished, the selected LSPs + are signaled as preempted and the new LSP is established (if a new + path satisfying the constraints can be found). The UB information is + then updated via flooding of an IGP-TE update and/or simply pruning + the link where preemption occurred. Figure 1 shows a flowchart that + summarizes how each LSP setup request is treated in a preemption- + enabled scenario. + + + + + + + + + + + + + + + + + +de Oliveira, et al. Informational [Page 5] + +RFC 4829 LSP Preemption Policies for MPLS-TE April 2007 + + + LSP Setup Request + (TE-Class i, bw=r) + | + | + v NO + UB[TE-Class i] >= r ? -------> Reject LSP + Setup and flood an updated IGP-TE + | LSA/LSP + |YES + v NO + Preemption Needed ? -------> Setup LSP/Update UB if a threshold is + | crossed + | YES + v + Preemption ----> Setup LSP/Reroute Preempted LSPs + Algorithm Update UB + + Figure 1: Flowchart for LSP setup procedure. + + In [DEC-PREP], the authors propose connection preemption policies + that optimize the discussed criteria in a given order of importance: + number of LSPs, bandwidth, and priority; bandwidth, priority, and + number of LSPs. The novelty in our approach is the use of an + objective function that can be adjusted by the service provider in + order to stress the desired criteria. No particular criteria order + is enforced. Moreover, a new criterion is added to the objective + function: optimize the blocking probability (to minimize the risk + that an LSP is not rerouted after being preempted). + +4. Preemption Cascading + + The decision of preempting an LSP may cause other preemptions in the + network. This is called preemption cascading effect and different + cascading levels may be achieved by the preemption of a single LSP. + The cascading levels are defined in the following manner: when an LSP + is preempted and rerouted without causing any further preemption, the + cascading is said to be of level zero. However, when a preempted LSP + is rerouted, and in order to be established in the new route it also + causes the preemption of other LSPs, the cascading is said to be of + level 1, and so on. + + Preemption cascading is not desirable and therefore policies that + minimize it are of interest. Typically, this can result in severe + network instabilities. In Section 5, a new versatile preemption + heuristic will be presented. In Section 6, preemption simulation + results will be discussed and the cascading effect will be analyzed. + + + + + +de Oliveira, et al. Informational [Page 6] + +RFC 4829 LSP Preemption Policies for MPLS-TE April 2007 + + +5. Preemption Heuristic + +5.1. Preempting Resources on a Path + + It is important to note that once a request for an LSP setup arrives, + each LSR along the TE LSP path checks the available bandwidth on its + outgoing link. For the links in which the available bandwidth is not + enough, the preemption policy needs to be activated in order to + guarantee the end-to-end bandwidth reservation for the new LSP. This + is a distributed approach, in which every node on the path is + responsible for running the preemption algorithm and determining + which LSPs would be preempted in order to fit the new request. A + distributed approach may not lead to an optimal solution. + + Alternatively, in a centralized approach, a manager entity runs the + preemption policy and determines the best LSPs to be preempted in + order to free the required bandwidth in all the links that compose + the path. The preemption policy would try to select LSPs that + overlap with the path being considered (preempt a single LSP that + overlaps with the route versus preempt a single LSP on every link + that belongs to the route). + + Both centralized and distributed approaches have advantages and + drawbacks. A centralized approach would be more precise, but require + that the whole network state be stored and updated accordingly, which + raises scalability issues. In a network where LSPs are mostly + static, an offline decision can be made to reroute LSPs and the + centralized approach could be appropriate. However, in a dynamic + network in which LSPs are set up and torn down in a frequent manner + because of new TE LSPs, bandwidth increase, reroute due to failure, + etc., the correctness of the stored network state could be + questionable. Moreover, the setup time is generally increased when + compared to a distributed solution. In this scenario, the + distributed approach would bring more benefits, even when resulting + in a non-optimal solution (The gain in optimality of a centralized + approach compared to a distributed approach depends on many factors: + network topology, traffic matrix, TE strategy, etc.). A distributed + approach is also easier to be implemented due to the distributed + nature of the current Internet protocols. + + Since the current Internet routing protocols are essentially + distributed, a decentralized approach was selected for the LSP + preemption policy. The parameters required by the new preemption + policies are currently available for OSPF and Intermediate System to + Intermediate System (IS-IS). + + + + + + +de Oliveira, et al. Informational [Page 7] + +RFC 4829 LSP Preemption Policies for MPLS-TE April 2007 + + +5.2. Preemption Heuristic Algorithm + + Consider a request for a new LSP setup with bandwidth b and setup + preemption priority p. When preemption is needed, due to lack of + available resources, the preemptable LSPs will be chosen among the + ones with lower holding preemption priority (higher numerical value) + in order to fit r=b-Abw(l). The variable r represents the actual + bandwidth that needs to be preempted (the requested, b, minus the + available bandwidth on link l: Abw(l)). + + L is the set of active LSPs having a holding preemption priority + lower (numerically higher) than p. So L is the set of candidates for + preemption. b(l) is the bandwidth reserved by LSP l in L, expressed + in bandwidth units, and p(l) is the holding preemption priority of + LSP l. + + In order to represent a cost for each preemption priority, an + associated cost y(l) inversely related to the holding preemption + priority p(l) is defined. For simplicity, a linear relation + y(l)=8-p(l) is chosen. y is a cost vector with L components, y(l). b + is a reserved bandwidth vector with dimension L, and components b(l). + + Concerning the objective function, four main objectives can be + reached in the selection of preempted LSPs: + + * minimize the priority of preempted LSPs, + * minimize the number of preempted LSPs, + * minimize the preempted bandwidth, + * minimize the blocking probability. + + To have the widest choice on the overall objective that each service + provider needs to achieve, the following equation was defined (for + simplicity chosen as a weighted sum of the above mentioned criteria): + + H(l)= alpha y(l) + beta 1/b(l) + gamma (b(l)-r)^2 + theta b(l) + + In this equation: + + - alpha y(l) captures the cost of preempting high priority LSPs. + + - beta 1/b(l) penalizes the preemption of low bandwidth LSPs, + capturing the cost of preempting a large number of LSPs. + + - gamma (b(l)-r)^2 captures the cost of preemption of LSPs that are + much larger or much smaller than r. + + - theta b(l) captures the cost of preempting large LSPs. + + + + +de Oliveira, et al. Informational [Page 8] + +RFC 4829 LSP Preemption Policies for MPLS-TE April 2007 + + + Coefficients alpha, beta, gamma, and theta can be chosen to emphasize + one or more components of H. + + The coefficient theta is defined such that theta = 0 if gamma > 0. + This is because when trying to minimize the blocking probability of + preempted LSPs, the heuristic gives preference to preempting several + small LSPs (therefore gamma, which is the weight for minimizing the + preempted bandwidth enforcing the selection of LSPs with similar + amount of bandwidth as the requested, needs to be set as zero). The + selection of several small LSPs in a normally loaded portion of the + network will increase the chance that such LSPs are successfully + rerouted. Moreover, the selection of several small LSPs may not + imply preempting much more than the required bandwidth (resulting in + low-bandwidth wastage), as it will be seen in the discussed examples. + When preemption is to happen in a heavy loaded portion of the + network, to minimize blocking probability, the heuristic will select + fewer LSPs for preemption in order to increase the chance of + rerouting. + + H is calculated for each LSP in L. The LSPs to be preempted are + chosen as the ones with smaller H that add enough bandwidth to + accommodate r. When sorting LSPs by H, LSPs with the same value for + H are ordered by bandwidth b, in increasing order. For each LSP with + repeated H, the algorithm checks whether the bandwidth b assigned to + only that LSP is enough to satisfy r. If there is no such LSP, it + checks whether the bandwidth of each of those LSPs added to the + previously preempted LSPs' bandwidth is enough to satisfy r. If that + is not true for any LSP in that repeated H-value sequence, the + algorithm preempts the LSP that has the larger amount of bandwidth in + the sequence, and keeps preempting in decreasing order of b until r + is satisfied or the sequence is finished. If the sequence is + finished and r is not satisfied, the algorithm again selects LSPs to + be preempted based on an increasing order of H. More details on the + algorithm are given in [PREEMPTION]. + + When the objective is to minimize blocking, the heuristic will follow + two options on how to calculate H: + + * If the link in which preemption is to happen is normally loaded, + several small LSPs will be selected for preemption using H(l)= + alpha y(l) + theta b(l). + + * If the link is overloaded, few LSPs are selected using H(l)= alpha + y(l) + beta 1/b(l). + + + + + + + +de Oliveira, et al. Informational [Page 9] + +RFC 4829 LSP Preemption Policies for MPLS-TE April 2007 + + +6. Examples + +6.1. Simple Case: Single Link + + We first consider a very simple case, in which the path considered + for preemption is composed by a single hop. The objective of this + example is to illustrate how the heuristic works. In the next + section, we will study a more complex case in which the preemption + policies are being tested on a network. + + Consider a link with 16 LSPs with reserved bandwidth b in Mbps, + preemption holding priority p, and cost y, as shown in Table 1. In + this example, 8 TE-Classes are active. The preemption here is being + performed on a single link as an illustrative example. + + ------------------------------------------------------------------ + LSP L1 L2 L3 L4 L5 L6 L7 L8 + ------------------------------------------------------------------ + Bandwidth (b) 20 10 60 25 20 1 75 45 + Priority (p) 1 2 3 4 5 6 7 5 + Cost (y) 7 6 5 4 3 2 1 3 + ------------------------------------------------------------------ + LSP L9 L10 L11 L12 L13 L14 L15 L16 + ------------------------------------------------------------------ + Bandwidth (b) 100 5 40 85 50 20 70 25 + Priority (p) 3 6 4 5 2 3 4 7 + Cost (y) 5 2 4 3 6 5 4 1 + ------------------------------------------------------------------ + Table 1: LSPs in the considered link. + + A request for an LSP establishment arrives with r=175 Mbps and p=0 + (highest possible priority, which implies that all LSPs with p>0 in + Table 1 will be considered when running the algorithm). Assume + Abw(l)=0. + + If priority is the only important criterion, the network operator + configures alpha=1, beta=gamma=theta=0. In this case, LSPs L6, L7, + L10, L12, and L16 are selected for preemption, freeing 191 bandwidth + units to establish the high-priority LSP. Note that 5 LSPs were + preempted, but all with a priority level between 5 and 7. + + In a network in which rerouting is an expensive task to perform (and + the number of rerouted TE LSPs should be as small as possible), one + might prefer to set beta=1 and alpha=gamma=theta=0. LSPs L9 and L12 + would then be selected for preemption, adding up to 185 bandwidth + units (less wastage than the previous case). The priorities of the + selected LSPs are 3 and 5 (which means that they might themselves + preempt some other LSPs when rerouted). + + + +de Oliveira, et al. Informational [Page 10] + +RFC 4829 LSP Preemption Policies for MPLS-TE April 2007 + + + Suppose the network operator decides that it is more appropriate to + configure alpha=1, beta=10, gamma=0, theta=0 (the parameters were set + to values that would balance the weight of each component, namely + priority and number, in the cost function), because in this network + rerouting is very expensive, LSP priority is important, but bandwidth + is not a critical issue. In this case, LSPs L7, L12, and L16 are + selected for preemption. This configuration results in a smaller + number of preempted LSPs when compared to the first case, and the + priority levels are kept between 5 and 7. + + To take into account the number of LSPs preempted, the preemption + priority, and the amount of bandwidth preempted, the network operator + may set alpha > 0, beta > 0, and gamma > 0. To achieve a balance + among the three components, the parameters need to be normalized. + Aiming for a balance, the parameters could be set as alpha=1, beta=10 + (bringing the term 1/b(l) closer to the other parameters), and + gamma=0.001 (bringing the value of the term (b(l)-r)^2 closer to the + other parameters). LSPs L7 and L9 are selected for preemption, + resulting in exactly 175 bandwidth units and with priorities 3 and 7 + (note that less LSP are preempted but they have a higher priority + which may result in a cascading effect). + + If the minimization of the blocking probability is the criterion of + most interest, the cost function could be configured with theta=1, + alpha=beta=gamma=0. In that case, several small LSPs are selected + for preemption: LSPs L2, L4, L5, L6, L7, L10, L14, and L16. Their + preemption will free 181 Mbps in this link, and because the selected + LSPs have small bandwidth requirement there is a good chance that + each of them will find a new route in the network. + + From the above example, it can be observed that when the priority was + the highest concern and the number of preempted LSPs was not an + issue, 5 LSPs with the lowest priority were selected for preemption. + When only the number of LSPs was an issue, the minimum number of LSPs + was selected for preemption: 2, but the priority was higher than in + the previous case. When priority and number were important factors + and a possible waste of bandwidth was not an issue, 3 LSPs were + selected, adding more bandwidth than requested, but still with low + preemption priority. When considering all the parameters but the + blocking probability, the smallest set of LSP was selected, 2, adding + just enough bandwidth, 175 Mbps, and with priority levels 3 and 7. + + When the blocking probability was the criterion of interest, several + (8) small LSPs were preempted. The bandwidth wastage is low, but the + number of rerouting events will increase. Given the bandwidth + requirement of the preempted LSPs, it is expected that the chances of + finding a new route for each LSP will be high. + + + + +de Oliveira, et al. Informational [Page 11] + +RFC 4829 LSP Preemption Policies for MPLS-TE April 2007 + + +6.2. Network Case + + For these experiments, we consider a 150 nodes topology with an + average network connectivity of 3. 10% of the nodes in the topology + have a degree of connectivity of 6. 10% of the links are OC3, 70% are + OC48, and 20% are OC192. + + Two classes of TE LSPs are in use: Voice LSPs and Data Internet/VPN + LSPs. For each class of TE LSP, the set of preemptions (and the + proportion of LSPs for each preemption) and the size distributions + are as follows (a total of T LSPs is considered): + + T: total number of TE LSPs in the network (T = 18,306 LSPs) + + Voice: + + Number: 20% of T + Preemption: 0, 1 and 2 + Size: uniform distribution between 30M and 50M + + Internet/VPN TE: + + Number: 4% of T + Preemption: 3 + Size: uniform distribution between 20M and 50M + + Number: 8% of T + Preemption 4 + Size: uniform distribution between 15M and 40M + + Number: 8% of T + Preemption 5 + Size: uniform distribution between 10M and 20M + + Number: 20% of T + Preemption 6 + Size: uniform distribution between 1M and 20M + + Number: 40% of T + Preemption 7 + Size: uniform distribution between 1K and 1M + + LSPs are set up mainly due to network failure: a link or a node + failed and LSPs are rerouted. + + + + + + + +de Oliveira, et al. Informational [Page 12] + +RFC 4829 LSP Preemption Policies for MPLS-TE April 2007 + + + The network failure events were simulated with two functions: + + - Constant: 1 failure chosen randomly among the set of links every 1 + hour. + + - Poisson process with interarrival average = 1 hour. + + Table 2 shows the results for simulations with constant failure. The + simulations were run with the preemption heuristic configured to + balance different criteria (left side of the table), and then with + different preemption policies that consider the criteria in a given + order of importance rather than balancing them (right side of the + table). + + The proposed heuristic was configured to balance the following + criteria: + + HPB: The heuristic with priority and bandwidth wastage as the most + important criteria (alpha=10, beta=0, gamma=0.001, theta=0). + + HBlock: The heuristic considering the minimization of blocking + probability (normal load links: alpha=1, beta=0, gamma=0, theta=0.01) + (heavy load links: alpha=1, beta=10). + + HNB: The heuristic with number of preemptions and wasted bandwidth in + consideration (alpha=0, beta=10, gamma=0.001, theta=0). + + Other algorithms that consider the criteria in a given order of + importance: + + P: Sorts candidate LSPs by priority only. + + PN: Sorts the LSPs by priority, and for cases in which the priority + is the same, orders those LSPs by decreasing bandwidth (selects + larger LSPs for preemption in order to minimize number of preempted + LSPs). + + PB: Sorts the LSPs by priority, and for LSPs with the same priority, + sorts those by increasing bandwidth (select smaller LSPs in order to + reduce bandwidth wastage). + + + + + + + + + + + +de Oliveira, et al. Informational [Page 13] + +RFC 4829 LSP Preemption Policies for MPLS-TE April 2007 + + + ------------------------------------------------- + | Heuristic | Other algorithms | + ------------------------------------------------- + | HPB | HBlock| HNB | P | PN | PB | + ----------------------------------------------------------------- + Need to be | 532 | 532 | 532 | 532 | 532 | 532 | + Rerouted | | | | | | | + ----------------------------------------------------------------- + Preempted | 612 | 483 | 619 | 504 | 477 | 598 | + ----------------------------------------------------------------- + Rerouted |467|76%|341|73%|475|77%|347|69%|335|70%|436|73%| + Blocked |145|24%|130|27%|144|23%|157|31%|142|30%|162|27%| + ----------------------------------------------------------------- + Max Cascading | 4.5 | 2 | 5 | 2.75 | 2 | 2.75 | + ----------------------------------------------------------------- + Wasted Bandwidth| | | | | | | + AVR (Mbps) | 6638 | 532 | 6479 | 8247 | 8955 | 6832 | + Worst Case(Mbps)| 35321 |26010 |36809 | 28501 | 31406 | 23449 | + ----------------------------------------------------------------- + Priority | | | | | | | + Average | 6 | 6.5 | 5.8 | 6.6 | 6.6 | 6.6 | + Worst Case | 1.5 | 3.8 | 1.2 | 3.8 | 3.8 | 3.8 | + ----------------------------------------------------------------- + Extra Hops | | | | | | | + Average | 0.23 | 0.25 | 0.22 | 0.25 | 0.25 | 0.23 | + Worst Case | 3.25 | 3 | 3.25 | 3 | 3 | 2.75 | + ----------------------------------------------------------------- + Table 2: Simulation results for constant network failure: + 1 random failure every hour. + + From Table 2, we can conclude that among the heuristic (HPB, HBlock, + HNB) results, HBlock resulted in the smaller number of LSPs being + preempted. More importantly, it also resulted in an overall smaller + rejection rate and smaller average wasted bandwidth (and second + overall smaller worst-case wasted bandwidth.) + + Although HBlock does not try to minimize the number of preempted + LSPs, it ends up doing so, because it preempts LSPs with lower + priority mostly, and therefore it does not propagate cascading much + further. Cascading was the overall lowest (preemption caused at most + two levels of preemption, which was also the case for the policy PN). + The average and worst preemption priority was very satisfactory + (preempting mostly lowest-priority LSPs, like the other algorithms P, + PN, and PB). + + When HPB was in use, more LSPs were preempted as a consequence of the + higher cascading effect. That is due to the heuristic's choice of + preempting LSPs that are very similar in bandwidth size to the + + + +de Oliveira, et al. Informational [Page 14] + +RFC 4829 LSP Preemption Policies for MPLS-TE April 2007 + + + bandwidth size of the preemptor LSP (which can result in preempting a + higher priority LSP and therefore causing cascading). The wasted + bandwidth was reduced when compared to the other algorithms (P, PN, + PB). + + When HNB was used, cascading was higher than the other cases, due to + the fact that LSPs with higher priority could be preempted. When + compared to P, PN, or PB, the heuristic HNB preempted more LSPs (in + fact, it preempted the largest number of LSPs overall, clearly + showing the cascading effect), but the average wasted bandwidth was + smaller, although not as small as HBlock's (the HNB heuristic tries + to preempt a single LSP, meaning it will preempt LSPs that have a + reserved bandwidth similar to the actual bandwidth needed. The + algorithm is not always successful, because such a match may not + exist, and in that case, the wasted bandwidth could be high). The + preempted priority was the highest on average and worse case, which + also shows why the cascading level was also the highest (the + heuristic tries to select LSPs for preemption without looking at + their priority levels). In summary, this policy resulted in a poor + performance. + + Policy PN resulted in the small number of preempted LSPs overall and + small number of LSPs not successfully rerouted. Cascading is low, + but bandwidth wastage is very high (overall highest bandwidth + wastage). Moreover, in several cases in which rerouting happened on + portions of the network that were underloaded, the heuristic HBlock + preempted a smaller number of LSPs than PN. + + Policy P selects a larger number of LSPs (when compared to PN) with + low priority for preemption, and therefore it is able to successfully + reroute less LSPs when compared to HBlock, HPB, HNB, or PN. The + bandwidth wastage is also higher when compared to any of the + heuristic results or to PB, and it could be worse if the network had + LSPs with a low priority and large bandwidth, which is not the case. + + Policy PB, when compared to PN, resulted in a larger number of + preempted LSPs and an overall larger number of blocked LSPs (not + rerouted), due to preemption. Cascading was slightly higher. Since + the selected LSPs have low priority, they are not able to preempt + much further and are blocked when the links are congested. Bandwidth + wastage was smaller since the policy tries to minimize wastage, and + preempted priority was about the same on average and worst case. + + The simulation results show that when preemption is based on + priority, cascading is not critical since the preempted LSPs will not + be able to propagate preemption much further. When the number of + LSPs is considered, fewer LSPs are preempted and the chances of + rerouting increases. When bandwidth wastage is considered, smaller + + + +de Oliveira, et al. Informational [Page 15] + +RFC 4829 LSP Preemption Policies for MPLS-TE April 2007 + + + LSPs are preempted in each link and the wasted bandwidth is low. The + heuristic seems to combine these features, yielding the best results, + especially in the case of blocking probability. When the heuristic + was configured to minimize blocking probability (HBlock), small LSPs + with low priority were selected for preemption on normally loaded + links and fewer (larger) LSPs with low priority were selected on + congested links. Due to their low priority, cascading was not an + issue. Several LSPs were selected for preemption, but the rate of + LSPs that were not successfully rerouted was the lowest. Since the + LSPs are small, it is easier to find a new route in the network. + When selecting LSPs on a congested link, fewer larger LSPs are + selected improving load balance. Moreover, the bandwidth wastage was + the overall lowest. In summary, the heuristic is very flexible and + can be configured according to the network provider's best interest + regarding the considered criteria. + + For several cases, the failure of a link resulted in no preemption at + all (all LSPs were able to find an alternate path in the network) or + resulted in preemption of very few LSPs and subsequent successfully + rerouting of the same with no cascading effect. + + It is also important to note that for all policies in use, the number + of extra hops when LSPs are rerouted was not critical, showing that + preempted LSPs can be rerouted on a path with the same length or a + path that is slightly longer in number of hops. + +7. Security Considerations + + The practice described in this document does not raise specific + security issues beyond those of existing TE. + +8. Acknowledgements + + We would like to acknowledge the input and helpful comments from + Francois Le Faucheur (Cisco Systems) and George Uhl (Swales + Aerospace). + + + + + + + + + + + + + + + +de Oliveira, et al. Informational [Page 16] + +RFC 4829 LSP Preemption Policies for MPLS-TE April 2007 + + +9. Informative References + + [ATM-PREP] Poretsky, S. and Gannon, T., "An Algorithm for + Connection Precedence and Preemption in Asynchronous + Transfer Mode (ATM) Networks", Proceedings of IEEE ICC + 1998. + + [DEC-PREP] Peyravian, M. and Kshemkalyani, A. D. , "Decentralized + Network Connection Preemption Algorithms", Computer + Networks and ISDN Systems, vol. 30 (11), pp. 1029-1043, + June 1998. + + [PREEMPTION] de Oliveira, J. C. et al., "A New Preemption Policy for + DiffServ-Aware Traffic Engineering to Minimize + Rerouting", Proceedings of IEEE INFOCOM 2002. + + [RFC2702] Awduche, D., Malcolm, J., Agogbua, J., O'Dell, M., and + J. McManus, "Requirements for Traffic Engineering Over + MPLS", RFC 2702, September 1999. + + [RFC3209] Awduche, D., Berger, L., Gan, D., Li, T., Srinivasan, + V., and G. Swallow, "RSVP-TE: Extensions to RSVP for + LSP Tunnels", RFC 3209, December 2001. + + [RFC3272] Awduche, D., Chiu, A., Elwalid, A., Widjaja, I., and X. + Xiao, "Overview and Principles of Internet Traffic + Engineering", RFC 3272, May 2002. + + [RFC3564] Le Faucheur, F. and W. Lai, "Requirements for Support + of Differentiated Services-aware MPLS Traffic + Engineering", RFC 3564, July 2003. + + [RFC4124] Le Faucheur, F., "Protocol Extensions for Support of + Diffserv-aware MPLS Traffic Engineering", RFC 4124, + June 2005. + + [RFC4126] Ash, J., "Max Allocation with Reservation Bandwidth + Constraints Model for Diffserv-aware MPLS Traffic + Engineering & Performance Comparisons", RFC 4126, + June 2005. + + [RFC4127] Le Faucheur, F., "Russian Dolls Bandwidth Constraints + Model for Diffserv-aware MPLS Traffic Engineering", + RFC 4127, June 2005. + + + + + + + +de Oliveira, et al. Informational [Page 17] + +RFC 4829 LSP Preemption Policies for MPLS-TE April 2007 + + + [RFC4128] Lai, W., "Bandwidth Constraints Models for + Differentiated Services (Diffserv)-aware MPLS Traffic + Engineering: Performance Evaluation", RFC 4128, + June 2005. + +Authors' Addresses + + Jaudelice C. de Oliveira (editor) + Drexel University + 3141 Chestnut Street (ECE Dept.) + Philadelphia, PA 19104 + USA + + EMail: jau@ece.drexel.edu + + + JP Vasseur (editor) + Cisco Systems, Inc. + 1414 Massachusetts Avenue + Boxborough, MA 01719 + USA + + EMail: jpv@cisco.com + + + Leonardo Chen + Verizon Laboratories + 40 Sylvan Rd. LA0MS55 + Waltham, MA 02451 + USA + + EMail: leonardo.c.chen@verizon.com + + + Caterina Scoglio + Kansas State University + 2061 Rathbone Hall + Manhattan, Kansas 66506-5204 + USA + + EMail: caterina@eece.ksu.edu + + + + + + + + + + +de Oliveira, et al. Informational [Page 18] + +RFC 4829 LSP Preemption Policies for MPLS-TE April 2007 + + +Full Copyright Statement + + Copyright (C) The IETF Trust (2007). + + This document is subject to the rights, licenses and restrictions + contained in BCP 78 and at www.rfc-editor.org/copyright.html, and + except as set forth therein, the authors retain all their rights. + + This document and the information contained herein are provided on an + "AS IS" basis and THE CONTRIBUTOR, THE ORGANIZATION HE/SHE REPRESENTS + OR IS SPONSORED BY (IF ANY), THE INTERNET SOCIETY, THE IETF TRUST AND + THE INTERNET ENGINEERING TASK FORCE DISCLAIM ALL WARRANTIES, EXPRESS + OR IMPLIED, INCLUDING BUT NOT LIMITED TO ANY WARRANTY THAT THE USE OF + THE INFORMATION HEREIN WILL NOT INFRINGE ANY RIGHTS OR ANY IMPLIED + WARRANTIES OF MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. + +Intellectual Property + + The IETF takes no position regarding the validity or scope of any + Intellectual Property Rights or other rights that might be claimed to + pertain to the implementation or use of the technology described in + this document or the extent to which any license under such rights + might or might not be available; nor does it represent that it has + made any independent effort to identify any such rights. Information + on the procedures with respect to rights in RFC documents can be + found in BCP 78 and BCP 79. + + Copies of IPR disclosures made to the IETF Secretariat and any + assurances of licenses to be made available, or the result of an + attempt made to obtain a general license or permission for the use of + such proprietary rights by implementers or users of this + specification can be obtained from the IETF on-line IPR repository at + http://www.ietf.org/ipr. + + The IETF invites any interested party to bring to its attention any + copyrights, patents or patent applications, or other proprietary + rights that may cover technology that may be required to implement + this standard. Please address the information to the IETF at + ietf-ipr@ietf.org. + +Acknowledgement + + Funding for the RFC Editor function is currently provided by the + Internet Society. + + + + + + + +de Oliveira, et al. Informational [Page 19] + |