summaryrefslogtreecommitdiff
path: root/doc/rfc/rfc3612.txt
blob: 2f81233761d6846878c24deed0636b4991c7c748 (plain) (blame)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
774
775
776
777
778
779
780
781
782
783
784
785
786
787
788
789
790
791
792
793
794
795
796
797
798
799
800
801
802
803
804
805
806
807
808
809
810
811
812
813
814
815
816
817
818
819
820
821
822
823
824
825
826
827
828
829
830
831
832
833
834
835
836
837
838
839
840
841
842
843
844
845
846
847
848
849
850
851
852
853
854
855
856
857
858
859
860
861
862
863
864
865
866
867
868
869
870
871
872
873
874
875
876
877
878
879
880
881
882
883
884
885
886
887
888
889
890
891
892
893
894
895
896
897
898
899
Network Working Group                                          A. Farrel
Request for Comments: 3612                            Old Dog Consulting
Category: Informational                                   September 2003


            Applicability Statement for Restart Mechanisms
               for the Label Distribution Protocol (LDP)

Status of this Memo

   This memo provides information for the Internet community.  It does
   not specify an Internet standard of any kind.  Distribution of this
   memo is unlimited.

Copyright Notice

   Copyright (C) The Internet Society (2003).  All Rights Reserved.

Abstract

   This document provides guidance on when it is advisable to implement
   some form of Label Distribution Protocol (LDP) restart mechanism and
   which approach might be more suitable.  The issues and extensions
   described in this document are equally applicable to RFC 3212,
   "Constraint-Based LSP Setup Using LDP".

1.  Introduction

   Multiprotocol Label Switching (MPLS) systems are used in core
   networks where system downtime must be kept to a minimum.  Similarly,
   where MPLS is at the network edges (e.g., in Provider Edge (PE)
   routers) [RFC2547], system downtime must also be kept to a minimum.
   Many MPLS Label Switching Routers (LSRs) may, therefore, exploit
   Fault Tolerant (FT) hardware or software to provide high availability
   of the core networks.

   The details of how FT is achieved for the various components of an FT
   LSR, including the switching hardware and the TCP stack, are
   implementation specific.  How the software module itself chooses to
   implement FT for the state created by the LDP is also implementation
   specific.  However, there are several issues in the LDP specification
   [RFC3036] that make it difficult to implement an FT LSR using the LDP
   protocols without some extensions to those protocols.

   Proposals have been made in [RFC3478] and [RFC3479] to address these
   issues.





Farrel                       Informational                      [Page 1]
^L
RFC 3612        Applicability for LDP Restart Mechanisms  September 2003


2.  Requirements of an LDP FT System

   Many MPLS LSRs may exploit FT hardware or software to provide high
   availability (HA) of core networks.  In order to provide HA, an MPLS
   system needs to be able to survive a variety of faults with minimal
   disruption to the Data Plane, including the following fault types:

   -  failure/hot-swap of the switching fabric in an LSR,

   -  failure/hot-swap of a physical connection between LSRs,

   -  failure of the TCP or LDP stack in an LSR,

   -  software upgrade to the TCP or LDP stacks in an LSR.

   The first two examples of faults listed above may be confined to the
   Data Plane.  Such faults can be handled by providing redundancy in
   the Data Plane which is transparent to LDP operating in the Control
   Plane.  However, the failure of the switching fabric or a physical
   link may have repercussions in the Control Plane since signaling may
   be disrupted.

   The third example may be caused by a variety of events including
   processor or other hardware failure, and software failure.

   Any of the last three examples may impact the Control Plane and will
   require action in the Control Plane to recover.  Such action should
   be designed to avoid disrupting traffic in the Data Plane.  Since
   many recent router architectures can separate the Control and Data
   Planes, it is possible that forwarding can continue unaffected by
   recovery action in the Control Plane.

   In other scenarios, the Data and Control Planes may be impacted by a
   fault, but the needs of HA require the coordinated recovery of the
   Data and Control Planes to a state that existed before the fault.

   The provision of protection paths for MPLS LSP and the protection of
   links, IP routes or tunnels through the use of protection LSPs is
   outside the scope of this document.  See [RFC3469] for further
   information.

3.  General Considerations

   In order for the Data and Control Plane states to be successfully
   recovered after a fault, procedures are required to ensure that the
   state held on a pair of LDP peers (at least one of which was affected





Farrel                       Informational                      [Page 2]
^L
RFC 3612        Applicability for LDP Restart Mechanisms  September 2003


   directly by the fault) are synchronized.  Such procedures must be
   implemented in the Control Plane software modules on the peers using
   Control Plane protocols.

   The required actions may operate fully after the failure (reactive
   recovery) or may contain elements that operate before the fault in
   order to minimize the actions taken after the fault (proactive
   recovery).  It is rare to implement actions that operate solely in
   advance of the failure and do not require any further processing
   after the failure (preventive recovery) - this is because of the
   dynamic nature of signaling protocols and the unpredictability of
   fault timing.

   Reactive recovery actions may include full re-signaling of state and
   re-synchronization of state between peers and synchronization based
   on checkpointing.

   Proactive recovery actions may include hand-shaking state transitions
   and checkpointing.

4.  Specific Issues with the LDP Protocol

   LDP uses TCP to provide reliable connections between LSRs to exchange
   protocol messages to distribute labels and to set up LSPs.  A pair of
   LSRs that have such a connection are referred to as LDP peers.

   TCP enables LDP to assume reliable transfer of protocol messages.
   This means that some of the messages do not need to be acknowledged
   (e.g., Label Release).

   LDP is defined such that if the TCP connection fails, the LSR should
   immediately tear down the LSPs associated with the session between
   the LDP peers, and release any labels and resources assigned to those
   LSPs.

   It is notoriously difficult to provide a Fault Tolerant
   implementation of TCP.  To do so might involve making copies of all
   data sent and received.  This is an issue familiar to implementers of
   other TCP applications, such as BGP.

   During failover affecting the TCP or LDP stacks, therefore, the TCP
   connection may be lost.  Recovery from this position is made worse by
   the fact that LDP control messages may have been lost during the
   connection failure.  Since these messages are unconfirmed, it is
   possible that LSP or label state information will be lost.






Farrel                       Informational                      [Page 3]
^L
RFC 3612        Applicability for LDP Restart Mechanisms  September 2003


   At the very least, the solution to this problem must include a change
   to the basic requirements of LDP so that the failure of an LDP
   session does not require that associated LDP or forwarding state be
   torn down.

   Any changes made to LDP in support of recovery processing must meet
   the following requirements:

   -  offer backward-compatibility with LSRs that do not implement the
      extensions to LDP,

   -  preserve existing protocol rules described in [RFC3036] for
      handling unexpected duplicate messages and for processing
      unexpected messages referring to unknown LSPs/labels.

   Ideally, any solution applicable to LDP should be equally applicable
   to CR-LDP.

5.  Summary of the Features of LDP FT

   LDP Fault Tolerance extensions are described in [RFC3479].  This
   approach involves:

   -  negotiation between LDP peers of the intent to support extensions
      to LDP that facilitate recovery from failover without loss of
      LSPs,

   -  selection of FT survival on a per LSP/label basis or for all
      labels on a session,

   -  sequence numbering of LDP messages to facilitate acknowledgement
      and checkpointing,

   -  acknowledgement of LDP messages to ensure that a full handshake is
      performed on those messages either frequently (such as per
      message) or less frequently as in checkpointing,

   -  solicitation of up-to-date acknowledgement (checkpointing) of
      previous LDP messages to ensure the current state is secured, with
      an additional option that allows an LDP partner to request that
      state is flushed in both directions if graceful shutdown is
      required,

   -  a timer to control how long LDP and forwarding state should be
      retained after the LDP session failure, but before being discarded
      if LDP communications are not re-established,





Farrel                       Informational                      [Page 4]
^L
RFC 3612        Applicability for LDP Restart Mechanisms  September 2003


   -  exchange of checkpointing information on LDP session recovery to
      establish what state has been retained by recovering LDP peers,

   -  re-issuing lost messages after failover to ensure that LSP/label
      state is correctly recovered after reconnection of the LDP
      session.

   The FT procedures in [RFC3479] concentrate on the preservation of
   label state for labels exchanged between a pair of adjacent LSRs when
   the TCP connection between those LSRs is lost.  There is no intention
   within these procedures to support end-to-end protection for LSPs.

6.  Summary of the Features of LDP Graceful Restart

   LDP graceful restart extensions are defined in [RFC3478].  This
   approach involves:

   -  negotiation between LDP peers of the intent to support extensions
      to LDP that facilitate recovery from failover without loss of
      LSPs,

   -  a mechanism whereby an LSR that restarts can relearn LDP state by
      resynchronization with its peers,

   -  use of the same mechanism to allow LSRs recovering from an LDP
      session failure to resynchronize LDP state with their peers
      provided that at least one of the LSRs has retained state across
      the failure or has itself resynchronized state with its peers,

   -  a timer to control how long LDP and forwarding state should be
      retained after the LDP session failure, but before being discarded
      if LDP communications are not re-established,

   -  a timer to control the length of the resynchronization period
      between adjacent peers should be completed.

   The procedures in [RFC3478] are applicable to all LSRs, both those
   with the ability to preserve forwarding state during LDP restart and
   those without.  LSRs that can not preserve their MPLS forwarding
   state across the LDP restart would impact MPLS traffic during
   restart.  However, by implementing a subset of the mechanisms in
   [RFC3478] they can minimize the impact if their neighbor(s) are
   capable of preserving their forwarding state across the restart of
   their LDP sessions or control planes by implementing the mechanism in
   [RFC3478].






Farrel                       Informational                      [Page 5]
^L
RFC 3612        Applicability for LDP Restart Mechanisms  September 2003


7.  Applicability Considerations

   This section considers the applicability of fault tolerance schemes
   within LDP networks and considers issues that might lead to the
   choice of one method or another.  Many of the points raised below
   should be viewed as implementation issues rather than specific
   drawbacks of either solution.

7.1.  General Applicability

   The procedures described in [RFC3478] and [RFC3479] are intended to
   cover two distinct scenarios.  In Session Failure, the LDP peers at
   the ends of a session remain active, but the session fails and is
   restarted.  Note that session failure does not imply failure of the
   data channel even when using an in-band control channel.  In Node
   Failure, the session fails because one of the peers has been
   restarted (or at least, the LDP component of the node has been
   restarted).  These two scenarios have different implications for the
   ease of retention of LDP state within an individual LSR, and are
   described in sections below.

   These techniques are only applicable in LDP networks where at least
   one LSR has the capability to retain LDP signaling state and the
   associated forwarding state across LDP session failure and recovery.
   In [RFC3478], the LSRs retaining state do not need to be adjacent to
   the failed LSR or session.

   If traffic is not to be impacted, both LSRs at the ends of an LDP
   session must at least preserve forwarding state.  Preserving LDP
   state is not a requirement to preserve traffic.

   [RFC3479] requires that the LSRs at both ends of the session
   implement the procedures that it describes.  Thus, either traffic is
   preserved and recovery resynchronizes state, or no traffic is
   preserved and the LSP fails.

   Further, to use the procedures of [RFC3479] to recover state on a
   session, both LSRs must have a mechanism for maintaining some session
   state and a way of auditing the forwarding state and the
   resynhcronized control state.

   [RFC3478] is scoped to support preservation of traffic if both LSRs
   implement the procedures that it describes.  Additionally, it
   functions if only one LSR on the failed session supports retention of
   forwarding state, and implements the mechanisms in the document.  In
   this case, traffic will be impacted by the session failure, but the
   forwarding state will be recovered on session recovery.  Further, in
   the event of simultaneous failures, [RFC3478] is capable of



Farrel                       Informational                      [Page 6]
^L
RFC 3612        Applicability for LDP Restart Mechanisms  September 2003


   relearning and redistributing state across multiple LSRs by combining
   its mechanisms with the usual LDP message exchanges of [RFC3036].

7.2.  Session Failure

   In Session Failure, an LDP session between two peers fails and is
   restarted.  There is no restart of the LSRs at either end of the
   session and LDP continues to function on those nodes.

   In these cases, it is simple for LDP implementations to retain the
   LDP state associated with the failed session and to associate the
   state with the new session when it is established.  Housekeeping may
   be applied to determine that the failed session is not returning and
   to release the old LDP state.  Both [RFC3478] and [RFC3479] handle
   this case.

   Applicability of [RFC3478] and [RFC3479] to the Session Failure
   scenario should be considered with respect to the availability of the
   data plane.

   In some cases the failure of the LDP session may be independent of
   any failure of the physical (or virtual) link(s) between adjacent
   peers; for example, it might represent a failure of the TCP/IP stack.
   In these cases, the data plane is not impacted and both [RFC3478] and
   [RFC3479] are applicable to preserve or restore LDP state.

   LDP signaling may also operate out of band; that is, it may use
   different links from the data plane.  In this case, a failure of the
   LDP session may be a result of a failure of the control channel, but
   there is no implied failure of the data plane.  For this scenario
   [RFC3478] and [RFC3479] are both applicable to preserve or restore
   LDP state.

   In the case where the failure of the LDP session also implies the
   failure of the data plane, it may be an implementation decision
   whether LDP peers retain forwarding state, and for how long.  In such
   situations, if forwarding state is retained, and if the LDP session
   is re-established, both [RFC3478] and [RFC3479] are applicable to
   preserve or restore LDP state.

   When the data plane has been disrupted an objective of a recovery
   implementation might be to restore data traffic as quickly as
   possible.








Farrel                       Informational                      [Page 7]
^L
RFC 3612        Applicability for LDP Restart Mechanisms  September 2003


7.3.  Controlled Session Failure

   In some circumstances, the LSRs may know in advance that an LDP
   session is going fail (e.g., perhaps a link is going to be taken out
   of service).

   [RFC3036] includes provision for controlled shutdown of a session.
   [RFC3478] and [RFC3479] allow resynchronization of LDP state upon
   re-establishment of the session.

   [RFC3479] offers the facility to both checkpoint all LDP states
   before the shut-down, and to quiesce the session so that no new state
   changes are attempted between the checkpoint and the shut-down.  This
   means that on recovery, resynchronization is simple and fast.

   [RFC3478] resynchronizes all state on recovery regardless of the
   nature of the shut-down.

7.4.  Node Failure

   Node Failure describes events where a whole node is restarted or
   where the component responsible for LDP signaling is restarted.  Such
   an event will be perceived by the LSR's peers as session failure, but
   the restarting node sees the restart as full re-initialization.

   The basic requirement is that the forwarding state is retained,
   otherwise the data plane will necessarily be interrupted.  If
   forwarding state is not retained, it may be relearned from the saved
   control state in [RFC3479].  [RFC3478] does not utilize or expect a
   saved control state.  If a node restarts without preserved forwarding
   state it informs its neighbors, which immediately delete all label-
   FEC bindings previously received from the restarted node.

   The ways to retain a forwarding and control state are numerous and
   implementation specific.  It is not the purpose of this document to
   espouse one mechanism or another, nor even to suggest how this might
   be done.  If state has been preserved across the restart,
   synchronization with peers can be carried out as though recovering
   from Session Failure as in the previous section.  Both [RFC3478] and
   [RFC3479] support this case.

   How much control state is retained is largely an implementation
   choice, but [RFC3479] requires that at least small amount of per-
   session control state be retained.  [RFC3478] does not require or
   expect control state to be retained.

   It is also possible that the restarting LSR has not preserved any
   state.  In this case, [RFC3479] is of no help.  [RFC3478] however,



Farrel                       Informational                      [Page 8]
^L
RFC 3612        Applicability for LDP Restart Mechanisms  September 2003


   allows the restarting LSR to relearn state from each adjacent peer
   through the processes for resynchronizing after Session Failure.
   Further, in the event of simultaneous failure of multiple adjacent
   nodes, the nodes at the edge of the failure zone can recover state
   from their active neighbors and distribute it to the other recovering
   LSRs without any failed LSR having to have saved state.

7.5.  Controlled Node Failure

   In some cases (hardware repair, software upgrade, etc.), node failure
   may be predictable.  In these cases all sessions with peers may be
   shutdown and existing state retention may be enhanced by special
   actions.

   [RFC3479] checkpointing and quiesce may be applied to all sessions so
   that state is up-to-date.

   As above, [RFC3478] does not require that state is retained by the
   restarting node, but can utilize it if it is.

7.6.  Speed of Recovery

   Speed of recovery is impacted by the amount of signaling required.

   If forwarding state is preserved on both LSRs on the failed session,
   then the recovery time is constrained by the time to resynchronize
   the state between the two LSRs.

   [RFC3479] may resynchronize very quickly.  In a stable network, this
   resolves to a handshake of a checkpoint.  At the most,
   resynchronization involves this handshake plus an exchange of
   messages to handle state changes since the checkpoint was taken.
   Implementations that support only the periodic checkpointing subset
   of [RFC3479] are more likely to have additional state to
   resynchronize.

   [RFC3478] must resynchronize state for all label mappings that have
   been retained.  At the same time, resources that have been retained
   by a restarting upstream LSR but are not actually required, because
   they have been released by the downstream LSR (perhaps because it was
   in the process of releasing the state), they must be held for the
   full resynchronization time to ensure that they are not needed.

   The impact of recovery time will vary according to the use of the
   network.  Both [RFC3478] and [RFC3479] allow advertisement of new
   labels while resynchronization is in progress.  Issues to consider
   are re-availability of falsely retained resources and conflict
   between retained label mappings and newly advertised ones.  This may



Farrel                       Informational                      [Page 9]
^L
RFC 3612        Applicability for LDP Restart Mechanisms  September 2003


   cause incorrect forwarding of data (since labels are advertised from
   downstream), an LSR upstream of a failure may continue to forward
   data for one FEC on an old label while the recovering downstream LSR
   might re-assign that label to another FEC and advertise it.  For this
   reason, restarting LSRs may choose to not advertise new labels until
   resynchronization with their peers has completed, or may decide to
   use special techniques to cover the short period of overlap between
   resynchronization and new LSP setup.

7.7.  Scalability

   Scalability is largely the same issue as speed of recovery and is
   governed by the number of LSPs managed through the failed session(s).

   Note that there are limits to how small the resynchronization time in
   [RFC3478] may be made given the capabilities of the LSRs, the
   throughput on the link between them, and the number of labels that
   must be resynchronized.

   Impact on normal operation should also be considered.

   [RFC3479] requires acknowledgement of all messages.  These
   acknowledgements may be deferred as for checkpointing described in
   section 4, or may be frequent.  Although acknowledgements can be
   piggy-backed on other state messages, an option for frequent
   acknowledgement is to send a message solely for the purpose of
   acknowledging a state change message.  Such an implementation would
   clearly be unwise in a busy network.

   [RFC3478] has no impact on normal operations.

7.8.  Rate of Change of LDP State

   Some networks do not show a high degree of change over time, such as
   those using targeted LDP sessions; others change the LDP forwarding
   state frequently, perhaps reacting to changes in routing information
   on LDP discovery sessions.

   Rate of change of LDP state exchanged over an LDP session depends on
   the application for which the LDP session is being used.  LDP
   sessions used for exchanging <FEC, label> bindings for establishing
   hop by hop LSPs will typically exchange state reacting to IGP
   changes.  Such exchanges could be frequent.  On the other hand, LDP
   sessions established for exchanging MPLS Layer 2 VPN FECs will
   typically exhibit a smaller rate of state exchange.






Farrel                       Informational                     [Page 10]
^L
RFC 3612        Applicability for LDP Restart Mechanisms  September 2003


   In [RFC3479], two options exist.  The first uses a frequent (up to
   per-message) acknowledgement system which is most likely to be
   applicable in a more dynamic system where it is desirable to preserve
   the maximum amount of state over a failure to reduce the level of
   resynchronization required and to speed the recovery time.

   The second option in [RFC3479] uses a less-frequent acknowledgement
   scheme known as checkpointing.  This is particularly suitable to
   networks where changes are infrequent or bursty.

   [RFC3478] resynchronizes all state on recovery regardless of the rate
   of change of the network before the failure.  This consideration is
   thus not relevant to the choice of [RFC3478].

7.9.  Label Distribution Modes

   Both [RFC3478] and [RFC3479] are suitable for use with Downstream
   Unsolicited label distribution.

   [RFC3478] describes Downstream-On-Demand as an area for future study
   and is therefore not applicable for a network in which this label
   distribution mode is used.  It is possible that future examination of
   this issue will reveal that once a label has been distributed in
   either distribution mode, it can be redistributed by [RFC3478] upon
   session recovery.

   [RFC3479] is suitable for use in a network that uses Downstream-On-
   Demand label distribution.

   In theory, and according to [RFC3036], even in networks configured to
   utilize Downstream Unsolicited label distribution, there may be
   occasions when the use of Downstream-On-Deman distribution is
   desirable.  The use of the Label Request message is not prohibited in
   a Downstream Unsolicited label distribution LDP network.

   Opinion varies as to whether there is a practical requirement for the
   use of the Label Request message in a Downstream Unsolicited label
   distribution LDP network.  Current deployment experience suggests
   that there is no requirement.

7.10.  Implementation Complexity

   Implementation complexity has consequences for the implementer and
   also for the deployer since complex software is more error prone and
   harder to manage.






Farrel                       Informational                     [Page 11]
^L
RFC 3612        Applicability for LDP Restart Mechanisms  September 2003


   [RFC3479] is a more complex solution than [RFC3478].  In particular,
   [RFC3478] does not require any modification to the normal signaling
   and processing of LDP state changing messages.

   [RFC3479] implementations may be simplified by implementing only the
   checkpointing subset of the functionality.

7.11.  Implementation Robustness

   In addition to the implication for robustness associated with
   complexity of the solutions, consideration should be given to the
   effects of state preservation on robustness.

   If state has become incorrect for whatever reason, then state
   preservation may retain incorrect state.  In extreme cases, it may be
   that the incorrect state is the cause of the failure in which case
   preserving that state would be inappropriate.

   When state is preserved, the precise amount that is retained is an
   implementation issue.  The basic requirement is that forwarding state
   is retained (to preserve the data path) and that that state can be
   accessed by the LDP software component.

   In both solutions, if the forwarding state is incorrect and is
   retained, it will continue to be incorrect.  Both solutions have a
   mechanism to housekeep and free the unwanted state after
   resynchronization is complete.  [RFC3478] may be better at
   eradicating incorrect forwarding state, because it replays all
   message exchanges that caused the state to be populated.

   In [RFC3478], no more data than the forwarding state needs to have
   been saved by the recovering node.  All LDP state may be relearned by
   message exchanges with peers.  Whether those exchanges may cause the
   same incorrect state to arise on the recovering node is an obvious
   concern.

   In [RFC3479], the forwarding state must be supplemented by a small
   amount of state specific to the protocol extensions.  LDP state may
   be retained directly or reconstructed from the forwarding state.  The
   same issues apply when reconstructing state but are mitigated by the
   fact that this is likely a different code path.  Errors in the
   retained state specific to the protocol extensions will persist.

7.12.  Interoperability and Backward Compatibility

   It is important that new additions to LDP interoperate with existing
   implementations at least in provision of the existing levels of
   function.



Farrel                       Informational                     [Page 12]
^L
RFC 3612        Applicability for LDP Restart Mechanisms  September 2003


   Both [RFC3478] and [RFC3479] do this through rules for handling the
   absence of the FT optional negotiation object during session
   initialization.

   Additionally, [RFC3478] is able to perform limited recovery (i.e.,
   redistribution of state) even when only one of the participating LSRs
   supports the procedures.  This may offer considerable advantages in
   interoperation with legacy implementations.

7.13.  Interaction With Other Label Distribution Mechanisms

   Many LDP LSRs also run other label distribution mechanisms.  These
   include management interfaces for configuration of static label
   mappings, other distinct instances of LDP, and other label
   distribution protocols.  The last example includes traffic
   engineering label distribution protocol that are used to construct
   tunnels through which LDP LSPs are established.

   As with re-use of individual labels by LDP within a restarting LDP
   system, care must be taken to prevent labels that need to be retained
   by a restarting LDP session or protocol component from being used by
   another label distribution mechanism.  This might compromise data
   security, amongst other things.

   It is a matter for implementations to avoid this issue through the
   use of techniques, such as a common label management component or
   segmented label spaces.

7.14.  Applicability to CR-LDP

   CR-LDP [RFC3212] utilizes Downstream-On-Demand label distribution.
   [RFC3478] describes Downstream-On-Demand as an area for future study
   and is therefore not applicable for CR-LDP.  [RFC3479] is suitable
   for use in a network entirely based on CR-LDP or in one that is mixed
   between LDP and CR-LDP.

8.  Security Considerations

   This document is informational and introduces no new security
   concerns.

   The security considerations pertaining to the original LDP protocol
   [RFC3036] remain relevant.

   [RFC3478] introduces the possibility of additional denial-of- service
   attacks.  All of these attacks may be countered by use of an
   authentication scheme between LDP peers, such as the MD5-based scheme
   outlined in [LDP].



Farrel                       Informational                     [Page 13]
^L
RFC 3612        Applicability for LDP Restart Mechanisms  September 2003


   In MPLS, a data mis-delivery security issue can arise if an LSR
   continues to use labels after expiration of the session that first
   caused them to be used.  Both [RFC3478] and [RFC3479] are open to
   this issue.

9.  Intellectual Property Statement

   The IETF takes no position regarding the validity or scope of any
   intellectual property or other rights that might be claimed to
   pertain to the implementation or use of the technology described in
   this document or the extent to which any license under such rights
   might or might not be available; neither does it represent that it
   has made any effort to identify any such rights.  Information on the
   IETF's procedures with respect to rights in standards-track and
   standards-related documentation can be found in BCP-11.  Copies of
   claims of rights made available for publication and any assurances of
   licenses to be made available, or the result of an attempt made to
   obtain a general license or permission for the use of such
   proprietary rights by implementors or users of this specification can
   be obtained from the IETF Secretariat.

   The IETF invites any interested party to bring to its attention any
   copyrights, patents or patent applications, or other proprietary
   rights which may cover technology that may be required to practice
   this standard.  Please address the information to the IETF Executive
   Director.

10.  References

10.1.  Normative References

   [RFC2026]    Bradner, S., "The Internet Standards Process -- Revision
                3", BCP 9, RFC 2026, October 1996.

   [RFC2119]    Bradner, S., "Key words for use in RFCs to Indicate
                Requirement Levels", BCP 14, RFC 2119, March 1997.

   [RFC3036]    Andersson, L., Doolan, P., Feldman, N., Fredette, A. and
                B. Thomas, "LDP Specification", RFC 3036, January 2001.

   [RFC3478]    Leelanivas, M., Rekhter, Y. and R. Aggarwal, "Graceful
                Restart Mechanism for LDP", RFC 3478, February 2003.

   [RFC3479]    Farrel, A., Editor, "Fault Tolerance for the Label
                Distribution Protocol (LDP)", RFC 3479, February 2003.






Farrel                       Informational                     [Page 14]
^L
RFC 3612        Applicability for LDP Restart Mechanisms  September 2003


10.2.  Informative References

   [RFC2547]    Rosen, E. and Y. Rekhter, "BGP/MPLS VPNs", RFC 2547,
                March 1999.

   [RFC3212]    Jamoussi, B., Editor, Andersson, L., Callon, R., Dantu,
                R., Wu, L., Doolan, P., Worster, T., Feldman, N.,
                Fredette, A., Girish, M., Gray, E., Heinanen, J., Kilty,
                T. and A. Malis, "Constraint-Based LSP Setup using LDP",
                RFC 3212, January 2002.

   [RFC3469]    Sharma, V., Ed., and F. Hellstrand, Ed., "Framework for
                Multi-Protocol Label Switching (MPLS)-based Recovery",
                RFC 3469, February 2003.

11.  Acknowledgements

   The author would like to thank the authors of [RFC3478] and [RFC3479]
   for their work on fault tolerance of LDP.  Many thanks to Yakov
   Rekhter, Rahul Aggarwal, Manoj Leelanivas and Andrew Malis for their
   considered input to this applicability statement.

12.  Author's Address

   Adrian Farrel
   Old Dog Consulting

   Phone:  +44 (0) 1978 860944
   EMail:  adrian@olddog.co.uk






















Farrel                       Informational                     [Page 15]
^L
RFC 3612        Applicability for LDP Restart Mechanisms  September 2003


13.  Full Copyright Statement

   Copyright (C) The Internet Society (2003).  All Rights Reserved.

   This document and translations of it may be copied and furnished to
   others, and derivative works that comment on or otherwise explain it
   or assist in its implementation may be prepared, copied, published
   and distributed, in whole or in part, without restriction of any
   kind, provided that the above copyright notice and this paragraph are
   included on all such copies and derivative works.  However, this
   document itself may not be modified in any way, such as by removing
   the copyright notice or references to the Internet Society or other
   Internet organizations, except as needed for the purpose of
   developing Internet standards in which case the procedures for
   copyrights defined in the Internet Standards process must be
   followed, or as required to translate it into languages other than
   English.

   The limited permissions granted above are perpetual and will not be
   revoked by the Internet Society or its successors or assignees.

   This document and the information contained herein is provided on an
   "AS IS" basis and THE INTERNET SOCIETY AND THE INTERNET ENGINEERING
   TASK FORCE DISCLAIMS ALL WARRANTIES, EXPRESS OR IMPLIED, INCLUDING
   BUT NOT LIMITED TO ANY WARRANTY THAT THE USE OF THE INFORMATION
   HEREIN WILL NOT INFRINGE ANY RIGHTS OR ANY IMPLIED WARRANTIES OF
   MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE.

Acknowledgement

   Funding for the RFC Editor function is currently provided by the
   Internet Society.



















Farrel                       Informational                     [Page 16]
^L