1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
774
775
776
777
778
779
780
781
782
783
784
785
786
787
788
789
790
791
792
793
794
795
796
797
798
799
800
801
802
803
804
805
806
807
808
809
810
811
812
813
814
815
816
817
818
819
820
821
822
823
824
825
826
827
828
829
830
831
832
833
834
835
836
837
838
839
840
841
842
843
844
845
846
847
848
849
850
851
852
853
854
855
856
857
858
859
860
861
862
863
864
865
866
867
868
869
870
871
872
873
874
875
876
877
878
879
880
881
882
883
884
885
886
887
888
889
890
891
892
893
894
895
896
897
898
899
900
901
902
903
904
905
906
907
908
909
910
911
912
913
914
915
916
917
918
919
920
921
922
923
924
925
926
927
928
929
930
931
932
933
934
935
936
937
938
939
940
941
942
943
944
945
946
947
948
949
950
951
952
953
954
955
956
957
958
959
960
961
962
963
964
965
966
967
968
969
970
971
972
973
974
975
976
977
978
979
980
981
982
983
984
985
986
987
988
989
990
991
992
993
994
995
996
997
998
999
1000
1001
1002
1003
1004
1005
1006
1007
1008
1009
1010
1011
1012
1013
1014
1015
1016
1017
1018
1019
1020
1021
1022
1023
1024
1025
1026
1027
1028
1029
1030
1031
1032
1033
1034
1035
1036
1037
1038
1039
1040
1041
1042
1043
1044
1045
1046
1047
1048
1049
1050
1051
1052
1053
1054
1055
1056
1057
1058
1059
1060
1061
1062
1063
1064
1065
1066
1067
1068
1069
1070
1071
1072
1073
1074
1075
1076
1077
1078
1079
1080
1081
1082
1083
1084
1085
1086
1087
1088
1089
1090
1091
1092
1093
1094
1095
1096
1097
1098
1099
1100
1101
1102
1103
1104
1105
1106
1107
1108
1109
1110
1111
1112
1113
1114
1115
1116
1117
1118
1119
1120
1121
1122
1123
1124
1125
1126
1127
1128
1129
1130
1131
1132
1133
1134
1135
1136
1137
1138
1139
1140
1141
1142
1143
1144
1145
1146
1147
1148
1149
1150
1151
1152
1153
1154
1155
1156
1157
1158
1159
1160
1161
1162
1163
1164
1165
1166
1167
1168
1169
1170
1171
1172
1173
1174
1175
1176
1177
1178
1179
1180
1181
1182
1183
1184
1185
1186
1187
1188
1189
1190
1191
1192
1193
1194
1195
1196
1197
1198
1199
1200
1201
1202
1203
1204
1205
1206
1207
1208
1209
1210
1211
1212
1213
1214
1215
1216
1217
1218
1219
1220
1221
1222
1223
1224
1225
1226
1227
1228
1229
1230
1231
1232
1233
1234
1235
1236
1237
1238
1239
1240
1241
1242
1243
1244
1245
1246
1247
1248
1249
1250
1251
1252
1253
1254
1255
1256
1257
1258
1259
1260
1261
1262
1263
1264
1265
1266
1267
1268
1269
1270
1271
1272
1273
1274
1275
1276
1277
1278
1279
1280
1281
1282
1283
1284
1285
1286
1287
1288
1289
1290
1291
1292
1293
1294
1295
1296
1297
1298
1299
1300
1301
1302
1303
1304
1305
1306
1307
1308
1309
1310
1311
1312
1313
1314
1315
1316
1317
1318
1319
1320
1321
1322
1323
1324
1325
1326
1327
1328
1329
1330
1331
1332
1333
1334
1335
1336
1337
1338
1339
1340
1341
1342
1343
1344
1345
1346
1347
1348
1349
1350
1351
1352
1353
1354
1355
1356
1357
1358
1359
1360
1361
1362
1363
1364
1365
1366
1367
1368
1369
1370
1371
1372
1373
1374
1375
1376
1377
1378
1379
1380
1381
1382
1383
1384
1385
1386
1387
1388
1389
1390
1391
1392
1393
1394
1395
1396
1397
1398
1399
1400
1401
1402
1403
1404
1405
1406
1407
1408
1409
1410
1411
1412
1413
1414
1415
1416
1417
1418
1419
1420
1421
1422
1423
1424
1425
1426
1427
1428
1429
1430
1431
1432
1433
1434
1435
1436
1437
1438
1439
1440
1441
1442
1443
1444
1445
1446
1447
1448
1449
1450
1451
1452
1453
1454
1455
1456
1457
1458
1459
1460
1461
1462
1463
1464
1465
1466
1467
1468
1469
1470
1471
1472
1473
1474
1475
1476
1477
1478
1479
1480
1481
1482
1483
1484
1485
1486
1487
1488
1489
1490
1491
1492
1493
1494
1495
1496
1497
1498
1499
1500
1501
1502
1503
1504
1505
1506
1507
1508
1509
1510
1511
1512
1513
1514
1515
1516
1517
1518
1519
1520
1521
1522
1523
1524
1525
1526
1527
1528
1529
1530
1531
1532
1533
1534
1535
1536
1537
1538
1539
1540
1541
1542
1543
1544
1545
1546
1547
1548
1549
1550
1551
1552
1553
1554
1555
1556
1557
1558
1559
1560
1561
1562
1563
1564
1565
1566
1567
1568
1569
1570
1571
1572
1573
1574
1575
1576
1577
1578
1579
1580
1581
1582
1583
1584
1585
1586
1587
1588
1589
1590
1591
1592
1593
1594
1595
1596
1597
1598
1599
1600
1601
1602
1603
1604
1605
1606
1607
1608
1609
1610
1611
1612
1613
1614
1615
1616
1617
1618
1619
1620
1621
1622
1623
1624
1625
1626
1627
1628
1629
1630
1631
1632
1633
1634
1635
1636
1637
1638
1639
1640
1641
1642
1643
1644
1645
1646
1647
1648
1649
1650
1651
1652
1653
1654
1655
1656
1657
1658
1659
1660
1661
1662
1663
1664
1665
1666
1667
1668
1669
1670
1671
1672
1673
1674
1675
1676
1677
1678
1679
1680
1681
1682
1683
1684
1685
1686
1687
1688
1689
1690
1691
1692
1693
1694
1695
1696
1697
1698
1699
1700
1701
1702
1703
1704
1705
1706
1707
1708
1709
1710
1711
1712
1713
1714
1715
1716
1717
1718
1719
1720
1721
1722
1723
1724
1725
1726
1727
1728
1729
1730
1731
1732
1733
1734
1735
1736
1737
1738
1739
|
Network Working Group E. Stokes
Request for Comments: 3384 IBM
Category: Informational R. Weiser
Digital Signature Trust
R. Moats
Lemur Networks
R. Huber
AT&T Laboratories
October 2002
Lightweight Directory Access Protocol (version 3)
Replication Requirements
Status of this Memo
This memo provides information for the Internet community. It does
not specify an Internet standard of any kind. Distribution of this
memo is unlimited.
Copyright Notice
Copyright (C) The Internet Society (2002). All Rights Reserved.
Abstract
This document discusses the fundamental requirements for replication
of data accessible via the Lightweight Directory Access Protocol
(version 3) (LDAPv3). It is intended to be a gathering place for
general replication requirements needed to provide interoperability
between informational directories.
Table of Contents
1 Introduction...................................................2
2 Terminology....................................................3
3 The Models.....................................................5
4 Requirements...................................................7
4.1 General........................................................7
4.2 Model..........................................................8
4.3 Protocol.......................................................9
4.4 Schema........................................................10
4.5 Single Master.................................................10
4.6 Multi-Master..................................................11
4.7 Administration and Management.................................11
4.8 Security......................................................12
5 Security Considerations.......................................13
6 Acknowledgements..............................................13
Stokes, et. al. Informational [Page 1]
^L
RFC 3384 LDAPv3 Replication Requirements October 2002
7 References....................................................13
A Appendix A - Usage Scenarios..................................15
A.1 Extranet Example..............................................15
A.2 Consolidation Example.........................................15
A.3 Replication Heterogeneous Deployment Example..................16
A.4 Shared Name Space Example.....................................16
A.5 Supplier Initiated Replication................................16
A.6 Consumer Initiated Replication................................17
A.7 Prioritized attribute replication.............................17
A.8 Bandwidth issues..............................................17
A.9 Interoperable Administration and Management...................18
A.10 Enterprise Directory Replication Mesh.........................18
A.11 Failure of the Master in a Master-Slave Replicated Directory..19
A.12 Failure of a Directory Holding Critical Service Information...19
B Appendix B - Rationale........................................20
B.1 Meta-Data Implications........................................20
B.2 Order of Transfer for Replicating Data........................20
B.3 Schema Mismatches and Replication.............................21
B.4 Detecting and Repairing Inconsistencies Among Replicas........22
B.5 Some Test Cases for Conflict Resolution in Multi-Master
Replication...................................................23
B.6 Data Confidentiality and Data Integrity During Replication....27
B.7 Failover in Single-Master Systems.............................27
B.8 Including Operational Attributes in Atomic Operations.........29
Authors' Addresses............................................30
Full Copyright Statement......................................31
1 Introduction
Distributing directory information throughout the network provides a
two-fold benefit: (1) it increases the reliability of the directory
through fault tolerance, and (2) it brings the directory content
closer to the clients using the data. LDAP's success as an access
protocol for directory information is driving the need to distribute
LDAP directory content within the enterprise and Internet.
Currently, LDAP does not define a replication mechanism, and mentions
LDAP shadow servers (see [RFC2251]) in passing. A standard mechanism
for directory replication in a multi-vendor environment is critical
to the continued success of LDAP in the market place.
This document sets out the requirements for replication between
multiple LDAP servers. While RFC 2251 and RFC 2252 [RFC2252] set
forth the standards for communication between LDAP clients and
servers there are additional requirements for server-to-server
communication. Some of these are covered here.
This document first introduces the terminology to be used, then
presents the different replication models being considered.
Stokes, et. al. Informational [Page 2]
^L
RFC 3384 LDAPv3 Replication Requirements October 2002
Requirements follow, along with security considerations. The
reasoning that leads to the requirements is presented in the
Appendices. This was done to provide a clean separation of the
requirements from their justification.
The key words "MUST", "MUST NOT", "REQUIRED", "SHALL", "SHALL NOT",
SHOULD", "SHOULD NOT", "RECOMMENDED", "MAY", and "OPTIONAL" in this
document are to be interpreted as described in [RFC2119].
2 Terminology
The following terms are used in this document:
Anonymous Replication - Replication where the endpoints are
identified to each other but not authenticated. Also known as
"unauthenticated replication".
Area of replication - A whole or portion of a Directory Information
Tree (DIT) that makes up a distinct unit of data to be replicated.
An area of replication is defined by a replication base entry and
includes all or some of the depending entries contained therein on a
single server. It divides directory data into partitions whose
propagation behavior may be independently configured from other
partitions. Areas of replication may overlap or be nested. This is
a subset of the definition of a "replicated area" in X.525 [X.525].
Atomic operation - A set of changes to directory data which the LDAP
standards guarantee will be treated as a unit; all changes will be
made or all the changes will fail.
Atomicity Information - Information about atomic operations passed as
part of replication.
Conflict - A situation that arises when changes are made to the same
directory data on different directory servers before replication can
synchronize the data on the servers. When the servers do
synchronize, they have inconsistent data - a conflict.
Conflict resolution - Deterministic procedures used to resolve change
information conflicts that may arise during replication.
Critical OID - Attributes or object classes defined in the
replication agreement as being critical to the operation of the
system. Changes affecting critical OIDs cause immediate initiation
of a replica cycle. An example of a critical OID might be a password
or certificate.
Stokes, et. al. Informational [Page 3]
^L
RFC 3384 LDAPv3 Replication Requirements October 2002
Fractional replication - The capability to filter a subset of
attributes for replication.
Incremental Update - An update that contains only those attributes or
entries that have changed.
Master Replica - A replica that may be directly updated via LDAP
operations. In a Master-Slave Replication system, the Master Replica
is the only directly updateable replica in the replica-group.
Master-Slave, or Single Master Replication - A replication model that
assumes only one server, the master, allows LDAP write access to the
replicated data. Note that Master-Slave replication can be
considered a proper subset of multi-master replication.
Meta-Data - Data collected by the replication system that describes
the status/state of replication.
Multi-Master Replication - A replication model where entries can be
written and updated on any of several master replica copies without
requiring communication with other master replicas before the write
or update is performed.
One-way Replication - The process of synchronization in a single
direction where the authoritative source information is provided to a
replica.
Partial Replication - Partial Replication is Fractional Replication,
Sparse Replication, or both.
Propagation Behavior - The behavior of the synchronization process
between a consumer and a supplier.
Replica - An instance of an area of replication on a server.
Replica-Group - The servers that hold instances of a particular area
of replication. A server may be part of several replica-groups.
Replica (or Replication) Cycle - The interval during which update
information is exchanged between two or more replicas. It begins
during an attempt to push data to, or pull data from, another replica
or set of replicas, and ends when the data has successfully been
exchanged or an error is encountered.
Replication - The process of synchronizing data distributed across
directory servers and rectifying update conflicts.
Stokes, et. al. Informational [Page 4]
^L
RFC 3384 LDAPv3 Replication Requirements October 2002
Replication Agreement - A collection of information describing the
parameters of replication between two or more servers in a replica-
group.
Replication Base Entry - The distinguished name of the root vertex of
a replicated area.
Replication Initiation Conflict - A Replication Initiation Conflict
is a situation where two sources want to update the same replica at
the same time.
Replication Session - A session set up between two servers in a
replica-group to pass update information as part of a replica cycle.
Slave (or Read-Only) Replica - A replica that cannot be directly
updated via LDAP requests. Changes may only be made via replication
from a master replica. Read-only replicas may occur in both single-
and multi-master systems.
Sparse Replication - The capability to filter some subset of entries
(other than a complete collection) of an area of replication.
Topology - The shape of the directed graph describing the
relationships between replicas.
Two-way Replication - The process of synchronization where change
information flows bi-directionally between two replicas.
Unauthenticated Replication - See Anonymous Replication.
Update Propagation - Protocol-based process by which directory
replicas are reconciled.
3 The Models
The objective is to provide an interoperable, LDAPv3 directory
synchronization protocol that is simple, efficient and flexible;
supporting both multi-master and master-slave replication. The
protocol must meet the needs of both the Internet and enterprise
environments.
There are five data consistency models.
Model 1: Transactional Consistency -- Environments that exhibit all
four of the ACID properties (Atomicity, Consistency, Isolation,
Durability) [ACID].
Stokes, et. al. Informational [Page 5]
^L
RFC 3384 LDAPv3 Replication Requirements October 2002
Model 2: Eventual (or Transient) Consistency -- Environments where
definite knowledge of the topology is provided through predetermined
replication agreements. Examples include X.500 Directories (the
X.500 model is single-master only) [X.501, X.525], Bayou [XEROX], and
NDS (Novell Directory Services) [NDS]. In this model, every update
propagates to every replica that it can reach via a path of stepwise
eventual connectivity.
Model 3: Limited Effort Eventual (or Probabilistic) Consistency --
Environments that provide a statistical probability of convergence
with knowledge of topology. An example is the Xerox Clearinghouse
[XEROX2]. This model is similar to "Eventual Consistency", except
where replicas may purge updates. Purging drops propagation changes
when some replica time boundary is exceeded, thus leaving some
changes replicated to only a portion of the topology. Transactional
consistency is not preserved, though some weaker constraints on
consistency are available.
Model 4: Loosest Consistency -- Environments where information is
provided from an opportunistic or simple cache until stale. Complete
knowledge of topology may not be shared among all replicas.
Model 5: Ad hoc -- A copy of a data store where no follow up checks
are made for the accuracy/freshness of the data.
Consistency models 1, 2 and 3 involve the use of prearranged
replication agreements among servers. While model 1 may simplify
support for atomicity in multi-master systems, the added complexity
of the distributed 2-phase commit required for Model 1 is
significant; therefor, model 1 will not be considered at this time.
Models 4 and 5 involve unregistered replicas that "pull" updates from
another directory server without that server's knowledge. These
models violate a directory's security policies.
Models 2 and 3 illustrate two replication scenarios that must be
handled: policy configuration through security management parameters
(model 2), and hosting relatively static data and address information
as in white-pages applications (model 3). Therefore, replication
requirements are presented for models 2 and 3.
Interoperability among directories using LDAP replication may be
limited for implementations that add semantics beyond those specified
by the LDAP core documents (RFC 2251-2256, 2829, 2830). In addition,
the "core" specifications include numerous features which are not
mandatory-to-implement (e.g., RECOMMENDED or OPTIONAL). There are
also numerous elective extensions. Thus LDAP replication
interoperability between independent implementations of LDAP which
support different options may be limited. Use of applicability
Stokes, et. al. Informational [Page 6]
^L
RFC 3384 LDAPv3 Replication Requirements October 2002
statements to improve interoperability in particular application
spaces is RECOMMENDED.
4 Requirements
4.1 General
G1. LDAP Replication MUST support models 2 (Eventual Consistency)
and 3 (Limited Effort Eventual Consistency) above.
G2. LDAP Replication SHOULD NOT preclude support for model 1
(Transactional Consistency) in the future.
G3. LDAP replication SHOULD have minimal impact on system
performance.
G4. The LDAP Replication Standard SHOULD NOT limit the replication
transaction rate.
G5. The LDAP replication standard SHOULD NOT limit the size of an
area of replication or a replica.
G6. Meta-data collected by the LDAP replication mechanism MUST NOT
grow without bound.
G7. All policy and state data pertaining to replication MUST be
accessible via LDAP.
G8. LDAP replication MUST be capable of replicating the following:
- all userApplication attribute types
- all directoryOperation and distributedOperation attribute
types defined in the LDAP "core" specifications (RFCs 2251-
2256, 2829-2830)
- attribute subtypes
- attribute description options (e.g., ";binary" and Language
Tags [RFC2596])
G9. LDAP replication SHOULD support replication of
directoryOperation and distributedOperation attribute types
defined in standards track LDAP extensions.
G10. LDAP replication MUST NOT support replication of dsaOperation
attribute types as such attributes are DSA-specific.
Stokes, et. al. Informational [Page 7]
^L
RFC 3384 LDAPv3 Replication Requirements October 2002
G11. The LDAP replication system should limit impact on the network
by minimizing the number of messages and the amount of traffic
sent.
4.2 Model
M1. The model MUST support the following triggers for initiation of
a replica cycle:
a) A configurable set of scheduled times
b) Periodically, with a configurable period between replica
cycles
c) A configurable maximum amount of time between replica cycles
d) A configurable number of accumulated changes
e) Change in the value of a critical OID
f) As the result of an automatic rescheduling after a
replication initiation conflict
g) A manual request for immediate replication
With the exception of manual request, the specific trigger(s)
and related parameters for a given server MUST be identified in
a well-known place defined by the standard, e.g., the
Replication Agreement(s).
M2. The replication model MUST support both master-slave and multi-
master relationships.
M3. An attribute in an entry MUST eventually converge to the same
set of values in every replica holding that entry.
M4. LDAP replication MUST encompass schema definitions, attribute
names and values, access control information, knowledge
information, and name space information.
M5. LDAP replication MUST NOT require that all copies of the
replicated information be complete, but MAY require that at
least one copy be complete. The model MUST support Partial
Replicas.
M6. The determination of which OIDs are critical MUST be
configurable in the replication agreement.
Stokes, et. al. Informational [Page 8]
^L
RFC 3384 LDAPv3 Replication Requirements October 2002
M7. The parameters of the replication process among the members of
the replica-group, including access parameters, necessary
authentication credentials, assurances of confidentiality
(encryption), and area(s) of replication MUST be defined in a
standard location (e.g., the replication agreements).
M8. The replication agreements SHOULD accommodate multiple servers
receiving the same area of replication under a single predefined
agreement.
M9. LDAP replication MUST provide scalability to both enterprise and
Internet environments, e.g., an LDAP server must be able to
provide replication services to replicas within an enterprise as
well as across the Internet.
M10. While different directory implementations can support
different/extended schema, schema mismatches between two
replicating servers MUST be handled. One way of handling such
mismatches might be to raise an error condition.
M11. There MUST be a facility that can update, or totally refresh, a
replica-group from a standard data format, such as LDIF format
[RFC2849].
M12. An update received by a consumer more than once MUST NOT produce
a different outcome than if the update were received only once.
4.3 Protocol
P1. The replication protocol MUST provide for recovery and
rescheduling of a replication session due to replication
initiation conflicts (e.g., consumer busy replicating with other
servers) and or loss of connection (e.g., supplier cannot reach
a replica).
P2. LDUP replication SHOULD NOT send an update to a consumer if the
consumer has previously acknowledged that update.
P3. The LDAP replication protocol MUST allow for full update to
facilitate replica initialization and reset loading utilizing a
standardized format such as LDIF [RFC2849] format.
P4. Incremental replication MUST be allowed.
P5. The replication protocol MUST allow either a master or slave
replica to initiate the replication process.
Stokes, et. al. Informational [Page 9]
^L
RFC 3384 LDAPv3 Replication Requirements October 2002
P6. The protocol MUST preserve atomicity of LDAP operations as
defined in RFC2251 [RFC2251]. In a multi-master environment
this may lead to an unresolvable conflict. MM5 and MM6 discuss
how to handle this situation.
P7. The protocol MUST support a mechanism to report schema
mismatches between replicas discovered during a replication
session.
4.4 Schema
SC1. A standard way to determine what replicas are held on a server
MUST be defined.
SC2. A standard schema for representing replication agreements MUST
be defined.
SC3. The semantics associated with modifying the attributes of
replication agreements MUST be defined.
SC4. A standard method for determining the location of replication
agreements MUST be defined.
SC5. A standard schema for publishing state information about a
given replica MUST be defined.
SC6. A standard method for determining the location of replica state
information MUST be defined.
SC7. It MUST be possible for appropriately authorized
administrators, regardless of their network location, to access
replication agreements in the DIT.
SC8. Replication agreements of all servers containing replicated
information MUST be accessible via LDAP.
SC9. An entry MUST be uniquely identifiable throughout its lifetime.
4.5 Single Master
SM1. A Single Master system SHOULD provide a fast method of
promoting a slave replica to become the master replica.
Stokes, et. al. Informational [Page 10]
^L
RFC 3384 LDAPv3 Replication Requirements October 2002
SM2. The master replica in a Single Master system SHOULD send all
changes to read-only replicas in the order in which the master
applied them.
4.6 Multi-Master
MM1. The replication protocol SHOULD NOT saturate the network with
redundant or unnecessary entry replication.
MM2. The initiator MUST be allowed to determine whether it will
become a consumer or supplier during the synchronization
startup process.
MM3. During a replica cycle, it MUST be possible for the two servers
to switch between the consumer and supplier roles.
MM4. When multiple master replicas want to start a replica cycle
with the same replica at the same time, the model MUST have an
automatic and deterministic mechanism for resolving or avoiding
replication initiation conflict.
MM5. Multi-master replication MUST NOT lose information during
replication. If conflict resolution would result in the loss
of directory information, the replication process MUST store
that information, notify the administrator of the nature of the
conflict and the information that was lost, and provide a
mechanism for possible override by the administrator.
MM6. Multi-master replication MUST support convergence of the values
of attributes and entries. Convergence may result in an event
as described in MM5.
MM7. Multi-master conflict resolution MUST NOT depend on the in-
order arrival of changes at a replica to assure eventual
convergence.
MM8. Multi-master replication MUST support read-only replicas as
well as read-write replicas.
4.7 Administration and Management
AM1. Replication agreements MUST allow the initiation of a replica
cycle to be administratively postponed to a more convenient
period.
AM2. Each copy of a replica MUST maintain audit history information
of which servers it has replicated with and which servers have
replicated with it.
Stokes, et. al. Informational [Page 11]
^L
RFC 3384 LDAPv3 Replication Requirements October 2002
AM3. Access to replication agreements, topologies, and policy
attributes MUST be provided through LDAP.
AM4. The capability to check the differences between two replicas
for the same information SHOULD be provided.
AM5. A mechanism to fix differences between replicas without
triggering new replica cycles SHOULD be provided.
AM6. The sequence of updates to access control information (ACI) and
the data controlled by that ACI MUST be maintained by
replication.
AM7. It MUST be possible to add a 'blank' replica to a replica-
group, and force a full update from (one of) the Master(s), for
the purpose of adding a new directory server to the system.
AM8. Vendors SHOULD provide tools to audit schema compatibility
within a potential replica-group.
4.8 Security
The terms "data confidentiality" and "data integrity" are defined in
the Internet Security Glossary [RFC2828].
S1. The protocol MUST support mutual authentication of the source
and the replica directories during initialization of a
replication session.
S2. The protocol MUST support mutual verification of authorization
of the source to send and the replica to receive replicated data
during initialization of a replication session.
S3. The protocol MUST also support the initialization of anonymous
replication sessions.
S4. The replication protocol MUST support transfer of data with data
integrity and data confidentiality.
S5. The replication protocol MUST support the ability during
initialization of a replication session for an authenticated
source and replica to mutually decide to disable data integrity
and data confidentiality within the context of and for the
duration of that particular replication session.
S6. To promote interoperability, there MUST be a mandatory-to-
implement data confidentiality mechanism.
Stokes, et. al. Informational [Page 12]
^L
RFC 3384 LDAPv3 Replication Requirements October 2002
S7. The transport for administrative access MUST permit assurance of
the integrity and confidentiality of all data transferred.
S8. To support data integrity, there must be a mandatory-to-
implement data integrity mechanism.
5 Security Considerations
This document includes security requirements (listed in section 4.8
above) for the replication model and protocol. As noted in Section
3, interoperability may be impacted when replicating among servers
that implement non-standard extensions to basic LDAP semantics.
Security-related and general LDAP interoperability will be
significantly impacted by the degree of consistency with which
implementations support existing and future standards detailing LDAP
security models, such as a future standard LDAP access control model.
6 Acknowledgements
This document is based on input from IETF members interested in LDUP
Replication.
7 References
[ACID] T. Haerder, A. Reuter, "Principles of Transaction-Oriented
Database Recovery", Computing Surveys, Vol. 15, No. 4
(December 1983), pp. 287-317.
[NDS] Novell, "NDS Technical Overview", 104-000223-001,
http://developer.novell.com/ndk/doc/ndslib/dsov_enu/data/
h6tvg4z7.html, September, 2000.
[RFC2119] Bradner, S., "Key Words for Use in RFCs to Indicate
Requirement Levels", BCP 14, RFC 2119, March 1997.
[RFC2251] Wahl, M., Howes, T. and S. Kille, "Lightweight Directory
Access Protocol", RFC 2251, December 1997.
[RFC2252] Wahl, M., Coulbeck, A., Howes, T. and S. Kille,
"Lightweight Directory Access Protocol (v3): Attribute
Syntax Definitions", RFC 2252, December 1997.
[RFC2253] Kille, S., Wahl, M. and T. Howes, "Lightweight Directory
Access Protocol (v3): UTF-8 String Representation of
Distinguished Names", RFC 2253, December 1997.
[RFC2254] Howes, T., "The String Representation of LDAP Search
Filters", RFC 2254, December 1997.
Stokes, et. al. Informational [Page 13]
^L
RFC 3384 LDAPv3 Replication Requirements October 2002
[RFC2255] Howes, T. and M. Smith, "The LDAP URL Format", RFC 2255,
December 1997.
[RFC2256] Wahl, M., "A Summary of the X.500(96) User Schema for use
with LDAPv3", RFC 2256, December 1997.
[RFC2596] Wahl, M. and T. Howes, "Use of Language Codes in LDAP", RFC
2596, May 1999.
[RFC2828] Shirey, R. "Internet Security Glossary", FYI 36, RFC 2828,
May 2000.
[RFC2829] Wahl, M., Alvestrand, H., Hodges, J. and R. Morgan,
"Authentication Methods for LDAP", RFC 2829, May 2000.
[RFC2830] Hodges, J., Morgan, R. and M. Wahl, "Lightweight Directory
Access Protocol (v3): Extension for Transport Layer
Security", RFC 2830, May 2000.
[RFC2849] Good, G., "The LDAP Data Interchange Format (LDIF)", RFC
2849, June 2000.
[X.501] ITU-T Recommendation X.501 (1993), | ISO/IEC 9594-2: 1993,
Information Technology - Open Systems Interconnection - The
Directory: Models.
[X.525] ITU-T Recommendation X.525 (1997), | ISO/IEC 9594-9: 1997,
Information Technology - Open Systems Interconnection - The
Directory: Replication.
[XEROX] C. Hauser, "Managing update conflicts in Bayou, a weakly
connected replicated storage system". Palo Alto, CA: Xerox
PARC, Computer Science Laboratory; 1995 August; CSL-95-4.
[XEROX2] Alan D. Demers, Mark Gealy, Daniel Greene, Carl Hauser,
Wesley Irish, John Larson, Sue Manning, Scott Shenker,
Howard Sturgis, Daniel Swinehart, Douglas Terry, Don Woods,
"Epidemic Algorithms for Replicated Database Maintenance".
Palo Alto, CA, Xerox PARC, January 1989.
Stokes, et. al. Informational [Page 14]
^L
RFC 3384 LDAPv3 Replication Requirements October 2002
A. APPENDIX A - Usage Scenarios
The following directory deployment examples are intended to validate
our replication requirements. A heterogeneous set of directory
implementations is assumed for all the cases below. This material is
intended as background; no requirements are presented in this
Appendix.
A.1. Extranet Example
A company has a trading partner with whom it wishes to share
directory information. This information may be as simple as a
corporate telephone directory, or as complex as an extranet workflow
application. For performance reasons, the company wishes to place a
replica of its directory within the Partner Company, rather than
exposing its directory beyond its firewall.
The requirements that follow from this scenario are:
- One-way replication, single mastered.
- Authentication of clients.
- Common access control and access control identification.
- Secure transmission of updates.
- Selective attribute replication (Fractional Replication), so that
only partial entries can be replicated.
A.2. Consolidation Example
Company A acquires company B. Each company has an existing
directory.
During the transition period, as the organizations are merged, both
directory services must coexist. Company A may wish to attach
company B's directory to its own.
The requirements that follow from this scenario are:
- Multi-Master replication.
- Common access control model. Access control model identification.
- Secure transmission of updates.
- Replication between DITs with potentially differing schema.
Stokes, et. al. Informational [Page 15]
^L
RFC 3384 LDAPv3 Replication Requirements October 2002
A.3. Replication Heterogeneous Deployment Example
An organization may choose to deploy directory implementations from
multiple vendors, to enjoy the distinguishing benefits of each.
In this case, multi-master replication is required to ensure that the
multiple replicas of the DIT are synchronized. Some vendors may
provide directory clients, which are tied to their own directory
service.
The requirements that follow from this scenario are:
- Multi-Master replication
- Common access control model and access control model
identification.
- Secure transmission of updates.
- Replication among DITs with potentially differing schemas.
A.4. Shared Name Space Example
Two organizations may choose to cooperate on some venture and need a
shared name space to manage their operation. Both organizations will
require administrative rights over the shared name space.
The requirements that follow from this scenario are:
- Multi-Master replication.
- Common access control model and access control model
identification.
- Secure transmission of updates.
A.5. Supplier Initiated Replication
This is a single master environment that maintains a number of
replicas of the DIT by pushing changes based on a defined schedule.
The requirements that follow from this scenario are:
- Single-master environment.
- Supplier-initiated replication.
- Secure transmission of updates.
Stokes, et. al. Informational [Page 16]
^L
RFC 3384 LDAPv3 Replication Requirements October 2002
A.6. Consumer Initiated Replication
Again a single mastered replication topology, but the slave replica
initiates the replication exchange rather than the master. An
example of this is a replica that resides on a laptop computer that
may run disconnected for a period of time.
The requirements that follow from this scenario are:
- Single-master environment.
- Consumer initiated replication.
- Open scheduling (anytime).
A.7. Prioritized attribute replication
The password attribute can provide an example of the requirement for
prioritized attribute replication. A user is working in Utah and the
administrator resides in California. The user has forgotten his
password. So the user calls or emails the administrator to request a
new password. The administrator provides the updated password (a
change).
Under normal conditions, the directory replicates to a number of
different locations overnight. But corporate security policy states
that passwords are critical and the new value must be available
immediately (e.g., shortly) after any change. Replication needs to
occur immediately for critical attributes/entries.
The requirements that follow from this scenario are:
- Incremental replication of changes.
- Immediate replication on change of certain attributes.
- Replicate based on time/attribute semantics.
A.8. Bandwidth issues
The replication of Server (A) R/W replica (a) in Kathmandu is handled
via a dial up phone link to Paris where server (B) R/W replica of (a)
resides. Server (C) R/W replica of (a) is connected by a T1
connection to server (B). Each connection has a different
performance characteristic.
Stokes, et. al. Informational [Page 17]
^L
RFC 3384 LDAPv3 Replication Requirements October 2002
The requirements that follow from this scenario are:
- Minimize repetitive updates when replicating from multiple
replication paths.
- Incremental replication of changes.
- Provide replication cycles to delay and/or retry when connections
cannot be reached.
- Allowances for consumer initiated or supplier initiated
replication.
A.9. Interoperable Administration and Management
The administrator with administrative authority of the corporate
directory which is replicated by numerous geographically dispersed
LDAP servers from different vendors notices that the replication
process is not completing correctly as the change log is continuing
to grow and/or error messages inform him. The administrator uses his
$19.95 RepCo LDAP directory replication diagnostic tools to look at
Root DSE replica knowledge on server 17 and determines that server 42
made by LDAP'RUS Inc. is not replicating properly due to an object
conflict. Using his Repco Remote repair tools he connects to server
42 and resolves the conflict on the remote server.
The requirements that follow from this scenario are:
- Provide replication audit history.
- Provide mechanisms for managing conflict resolution.
- Provide LDAP access to predetermined agreements, topology and
policy attributes.
- Provide operations for comparing replica's content for validity.
- Provide LDAP access to status and audit information.
A.10. Enterprise Directory Replication Mesh
A Corporation builds a mesh of directory servers within the
enterprise utilizing LDAP servers from various vendors. Five servers
are holding the same area of replication. The predetermined
replication agreement(s) for the enterprise mesh are under a single
management, and the security domain allows a single predetermined
replication agreement to manage the 5 servers' replication.
Stokes, et. al. Informational [Page 18]
^L
RFC 3384 LDAPv3 Replication Requirements October 2002
The requirements that follow from this scenario are:
- One predefined replication agreement that manages a single area of
replication that is held on numerous servers.
- Common support of replication management knowledge across vendor
implementation.
- Rescheduling and continuation of a replication cycle when one
server in a replica-group is busy and/or unavailable.
A.11. Failure of the Master in a Master-Slave Replicated Directory
A company has a corporate directory that is used by the corporate
email system. The directory is held on a mesh of servers from
several vendors. A corporate relocation results in the closing of
the location where the master copy of the directory is located.
Employee information (such as mailbox locations and employee
certificate information) must be kept up to date or mail cannot be
delivered.
The requirements that follow from this scenario are:
- An existing slave replica must be "promote-able" to become the new
master.
- The "promotion" must be done without significant downtime, since
updates to the directory will continue.
A.12. Failure of a Directory Holding Critical Service Information
An ISP uses a policy management system that uses a directory as the
policy data repository. The directory is replicated in several
different sites on different vendors' products to avoid single points
of failure. It is imperative that the directory be available and be
updateable even if one site is disconnected from the network.
Changes to the data must be traceable, and it must be possible to
determine how changes made from different sites interacted.
The requirements that follow from this scenario are:
- Multi-master replication.
- Ability to reschedule replication sessions.
- Support for manual review and override of replication conflict
resolution.
Stokes, et. al. Informational [Page 19]
^L
RFC 3384 LDAPv3 Replication Requirements October 2002
B. APPENDIX B - Rationale
This Appendix gives some of the background behind the requirements.
It is included to help the protocol designers understand the thinking
behind some of the requirements and to present some of the issues
that should be considered during design. With the exception of
section B.8, which contains a suggested requirement for the update to
RFC 2251, this Appendix does not state any formal requirements.
B.1. Meta-Data Implications
Requirement G4 states that meta-data must not grow without bound.
This implies that meta-data must, at some point, be purged from the
system. This, in turn, raises concerns about stability. Purging
meta-data before all replicas have been updated may lead to
incomplete replication of change information and inconsistencies
among replicas. Therefore, care must be taken setting up the rules
for purging meta-data from the system while still ensuring that
meta-data will not grow forever.
B.2. Order of Transfer for Replicating Data
Situations may arise where it would be beneficial to replicate data
out-of-order (e.g., send data to consumer replicas in a different
order than it was processed at the supplier replica). One such case
might occur if a large bulk load was done on the master server in a
single-master environment and then a single change to a critical OID
(a password change, for example) was then made. Rather than wait for
all the bulk data to be sent to the replicas, the password change
might be moved to the head of the queue and be sent before all the
bulk data was transferred. Other cases where this might be
considered are schema changes or changes to critical policy data
stored in the directory.
While there are practical benefits to allowing out-of-order transfer,
there are some negative consequences as well. Once out-of-order
transfers are permitted, all receiving replicas must be prepared to
deal with data and schema conflicts that might arise.
As an example, assume that schema changes are critical and must be
moved to the front of the replication queue. Now assume that a
schema change deletes an attribute for some object class. It is
possible that some of the operations ahead of the schema change in
the queue are operations to delete values of the soon-to-be-deleted
Stokes, et. al. Informational [Page 20]
^L
RFC 3384 LDAPv3 Replication Requirements October 2002
attribute so that the schema change can be done with no problems. If
the schema change moves to the head of the queue, the consumer
servers might have to delete an attribute that still has values, and
then receive requests to delete the values of an attribute that is no
longer defined.
In the multi-master case, similar situations can arise when
simultaneous changes are made to different replicas. Thus, multi-
master systems must have conflict resolution algorithms in place to
handle such situations. But in the single-master case conflict
resolution is not needed unless the master is allowed to send data
out-of-order. This is the reasoning behind requirement SM2, which
recommends that data always be sent in order in single-master
replication.
Note that even with this restriction, the concept of a critical OID
is still useful in single-master replication. An example of its
utility can be found in section A.7.
B.3. Schema Mismatches and Replication
Multi-vendor environments are the primary area of interest for LDAP
replication standards. Some attention must thus be paid to the issue
of schema mismatches, since they can easily arise when vendors
deliver slightly different base schema with their directory products.
Even when both products meet the requirements of the standards
[RFC2252], the vendors may have included additional attributes or
object classes with their products. When two different vendors'
products attempt to replicate, these additions can cause schema
mismatches. Another potential cause of schema mismatches is
discussed in section A.3.
There are only a few possible responses when a mismatch is
discovered.
- Raise an error condition and ignore the data. This should always
be allowed and is the basis for requirement P8 and the comment on
M10.
- Map/convert the data to the form required by the consuming replica.
A system may choose this course; requirement M10 is intended to
allow this option. The extent of the conversion is up to the
implementation; in the extreme it could support use of the
replication protocol in meta-directories.
- Quietly ignore (do not store on the consumer replica and do not
raise an error condition) any data that does not conform to the
schema at the consumer.
Stokes, et. al. Informational [Page 21]
^L
RFC 3384 LDAPv3 Replication Requirements October 2002
Requirement M10 is intended to exclude the last option.
Requirement AM8 suggests that vendors should provide tools to help
discover schema mismatches when replication is being set up. But
schema will change after the initial setup, so the replication system
must be prepared to handle unexpected mismatches.
Normal IETF practice in protocol implementation suggests that one be
strict in what one sends and be flexible in what one receives. The
parallel in this case is that a supplier should be prepared to
receive an error notification for any schema mismatch, but a consumer
may choose to do a conversion instead.
The other option that can be considered in this situation is the use
of fractional replication. If replication is set up so only the
common attributes are replicated, mismatches can be avoided.
One additional consideration here is replication of the schema
itself. M4 requires that it be possible to replicate schema. If a
consumer replica is doing conversion, extreme care should be taken if
schema elements are replicated since some attributes are intended to
have different definitions on different replicas.
For fractional replication, the protocol designers and implementors
should give careful consideration to the way they handle schema
replication. Some options for schema replication include:
- All schema elements are replicated.
- Schema elements are replicated only if they are used by attributes
that are being replicated.
- Schema are manually configured on the servers involved in
fractional replication; schema elements are not replicated via the
protocol.
B.4. Detecting and Repairing Inconsistencies Among Replicas
Despite the best efforts of designers, implementors, and operators,
inconsistencies will occasionally crop up among replicas in
production directories. Tools will be needed to detect and to
correct these inconsistencies.
Stokes, et. al. Informational [Page 22]
^L
RFC 3384 LDAPv3 Replication Requirements October 2002
A special client may accomplish detection through periodic
comparisons of replicas. This client would typically read two
replicas of the same replication base entry and compare the answers,
possibly by BINDing to each of the two replicas to be compared and
reading them both. In cases where the directory automatically
reroutes some requests (e.g., chaining), mechanisms to force access
to a particular replica should be supplied.
Alternatively, the server could support a special request to handle
this situation. A client would invoke an operation at some server.
It would cause that server to extract the contents from some other
server it has a replication agreement with and report the differences
back to the client as the result.
If an inconsistency is found, it needs to be repaired. To determine
the appropriate repair, the administrator will need access to the
replication history to figure out how the inconsistency occurred and
what the correct repair should be.
When a repair is made, it should be restricted to the replica that
needs to be fixed; the repair should not cause new replication events
to be started. This may require special tools to change the local
data store without triggering replication.
Requirements AM2, AM4, and AM5 address these needs.
B.5. Some Test Cases for Conflict Resolution in Multi-Master Replication
Use of multi-master replication inevitably leads to the possibility
that incompatible changes will be made simultaneously on different
servers. In such cases, conflict resolution algorithms must be
applied.
As a guiding principle, conflict resolution should avoid surprising
the user. One way to do this is to adopt the principle that, to the
extent possible, conflict resolution should mimic the situation that
would happen if there were a single server where all the requests
were handled.
While this is a useful guideline, there are some situations where it
is impossible to implement. Some of these cases are examined in this
section. In particular, there are some cases where data will be
"lost" in multi-master replication that would not be lost in a
single-server configuration.
Stokes, et. al. Informational [Page 23]
^L
RFC 3384 LDAPv3 Replication Requirements October 2002
In the examples below, assume that there are three replicas, A, B,
and C. All three replicas are updateable. Changes are made to
replicas A and B before replication allows either replica to see the
change made on the other. In discussion of the multi-master cases,
we assume that the change to A takes precedence using whatever rules
are in force for conflict resolution.
B.5.1. Create-Create
A user creates a new entry with distinguished name DN on A. At the
same time, a different user adds an entry with the same distinguished
name on B.
In the single-server case, one of the create operations would have
occurred before the other, and the second request would have failed.
In the multi-master case, each create was successful on its
originating server. The problem is not detected until replication
takes place. When a replication request to create a DN that already
exists arrives at one of the servers, conflict resolution is invoked.
(Note that the two requests can be distinguished even though they
have the same DN because every entry has some sort of unique
identifier per requirement SC9.)
As noted above, in these discussions we assume that the change from
replica A has priority based on the conflict resolution algorithm.
Whichever change arrives first, requirement MM6 says that the values
from replica A must be those in place on all replicas at the end of
the replication cycle. Requirement MM5 states that the system cannot
quietly ignore the values from replica B.
The values from replica B might be logged with some notice to the
administrators, or they might be added to the DIT with a machine
generated DN (again with notice to the administrators). If they are
stored with a machine generated DN, the same DN must be used on all
servers in the replica-group (otherwise requirement M3 would be
violated). Note that in the case where the entry in question is a
container, storage with a machine generated DN provides a place where
descendent entries may be stored if any descendents were generated
before the replication cycle was completed.
In any case, some mechanism must be provided to allow the
administrator to reverse the conflict resolution algorithm and force
the values originally created on B into place on all replicas if
desired.
Stokes, et. al. Informational [Page 24]
^L
RFC 3384 LDAPv3 Replication Requirements October 2002
B.5.2. Rename-Rename
On replica A, an entry with distinguished name DN1 is renamed to DN.
At the same time on replica B, an entry with distinguished name DN2
is renamed to DN.
In the single-server case, one rename operation would occur before
the other and the second would fail since the target name already
exists.
In the multi-master case, each rename was successful on its
originating server. Assuming that the change on A has priority in
the conflict resolution sense, DN will be left with the values from
DN1 in all replicas and DN1 will no longer exist in any replica. The
question is what happens to DN2 and its original values.
Requirement MM5 states that these values must be stored somewhere.
They might be logged, they might be left in the DIT as the values of
DN2, or they might be left in the DIT as the values of some machine
generated DN. Leaving them as the values of DN2 is attractive since
it is the same as the single-server case, but if a new DN2 has
already been created before the replica cycle finishes, there are
some very complex cases to resolve. Any of the solutions described
in this paragraph would be consistent with requirement MM5.
B.5.3. Locking Based on Atomicity of ModifyRequest
There is an entry with distinguished name DN that contains attributes
X, Y, and Z. The value of X is 1. On replica A, a ModifyRequest is
processed which includes modifications to change that value of X from
1 to 0 and to set the value of Y to "USER1". At the same time,
replica B processes a ModifyRequest which includes modifications to
change the value of X from 1 to 0 and to set the value of Y to
"USER2" and the value of Z to 42. The application in this case is
using X as a lock and is depending on the atomic nature of
ModifyRequests to provide mutual exclusion for lock access.
In the single-server case, the two operations would have occurred
sequentially. Since a ModifyRequest is atomic, the entire first
operation would succeed. The second ModifyRequest would fail, since
the value of X would be 0 when it was attempted, and the modification
changing X from 1 to 0 would thus fail. The atomicity rule would
cause all other modifications in the ModifyRequest to fail as well.
In the multi-master case, it is inevitable that at least some of the
changes will be reversed despite the use of the lock. Assuming the
changes from A have priority per the conflict resolution algorithm,
the value of X should be 0 and the value of Y should be "USER1" The
Stokes, et. al. Informational [Page 25]
^L
RFC 3384 LDAPv3 Replication Requirements October 2002
interesting question is the value of Z at the end of the replication
cycle. If it is 42, the atomicity constraint on the change from B
has been violated. But for it to revert to its previous value,
grouping information must be retained and it is not clear when that
information can be safely discarded. Thus, requirement G6 may be
violated.
B.5.4. General Principles
With multi-master replication there are a number of cases where a
user or application will complete a sequence of operations with a
server but those actions are later "undone" because someone else
completed a conflicting set of operations at another server.
To some extent, this can happen in any multi-user system. If a user
changes the value of an attribute and later reads it back,
intervening operations by another user may have changed the value.
In the multi-master case, the problem is worsened, since techniques
used to resolve the problem in the single-server case won't work as
shown in the examples above.
The major question here is one of intended use. In LDAP standards
work, it has long been said that replication provides "loose
consistency" among replicas. At several IETF meetings and on the
mailing list, usage examples from finance where locking is required
have been declared poor uses for LDAP. Requirement G1 is consistent
with this history. But if loose consistency is the goal, the locking
example above is an inappropriate use of LDAP, at least in a
replicated environment.
B.5.5. Avoiding the Problem
The examples above discuss some of the most difficult problems that
can arise in multi-master replication. While they can be dealt with,
dealing with them is difficult and can lead to situations that are
quite confusing to the application and to users.
The common characteristics of the examples are:
- Several directory users/applications are changing the same data.
- They are changing the data before previous changes have replicated.
- They are using different directory servers to make these changes.
- They are changing data that are parts of a distinguished name or
they are using ModifyRequest to both read and write a given
attribute value in a single atomic request.
Stokes, et. al. Informational [Page 26]
^L
RFC 3384 LDAPv3 Replication Requirements October 2002
If any one of these conditions is reversed, the types of problems
described above will not occur. There are many useful applications
of multi-master directories where at least one of the above
conditions does not occur. For cases where all four do occur,
application designers should be aware of the possible consequences.
B.6. Data Confidentiality and Data Integrity During Replication
Directories will frequently hold proprietary information. Policy
information, name and address information, and customer lists can be
quite proprietary and are likely to be stored in directories. Such
data must be protected against intercept or modification during
replication.
In some cases, the network environment (e.g., a private network) may
provide sufficient data confidentiality and integrity for the
application. In other cases, the data in the directory may be public
and not require protection. For these reasons data confidentiality
and integrity were not made requirements for all replication
sessions. But there are a substantial number of applications that
will need data confidentiality and integrity for replication, so
there is a requirement (S4) that the protocol allow for data
confidentiality and integrity in those cases where they are needed.
Typically, the policy on the use of confidentiality and integrity
measures would be held in the replication agreement per requirement
M7.
This leaves the question of what mechanism(s) to use. While this is
ultimately a design/implementation decision, replication across
different vendors' directory products is an important goal of the
LDAP replication work at the IETF. If different vendors choose to
support different data confidentiality and integrity mechanisms, the
advantages of a standard replication protocol would be lost. Thus
there is a requirement (S6) for mandatory-to-implement data
confidentiality and integrity mechanisms.
Anonymous replication (requirement S3) is supported since it may be
useful in the same sorts of situations where data integrity and data
confidentiality protection are not needed.
B.7. Failover in Single-Master Systems
In a single-master system, all modifications must originate at the
master. The master is therefore a single point of failure for
modifications. This can cause concern when high availability is a
requirement for the directory system.
Stokes, et. al. Informational [Page 27]
^L
RFC 3384 LDAPv3 Replication Requirements October 2002
One way to reduce the problem is to provide a failover process that
converts a slave replica to master when the original master fails.
The time required to execute the failover process then becomes a
major factor in availability of the system as a whole.
Factors that designers and implementors should consider when working
on failover include:
- If the master replica contains control information or meta-data
that is not part of the slave replica(s), this information will
have to be inserted into the slave that is being "promoted" to
master as part of the failover process. Since the old master is
presumably unavailable at this point, it may be difficult to obtain
this data. For example, if the master holds the status information
of all replicas, but each slave replica only holds its own status
information, failover would require that the new master get the
status of all existing replicas, presumably from those replicas.
Similar issues could arise for replication agreements if the master
is the only system that holds a complete set.
- If data privacy mechanisms (e.g., encryption) are in use during
replication, the new master would need to have the necessary key
information to talk to all of the slave replicas.
- It is not only the new master that needs to be reconfigured. The
slaves also need to have their configurations updated so they know
where updates should come from and where they should refer
modifications.
- The failover mechanism should be able to handle a situation where
the old master is "broken" but not "dead". The slave replicas
should ignore updates from the old master after failover is
initiated.
- The old master will eventually be repaired and returned to the
replica-group. It might join the group as a slave and pick up the
changes it has "missed" from the new master, or there might be some
mechanism to bring it into sync with the new master and then let it
take over as master. Some resynchronization mechanism will be
needed.
- Availability would be maximized if the whole failover process could
be automated (e.g., failover is initiated by an external system
when it determines that the original master is not functioning
properly).
Stokes, et. al. Informational [Page 28]
^L
RFC 3384 LDAPv3 Replication Requirements October 2002
B.8. Including Operational Attributes in Atomic Operations
LDAPv3 [RFC2251] declares that some operations are atomic (e.g., all
of the modifications in a single ModifyRequest). It also defines
several operational attributes that store information about when
changes are made to the directory (createTimestamp, etc.) and which
ID was responsible for a given change (modifiersName, etc.).
Currently, there is no statement in RFC2251 requiring that changes to
these operational attributes be atomic with the changes to the data.
It is RECOMMENDED that this requirement be added during the revision
of RFC2251. In the interim, replication SHOULD treat these
operations as though such a requirement were in place.
Stokes, et. al. Informational [Page 29]
^L
RFC 3384 LDAPv3 Replication Requirements October 2002
Authors' Addresses
Russel F. Weiser
Digital Signature Trust Co.
1095 East 2100 South
Suite #201
Salt Lake City, UT 84106
Phone: +1 801 326 5421
Fax: +1 801 326 5421
EMail: rweiser@trustdst.com
Ellen J. Stokes
IBM
11400 Burnet Rd.
Austin, TX 78758
Phone: +1 512 436 9098
Fax: +1 512 436 1193
EMail: stokese@us.ibm.com
Ryan D. Moats
Lemur Networks
15621 Drexel Circle
Omaha, NE 68135
Phone: +1 402 894 9456
EMail: rmoats@lemurnetworks.net
Richard V. Huber
Room C3-3B30
AT&T Laboratories
200 Laurel Avenue South
Middletown, NJ 07748
Phone: +1 732 420 2632
Fax: +1 732 368 1690
EMail: rvh@att.com
Stokes, et. al. Informational [Page 30]
^L
RFC 3384 LDAPv3 Replication Requirements October 2002
Full Copyright Statement
Copyright (C) The Internet Society (2002). All Rights Reserved.
This document and translations of it may be copied and furnished to
others, and derivative works that comment on or otherwise explain it
or assist in its implementation may be prepared, copied, published
and distributed, in whole or in part, without restriction of any
kind, provided that the above copyright notice and this paragraph are
included on all such copies and derivative works. However, this
document itself may not be modified in any way, such as by removing
the copyright notice or references to the Internet Society or other
Internet organizations, except as needed for the purpose of
developing Internet standards in which case the procedures for
copyrights defined in the Internet Standards process must be
followed, or as required to translate it into languages other than
English.
The limited permissions granted above are perpetual and will not be
revoked by the Internet Society or its successors or assigns.
This document and the information contained herein is provided on an
"AS IS" basis and THE INTERNET SOCIETY AND THE INTERNET ENGINEERING
TASK FORCE DISCLAIMS ALL WARRANTIES, EXPRESS OR IMPLIED, INCLUDING
BUT NOT LIMITED TO ANY WARRANTY THAT THE USE OF THE INFORMATION
HEREIN WILL NOT INFRINGE ANY RIGHTS OR ANY IMPLIED WARRANTIES OF
MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE.
Acknowledgement
Funding for the RFC Editor function is currently provided by the
Internet Society.
Stokes, et. al. Informational [Page 31]
^L
|