diff options
author | Thomas Voss <mail@thomasvoss.com> | 2024-11-27 20:54:24 +0100 |
---|---|---|
committer | Thomas Voss <mail@thomasvoss.com> | 2024-11-27 20:54:24 +0100 |
commit | 4bfd864f10b68b71482b35c818559068ef8d5797 (patch) | |
tree | e3989f47a7994642eb325063d46e8f08ffa681dc /doc/rfc/rfc2191.txt | |
parent | ea76e11061bda059ae9f9ad130a9895cc85607db (diff) |
doc: Add RFC documents
Diffstat (limited to 'doc/rfc/rfc2191.txt')
-rw-r--r-- | doc/rfc/rfc2191.txt | 675 |
1 files changed, 675 insertions, 0 deletions
diff --git a/doc/rfc/rfc2191.txt b/doc/rfc/rfc2191.txt new file mode 100644 index 0000000..1e5bb2d --- /dev/null +++ b/doc/rfc/rfc2191.txt @@ -0,0 +1,675 @@ + + + + + + +Network Working Group G. Armitage +Request for Comments: 2191 Lucent Technologies +Category: Informational September 1997 + + + VENUS - Very Extensive Non-Unicast Service + +Status of this Memo + + This memo provides information for the Internet community. This memo + does not specify an Internet standard of any kind. Distribution of + this memo is unlimited. + +Abstract + + The MARS model (RFC2022) provides a solution to intra-LIS IP + multicasting over ATM, establishing and managing the use of ATM pt- + mpt SVCs for IP multicast packet forwarding. Inter-LIS multicast + forwarding is achieved using Mrouters, in a similar manner to which + the "Classical IP over ATM" model uses Routers to inter-connect LISes + for unicast traffic. The development of unicast IP shortcut + mechanisms (e.g. NHRP) has led some people to request the + development of a Multicast equivalent. There are a number of + different approaches. This document focuses exclusively on the + problems associated with extending the MARS model to cover multiple + clusters or clusters spanning more than one subnet. It describes a + hypothetical solution, dubbed "Very Extensive NonUnicast Service" + (VENUS), and shows how complex such a service would be. It is also + noted that VENUS ultimately has the look and feel of a single, large + cluster using a distributed MARS. This document is being issued to + help focus ION efforts towards alternative solutions for establishing + ATM level multicast connections between LISes. + +1. Introduction + + The classical model of the Internet running over an ATM cloud + consists of multiple Logical IP Subnets (LISs) interconnected by IP + Routers [1]. The evolving IP Multicast over ATM solution (the "MARS + model" [2]) retains the classical model. The LIS becomes a "MARS + Cluster", and Clusters are interconnected by conventional IP + Multicast routers (Mrouters). + + The development of NHRP [3], a protocol for discovering and managing + unicast forwarding paths that bypass IP routers, has led to some + calls for an IP multicast equivalent. Unfortunately, the IP + multicast service is a rather different beast to the IP unicast + service. This document aims to explain how much of what has been + learned during the development of NHRP must be carefully scrutinized + + + +Armitage Informational [Page 1] + +RFC 2191 VENUS September 1997 + + + before being re-applied to the multicast scenario. Indeed, the + service provided by the MARS and MARS Clients in [2] are almost + orthogonal to the IP unicast service over ATM. + + For the sake of discussion, let's call this hypothetical multicast + shortcut discovery protocol the "Very Extensive Non-Unicast Service" + (VENUS). A "VENUS Domain" is defined as the set of hosts from two or + more participating Logical IP Subnets (LISs). A multicast shortcut + connection is a point to multipoint SVC whose leaf nodes are + scattered around the VENUS Domain. (It will be noted in section 2 + that a VENUS Domain might consist of a single MARS Cluster spanning + multiple LISs, or multiple MARS Clusters.) + + VENUS faces a number of fundamental problems. The first is exploding + the scope over which individual IP/ATM interfaces must track and + react to IP multicast group membership changes. Under the classical + IP routing model Mrouters act as aggregation points for multicast + traffic flows in and out of Clusters [4]. They also act as + aggregators of group membership change information - only the IP/ATM + interfaces within each Cluster need to know the specific identities + of their local (intra-cluster) group members at any given time. + However, once you have sources within a VENUS Domain establishing + shortcut connections the data and signaling plane aggregation of + Mrouters is lost. In order for all possible sources throughout a + VENUS Domain to manage their outgoing pt-mpt SVCs they must be kept + aware of MARS_JOINs and MARS_LEAVEs occuring in every MARS Cluster + that makes up a VENUS Domain. The nett effect is that a VENUS domain + looks very similar to a single, large distributed MARS Cluster. + + A second problem is the impact that shortcut connections will have on + IP level Inter Domain Multicast Routing (IDMR) protocols. Multicast + groups have many sources and many destinations scattered amongst the + participating Clusters. IDMR protocols assume that they can calculate + efficient inter-Cluster multicast trees by aggregating individual + sources or group members in any given Cluster (subnet) behind the + Mrouter serving that Cluster. If sources are able to simply bypass an + Mrouter we introduce a requirement that the existence of each and + every shortcut connection be propagated into the IDMR decision making + processes. The IDMR protocols may need to adapt when a source's + traffic bypasses its local Mrouter(s) and is injected into Mrouters + at more distant points on the IP-level multicast distribution tree. + (This issue has been looked at in [7], focussing on building + forwarding trees within networks where the termination points are + small in number and sparsely distributed. VENUS introduces tougher + requirements by assuming that multicast group membership may be dense + across the region of interest.) + + + + + +Armitage Informational [Page 2] + +RFC 2191 VENUS September 1997 + + + This document will focus primarily on the internal problems of a + VENUS Domain, and leave the IDMR interactions for future analysis. + +2. What does it mean to "shortcut" ? + + Before going further it is worth considering both the definition of + the Cluster, and two possible definitions of "shortcut". + +2.1 What is a Cluster? + + In [2] a MARS Cluster is defined as the set of IP/ATM interfaces that + are willing to engage in direct, ATM level pt-mpt SVCs to perform IP + multicast packet forwarding. Each IP/ATM interface (a MARS Client) + must keep state information regarding the ATM addresses of each leaf + node (recipient) of each pt-mpt SVC it has open. In addition, each + MARS Client receives MARS_JOIN and MARS_LEAVE messages from the MARS + whenever there is a requirement that Clients around the Cluster need + to update their pt-mpt SVCs for a given IP multicast group. + + It is worth noting that no MARS Client has any concept of how big its + local cluster is - this knowledge is kept only by the MARS that a + given Client is registered with. + + Fundamentally the Cluster (and the MARS model as a whole) is a + response to the requirement that any multicast IP/ATM interface using + pt-mpt SVCs must, as group membership changes, add and drop leaf + nodes itself. This means that some mechanism, spanning all possible + group members within the scopes of these pt-mpt SVCs, is required to + collect group membership information and distribute it in a timely + fashion to those interfaces. This is the MARS Cluster, with certain + scaling limits described in [4]. + +2.2 LIS/Cluster boundary "shortcut" + + The currently popular definition of "shortcut" is based on the + existence of unicast LIS boundaries. It is tied to the notion that + LIS boundaries have physical routers, and cutting through a LIS + boundary means bypassing a router. Intelligently bypassing routers + that sit at the edges of LISs has been the goal of NHRP. Discovering + the ATM level identity of an IP endpoint in a different LIS allows a + direct SVC to be established, thus shortcutting the logical IP + topology (and very real routers) along the unicast path from source + to destination. + + For simplicity of early adoption RFC2022 recommends that a Cluster's + scope be made equivalent to that of a LIS. Under these circumstances + the "Classical IP" routing model places Mrouters at LIS/Cluster + boundaries, and multicast shortcutting must involve bypassing the + + + +Armitage Informational [Page 3] + +RFC 2191 VENUS September 1997 + + + same physical routing entities as unicast shortcutting. Each MARS + Cluster would be independent and contain only those IP/ATM interfaces + that had been assigned to the same LIS. + + As a consequence, a VENUS Domain covering the hosts in a number of + LIS/Clusters would have to co-ordinate each individual MARS from each + LIS/Cluster (to ensure group membership updates from around the VENUS + Domain were propagated correctly). + +2.3 Big Cluster, LIS boundary "shortcut" + + The MARS model's fundamental definition of a Cluster was deliberately + created to be independent of unicast terminology. Although not + currently well understood, it is possible to build a single MARS + Cluster that encompasses the members of multiple LISs. As expected, + inter-LIS unicast traffic would pass through (or bypass, if using + NHRP) routers on the LIS boundaries. Also as expected, each IP/ATM + interface, acting as a MARS Client, would forward their IP multicast + packets directly to intra-cluster group members. However, because the + direct intra-cluster SVCs would exist between hosts from the + different LISs making up the cluster, this could be considered a + "shortcut" of the unicast LIS boundaries. + + This approach immediately brings up the problem of how the IDMR + protocols will react. Mrouters only need to exist at the edges of + Clusters. In the case of a single Cluster spanning multiple LISs, + each LIS becomes hidden behind the Mrouter at the Cluster's edge. + This is arguably not a big problem if the Cluster is a stub on an + IDMR protocol's multicast distribution tree, and if there is only a + single Mrouter in or out of the Cluster. Problems arise when two or + more Mrouters are attached to the edges of the Cluster, and the + Cluster is used for transit multicast traffic. Each Mrouter's + interface is assigned a unicast identity (e.g. that of the unicast + router containing the Mrouter). IDMR protocols that filter packets + based on the correctness of the upstream source may be confused at + receiving IP multicast packets directly from another Mrouter in the + same cluster but notionally "belonging" to an LIS multiple unicast IP + hops away. + + Adjusting the packet filtering algorithms of Mrouters is something + that needs to be addressed by any multicast shortcut scheme. It has + been noted before and a solution proposed in [7]. For the sake of + argument this document will assume the problem solvable. (However, it + is important that any solution scales well under general topologies + and group membership densities.) + + + + + + +Armitage Informational [Page 4] + +RFC 2191 VENUS September 1997 + + + A multi-LIS MARS Cluster can be considered a simple VENUS Domain. + Since it is a single Cluster it can be scaled using the distributed + MARS solutions currently being developed within the IETF [5,6]. + +3. So what must VENUS look like? + + A number of functions that occur in the MARS model are fundamental to + the problem of managing root controlled, pt-mpt SVCs. The initial + setup of the forwarding SVC by any one MARS Client requires a + query/response exchange with the Client's local MARS, establishing + who the current group members are (i.e. what leaf nodes should be on + the SVC). Following SVC establishment comes the management phase - + MARS Clients need to be kept informed of group membership changes + within the scopes of their SVCs, so that leaf nodes may be added or + dropped as appropriate. + + For intra-cluster multicasting the current MARS approach is our + solution for these two phases. + + For the rest of this document we will focus on what VENUS would look + like when a VENUS Domain spans multiple MARS Clusters. Under such + circumstances VENUS is a mechanism co-ordinating the MARS entities of + each participating cluster. Each MARS is kept up to date with + sufficient domain-wide information to support both phases of client + operation (SVC establishment and SVC management) when the SVC's + endpoints are outside the immediate scope of a client's local MARS. + Inside a VENUS Domain a MARS Client is supplied information on group + members from all participating clusters. + + The following subsections look at the problems associated with both + of these phases independently. To a first approximation the problems + identified are independent of the possible inter-MARS mechanisms. The + reader may assume the MARS in any cluster has some undefined + mechanism for communicating with the MARSs of clusters immediately + adjacent to its own cluster (i.e. connected by a single Mrouter hop). + +3.1 SVC establishment - answering a MARS_REQUEST. + + The SVC establishment phase contains a number of inter-related + problems. + + First, the target of a MARS_REQUEST (an IP multicast group) is an + abstract entity. Let us assume that VENUS does not require every MARS + to know the entire list of group members across the participating + clusters. In this case each time a MARS_REQUEST is received by a + MARS from a local client, the MARS must construct a sequence of + MARS_MULTIs based on locally held information (on intra-cluster + members) and remotely solicited information. + + + +Armitage Informational [Page 5] + +RFC 2191 VENUS September 1997 + + + So how does it solicit this information? Unlike the unicast + situation, there is no definite, single direction to route a + MARS_REQUEST across the participating clusters. The only "right" + approach is to send the MARS_REQUEST to all clusters, since group + members may exist anywhere and everywhere. Let us allow one obvious + optimization - the MARS_REQUEST is propagated along the IP multicast + forwarding tree that has been established for the target group by + whatever IDMR protocol is running at the time. + + As noted in [4] there are various reasons why a Cluster's scope be + kept limited. Some of these (MARS Client or ATM NIC limitations) + imply that the VENUS discovery process not return more group members + in the MARS_MULTIs that the requesting MARS Client can handle. This + provides VENUS with an interesting problem of propagating out the + original MARS_REQUEST, but curtailing the MARS_REQUESTs propagation + when a sufficient number of group members have been identified. + Viewed from a different perspective, this means that the scope of + shortcut achievable by any given MARS Client may depend greatly on + the shape of the IP forwarding tree away from its location (and the + density of group members within clusters along the tree) at the time + the request was issued. + + How might we limit the number of group members returned to a given + MARS Client? Adding a limit TLV to the MARS_REQUEST itself is + trivial. At first glance it might appear that when the limit is being + reached we could summarize the next cluster along the tree by the ATM + address of the Mrouter into that cluster. The nett effect would be + that the MARS Client establishes a shortcut to many hosts that are + inside closer clusters, and passes its traffic to more distant + clusters through the distant Mrouter. However, this approach only + works passably well for a very simplistic multicast topology (e.g. a + linear concatenation of clusters). + + In a more general topology the IP multicast forwarding tree away from + the requesting MARS Client will branch a number of times, requiring + the MARS_REQUEST to be replicated along each branch. Ensuring that + the total number of returned group members does not exceed the + client's limit becomes rather more difficult to do efficiently. + (VENUS could simply halve the limit value each time it split a + MARS_REQUEST, but this might cause group member discovery on one + branch to end prematurely while all the group members along another + branch are discovered without reaching the subdivided limit.) + + Now consider this decision making process scattered across all the + clients in all participating clusters. Clients may have different + limits on how many group members they can handle - leading to + situations where different sources can shortcut to different + (sub)sets of the group members scattered across the participating + + + +Armitage Informational [Page 6] + +RFC 2191 VENUS September 1997 + + + clusters (because the IP multicast forwarding trees from senders in + different clusters may result in different discovery paths being + taken by their MARS_REQUESTs.) + + Finally, when the MARS_REQUEST passes a cluster where the target + group is MCS supported, VENUS must ensure the ATM address of the MCS + is collected rather than the addresses of the actual group members. + (To do otherwise would violate the remote cluster's intra-cluster + decision to use an MCS. The shortcut in this case must be content to + directly reach the remote cluster's MCS.) + + (A solution to part of this problem would be to ensure that a VENUS + Domain never has more MARS Clients throughout than the clients are + capable of adding as leaf nodes. This may or may not appeal to + people's desire for generality of a VENUS solution. It also would + appear to beg the question of why the problem of multiple-LIS + multicasting isn't solved simply by creating a single big MARS + Cluster.) + +3.2 SVC management - tracking group membership changes. + + Once a client's pt-mpt SVC is established, it must be kept up to + date. The consequence of this is simple, and potentially + devastating: The MARS_JOINs and MARS_LEAVEs from every MARS Client in + every participating cluster must be propagated to every possible + sender in every participating cluster (this applies to groups that + are VC Mesh supported - groups that are MCS supported in some or all + participating clusters introduce complications described below). + Unfortunately, the consequential signaling load (as all the + participating MARSs start broadcasting their MARS_JOIN/LEAVE + activity) is not localized to clusters containing MARS Clients who + have established shortcut SVCs. Since the IP multicast model is Any + to Multipoint, and you can never know where there may be source MARS + Clients, the JOINs and LEAVEs must be propagated everywhere, always, + just in case. (This is simply a larger scale version of sending JOINs + and LEAVEs to every cluster member over ClusterControlVC, and for + exactly the same reason.) + + The use of MCSs in some clusters instead of VC Meshes significantly + complicates the situation, as does the initial scoping of a client's + shortcut during the SVC establishment phase (described in the + preceding section). + + In Clusters where MCSs are supporting certain groups, MARS_JOINs or + MARS_LEAVEs are only propagated to MARS Clients when an MCS comes or + goes. However, it is not clear how to effectively accommodate the + current MARS_MIGRATE functionality (that allows a previously VC Mesh + based group to be shifted to an MCS within the scope of a single + + + +Armitage Informational [Page 7] + +RFC 2191 VENUS September 1997 + + + cluster). If an MCS starts up within a single Cluster, it is possible + to shift all the intra-cluster senders to the MCS using MARS_MIGRATE + as currently described in the MARS model. However, MARS Clients in + remote clusters that have shortcut SVCs into the local cluster also + need some signal to shift (otherwise they will continue to send their + packets directly to the group members in the local cluster). + + This is a non-trivial requirement, since we only want to force the + remote MARS Clients to drop some of their leaf nodes (the ones to + clients within the Cluster that now has an MCS), add the new MCS as a + leaf node, and leave all their other leaf nodes untouched (the cut- + through connections to other clusters). Simply broadcasting the + MARS_MIGRATE around all participating clusters would certainly not + work. VENUS needs a new control message with semantics of "replaced + leaf nodes {x, y, z} with leaf node {a}, and leave the rest alone". + Such a message is easy to define, but harder to use. + + Another issue for SVC management is that the scope over which a MARS + Client needs to receive JOINs and LEAVEs needs to respect the + Client's limited capacity for handling leaf nodes on its SVC. If the + MARS Client initially issued a MARS_REQUEST and indicated it could + handle 1000 leaf nodes, it is not clear how to ensure that subsequent + joins of new members wont exceed that limit. Furthermore, if the SVC + establishment phase decided that the SVC would stop at a particular + Mrouter (due to leaf node limits being reached), the Client probably + should not be receiving direct MARS_JOIN or MARS_LEAVE messages + pertaining to activity in the cluster "behind" this Mrouter. (To do + otherwise could lead to multiple copies of the source client's + packets reaching group members inside the remote cluster - one + version through the Mrouter, and another on the direct SVC connection + that the source client would establish after receiving a subsequent, + global MARS_JOIN regarding a host inside the remote cluster.) + + Another scenario involves the density of group members along the IDMR + multicast tree increasing with time after the initial MARS_REQUEST is + answered. Subsequent JOINs from Cluster members may dictate that a + "closer" Mrouter be used to aggregate the source's outbound traffic + (so as not to exceed the source's leaf node limitations). How to + dynamically shift between terminating on hosts within a Cluster, and + terminating on a cluster's edge Mrouter, is an open question. + + To complicate matters further, this scoping of the VENUS domain-wide + propagation of MARS_JOINs and MARS_LEAVEs needs to be on a per- + source- cluster basis, at least. If MARS Clients within the same + cluster have different leaf node limits, the problem worsens. Under + such circumstances, one client may have been able to establish a + shortcut SVC directly into a remote cluster while a second client - + in the same source cluster - may have been forced to terminate its + + + +Armitage Informational [Page 8] + +RFC 2191 VENUS September 1997 + + + shortcut on the remote cluster's Mrouter. The first client obviously + needs to know about group membership changes in the remote cluster, + whilst the second client does not. Propagating these JOIN/LEAVE + messages on ClusterControlVC in the source cluster will not work - + the MARS for the source cluster will need to explicitly send copies + of the JOIN/LEAVE messages only to those MARS Clients whose prior SVC + establishment phase indicates they need them. Propagation of messages + to indicate a VC Mesh to MCS transition within clusters may also need + to take account of the leaf node limitations of MARS Clients. The + scaling characteristics of this problem are left to the readers + imagination. + + It was noted in the previous section that a VENUS domain could be + limited to ensure there are never more MARS Clients than any one + client's leaf node limit. This would certainly avoid the need to for + complicated MARS_JOIN/LEAVE propagation mechanisms. However, it begs + the question of how different the VENUS domain then becomes from a + single, large MARS Cluster. + +4. What is the value in bypassing Mrouters? + + This is a good question, since the whole aim of developing a shortcut + connection mechanism is predicated on the assumption that bypassing + IP level entities is always a "win". However, this is arguably not + true for multicast. + + The most important observation that should be made about shortcut + connection scenarios is that they increase the exposure of any given + IP/ATM interface to externally generated SVCs. If there are a + potential 1000 senders in a VENUS Domain, then you (as a group + member) open yourself up to a potential demand for 1000 instances of + your re-assembly engine (and 1000 distinct incoming SVCs, when you + get added as a leaf node to each sender's pt-mpt SVC, which your + local switch port must be able to support). + + It should be no surprise that the ATM level scaling limits applicable + to a single MARS Cluster [4] will also apply to a VENUS Domain. Again + we're up against the question of why you'd bypass an Mrouter. As + noted in [4] Mrouters perform a useful function of data path + aggregation - 100 senders in one cluster become 1 pt-mpt SVC out of + the Mrouter into the next cluster along the tree. They also hide MARS + signaling activity - individual group membership changes in one + cluster are hidden from IP/ATM interfaces in surrounding clusters. + The loss of these benefits must be factored into any network designed + to utilize multicast shortcut connections. + + + + + + +Armitage Informational [Page 9] + +RFC 2191 VENUS September 1997 + + + (For the sake of completeness, it must be noted that extremely poor + mismatches of IP and ATM topologies may make Mrouter bypass + attractive if it improves the use of the underlying ATM cloud. There + may also be benefits in removing the additional re- + assembly/segmentation latencies of having packets pass through an + Mrouter. However, a VENUS Domain ascertained to be small enough to + avoid the scaling limits in [4] might just as well be constructed as + a single large MARS Cluster. A large cluster also avoids a + topological mismatch between IP Mrouters and ATM switches.) + +5. Relationship to Distributed MARS protocols. + + The ION working group is looking closely at the development of + distributed MARS architectures. An outline of some issues is provided + in [5,6]. As noted earlier in this document the problem space looks + very similar that faced by our hypothetical VENUS Domain. For + example, in the load-sharing distributed MARS model: + + - The Cluster is partitioned into sub-clusters. + + - Each Active MARS is assigned a particular sub-cluster, and uses + its own sub-ClusterControlVC to propagate JOIN/LEAVE messages to + members of its sub-cluster. + + - The MARS_REQUEST from any sub-cluster member must return + information from all the sub-clusters, so as to ensure that all a + group's members across the cluster are identified. + + - Group membership changes in any one sub-cluster must be + immediately propagated to all the other sub-clusters. + + There is a clear analogy to be made between a distributed MARS + Cluster, and a VENUS Domain made up of multiple single-MARS Clusters. + The information that must be shared between sub-clusters in a + distributed MARS scenario is similar to the information that must be + shared between Clusters in a VENUS Domain. + + The distributed MARS problem is slightly simpler than that faced by + VENUS: + + - There are no Mrouters (IDMR nodes) within the scope of the + distributed Cluster. + + - In a distributed MARS Cluster an MCS supported group uses the + same MCS across all the sub-clusters (unlike the VENUS Domain, + where complete generality makes it necessary to cope with mixtures + of MCS and VC Mesh based Clusters). + + + + +Armitage Informational [Page 10] + +RFC 2191 VENUS September 1997 + + +6. Conclusion. + + This document has described a hypothetical multicast shortcut + connection scheme, dubbed "Very Extensive NonUnicast Service" + (VENUS). The two phases of multicast support - SVC establishment, + and SVC management - are shown to be essential whether the scope is a + Cluster or a wider VENUS Domain. It has been shown that once the + potential scope of a pt-mpt SVC at establishment phase has been + expanded, the scope of the SVC management mechanism must similarly be + expanded. This means timely tracking and propagation of group + membership changes across the entire scope of a VENUS Domain. + + It has also been noted that there is little difference in result + between a VENUS Domain and a large MARS Cluster. Both suffer from the + same fundamental scaling limitations, and both can be arranged to + provide shortcut of unicast routing boundaries. However, a completely + general multi-cluster VENUS solution ends up being more complex. It + needs to deal with bypassed Mrouter boundaries, and dynamically + changing group membership densities along multicast distribution + trees established by the IDMR protocols in use. + + No solutions have been presented. This document's role is to provide + context for future developments. + +Acknowledgment + + This document was prepared while the author was with the + Internetworking Research group at Bellcore. + +Security Considerations + + This memo addresses specific scaling issues associated with the + extension of the MARS architecture beyond that described in RFC 2022. + It is an Informational memo, and does not mandate any additional + protocol behaviors beyond those described in RFC 2022. As such, the + security implications are no greater or less than the implications + inherent in RFC 2022. Should enhancements to security be required, + they would need to be added as an extension to the base architecture + in RFC 2022. + + + + + + + + + + + + +Armitage Informational [Page 11] + +RFC 2191 VENUS September 1997 + + +Author's Address + + Grenville Armitage + Bell Labs, Lucent Technologies. + 101 Crawfords Corner Rd, + Holmdel, NJ, 07733 + USA + + EMail: gja@dnrc.bell-labs.com + + +References + + [1] Laubach, M., "Classical IP and ARP over ATM", RFC 1577, Hewlett- + Packard Laboratories, December 1993. + + [2] Armitage, G., "Support for Multicast over UNI 3.0/3.1 based ATM + Networks.", Bellcore, RFC 2022, November 1996. + + [3] Luciani, J., et al, "NBMA Next Hop Resolution Protocol (NHRP)", + Work in Progress, February 1997. + + [4] Armitage, G., "Issues affecting MARS Cluster Size", Bellcore, RFC + 2121, March 1997. + + [5] Armitage, G., "Redundant MARS architectures and SCSP", Bellcore, + Work in Progress, November 1996. + + [6] Luciani, J., G. Armitage, J. Jalpern, "Server Cache + Synchronization Protocol (SCSP) - NBMA", Work in Progress, March 1997. + + [7] Rekhter, Y., D. Farinacci, " Support for Sparse Mode PIM over + ATM", Cisco Systems, Work in Progress, April 1996. + + + + + + + + + + + + + + + + + + +Armitage Informational [Page 12] + |