summaryrefslogtreecommitdiff
path: root/doc/rfc/rfc2191.txt
diff options
context:
space:
mode:
authorThomas Voss <mail@thomasvoss.com> 2024-11-27 20:54:24 +0100
committerThomas Voss <mail@thomasvoss.com> 2024-11-27 20:54:24 +0100
commit4bfd864f10b68b71482b35c818559068ef8d5797 (patch)
treee3989f47a7994642eb325063d46e8f08ffa681dc /doc/rfc/rfc2191.txt
parentea76e11061bda059ae9f9ad130a9895cc85607db (diff)
doc: Add RFC documents
Diffstat (limited to 'doc/rfc/rfc2191.txt')
-rw-r--r--doc/rfc/rfc2191.txt675
1 files changed, 675 insertions, 0 deletions
diff --git a/doc/rfc/rfc2191.txt b/doc/rfc/rfc2191.txt
new file mode 100644
index 0000000..1e5bb2d
--- /dev/null
+++ b/doc/rfc/rfc2191.txt
@@ -0,0 +1,675 @@
+
+
+
+
+
+
+Network Working Group G. Armitage
+Request for Comments: 2191 Lucent Technologies
+Category: Informational September 1997
+
+
+ VENUS - Very Extensive Non-Unicast Service
+
+Status of this Memo
+
+ This memo provides information for the Internet community. This memo
+ does not specify an Internet standard of any kind. Distribution of
+ this memo is unlimited.
+
+Abstract
+
+ The MARS model (RFC2022) provides a solution to intra-LIS IP
+ multicasting over ATM, establishing and managing the use of ATM pt-
+ mpt SVCs for IP multicast packet forwarding. Inter-LIS multicast
+ forwarding is achieved using Mrouters, in a similar manner to which
+ the "Classical IP over ATM" model uses Routers to inter-connect LISes
+ for unicast traffic. The development of unicast IP shortcut
+ mechanisms (e.g. NHRP) has led some people to request the
+ development of a Multicast equivalent. There are a number of
+ different approaches. This document focuses exclusively on the
+ problems associated with extending the MARS model to cover multiple
+ clusters or clusters spanning more than one subnet. It describes a
+ hypothetical solution, dubbed "Very Extensive NonUnicast Service"
+ (VENUS), and shows how complex such a service would be. It is also
+ noted that VENUS ultimately has the look and feel of a single, large
+ cluster using a distributed MARS. This document is being issued to
+ help focus ION efforts towards alternative solutions for establishing
+ ATM level multicast connections between LISes.
+
+1. Introduction
+
+ The classical model of the Internet running over an ATM cloud
+ consists of multiple Logical IP Subnets (LISs) interconnected by IP
+ Routers [1]. The evolving IP Multicast over ATM solution (the "MARS
+ model" [2]) retains the classical model. The LIS becomes a "MARS
+ Cluster", and Clusters are interconnected by conventional IP
+ Multicast routers (Mrouters).
+
+ The development of NHRP [3], a protocol for discovering and managing
+ unicast forwarding paths that bypass IP routers, has led to some
+ calls for an IP multicast equivalent. Unfortunately, the IP
+ multicast service is a rather different beast to the IP unicast
+ service. This document aims to explain how much of what has been
+ learned during the development of NHRP must be carefully scrutinized
+
+
+
+Armitage Informational [Page 1]
+
+RFC 2191 VENUS September 1997
+
+
+ before being re-applied to the multicast scenario. Indeed, the
+ service provided by the MARS and MARS Clients in [2] are almost
+ orthogonal to the IP unicast service over ATM.
+
+ For the sake of discussion, let's call this hypothetical multicast
+ shortcut discovery protocol the "Very Extensive Non-Unicast Service"
+ (VENUS). A "VENUS Domain" is defined as the set of hosts from two or
+ more participating Logical IP Subnets (LISs). A multicast shortcut
+ connection is a point to multipoint SVC whose leaf nodes are
+ scattered around the VENUS Domain. (It will be noted in section 2
+ that a VENUS Domain might consist of a single MARS Cluster spanning
+ multiple LISs, or multiple MARS Clusters.)
+
+ VENUS faces a number of fundamental problems. The first is exploding
+ the scope over which individual IP/ATM interfaces must track and
+ react to IP multicast group membership changes. Under the classical
+ IP routing model Mrouters act as aggregation points for multicast
+ traffic flows in and out of Clusters [4]. They also act as
+ aggregators of group membership change information - only the IP/ATM
+ interfaces within each Cluster need to know the specific identities
+ of their local (intra-cluster) group members at any given time.
+ However, once you have sources within a VENUS Domain establishing
+ shortcut connections the data and signaling plane aggregation of
+ Mrouters is lost. In order for all possible sources throughout a
+ VENUS Domain to manage their outgoing pt-mpt SVCs they must be kept
+ aware of MARS_JOINs and MARS_LEAVEs occuring in every MARS Cluster
+ that makes up a VENUS Domain. The nett effect is that a VENUS domain
+ looks very similar to a single, large distributed MARS Cluster.
+
+ A second problem is the impact that shortcut connections will have on
+ IP level Inter Domain Multicast Routing (IDMR) protocols. Multicast
+ groups have many sources and many destinations scattered amongst the
+ participating Clusters. IDMR protocols assume that they can calculate
+ efficient inter-Cluster multicast trees by aggregating individual
+ sources or group members in any given Cluster (subnet) behind the
+ Mrouter serving that Cluster. If sources are able to simply bypass an
+ Mrouter we introduce a requirement that the existence of each and
+ every shortcut connection be propagated into the IDMR decision making
+ processes. The IDMR protocols may need to adapt when a source's
+ traffic bypasses its local Mrouter(s) and is injected into Mrouters
+ at more distant points on the IP-level multicast distribution tree.
+ (This issue has been looked at in [7], focussing on building
+ forwarding trees within networks where the termination points are
+ small in number and sparsely distributed. VENUS introduces tougher
+ requirements by assuming that multicast group membership may be dense
+ across the region of interest.)
+
+
+
+
+
+Armitage Informational [Page 2]
+
+RFC 2191 VENUS September 1997
+
+
+ This document will focus primarily on the internal problems of a
+ VENUS Domain, and leave the IDMR interactions for future analysis.
+
+2. What does it mean to "shortcut" ?
+
+ Before going further it is worth considering both the definition of
+ the Cluster, and two possible definitions of "shortcut".
+
+2.1 What is a Cluster?
+
+ In [2] a MARS Cluster is defined as the set of IP/ATM interfaces that
+ are willing to engage in direct, ATM level pt-mpt SVCs to perform IP
+ multicast packet forwarding. Each IP/ATM interface (a MARS Client)
+ must keep state information regarding the ATM addresses of each leaf
+ node (recipient) of each pt-mpt SVC it has open. In addition, each
+ MARS Client receives MARS_JOIN and MARS_LEAVE messages from the MARS
+ whenever there is a requirement that Clients around the Cluster need
+ to update their pt-mpt SVCs for a given IP multicast group.
+
+ It is worth noting that no MARS Client has any concept of how big its
+ local cluster is - this knowledge is kept only by the MARS that a
+ given Client is registered with.
+
+ Fundamentally the Cluster (and the MARS model as a whole) is a
+ response to the requirement that any multicast IP/ATM interface using
+ pt-mpt SVCs must, as group membership changes, add and drop leaf
+ nodes itself. This means that some mechanism, spanning all possible
+ group members within the scopes of these pt-mpt SVCs, is required to
+ collect group membership information and distribute it in a timely
+ fashion to those interfaces. This is the MARS Cluster, with certain
+ scaling limits described in [4].
+
+2.2 LIS/Cluster boundary "shortcut"
+
+ The currently popular definition of "shortcut" is based on the
+ existence of unicast LIS boundaries. It is tied to the notion that
+ LIS boundaries have physical routers, and cutting through a LIS
+ boundary means bypassing a router. Intelligently bypassing routers
+ that sit at the edges of LISs has been the goal of NHRP. Discovering
+ the ATM level identity of an IP endpoint in a different LIS allows a
+ direct SVC to be established, thus shortcutting the logical IP
+ topology (and very real routers) along the unicast path from source
+ to destination.
+
+ For simplicity of early adoption RFC2022 recommends that a Cluster's
+ scope be made equivalent to that of a LIS. Under these circumstances
+ the "Classical IP" routing model places Mrouters at LIS/Cluster
+ boundaries, and multicast shortcutting must involve bypassing the
+
+
+
+Armitage Informational [Page 3]
+
+RFC 2191 VENUS September 1997
+
+
+ same physical routing entities as unicast shortcutting. Each MARS
+ Cluster would be independent and contain only those IP/ATM interfaces
+ that had been assigned to the same LIS.
+
+ As a consequence, a VENUS Domain covering the hosts in a number of
+ LIS/Clusters would have to co-ordinate each individual MARS from each
+ LIS/Cluster (to ensure group membership updates from around the VENUS
+ Domain were propagated correctly).
+
+2.3 Big Cluster, LIS boundary "shortcut"
+
+ The MARS model's fundamental definition of a Cluster was deliberately
+ created to be independent of unicast terminology. Although not
+ currently well understood, it is possible to build a single MARS
+ Cluster that encompasses the members of multiple LISs. As expected,
+ inter-LIS unicast traffic would pass through (or bypass, if using
+ NHRP) routers on the LIS boundaries. Also as expected, each IP/ATM
+ interface, acting as a MARS Client, would forward their IP multicast
+ packets directly to intra-cluster group members. However, because the
+ direct intra-cluster SVCs would exist between hosts from the
+ different LISs making up the cluster, this could be considered a
+ "shortcut" of the unicast LIS boundaries.
+
+ This approach immediately brings up the problem of how the IDMR
+ protocols will react. Mrouters only need to exist at the edges of
+ Clusters. In the case of a single Cluster spanning multiple LISs,
+ each LIS becomes hidden behind the Mrouter at the Cluster's edge.
+ This is arguably not a big problem if the Cluster is a stub on an
+ IDMR protocol's multicast distribution tree, and if there is only a
+ single Mrouter in or out of the Cluster. Problems arise when two or
+ more Mrouters are attached to the edges of the Cluster, and the
+ Cluster is used for transit multicast traffic. Each Mrouter's
+ interface is assigned a unicast identity (e.g. that of the unicast
+ router containing the Mrouter). IDMR protocols that filter packets
+ based on the correctness of the upstream source may be confused at
+ receiving IP multicast packets directly from another Mrouter in the
+ same cluster but notionally "belonging" to an LIS multiple unicast IP
+ hops away.
+
+ Adjusting the packet filtering algorithms of Mrouters is something
+ that needs to be addressed by any multicast shortcut scheme. It has
+ been noted before and a solution proposed in [7]. For the sake of
+ argument this document will assume the problem solvable. (However, it
+ is important that any solution scales well under general topologies
+ and group membership densities.)
+
+
+
+
+
+
+Armitage Informational [Page 4]
+
+RFC 2191 VENUS September 1997
+
+
+ A multi-LIS MARS Cluster can be considered a simple VENUS Domain.
+ Since it is a single Cluster it can be scaled using the distributed
+ MARS solutions currently being developed within the IETF [5,6].
+
+3. So what must VENUS look like?
+
+ A number of functions that occur in the MARS model are fundamental to
+ the problem of managing root controlled, pt-mpt SVCs. The initial
+ setup of the forwarding SVC by any one MARS Client requires a
+ query/response exchange with the Client's local MARS, establishing
+ who the current group members are (i.e. what leaf nodes should be on
+ the SVC). Following SVC establishment comes the management phase -
+ MARS Clients need to be kept informed of group membership changes
+ within the scopes of their SVCs, so that leaf nodes may be added or
+ dropped as appropriate.
+
+ For intra-cluster multicasting the current MARS approach is our
+ solution for these two phases.
+
+ For the rest of this document we will focus on what VENUS would look
+ like when a VENUS Domain spans multiple MARS Clusters. Under such
+ circumstances VENUS is a mechanism co-ordinating the MARS entities of
+ each participating cluster. Each MARS is kept up to date with
+ sufficient domain-wide information to support both phases of client
+ operation (SVC establishment and SVC management) when the SVC's
+ endpoints are outside the immediate scope of a client's local MARS.
+ Inside a VENUS Domain a MARS Client is supplied information on group
+ members from all participating clusters.
+
+ The following subsections look at the problems associated with both
+ of these phases independently. To a first approximation the problems
+ identified are independent of the possible inter-MARS mechanisms. The
+ reader may assume the MARS in any cluster has some undefined
+ mechanism for communicating with the MARSs of clusters immediately
+ adjacent to its own cluster (i.e. connected by a single Mrouter hop).
+
+3.1 SVC establishment - answering a MARS_REQUEST.
+
+ The SVC establishment phase contains a number of inter-related
+ problems.
+
+ First, the target of a MARS_REQUEST (an IP multicast group) is an
+ abstract entity. Let us assume that VENUS does not require every MARS
+ to know the entire list of group members across the participating
+ clusters. In this case each time a MARS_REQUEST is received by a
+ MARS from a local client, the MARS must construct a sequence of
+ MARS_MULTIs based on locally held information (on intra-cluster
+ members) and remotely solicited information.
+
+
+
+Armitage Informational [Page 5]
+
+RFC 2191 VENUS September 1997
+
+
+ So how does it solicit this information? Unlike the unicast
+ situation, there is no definite, single direction to route a
+ MARS_REQUEST across the participating clusters. The only "right"
+ approach is to send the MARS_REQUEST to all clusters, since group
+ members may exist anywhere and everywhere. Let us allow one obvious
+ optimization - the MARS_REQUEST is propagated along the IP multicast
+ forwarding tree that has been established for the target group by
+ whatever IDMR protocol is running at the time.
+
+ As noted in [4] there are various reasons why a Cluster's scope be
+ kept limited. Some of these (MARS Client or ATM NIC limitations)
+ imply that the VENUS discovery process not return more group members
+ in the MARS_MULTIs that the requesting MARS Client can handle. This
+ provides VENUS with an interesting problem of propagating out the
+ original MARS_REQUEST, but curtailing the MARS_REQUESTs propagation
+ when a sufficient number of group members have been identified.
+ Viewed from a different perspective, this means that the scope of
+ shortcut achievable by any given MARS Client may depend greatly on
+ the shape of the IP forwarding tree away from its location (and the
+ density of group members within clusters along the tree) at the time
+ the request was issued.
+
+ How might we limit the number of group members returned to a given
+ MARS Client? Adding a limit TLV to the MARS_REQUEST itself is
+ trivial. At first glance it might appear that when the limit is being
+ reached we could summarize the next cluster along the tree by the ATM
+ address of the Mrouter into that cluster. The nett effect would be
+ that the MARS Client establishes a shortcut to many hosts that are
+ inside closer clusters, and passes its traffic to more distant
+ clusters through the distant Mrouter. However, this approach only
+ works passably well for a very simplistic multicast topology (e.g. a
+ linear concatenation of clusters).
+
+ In a more general topology the IP multicast forwarding tree away from
+ the requesting MARS Client will branch a number of times, requiring
+ the MARS_REQUEST to be replicated along each branch. Ensuring that
+ the total number of returned group members does not exceed the
+ client's limit becomes rather more difficult to do efficiently.
+ (VENUS could simply halve the limit value each time it split a
+ MARS_REQUEST, but this might cause group member discovery on one
+ branch to end prematurely while all the group members along another
+ branch are discovered without reaching the subdivided limit.)
+
+ Now consider this decision making process scattered across all the
+ clients in all participating clusters. Clients may have different
+ limits on how many group members they can handle - leading to
+ situations where different sources can shortcut to different
+ (sub)sets of the group members scattered across the participating
+
+
+
+Armitage Informational [Page 6]
+
+RFC 2191 VENUS September 1997
+
+
+ clusters (because the IP multicast forwarding trees from senders in
+ different clusters may result in different discovery paths being
+ taken by their MARS_REQUESTs.)
+
+ Finally, when the MARS_REQUEST passes a cluster where the target
+ group is MCS supported, VENUS must ensure the ATM address of the MCS
+ is collected rather than the addresses of the actual group members.
+ (To do otherwise would violate the remote cluster's intra-cluster
+ decision to use an MCS. The shortcut in this case must be content to
+ directly reach the remote cluster's MCS.)
+
+ (A solution to part of this problem would be to ensure that a VENUS
+ Domain never has more MARS Clients throughout than the clients are
+ capable of adding as leaf nodes. This may or may not appeal to
+ people's desire for generality of a VENUS solution. It also would
+ appear to beg the question of why the problem of multiple-LIS
+ multicasting isn't solved simply by creating a single big MARS
+ Cluster.)
+
+3.2 SVC management - tracking group membership changes.
+
+ Once a client's pt-mpt SVC is established, it must be kept up to
+ date. The consequence of this is simple, and potentially
+ devastating: The MARS_JOINs and MARS_LEAVEs from every MARS Client in
+ every participating cluster must be propagated to every possible
+ sender in every participating cluster (this applies to groups that
+ are VC Mesh supported - groups that are MCS supported in some or all
+ participating clusters introduce complications described below).
+ Unfortunately, the consequential signaling load (as all the
+ participating MARSs start broadcasting their MARS_JOIN/LEAVE
+ activity) is not localized to clusters containing MARS Clients who
+ have established shortcut SVCs. Since the IP multicast model is Any
+ to Multipoint, and you can never know where there may be source MARS
+ Clients, the JOINs and LEAVEs must be propagated everywhere, always,
+ just in case. (This is simply a larger scale version of sending JOINs
+ and LEAVEs to every cluster member over ClusterControlVC, and for
+ exactly the same reason.)
+
+ The use of MCSs in some clusters instead of VC Meshes significantly
+ complicates the situation, as does the initial scoping of a client's
+ shortcut during the SVC establishment phase (described in the
+ preceding section).
+
+ In Clusters where MCSs are supporting certain groups, MARS_JOINs or
+ MARS_LEAVEs are only propagated to MARS Clients when an MCS comes or
+ goes. However, it is not clear how to effectively accommodate the
+ current MARS_MIGRATE functionality (that allows a previously VC Mesh
+ based group to be shifted to an MCS within the scope of a single
+
+
+
+Armitage Informational [Page 7]
+
+RFC 2191 VENUS September 1997
+
+
+ cluster). If an MCS starts up within a single Cluster, it is possible
+ to shift all the intra-cluster senders to the MCS using MARS_MIGRATE
+ as currently described in the MARS model. However, MARS Clients in
+ remote clusters that have shortcut SVCs into the local cluster also
+ need some signal to shift (otherwise they will continue to send their
+ packets directly to the group members in the local cluster).
+
+ This is a non-trivial requirement, since we only want to force the
+ remote MARS Clients to drop some of their leaf nodes (the ones to
+ clients within the Cluster that now has an MCS), add the new MCS as a
+ leaf node, and leave all their other leaf nodes untouched (the cut-
+ through connections to other clusters). Simply broadcasting the
+ MARS_MIGRATE around all participating clusters would certainly not
+ work. VENUS needs a new control message with semantics of "replaced
+ leaf nodes {x, y, z} with leaf node {a}, and leave the rest alone".
+ Such a message is easy to define, but harder to use.
+
+ Another issue for SVC management is that the scope over which a MARS
+ Client needs to receive JOINs and LEAVEs needs to respect the
+ Client's limited capacity for handling leaf nodes on its SVC. If the
+ MARS Client initially issued a MARS_REQUEST and indicated it could
+ handle 1000 leaf nodes, it is not clear how to ensure that subsequent
+ joins of new members wont exceed that limit. Furthermore, if the SVC
+ establishment phase decided that the SVC would stop at a particular
+ Mrouter (due to leaf node limits being reached), the Client probably
+ should not be receiving direct MARS_JOIN or MARS_LEAVE messages
+ pertaining to activity in the cluster "behind" this Mrouter. (To do
+ otherwise could lead to multiple copies of the source client's
+ packets reaching group members inside the remote cluster - one
+ version through the Mrouter, and another on the direct SVC connection
+ that the source client would establish after receiving a subsequent,
+ global MARS_JOIN regarding a host inside the remote cluster.)
+
+ Another scenario involves the density of group members along the IDMR
+ multicast tree increasing with time after the initial MARS_REQUEST is
+ answered. Subsequent JOINs from Cluster members may dictate that a
+ "closer" Mrouter be used to aggregate the source's outbound traffic
+ (so as not to exceed the source's leaf node limitations). How to
+ dynamically shift between terminating on hosts within a Cluster, and
+ terminating on a cluster's edge Mrouter, is an open question.
+
+ To complicate matters further, this scoping of the VENUS domain-wide
+ propagation of MARS_JOINs and MARS_LEAVEs needs to be on a per-
+ source- cluster basis, at least. If MARS Clients within the same
+ cluster have different leaf node limits, the problem worsens. Under
+ such circumstances, one client may have been able to establish a
+ shortcut SVC directly into a remote cluster while a second client -
+ in the same source cluster - may have been forced to terminate its
+
+
+
+Armitage Informational [Page 8]
+
+RFC 2191 VENUS September 1997
+
+
+ shortcut on the remote cluster's Mrouter. The first client obviously
+ needs to know about group membership changes in the remote cluster,
+ whilst the second client does not. Propagating these JOIN/LEAVE
+ messages on ClusterControlVC in the source cluster will not work -
+ the MARS for the source cluster will need to explicitly send copies
+ of the JOIN/LEAVE messages only to those MARS Clients whose prior SVC
+ establishment phase indicates they need them. Propagation of messages
+ to indicate a VC Mesh to MCS transition within clusters may also need
+ to take account of the leaf node limitations of MARS Clients. The
+ scaling characteristics of this problem are left to the readers
+ imagination.
+
+ It was noted in the previous section that a VENUS domain could be
+ limited to ensure there are never more MARS Clients than any one
+ client's leaf node limit. This would certainly avoid the need to for
+ complicated MARS_JOIN/LEAVE propagation mechanisms. However, it begs
+ the question of how different the VENUS domain then becomes from a
+ single, large MARS Cluster.
+
+4. What is the value in bypassing Mrouters?
+
+ This is a good question, since the whole aim of developing a shortcut
+ connection mechanism is predicated on the assumption that bypassing
+ IP level entities is always a "win". However, this is arguably not
+ true for multicast.
+
+ The most important observation that should be made about shortcut
+ connection scenarios is that they increase the exposure of any given
+ IP/ATM interface to externally generated SVCs. If there are a
+ potential 1000 senders in a VENUS Domain, then you (as a group
+ member) open yourself up to a potential demand for 1000 instances of
+ your re-assembly engine (and 1000 distinct incoming SVCs, when you
+ get added as a leaf node to each sender's pt-mpt SVC, which your
+ local switch port must be able to support).
+
+ It should be no surprise that the ATM level scaling limits applicable
+ to a single MARS Cluster [4] will also apply to a VENUS Domain. Again
+ we're up against the question of why you'd bypass an Mrouter. As
+ noted in [4] Mrouters perform a useful function of data path
+ aggregation - 100 senders in one cluster become 1 pt-mpt SVC out of
+ the Mrouter into the next cluster along the tree. They also hide MARS
+ signaling activity - individual group membership changes in one
+ cluster are hidden from IP/ATM interfaces in surrounding clusters.
+ The loss of these benefits must be factored into any network designed
+ to utilize multicast shortcut connections.
+
+
+
+
+
+
+Armitage Informational [Page 9]
+
+RFC 2191 VENUS September 1997
+
+
+ (For the sake of completeness, it must be noted that extremely poor
+ mismatches of IP and ATM topologies may make Mrouter bypass
+ attractive if it improves the use of the underlying ATM cloud. There
+ may also be benefits in removing the additional re-
+ assembly/segmentation latencies of having packets pass through an
+ Mrouter. However, a VENUS Domain ascertained to be small enough to
+ avoid the scaling limits in [4] might just as well be constructed as
+ a single large MARS Cluster. A large cluster also avoids a
+ topological mismatch between IP Mrouters and ATM switches.)
+
+5. Relationship to Distributed MARS protocols.
+
+ The ION working group is looking closely at the development of
+ distributed MARS architectures. An outline of some issues is provided
+ in [5,6]. As noted earlier in this document the problem space looks
+ very similar that faced by our hypothetical VENUS Domain. For
+ example, in the load-sharing distributed MARS model:
+
+ - The Cluster is partitioned into sub-clusters.
+
+ - Each Active MARS is assigned a particular sub-cluster, and uses
+ its own sub-ClusterControlVC to propagate JOIN/LEAVE messages to
+ members of its sub-cluster.
+
+ - The MARS_REQUEST from any sub-cluster member must return
+ information from all the sub-clusters, so as to ensure that all a
+ group's members across the cluster are identified.
+
+ - Group membership changes in any one sub-cluster must be
+ immediately propagated to all the other sub-clusters.
+
+ There is a clear analogy to be made between a distributed MARS
+ Cluster, and a VENUS Domain made up of multiple single-MARS Clusters.
+ The information that must be shared between sub-clusters in a
+ distributed MARS scenario is similar to the information that must be
+ shared between Clusters in a VENUS Domain.
+
+ The distributed MARS problem is slightly simpler than that faced by
+ VENUS:
+
+ - There are no Mrouters (IDMR nodes) within the scope of the
+ distributed Cluster.
+
+ - In a distributed MARS Cluster an MCS supported group uses the
+ same MCS across all the sub-clusters (unlike the VENUS Domain,
+ where complete generality makes it necessary to cope with mixtures
+ of MCS and VC Mesh based Clusters).
+
+
+
+
+Armitage Informational [Page 10]
+
+RFC 2191 VENUS September 1997
+
+
+6. Conclusion.
+
+ This document has described a hypothetical multicast shortcut
+ connection scheme, dubbed "Very Extensive NonUnicast Service"
+ (VENUS). The two phases of multicast support - SVC establishment,
+ and SVC management - are shown to be essential whether the scope is a
+ Cluster or a wider VENUS Domain. It has been shown that once the
+ potential scope of a pt-mpt SVC at establishment phase has been
+ expanded, the scope of the SVC management mechanism must similarly be
+ expanded. This means timely tracking and propagation of group
+ membership changes across the entire scope of a VENUS Domain.
+
+ It has also been noted that there is little difference in result
+ between a VENUS Domain and a large MARS Cluster. Both suffer from the
+ same fundamental scaling limitations, and both can be arranged to
+ provide shortcut of unicast routing boundaries. However, a completely
+ general multi-cluster VENUS solution ends up being more complex. It
+ needs to deal with bypassed Mrouter boundaries, and dynamically
+ changing group membership densities along multicast distribution
+ trees established by the IDMR protocols in use.
+
+ No solutions have been presented. This document's role is to provide
+ context for future developments.
+
+Acknowledgment
+
+ This document was prepared while the author was with the
+ Internetworking Research group at Bellcore.
+
+Security Considerations
+
+ This memo addresses specific scaling issues associated with the
+ extension of the MARS architecture beyond that described in RFC 2022.
+ It is an Informational memo, and does not mandate any additional
+ protocol behaviors beyond those described in RFC 2022. As such, the
+ security implications are no greater or less than the implications
+ inherent in RFC 2022. Should enhancements to security be required,
+ they would need to be added as an extension to the base architecture
+ in RFC 2022.
+
+
+
+
+
+
+
+
+
+
+
+
+Armitage Informational [Page 11]
+
+RFC 2191 VENUS September 1997
+
+
+Author's Address
+
+ Grenville Armitage
+ Bell Labs, Lucent Technologies.
+ 101 Crawfords Corner Rd,
+ Holmdel, NJ, 07733
+ USA
+
+ EMail: gja@dnrc.bell-labs.com
+
+
+References
+
+ [1] Laubach, M., "Classical IP and ARP over ATM", RFC 1577, Hewlett-
+ Packard Laboratories, December 1993.
+
+ [2] Armitage, G., "Support for Multicast over UNI 3.0/3.1 based ATM
+ Networks.", Bellcore, RFC 2022, November 1996.
+
+ [3] Luciani, J., et al, "NBMA Next Hop Resolution Protocol (NHRP)",
+ Work in Progress, February 1997.
+
+ [4] Armitage, G., "Issues affecting MARS Cluster Size", Bellcore, RFC
+ 2121, March 1997.
+
+ [5] Armitage, G., "Redundant MARS architectures and SCSP", Bellcore,
+ Work in Progress, November 1996.
+
+ [6] Luciani, J., G. Armitage, J. Jalpern, "Server Cache
+ Synchronization Protocol (SCSP) - NBMA", Work in Progress, March 1997.
+
+ [7] Rekhter, Y., D. Farinacci, " Support for Sparse Mode PIM over
+ ATM", Cisco Systems, Work in Progress, April 1996.
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+Armitage Informational [Page 12]
+