diff options
Diffstat (limited to 'doc/rfc/rfc1077.txt')
-rw-r--r-- | doc/rfc/rfc1077.txt | 2579 |
1 files changed, 2579 insertions, 0 deletions
diff --git a/doc/rfc/rfc1077.txt b/doc/rfc/rfc1077.txt new file mode 100644 index 0000000..75a5a64 --- /dev/null +++ b/doc/rfc/rfc1077.txt @@ -0,0 +1,2579 @@ + + + + + + +Network Working Group Gigabit Working Group +Request for Comments: 1077 B. Leiner, Editor + November 1988 + + + Critical Issues in High Bandwidth Networking + + +Status of this Memo + + This memo presents the results of a working group on High Bandwidth + Networking. This RFC is for your information and you are encouraged + to comment on the issues presented. Distribution of this memo is + unlimited. + +ABSTRACT + + At the request of Maj. Mark Pullen and Maj. Brian Boesch of DARPA, an + ad-hoc working group was assembled to develop a set of + recommendations on the research required to achieve a ubiquitous + high-bandwidth network as discussed in the FCCSET recommendations for + Phase III. + + This report outlines a set of research topics aimed at providing the + technology base for an interconnected set of networks that can + provide highbandwidth capabilities. The suggested research focus + draws upon ongoing research and augments it with basic and applied + components. The major activities are the development and + demonstration of a gigabit backbone network, the development and + demonstration of an interconnected set of networks with gigabit + throughput and appropriate management techniques, and the development + and demonstration of the required overall architecture that allows + users to gain access to such high bandwidth. + + + + + + + + + + + + + + + + + + +Gigabit Working Group [Page 1] + +RFC 1077 November 1988 + + + 1. Introduction and Summary + + + + 1.1. Background + + + The computer communications world is evolving toward both high- + bandwidth capability and high-bandwidth requirements. The recent + workshop conducted under the auspices of the FCCSET Committee on High + Performance Computing [1] identified a number of areas where + extremely high-bandwidth networking is required to support the + scientific research community. These areas range from remote + graphical visualization of supercomputer results through the movement + of high rate sensor data from space to the ground-based scientific + investigator. Similar requirements exist for other applications, + such as military command and control (C2) where there is a need to + quickly access and act on data obtained from real-time sensors. The + workshop identified requirements for switched high-bandwidth service + in excess of 300 Mbit/s to a single user, and the need to support + service in the range of a Mbit/s on a low-duty-cycle basis to + millions of researchers. When added to the needs of the military and + commercial users, the aggregate requirement for communications + service adds up to many billions of bits per second. The results of + this workshop were incorporated into a report by the FCCSET [2]. + + Fortunately, technology is also moving rapidly. Even today, the + installed base of fiber optics communications allows us to consider + aggregate bandwidths in the range of Gbit/s and beyond to limited + geographical regions. Estimates arrived at in the workshop lead one + to believe that there will be available raw bandwidth approaching + terabits per second. + + The critical question to be addressed is how this raw bandwidth can + be used to satisfy the requirements identified in the workshop: 1) + provide bandwidth on the order of several Gbit/s to individual users, + and 2) provide modest bandwidth on the order of several Mbit/s to a + large number of users in a cost-effective manner through the + aggregation of their traffic. + + Through its research funding, the Defense Advanced Research Projects + Agency (DARPA) has played a central role in the development of + packet-oriented communications, which has been of tremendous benefit + to the U.S. military in terms of survivability and interoperability. + DARPA-funded research has resulted in the ARPANET, the first packet- + switched network; the SATNET, MATNET and Wideband Network, which + demonstrated the efficient utilization of shared-access satellite + channels for communications between geographically diverse sites; + + + +Gigabit Working Group [Page 2] + +RFC 1077 November 1988 + + + packet radio networks for mobile tactical environments; the Internet + and TCP/IP protocols for interconnection and interoperability between + heterogeneous networks and computer systems; the development of + electronic mail; and many advances in the areas of network security, + privacy, authentication and access control for distributed computing + environments. Recognizing DARPA's past accomplishments and its + desire to continue to take a leading role in addressing these issues, + this document provides a recommendation for research topics in + gigabit networking. It is meant to be an organized compendium of the + critical research issues to be addressed in developing the technology + base needed for such a high bandwidth ubiquitous network. + + + 1.2. Ongoing Activities + + + The OSTP report referred to above recommended a three-phase approach + to achieving the required high-bandwidth networking for the + scientific and research community. Some of this work is now well + underway. An ad-hoc committee, the Federal Research Internet + Coordinating Committee (FRICC) is coordinating the interconnection of + the current wide area networking systems in the government; notably + those of DARPA, Department of Energy (DoE), National Science + Foundation (NSF), National Aeronautics and Space Administration + (NASA), and the Department of Health and Human Services (HHS). In + accordance with Phases I and II of the OSTP report, this activity + will provide for an interconnected set of networks to support + research and other scholarly pursuits, and provide a basis for future + networking for this community. The networking is being upgraded + through shared increased bandwidth (current plans are to share a 45 + Mbit/s backbone) and coordinated interconnection with the rest of the + world. In particular, the FRICC is working with the European + networking community under the auspices of another ad-hoc group, the + Coordinating Committee for Intercontinental Research Networks + (CCIRN), to establish effective US-Europe networking. + + However, as the OSTP recommendations note, the required bandwidth for + the future is well beyond currently planned public, private, and + government networks. Achieving the required gigabit networking + capabilities will require a strong research activity. There is + considerable ongoing research in relevant areas that can be drawn + upon; particularly in the areas of high-bandwidth communication + links, high-speed computer switching, and high-bandwidth local area + networks. Appendix A provides some pointers to current research + efforts. + + + + + + +Gigabit Working Group [Page 3] + +RFC 1077 November 1988 + + + 1.3. Document Overview + + + This report outlines a set of research topics aimed at providing the + technology base for an interconnected set of networks that can + provide the required high-bandwidth capabilities discussed above. + The suggested research focus draws upon ongoing research and augments + it with basic and applied components. The major activities are the + development and demonstration of a Gigabit Backbone network (GB) [3], + the development and demonstration of an interconnected set of + networks with gigabit throughput and appropriate management + techniques, and the development and demonstration of the required + overall architecture that allows users to gain access to such high + bandwidth. Section 2 discusses functional and performance goals + along with the anticipated benefits to the ultimate users of such a + system. Section 3 provides the discussion of the critical research + issues needed to achieve these goals. It is organized into the major + areas of technology that need to be addressed: general architectural + issues, high-bandwidth switching, high-bandwidth host interfaces, + network management algorithms, and network services. The discussion + in some cases contains examples of ongoing relevant research or + potential approaches. These examples are intended to clarify the + issues and not to propose that particular approach. A discussion of + the relationship of the suggested research to other ongoing + activities and optimal methods for pursuing this research is provided + in Section 4. + + + 2. Functional and Performance Goals + + + In this section, we provide an assessment of the types of services a + GN (four or five orders of magnitude faster than the current + networks) should provide to its users. In instances where we felt + there would be a significant impact on performance, we have provided + an estimate of the amount of bandwidth needed and delay allowable to + provide these services. + + + 2.1. Networking Application Support + + + It is envisioned that the GN will be capable of supporting all of the + following types of networking applications. + + + + + + + +Gigabit Working Group [Page 4] + +RFC 1077 November 1988 + + + Currently Provided Packet Services + + It is important that the network provide the users with the + equivalent of services that are already available in packet- + switched networks, such as interactive data exchange, mail + service, file transfer, on-line access to remote computing + resources, etc., and allow them to expand to other more advanced + services to meet their needs as they become available. + + Multi-Media Mail + + This capability will allow users to take advantage of different + media types (e.g., graphics, images, voice, and video as well as + text and computer data) in the transfer of messages, thereby + increasing the effectiveness of message exchange. + + Multi-Media Conferencing + + Such conferencing requires the exchange of large amounts of + information in short periods of time. Hence the requirement for + high bandwidth at low delay. We estimate that the bandwidth would + range from 1.5 to 100 Mbit/s, with an end-to-end delay of no more + than a few hundred msec. + + Computer-Generated Real-time Graphics + + Visualizing computer results in the modern world of supercomputers + requires large amounts of real time graphics. This in turn will + require about 1.5 Mbit/s of bandwidth and no more than several + hundred msec. delay. + + High-Speed Transaction Processing + + One of the most important reasons for having an ultra-high-speed + network is to take advantage of supercomputing capability. There + are several scenarios in which this capability could be utilized. + For example, there could be instances where a non-supercomputer + may require a supercomputer to perform some processing and provide + some intermediate results that will be used to perform still + further processing, or the exchange may be between several + supercomputers operating in tandem and periodically exchanging + results, such as in a battle management, war gaming, or process + control applications. In such cases, extremely short response + times are necessary to accomplish as many as hundreds of + interactions in real time. This requires very high bandwidth, on + the order of 100 Mbit/s, and minimum delay, on the order of + hundreds of msec. + + + + +Gigabit Working Group [Page 5] + +RFC 1077 November 1988 + + + Wide-Area Distributed Data/Knowledge Base Management Systems + + Computer-stored data, information, and knowledge is distributed + around the country for a variety of reasons. The ability to + perform complex queries, updates, and report generation as though + many large databases are one system would be extremely powerful, + yet requires low-delay, high-bandwidth communication for + interactive use. The Corporation for National Research + Initiatives (NRI) has promoted the notion of a National Knowledge + base with these characteristics. In particular, an attractive + approach is to cache views at the user sites, or close by to allow + efficient repeated queries and multi-relation processing for + relations on different nodes. However, with caching, a processing + activity may incur a miss in the midst of a query or update, + causing it to be delayed by the time required to retrieve the + missing relation or portion of relation. To minimize the overhead + for cache directories, both at the server and client sites, the + unit of caching should be large---say a megabyte or more. In + addition, to maintain consistency at the caching client sites, + server sites need to multicast invalidations and/or updates. + Communication requirements are further increased by replication of + the data. The critical parameter is latency for cache misses and + consistency operations. Taking the distance between sites to be + on average 1/4 the diameter of the country, a one Gbit/s data rate + is required to reduce the transmission time to be roughly the same + as the propagation delay, namely around 8 milliseconds for this + size of unit. Note that this application is supporting far more + sophisticated queries and updates than normally associated with + transaction processing, thus requiring larger amount of data to be + transferred. + + + 2.2. Types of Traffic and Communications Modes + + + Different types of traffic may impose different constraints in terms + of throughput, delay, delay dispersion, reliability and sequenced + delivery. Table 1 summarizes some of the main characteristics of + several different types of traffic. + + + + + + + + + + + + +Gigabit Working Group [Page 6] + +RFC 1077 November 1988 + + + Table 1: Communication Traffic Requirements + + +------------------------+-------------+-------------+-------------+ + | | | | Error-free | + | Traffic | Delay | Throughput | Sequenced | + | Type | Requirement | Requirement | Delivery | + +------------------------+-------------+-------------+-------------+ + | Interactive Simulation | Low |Moderate-High| No | + +------------------------+-------------+-------------+-------------+ + | Network Monitoring | Moderate | Low | No | + +------------------------+-------------+-------------+-------------+ + | Virtual Terminal | Low | Low | Yes | + +------------------------+-------------+-------------+-------------+ + | Bulk Transfer | High | High | Yes | + +------------------------+-------------+-------------+-------------+ + | Message | Moderate | Moderate | Yes | + +------------------------+-------------+-------------+-------------+ + | Voice |Low, constant| Moderate | No | + +------------------------+-------------+-------------+-------------+ + | Video |Low, constant| High | No | + +------------------------+-------------+-------------+-------------+ + | Facsimile | Moderate | High | No | + +------------------------+-------------+-------------+-------------+ + | Image Transfer | Variable | High | No | + +------------------------+-------------+-------------+-------------+ + | Distributed Computing | Low | Variable | Yes | + +------------------------+-------------+-------------+-------------+ + | Network Control | Moderate | Low | Yes | + +------------------------+-------------+-------------+-------------+ + + The topology among users can be of three types: point-to-point (one- + to-one connectivity), multicast (one sender and multiple receivers), + and conferencing (multiple senders and multiple receivers). There + are three types of transfers that can take place among users. They + are connection-oriented network service, connectionless network + service, and stream or synchronous traffic. Connection and + connectionless services are asynchronous. A connection-oriented + service assumes and provides for relationships among the multiple + packets sent over the connection (e.g., to a common destination) + while connectionless service assumes each packet is a complete and + separate entity unto itself. For stream or synchronous service a + reservation scheme is used to set up and guarantee a constant and + steady amount of bandwidth between any two subscribers. + + + + + + + + +Gigabit Working Group [Page 7] + +RFC 1077 November 1988 + + + 2.3. Network Backbone + + + The GB needs to be of high bandwidth to support a large population of + users, and additionally to provide high-speed connectivity among + certain subscribers who may need such capability (e.g., between two + supercomputers). These users may access the GN from local area + networks (LANs) directly connected to the backbone or via high-speed + intermediate regional networks. The backbone must also minimize + end-to-end delay to support highly interactive high-speed + (supercomputer) activities. + + It is important that the LANs that will be connected to the GN be + permitted data rates independent of the data rates of the GB. LAN + speeds should be allowed to change without affecting the GB, and the + GB speeds should be allowed to change without affecting the LANs. In + this way, development of the technology for LANs and the GB can + proceed independently. + + Access rate requirements to the GB and the GN will vary depending on + user requirements and local environments. The users may require + access rates ranging from multi-kbit/s in the case of terminals or + personal computers connected by modems up to multi-Mbit/s and beyond + for powerful workstations up to the Gbit/s range for high-speed + computing and data resources. + + + 2.4. Directory Services + + + Directory services similar to those found in CCITT X.500/ISO DIS 9594 + need to be provided. These include mapping user names to electronic + mail addresses, distribution lists, support for authorization + checking, access control, and public key encryption schemes, + multimedia mail capabilities, and the ability to keep track of mobile + users (those who move from place to place and host computer to host + computer). The directory services may also list facilities available + to users via the network. Some examples are databases, + supercomputing or other special-purpose applications, and on-line + help or telephone hotlines. + + The services provided by X.500 may require some extension for GN. + For example, there is no provision for multilevel security, and the + approach taken to authentication must be studied to ensure that it + meets the requirements of GN and its user community. + + + + + + +Gigabit Working Group [Page 8] + +RFC 1077 November 1988 + + + 2.5. Network Management and Routing + + + The objective of network management is to ensure that the network + functions smoothly and efficiently, and consists of the following: + accounting, security, performance monitoring, fault isolation and + configuration control. + + Accounting ensures that users are properly billed for the services + that the network provides. Accounting enforces a tariff; a tariff + expresses a usage policy. The network need only keep track of those + items addressed by the tariff, such as allocated bandwidth, number of + packets sent, number of ports used, etc. Another type of accounting + may need to be supported by the network to support resource sharing, + namely accounting analogous to telephone "900" numbers. This + accounting performed by the network on behalf of resource providers + and consumers is a pragmatic solution to the problem of getting the + users and consumers into a financial relationship with each other + which has stymied previous attempts to achieve widespread use of + specialized resources. + + Performance monitoring is needed so that the managers can tell how + the network is performing and take the necessary actions to keep its + performance at a level that will provide users with satisfactory + service. Fault isolation using technical control mechanisms is + needed for network maintenance. Configuration management allows the + network to function efficiently. + + Several new types of routing will be required by GN. In addition to + true type-of-service, needed to support diverse distributed + applications, real-time applications, interactive applications, and + bulk data transfer, there will be need for traffic controls to + enforce various routing policies. For example, policy may dictate + that traffic from certain users, applications, or hosts may not be + permitted to traverse certain segments of the network. + Alternatively, traffic controls may be used to promote fairness; that + is, to make sure that busy link or network segment isn't dominated by + a particular source or destination. The ability of applications to + reserve network bandwidth in advance of its use, and the use of + strategies such as soft connections, will also require development of + new routing algorithms. + + + 2.6. Network Security Requirements + + + Security is a critical factor within the GN and one of those features + that are difficult to provide. It is envisioned that both + + + +Gigabit Working Group [Page 9] + +RFC 1077 November 1988 + + + unclassified and classified traffic will utilize the GN, so + protection mechanisms must be an integral part of the network access + strategy. Features such as authentication, integrity, + confidentiality, access control, and nonrepudiation are essential to + provide trusted and secure communication services for network users. + + A subscriber must have assurance that the person or system he is + exchanging information with is indeed who he says he is. + Authentication provides this assurance by verifying that the claimed + source of a query request, control command, response, etc., is the + actual source. Integrity assures that the subscriber's information + (such as requests, commands, data, responses, etc.) is not changed, + intentionally or unintentionally, while in transit or by replays of + earlier traffic. Unauthorized users (e.g., intruders or network + viruses) would be denied use of GN assets through access control + mechanisms which verify that the authenticated source is authorized + to receive the requested information or to initiate the specified + command. In addition, nonrepudiation services can be offered to + assure a third party that the transmitted information has not been + altered. And finally, confidentiality will ensure that the contents + of a message are not divulged to unauthorized individuals. + Subscribers can decide, based upon their own security needs and + particular activities, which of these services are necessary at a + given time. + + + 3. Critical Research Issues + + + In the section above, we discussed the goals of a research program in + gigabit networking; namely to provide the technology base for a + network that will allow gigabit service to be provided in an + effective way. In this section, we discuss those issues which we + feel are critical to address in a research program to achieve such + goals. + + + 3.1. General Architectural Issues + + + In the last generation of networks, it was assumed that bandwidth was + the scarce resource and the design of the switch was dictated by the + need to manage and allocate the bandwidth effectively. The most + basic change in the next generation network is that the speeds of the + trunks are rising faster than the speeds of the switching elements. + + This change in the balance of speeds has manifested itself in several + ways. In most current designs for local area networks, where + + + +Gigabit Working Group [Page 10] + +RFC 1077 November 1988 + + + bandwidth is not expensive, the design decision was to trade off + effective use of the bandwidth for a simplified switching technique. + In particular, networks such as Ethernet use broadcast as the normal + distribution method, which essentially eliminates the need for a + switching element. + + As we look at still higher speed networks, and in particular networks + in which the bandwidth is still the expensive component, we must + design new options for switching which will permit effective use of + bandwidth without the switch itself becoming the bottleneck. + + The central thrust of new research must thus be to explore new + network architectures that are consistent with these very different + speed assumptions. + + The development of computer communications has been tremendously + distorted by the characteristics of wide-area networking: normally + high cost, low speed, high error rate, large delay. The time is ripe + for a revolution in thinking, technology, and approaches, analogous + to the revolution caused by VCR technology over 8 and 16 mm. film + technology. + + Fiber optics is clearly the enabling technology for high-speed + transmission, in fact, so much so that there is an expectation that + the switching elements will now hold down the data rates. Both + conventional circuit switching and packet switching have significant + problems at higher data rates. For instance, circuit switching + requires increasing delays for FTDM synchronization to handle skew. + In the case of packet switching, traditional approaches require too + much processing per packet to handle the tremendous data flow. The + problem for both switching regimes is the "intelligence" in the + switches, which in turn requires electronics technology. + + Besides intelligence, another problem for wide-area networks is + storage, both because it ties us to electronics (for the foreseeable + future) and because it produces instabilities in a large-scale + system. (See, for instance, the work by Van Jacobson on self- + organizing phenomena for self-destruction in the Internet.) + Techniques are required to eliminate dependence on storage, such as + cut-through routing. + + Overall, high-speed WANs are the greatest agents of change, the + greatest catalyst both commercially and militarily, and the area ripe + for revolution. Judging by the attributes of current high-speed + network research prototypes, WANs of the future will be photonic, + multi-gigabit networks with enormous throughput, low delay, and low + error rate. + + + + +Gigabit Working Group [Page 11] + +RFC 1077 November 1988 + + + A zero-based budgeting approach is required to develop the new high- + speed internetwork architecture. That is, the time is ripe to + significantly rethink the Internet, building on experience with this + system. Issues of concern are manageability, understanding + evolvability and support for the new communication requirements, + including remote procedure call, real-time, security and fault- + tolerance. + + The GN must be able to deal with two sources of high-bandwidth + requirements. There will be some end devices (computers) connected + more or less directly to the GN because of their individual + requirements for high bandwidth (e.g., supercomputers needing to + drive remote high-bandwidth graphics devices). In addition, the + aggregate traffic due to large numbers of moderate rate users + (estimates are roughly up to a million potential users needing up to + 1 Mbit/s at any given time) results in a high-bandwidth requirement + in total on the GN. The statistics of such traffic are different and + there are different possible technical approaches for dealing with + them. Thus, an architectural approach for dealing with both must be + developed. + + Overall, the next-generation architecture has to be, first and + foremost, a management architecture. The directions in link speeds, + processor speeds and memory solve the performance problems for many + communication situations so well that manageability becomes the + predominant concern. (In fact, fast communication makes large + systems more prone to performance, reliability, and security + problems.) In many ways, the management system of the internetwork + is the ultimate distributed system. The solution to this tough + problem may well require the best talents from the communications, + operating systems and distributed systems communities, perhaps even + drawing on database and parallelism research. + + + 3.1.1. High-Speed Internet using High-Speed Networks + + + The GN will need to take advantage of a multitude of different and + heterogeneous networks, all of high speed. In addition to networks + based on the technology of the GB, there will be high-speed LANs. A + key issue in the development of the GN will be the development of a + strategy for interconnecting such networks to provide gigabit service + on an end to end basis. This will involve techniques for switching, + interfacing, and management (as discussed in the sections below) + coupled with an architecture that allows the GN to take full + advantage of the performance of the various high-speed networks. + + + + + +Gigabit Working Group [Page 12] + +RFC 1077 November 1988 + + + 3.1.2. Network Organization + + + The GN will need an architecture that supports the need to manage the + system as well as obtain high performance. We note that almost all + human-engineered systems are hierarchically structured from the + standpoint of control, monitoring, and information flow. A + hierarchical design may be the key to manageability in the next- + generation architecture. + + One approach is to use a general three-level structure, corresponding + to interadministrational, intraadministrational, and cluster + networks. The first level interconnects communication facilities of + truly separate administrations where there is significant separation + of security, accounting, and goals. The second level interconnects + subadministrations which exist for management convenience in large + organizations. For example, a research group within a university may + function as a subadministration. The cluster level consists of + networks configured to provides maximal performance among hosts which + are in frequent communication, such as a set of diskless workstations + and their common file server. These hosts are typically, but not + necessarily, geographically collocated. For example, two remote + networks may be tightly coupled by a fiber optic link that bridges + between the two physical networks, making them function as one. + + Research along these lines should study the interorganizational + characteristics of communications, such as those being investigated + by the IAB Task Force on Autonomous Networks. Based on current + results, we expect that such work would clearly demonstrate that + considerable communication takes place between particular + subadministrations in different administrations; communication + patterns are not strictly hierarchical. For example, there might be + intense direct communication between the experimental physics + departments of two independent universities, or between the computer + support group of one company and the operating system development + group of another. In addition, (sub)administrations may well also + require divisions into public information and private information. + + + 3.1.3. Fault-Tolerant System + + + Although the GN will be developed as part of an experimental research + program, it will also serve as part of the infrastructure for + researchers who are experimenting with applications which will use + such a network. The GN must have reasonably high availability to + support these research activities. In addition to facilitate the + transfer of this technology to future operational military and + + + +Gigabit Working Group [Page 13] + +RFC 1077 November 1988 + + + commercial users, it will need to be designed to become highly + reliable. This can be accomplished through diversity of transmission + paths, the development of fault-tolerant switches, use of a + distributed control structure with self-correcting algorithms, and + the protection of network control traffic. The architecture of a GN + should support and allow for all of these things. + + + 3.1.4. Functional Division of Control Between Network Elements + + + Current protocol architectures use the layered model of functional + decomposition first developed in the early work on ARPANET protocols. + The concept of layering has been a powerful concept which has allowed + dramatic variation in network technologies without requiring the + complete reimplementation of applications. The concept of layering + has had a first-order impact on the development of international + standards for data communication---witness the ISO "Reference Model + for Open Systems Interconnection." + + Unfortunately, however, the powerful concept of layering has been + paired, both in the DoD Internet work and the ISO work, with an + extremely weak concept of the interface between layers. The + interface designs are all organized around the idea of commands and + responses plus an error indicator. For example, the TCP service + interface provides the user with commands to set up or close a TCP + connection and commands to send and receive datagrams. The user may + well "know" whether they are using a file transfer service or a + character-at-a- time virtual terminal, but can't tell the TCP. The + underlying network may "know" that failures have reduced the path to + the user's destination to a single 9.6 kbit/s link, but it also can't + tell the TCP implementation. + + All of the information that an analyst would consider crucial in + diagnosing system performance is carefully hidden from adjacent + layers. One "solution" often discussed (but rarely implemented) is + to condense all of this information into a few bits of "Type of + Service" or "Quality of Service" request flowing in one direction + only---from application to network. It seems likely that this + approach cannot succeed, both because it applies too much compression + to the knowledge available and because it does not provide two-way + flow. + + We believe it to be likely that the next-generation network will + require a much richer interface between every pair of adjacent layers + if adequate performance is to be achieved. Research is needed into + the conceptual mechanisms, both indicators and controls, that can be + implemented at these interfaces and that, when used, will result in + + + +Gigabit Working Group [Page 14] + +RFC 1077 November 1988 + + + better performance. If real differences in performance can be + observed, then the implementors of every layer will have a strong + incentive to make use of the mechanisms. + + We can observe the first glimmers of this sort of coordination + between layers in current work. For example, in the ISO work there + are 5 classes of transport protocol which are supposed to provide a + range of possible matches between application needs and network + capabilities. Unfortunately, it is the case today that the class of + transport protocol is chosen statically, by the implementer, rather + than dynamically. The DARPA Wideband net offers a choice of stream + or datagram service, but typically a given host uses all one or all + the other---again, a static rather than a dynamic choice. The + research that we believe is needed, therefore, is not how to provide + alternatives, but how to provide them and choose among them on a + dynamic, real-time basis. + + + 3.1.5. Different Switch Technologies + + + One approach to high-performance networking is to design a technology + that is expected to work as a stand-alone demonstration, without + addressing the need for interconnection to other networks. Such an + experiment may be very valuable for rapid exploration of the design + space. However, our experience with the Internet project suggests + that a primary research goal should be the development of a network + architecture that permits the interconnection of a number of + different switching technologies. + + The Internet project was successful to a large extent because it + could incorporate a number of new and preexisting network + technologies: various local area networks, store and forward + switching networks, broadcast satellite nets, packet radio networks, + and so on. In this way, it decoupled the use of the protocols from a + particular technology base. In fact, the technology base evolved + rapidly, but the Internet protocols themselves provided a stability + that led to their success. + + The next-generation architecture must similarly deal with a diverse + and evolving technology base. We see "fast-packet" switching now + being developed (for example in B-ISDN); we see photonic switching + and wavelength division multiplexing as more advanced technologies. + We must divorce our architecture from dependence on any one of these. + + At the host interface, we must divorce the multiplexing of the medium + from the form of data that the host sees. Today the packet is used + both as multiplexing and interface element. In the future, the host + + + +Gigabit Working Group [Page 15] + +RFC 1077 November 1988 + + + may see the network as a message-passing system, or as memory. At + the same time, the network may use classic packets, wavelength + division, or space division switching. + + A number of basic functions must be rethought to provide an + architecture that is not dependent on the underlying switching model. + For example, our transport protocols assume that data will be lost in + units of a packet. If part of a packet is lost, we discard the whole + thing. And if several packets are systematically lost in sequence, + we may not recover effectively. There must be a host-level unit of + error recovery that is independent of the network. This sort of + abstraction must be applied to all the aspects of service + specification: error recovery, flow control, addressing, and so on. + + + 3.1.6. Network Operations, Monitoring, and Control + + + There is a hierarchy of progressively more effective and + sophisticated techniques for network management that applies + regardless of network bandwidth and application considerations: + + 1. Reactive problem management + + 2. Reactive resource management + + 3. Proactive problem management + + 4. Proactive resource management. + + Today's network management strategies are primarily reactive rather + than proactive: Problem management is initiated in response to user + complaints about service outages; resource allocation decisions are + made when users complain about deterioration of quality of service. + Today's network management systems are stuck at step 1 or perhaps + step 2 of the hierarchy. + + Future network management systems will provide proactive problem + management---problem diagnosis and restoral of service before users + become aware that there was a problem; and proactive resource + management---dynamic allocation of network bandwidth and switching + resources to ensure that an acceptable level of service is + continuously maintained. + + The GN management system should be expected to provide proactive + problem and resource management capabilities. It will have to do so + while contending with three important changes in the managed network + environment: + + + +Gigabit Working Group [Page 16] + +RFC 1077 November 1988 + + + 1. More complicated devices under management + + 2. More diverse types of devices + + 3. More variety of application protocols. + + Performance under these conditions will require that we seriously + re-think how a network management system handles the expected high + volumes of raw management-related data. It will become especially + important for the system to provide thresholding, filtering, and + alerting mechanisms that can save the human operator from drowning in + data, while still permitting access to details when diagnostic or + fault isolation modes are invoked. + + The presence of expert assistant capabilities for early fault + detection, diagnosis, and problem resolution will be mandatory. + These capabilities are highly desirable today, but they will be + essential to contend with the complexity and diversity of devices and + applications in the Gigabit Network. + + In addition to its role in dealing with complexity, automation + provides the only hope of controlling and reducing the high costs of + daily management and operation of a GN. + + Proactive resource management in GNs must be better understood and + practiced, initially as an effort requiring human intervention and + direction. Once this is achieved, it too must become automated to a + high degree in the GN. + + + 3.1.7. Naming and Addressing Strategies + + + Current networks, both voice (telephone) and data, use addressing + structures which closely tie the address to the physical location on + the network. That is, the address identifies a physical access + point, rather than the higher-level entity (computer, process, human) + attached to that access point. In future networks, this physical + aspect of addressing must be removed. + + Consider, for example, finding the desired party in the telephone + network of today. For a person not at his listed number, finding the + number of the correct telephone may require preliminary calls, in + which advice is given to the person placing the call. This works + well when a human is placing the call, since humans are well equipped + to cope with arbitrary conversations. But if a computer is placing + the call, the process of obtaining the correct address will have to + be incorporated in the architecture as a core service of the network. + + + +Gigabit Working Group [Page 17] + +RFC 1077 November 1988 + + + Since it is reasonable to expect mobile hosts, hosts that are + connected to multiple networks, and replicated hosts, the issue of + mapping to the physical address must be properly resolved. + + To permit the network to maintain the dynamic mapping to current + physical address, it is necessary that high-level entities have a + name (or logical address) that identifies them independently of + location. The name is maintained by the network, and mapped to the + current physical location as a core network service. For example, + mobile hosts, hosts that are connected to multiple networks, and + replicated hosts would have static names whose mapping to physical + addresses (many-to-one, in some cases) would change with time. + + Hosts are not the only entities whose physical location varies. + Users' electronic mail addresses change. Within distributed systems, + processes and files migrate from host to host. In a computing + environment where robustness and survivability are important, entire + applications may move about, or they may be redundant. + + The needed function must be considered in the context of the mobility + and address resolution rates if all addresses in a global data + network were of this sort. The distributed network directory + discussed elsewhere in this report should be designed to provide the + necessary flexibility, and responsiveness. The nature and + administration of names must also be considered. + + Names that are arbitrary or unwieldy would be barely better than the + addresses used now. The name space should be designed so that it can + easily be partitioned among the agencies that will assign names. The + structure of names should facilitate, rather than hinder, the mapping + function. For example, it would be hard to optimize the mapping + function if names were flat and unstructured. + + + 3.2. High-Speed Switching + + + The term "high-speed switching" refers to changing the switching at a + high rate, rather than switching high-speed links, because the latter + is not difficult at low speeds. (Consider, for example, manual + switching of fiber connections). The switching regime chosen for the + network determines various aspects of its performance, its charging + policies, and even its effective capabilities. As an example of the + latter, it is difficult to expect a circuit-switched network to + provide strong multicast support. + + A major area of debate lies in the choice between packet switching + and circuit switching. This is a key research issue for the GN, + + + +Gigabit Working Group [Page 18] + +RFC 1077 November 1988 + + + considering also the possibility of there being combinations of the + two approaches that are feasible. + + + 3.2.1. Unit of Management vs. Multiplexing + + + With very high data rates, either the unit of management and + switching must be larger or the speed of the processor elements for + management and switching must be faster. For example, at a gigabit, + a 576 byte packet takes roughly 5 microseconds to be received so a + packet switch must act extremely fast to avoid being the dominant + delay in packet times. Moreover, the storage time for the packet in + a conventional store and forward implementation also becomes a + significant component of the delay. Thus, for packet switching to + remain attractive in this environment, it appears necessary to + increase the size of packets (or switch on packet groups), do so- + called virtual cut-through and use high-speed routing techniques, + such as high-speed route caches and source routing. + + Alternatively, for circuit switching to be attractive, it must + provide very fast circuit setup and tear-down to support the bursty + nature of most computer communication. This problem is rendered + difficult (and perhaps impossible for certain traffic loads) because + the delay across the country is so large relative to the data rate. + That is, even with techniques such as so-called fast select, + bandwidth is reserved by the circuit along the path for almost twice + the propagation time before being used. + + With gigabit circuit switching, because it is not feasible to + physically switch channels, the low-level switching is likely doing + FTDM on micro-packets, as is currently done in telephony. Performing + FTDM at gigabit data rates is a challenging research problem if the + skew introduced by wide-area communication is to be handled with + reasonable overhead for spacing of this micro-packets. Given the + lead and resources of the telephone companies, this area of + investigation should, if pursued, be pursued cooperatively. + + + 3.2.2. Bandwidth Reservation Algorithms + + + Some applications, such as real-time video, require sustained high + data rate streams over a significant period of time, such as minutes + if not hours. Intuitively, it is appealing for such applications to + pre-allocate the bandwidth they require to minimize the switching + load on the network and guarantee that the required bandwidth is + available. Research is required to determine the merits of bandwidth + + + +Gigabit Working Group [Page 19] + +RFC 1077 November 1988 + + + reservation, particular in conjunction with the different switching + technologies. There is some concern to raise that bandwidth + reservation may require excessive intelligence in the network, + reducing the performance and reliability of the network. In + addition, bandwidth reservation opens a new option for denial of + service by an intruder or malicious user. Thus, investigations in + this area need to proceed in concert with work on switching + technologies and capabilities and security and reliability + requirements. + + + 3.2.3. Multicast Capabilities + + + It is now widely accepted that multicast should be provided as a + user-level service, as described in RFC 1054 for IP, for example. + However, further research is required to determine the best way to + support this facility at the network layer and lower. It is fairly + clear that the GN will be built from point-to-point fiber links that + do not provide multicast/broadcast for free. At the most + conservative extreme, one could provide no support and require that + each host or gateway simulate multicast by sending multiple, + individually addressed packets. However, there are significant + advantages to providing very low level multicast support (besides the + obvious performance advantages). For example, multicast routing in a + flooding form provides the most fault-tolerant, lowest-delay form of + delivery which, if reserved for very high priority messages, provides + a good emergency facility for high-stress network applications. + Multicast may also be useful as an approach to defeat traffic + analysis. + + Another key issue arises with the distinction between so-called open + group multicast and closed group multicast. In the former, any host + can multicast to the group, whereas in the latter, only members of + the group can multicast to it. The latter is easier to support and + adequate for conferencing, for example. However, for more client- + server structured applications, such as using file/database server, + computation servers, etc. as groups, open multicast is required. + Research is needed to address both forms of multicast. In addition, + security issues arise in controlling the membership of multicast + groups. This issue should be addressed in concert with work on + secure forms of routing in general. + + + + + + + + + +Gigabit Working Group [Page 20] + +RFC 1077 November 1988 + + + 3.2.4. Gateway Technologies + + + With the wide-area interconnection of local networks by the GN, + gateways are expected to become a significant performance bottleneck + unless significant advances are made in gateway performance. In + addition, many network management concerns suggest putting more + functionality (such as access control) in the gateways, further + increasing their load and the need for greater capacity. This would + then raise the issue of the trade-off between general-purpose + hardware and special-purpose hardware. + + On the general-purpose side, it may be feasible to use a general- + purpose multiprocessor based on high-end microprocessors (perhaps as + exotic as the GaAs MIPS) in conjunction with a high-speed block + transfer bus, as proposed as part of the FutureBus standard (which is + extendible to higher speeds than currently commercially planned) and + intelligent high-speed network adaptors. This would also allow the + direct use of hardware, operating systems, and software tools + developed as part of other DARPA programs, such as Strategic + Computing. It also appears to make this gateway software more + portable to commercial machines as they become available in this + performance range. + + The specialized hardware approach is based on the assumption that + general-purpose hardware, particularly the interconnection bus, + cannot be fast enough to support the level of performance required. + The expected emphasis is on various interconnection network + techniques. These approaches appear to require greater expense, less + commercial availability and more specialized software. They need to + be critically evaluated with respect to the general-purpose gateway + hardware approach, especially if the latter is using multiple buses + for fault-tolerance as well as capacity extension (in the absence of + failure). + + The same general-purpose vs. special-purpose contention is an issue + with operating system software. Conventionally, gateways run + specialized run-time executives that are designed specifically for + the gateway and gateway functions. However, the growing + sophistication of the gateway makes this approach less feasible. It + appears important to investigate the feasibility of using a standard + operating system foundation on the gateways that is known to provide + the required security and reliability properties (as well as real- + time performance properties). + + + + + + + +Gigabit Working Group [Page 21] + +RFC 1077 November 1988 + + + 3.2.5. VLSI and Optronics Implementations + + + It appears fairly clear that gigabit communication will use fiber + optics for at least the near future. Without major advances in + optronics to allow effectively for optical computers, communication + must cross the optical-electronic boundary two or more times. There + are significant cost, performance, reliability, and security benefits + for minimizing the number of such crossings. (As an example of a + security benefit, optics is not prone to electronic surveillance or + jamming while electronics clearly is, so replacing an optic- + electronic-optic node with a pure optic node eliminates that + vulnerability point.) + + The benefits of improved technology in optronics is so great that its + application here is purely another motivation for an already active + research area (that deserves strong continued support). Therefore, + we focus here in the issue of matching current (and near-term + expected) optronics capabilities with network requirements. + + The first and perhaps greatest area of opportunity is to achieve + totally (or largely) photonic switches in the network switching + nodes. That is, most packets would be switched without crossing the + optics-electronics boundary at all. For this to be feasible, the + switch must use very simple switching logic, require very little + storage and operate on packets of a significant size. The source- + routed packet switches with loopback on blockage of Blazenet + illustrate the type of techniques that appear required to achieve + this goal. + + Research is required to investigate the feasibility of optronic + implementation of switches. It appears highly likely that networks + will at some point in the future be totally photonically switched, + having the impact on networking comparable to the effect of + integrated circuits on processors and memories. + + A next level of focus is to achieve optical switching in the common + case in gateways. One model is a multiprocessor with an optical + interconnect. Packets associated with established paths through the + gateway are optically switched and processed through the + interconnect. Other packets are routed to the multiprocessor, + crossing into the electronics domain. Research is required to marry + the networking requirements and technology with optronics technology, + pushing the state of the art in both areas in the process. + + Given the long-term presence of the optic-electronic boundary, + improvements in technology in this area are also important. However, + it appears that there is already enormous commercial research + + + +Gigabit Working Group [Page 22] + +RFC 1077 November 1988 + + + activity in this area, particularly within the telephone companies. + This is another area in which collaborative investigation appears far + better than an new independent research effort. + + VLSI technology is an established technology with active research + support. The GN effort does not appear to require major new + initiatives in the VLSI area, yet one should be open to significant + novel opportunities not identified here. + + + 3.2.6. High-Speed Transfer Protocols + + + To achieve the desired speeds, it will be necessary to rethink the + form of protocols. + + 1. The simple idea of a stateless gateway must be replaced by a + more complex model in which the gateway understands the + desired function of the end point and applies suitable + optimizations to the flow. + + 2. If multiplexing is done in the time domain, the elements of + multiplexing are probably so small that no significant + processing can be performed on each individually. They must + be processed as an aggregate. This implies that the unit of + multiplexing is not the same as the unit of processing. + + 3. The interfaces between the structural layers of the + communication system must change from a simple + command/response style to a richer system which includes + indications and controls. + + 4. An approach must be developed that couples the memory + management in the host and the structure of the transmitted + data, to allow efficient transfers into host memory. + + The result of rethinking these problems will be a new style of + communications and protocols, in which there is a much higher degree + of shared responsibility among the components (hosts, switches, + gateways). This may have little resemblance to previous work either + in the DARPA or commercial communities. + + + 3.3. High-Speed Host Interfaces + + + As networks get faster, the most significant bottleneck will turn out + to be the packet processing overhead in the host. While this does + + + +Gigabit Working Group [Page 23] + +RFC 1077 November 1988 + + + not restrict the aggregate rates we can achieve over trunks, it + prevents delivery of high data rate flows to the host-based + applications, which will prevent the development of new applications + needing high bandwidth. The host bottleneck is thus a serious + impediment to networked use of supercomputers. + + To build a GN we need to create new ways for hosts and their high + bandwidth peripherals to connect to networks. We believe that + pursuing research in the ways to most effectively isolate host and + LAN development paths from the GN is the most productive way to + proceed. By decoupling the development paths, neither is restricted + by the momentary performance of capability bottlenecks of the other. + The best context in which to view this separation is with the notion + of a network front end (NFE). The NFE can take the electronic input + data at many data rates and transform it into gigabit light data + appropriately packetized to traverse the GN. The NFE can accept + inputs from many types of gateways, hosts, host peripherals, and LANS + and provide arbitration and path set-up facilities as needed. Most + importantly, the NFE can perform protocol arbitration to retain + upward compatibility with the existing Internet protocols while + enabling those sophisticated network input sources to execute GN + specific high-throughput protocols. Of course, this introduces the + need for research into high-speed NFEs to avoid the NFE becoming a + bottleneck. + + + 3.3.1. VLSI and Optronics Implementations + + + In a host interface, unless the host is optical (an unlikely prospect + in the near-term), the opportunities for optronic support are + limited. In fact, with a serial-to-parallel conversion on reception + stepping the clock rate down by a factor of 32 (assuming a 32-bit + data path on the host interface), optronic speeds are not required in + the immediate future. + + One exception may be for encryption. Current VLSI implementations of + standard encryption algorithms run in the 10 Mbit/s range. Optronic + implementation of these encryption techniques and encryption + techniques specifically oriented to, or taking advantage of, optronic + capabilities appears to be an area of some potential (and enormous + benefit if achieved). + + The potential of targeted VLSI research in this area appears limited + for similar reasons discussed above with its application in high- + speed switching. The major benefits will arise from work that is + well-motivated by other research (such as high-performance + parallelism) and by strong commercial interest. Again, we need to be + + + +Gigabit Working Group [Page 24] + +RFC 1077 November 1988 + + + open to imaginative opportunities not foreseen here while keeping + ourselves from being diverted into low-impact research without + further insights being put forward. + + + 3.3.2. High-Performance Transport Protocols + + + Current transport protocols exhibit some severe problems for maximal + performance, especially for using hardware support. For example, TCP + places the checksum in the packet header, forcing the packet to be + formed and read fully before transmission begins. ISO TP4 is even + worse, locating the checksum in a variable portion of the header at + an indeterminate offset, making hardware implementation extremely + difficult. + + The current Internet has thrived and grown due to the existence of + TCP implementations for a wide variety of classes of host computers. + These various TCP implementations achieve robust interoperability by + a "least common denominator" approach to features and options. Some + applications have arisen in the current Internet, and analogs can be + envisioned for the GN environment, which need qualities of service + not generally supported by the ubiquitous generic TCP, and therefore + special purpose transport protocols have been developed. Examples + include special purpose transport protocols such as UDP (user + datagram protocol), RDP (reliable datagram protocol), LDP + (loader/debugger protocol), NETBLT (high-speed block transfer + protocol), NVP (network voice protocol) and PVP (packet video + protocol). Efforts are also under way to develop a new generic + transport protocol VMTP (versatile message transaction protocol) + which will remedy some of deficiencies of TCP, without the need to + resort to special purpose protocols for some applications. Research + is needed in this area to understand how transport level protocols + should be constructed for a GN which provide adequate qualities of + service and ease of implementation. + + A new transport protocol of reasonable success can be expected to + last for ten years more. Therefore, a new protocol should not be + over optimized for current networks and must not ignore the + functional deficiencies of current protocols. These deficiencies are + essential to remedy before it is feasible to deploy even current + distributed systems technology for military and commercial + applications. + + Forward Error Correction (FEC) is a useful approach when the + bandwidth/delay ratio of the physical medium is high, as can be + expected in transcontinental photonic links. A degenerate form of + FEC is to simply transmit multiple copies of the data; this allows + + + +Gigabit Working Group [Page 25] + +RFC 1077 November 1988 + + + one to trade bandwidth for delay and reliability, without requiring + much intelligence. In fact, it is generally true that reliability, + bandwidth, and delay are interrelated and an improvement in one + generally comes at the expense of the others for a given technology. + Research is required to find appropriate operating points in networks + using transmission components which offer extremely high bandwidth + with very good bit-error-rate performance. + + + 3.3.3. Network Adaptors + + + With the promised speed of networks, the future network adaptor must + be viewed as a memory interconnect, tying the memory in one host to + another, at least if the data rate and the low latency made possible + by the network is to be realized at the host-to-host or process-to- + process level. The challenge is too great to be met by just + implementing protocols in custom VLSI. + + Research is required to investigate the impact of network + interconnection on a machine architecture and to define and evaluate + new network adaptor architectures. Of key importance is integration + of network adaptor into the operating system so that process-to- + process communications performance matches that offered by the + network. In particular, we conjecture that the transport level will + be implemented largely, if not entirely, in the network adaptor, + providing the host with reliable memory-to-memory transfer at memory + speeds with a minimum of interrupt processing bus overhead and packet + processing. + + Drawing an analogy to RISC technology again, maximal performance + requires a well-designed and coordinated protocol, software, and + hardware (network adaptor) design. Current standard protocols are + significantly flawed for hardware compatibility, suggesting a need + for considerable further research on high-performance protocol + design. + + + 3.3.4. Host Operating System Software + + + Conventionally, communication has been an add-on to an operating + system. With the GN, the network may well become the fastest + "peripheral" connected to most nodes. High-performance process-to- + process (or application to application) communication will not be + achieved until the operating system is well designed for fast access + to and from the network. For example, incorporating templates of the + network packet header directly in the process descriptor may allow a + + + +Gigabit Working Group [Page 26] + +RFC 1077 November 1988 + + + process to initiate communications with minimal overhead. Similarly, + memory mapping can be used to eliminate copies between data arriving + from the network and it being delivered to the applications. With a + GN, an extra copy forced by the operating system may easily double + the perceived transfer time for a packet between applications. + + Besides matching data transfer mechanisms, operating systems must be + well-matched in security design to that supported by the host + interface and network as well. Otherwise, all but the most trivial + additional security actions by the operating system in common case + communication can easily eliminate the performance benefits of the + GN. For example, if the host has to do further encryption or + decryption, the throughput is likely to be at least halved and the + latency doubled. + + Research effort is required to further refine operating systems for + the level of performance offered by the GN. This effort may well be + best realized with coupling existing efforts in distributed systems + with the GN activities, as opposed to starting new separate efforts. + + + 3.4. Advanced Network Management Algorithms + + + An important emphasis for research into network management should be + on decentralized approaches. The ratio of propagation delay across + the country to data rates in a GN appear to be too great to deal + effectively with resource management centrally when traffic load is + bursty and unstable (and if it is not, one might argue there is no + problem). In addition, important principles of fault containment and + minimal privilege for reliability and security suggest that a + centralized management approach is infeasible. In particular, + compromising the security of one portion of the network should not + compromise the security of the whole network. Similarly, a failure + or fault should affect at most a local region of the network. + + The challenge is clearly to provide decentralized management + techniques that lead to good global behavior in the normal case and + acceptable behavior in expected worst-case failures, traffic + variations and security intrusions. + + + 3.4.1. Control Flow vs. Data Flow + + + Network operational communications can be separated into flow of user + data and flow of management/control data. However, the user data + must contain some amount of control data. One question that needs to + + + +Gigabit Working Group [Page 27] + +RFC 1077 November 1988 + + + be explored in light of changes in communications and computing costs + and performance is the trade-off between these two flows. An example + of a potential approach is to use data units which contain predefined + path indicators. The switch can perform a simple table look-up which + maps the path indicator onto the preferred outbound link and + transmits the packet immediately. There is a path set-up packet + which fills in the appropriate tables. Path set-up occurs before the + first data packet flows and then, while data is flowing, to improve + the routes during the lifetime of the connection. This concept has + been discussed in the Internet engineering group under the name of + soft connections. + + We note that separating the data flow from the control flow in the GN + has security and reliability advantages as well. We could encrypt + most of the packet header to provide confidentiality within the GN + and to limit the ability of intruders to perform traffic analysis. + And, by separating the control flow, we can encrypt all the control + exchanges between switches and the host front ends thereby offering + confidentiality and integrity. No unauthorized entity will be able + to alter or examine the control traffic. By employing a path set-up + procedure, we can assure that the GN NFE-to-NFE path is functioning + and also include user-specific requirements in the route. For + example, we could request a certain bandwidth allocation and simplify + the job of the switches in handling flow control. We could also set + up backup paths in case the output link will be busy for so many + microseconds that the packet cannot be stored until the link is + freed. + + + 3.4.2. Resource Management Algorithms + + + Most current networks deliver one quality of service. X.25 networks + deliver a reliable byte-stream. Most LANs deliver a best-effort + unreliable service. There are few networks today that can support + multiple types of service, and allocate their resources among them. + Indeed, for many networks, such as best-effort unreliable service, + there is little management of resources at all. The next generation + of network will require a much more controlled allocation of + resources. + + There will be a much wider range of desired types of service, with + current services such as remote procedure call mixing with new + services such as video streams. Unless these are separately + recognized and controlled, there is little reason to believe that + effective service can be delivered unless the network is very lightly + loaded. + + + + +Gigabit Working Group [Page 28] + +RFC 1077 November 1988 + + + In order to support multiple types of service, two things must + happen, both a change from current practice. First, the application + must describe to the network what type of service is required. + Second, the network must use this information to make resource + allocation decisions. Both of these practices present difficulties. + + Past experience suggests that application code is not prepared to + know or specify what service it needs. By custom, operating systems + provide a virtual world, and the applications in this world are + unaware of the relation between this and the reality of time and + space. Resource requests must be in real terms. Allocation of + resources in the network is difficult, because it requires that + decisions be made in the network, but as network packet throughput + increases, there is less time for decisions. + + The resolution of this latter conflict is to observe that decisions + must be made on larger units than the unit of multiplexing such as + the packet. This in turn implies that packets must be visible to the + network as being part of a sequence, as opposed to the pure datagram + model previously exploited. As suggested earlier in this report, + research is required to support this more complex form of switch + without compromising robustness. + + To permit the application to specify the service it needs, it will be + necessary to propose some abstraction of service class. By clever + design of this abstraction, it should be possible to allow the + application to describe its needs effectively. For example, an + application such as file transfer or mail has two modes of operation; + bulk data transfer and remote procedure call. The application may + not be able to predict when it will be in which mode, but if it just + describes both of them, the system may be able to adapt by observing + its current operation. + + Experimentation needs to be done to determine a suitable service + specification interface. This experimentation could be done in the + context of the current protocols, and could thus be undertaken at + once. + + + 3.4.3. Adaptive Protocols + + + Network operating conditions can vary quickly and over a wide range. + This is true of the current Internet, and is likely to affect the GN + too. Protocols that can adapt to changing circumstances would + provide more even and robust service than is currently possible. For + example, when error rates increased, a protocol implementation might + decide to use smaller packets, thus reducing the burden caused by + + + +Gigabit Working Group [Page 29] + +RFC 1077 November 1988 + + + retransmissions. + + The environment in which a protocol operates can be described in + terms of the service it is getting from the next lower layer. A + protocol implementation can adapt to changes in that service by + tuning its internal mechanisms (time-outs, retransmission strategies, + etc.). Therefore, to design adaptive protocols, we must understand + the interaction between protocol layers and the mechanisms used + within them. There has been some work done in this area. For + example, the SATNET measurement task force has looked at the + interactions between the protocol used by the SIMP, IP, and TCP. + What is needed is a more complete characterization of the + interactions at various layer boundaries, and the development of + appropriate protocol designs and mechanisms to provide for necessary + adaptations and renegotiations. + + + 3.4.4. Error Recovery Mechanisms + + + Being large and complex, GNs will experience a variety of faults such + as link or nodal failure, excessive buffer overflow due to faulty + flow and congestion control, and partial failure of switching fabric. + These failures, which also exist in today's networks, will have a + stronger effect in GNs where a large amount of data will be "stored" + in transit and, to expedite the switching, nodes will apply only + minimal processing to the packets traversing them. In source + routing, for example, a link failure may cause the loss of all + packets sent until the source is notified about the change in + topology. The longer is the delay in recovering from failures, the + higher is the degradation in performance observed by the users. + + To minimize the effects of failures, GNs will need to employ error + recovery mechanisms whereby the network detects failures and error + conditions, reconfigures itself to adapt to the new network state, + and notifies peripheral devices of the new configuration. Such + protocols, which have to be developed, will respond quickly, will be + decentralized or distributed to minimize the possibility of fatal + failures, and will complement, rather than replicate, the error + correction mechanisms of the end-to-end protocols, and the two must + operate in coordinated manner. To this end, the peripheral devices + will have to be knowledgeable about the intranet recovery mechanisms + and interact continuously with them to minimize the effect on the + connections they manage. + + + + + + + +Gigabit Working Group [Page 30] + +RFC 1077 November 1988 + + + 3.4.5. Flow Control + + + As networks become faster, two related problems arise. First, + existing flow control mechanisms such as windows do not work well, + because the window must be opened to such an extent to achieve + desired bandwidth that effective flow control cannot be achieved. + Second, especially for long-haul networks, the larger number of bits + in transit at one time becomes so large that most computer messages + will fit into one window. This means that traditional congestion + control schemes will cease to work well. + + What is needed is a combination of two approaches, both new. First, + for messages that are small (most messages generated by computers + today will be small, since they will fit into one round-trip time of + future networks), open-loop controls on flow and congestion are + needed. For longer messages (voice or video streams, for example), + some explicit resource commitment will be required. + + + 3.4.6. Latency Control and Real-Time Operations + + + Currently, there are several distinct approaches to latency control. + First, there are some networks which are physically short, more like + multiprocessor buses. Applications in these networks are built + assuming that delays will be short. + + Second, there are networks where the physical length is not + constrained by the design and may differ by orders of magnitude, + depending on the scope of the network. Most general purpose networks + fall in this category. In these networks, one of two things happens. + Either the application takes special steps to deal with variable + latency, such as echo suppression in voice networks, or these + applications are not supported. + + For most applications today, the latency in the network is not an + obvious issue so long as the network is not overloaded (which leads + to losses and long queues), because the protocol overhead masks the + variation in the network latency. This balance will change. The + latency due to the speed of light will obviously remain the same, but + the overhead will drop (of necessity if we are to achieve high + performance) which will leave speed of light and queueing as the most + critical sources of delay. + + This conclusion implies that if queueing delay can be controlled, it + will be possible to build networks with stable and controlled + latency. If applications exist that require this class of service, + + + +Gigabit Working Group [Page 31] + +RFC 1077 November 1988 + + + it can be supported. Either the network must be underloaded, so that + queues do not develop at all, or a specific class of service must be + supported in which resources are allocated to stabilize the delay. + + If this service is provided, it will still leave the application with + delays that can vary by several orders of magnitude, depending on the + physical size of the network. Research at the application level will + be required to see how applications can be designed to cope with this + variation. + + + 3.4.7. High-Speed Internetworking and Administrational Domains + + + Internetworking recognized that the value of communication services + increases significantly with wider interconnection but ignored + management and the role of administrations. As a consequence we see + that: + + 1. The Internet is more or less unmanageable, as evidenced by + performance, reliability, and security problems. + + 2. The Internet is being stressed by administrators that are + building networks to match their organization rather than the + geography. An example is a set of Ethernets at different + company locations operating as a single Internet network but + geographically dispersed and connected by satellite or leased + lines. + + The next generation of internetworking must focus on administration + and management. Internetworking must support cohesion within an + administration and a healthy separation between administrations. To + illustrate by analogy, the American and Soviet embassies in Mexico + City are geographically closer to each other than to their respective + home countries but further in administrational distance, including + security, accounting, etc. The emerging revolution in WANs makes + this issue that much more critical. The amount of communication to + exchange the state of systems is bound to increase enormously. The + potential cost of failures and security violations is frightening. + + A promising approach appears to be high-level gateways that guard + between administrations and require negotiations to set up access + paths between administrations. These paths are set up, and labeled + with agreements on authorization, security, accounting, and possible + resource limits. These administrative virtual circuits provide + transparency to the physical and geographical interconnection, but + need not support more than datagram packet delivery. One view is + that of communication contracts with high-level gateways acting as + + + +Gigabit Working Group [Page 32] + +RFC 1077 November 1988 + + + contract monitors at each end. The key is the focus on controlled + interadministrational connectivity, not the conventional protocol + concerns. + + Focus is required on developing an (inter)network management + architecture and the specifics of high-level gateways. The + structures of such gateways will have to take advantage of advances + in multi-processor architectures to handle the processing load. + Moreover, a key issue is being able to optimize communication between + administrations once the contract is in place, but without losing + control. Related is the issue of allowing high-speed interconnection + within a single administration, although geographical dispersed. + Another issue is fault-tolerance. High-level gateways contain state + information whose loss typically disrupts communication. How does + one minimize this problem? + + A key goal of these administrational gateways has to be failure + containment: How to protect against external (to administration) + problems and how to prevent local problems imposing liability on + others. A particular area of concern is the self-organizing problems + of large-scale systems, observed by Van Jacobson in the Internet. + Gateways must serve to damp out oscillations and control wide load + swings. Rate control appears to be a key area to investigate as a + basis for buffer management and for congestion control, as well as to + control offered load. + + Given the speed of new networks, and the sophistication of the + gateways suggested above, another key area to investigate is the + provision of high-speed network interface adaptors. + + + 3.4.8. Policy-Based Algorithms + + + Networks of today generally select routes based on minimizing some + measure such as delay. However, in the real world, route selection + will commonly be constrained at the global level by policy issues, + such as access rights to resources and accounting and billing for + usage. + + It is difficult for connectionless protocols such as Internet to deal + with policy controls, because a lack of state in the gateway implies + that a separate policy decision must be made for each packet in + isolation. As networks get faster, the cost of this processing will + be intolerable. One possible approach, discussed above, is to move + to a more sophisticated model in which there is knowledge in the + gateways of the ongoing flows. Alternatively, it may be possible to + design gateways that simply cache recent policy evaluations and apply + + + +Gigabit Working Group [Page 33] + +RFC 1077 November 1988 + + + them to successive packets. + + Routing based on policy is particularly difficult because a route + must be globally consistent to be useful; otherwise it may loop. + This implies that the every policy decision must be propagated + globally. Since there can be expected to be a large number of + policies, this global passing of information might easily lead to an + information explosion. + + There are at least two solutions. One is to restrict the possible + classes of policy. Another is to use some form of source route, so + that the route consistent with some set of policies is computed at + one point only, and then attached to the packet. Both of these + approaches have problems. A two-pronged research program is needed, + in which mechanisms are proposed, and at the same time the needed + policies are defined. + + The same trade-off can be seen for accounting and billing. A single + accounting metric, such as "bytes times distance", could be proposed. + This might be somewhat simple to implement, but would not permit the + definition of individual billing policies, as is now done in the + parts of the telephone system. The current connectionless transport + architectures such as TCP/IP or the connectionless ISO configuration + using TP4 do not have good tools for accounting for traffic, or for + restricting traffic from certain resources. Building these tools is + difficult in a connectionless environment, because an accounting or + control facility must deal with each packet in isolation, which + implies a significant processing burden as part of packet forwarding. + This burden is an increasing problem as switches are expected to + operate faster. + + The lack of these tools is proving a significant problem for network + design. Not only are accounting and control needed to support + management requirements, they are needed as a building block to + support enforcement of such things as multiple qualities of service, + as discussed above. + + Network accounting is generally considered to be simply a step that + leads to billing, and thus is often evaluated in terms of how simple + or difficult it will be to implement. Yet an accounting and billing + procedure is a mechanism for implementing a policy considered to be + desirable for reasons beyond the scope of accounting per se. For + example, a policy might be established either to encourage or + discourage network use, while fully recovering operational cost. A + policy of encouraging use could be implemented by a relatively high + monthly attachment charge and a relatively low per-packet charge. A + policy of discouraging use could be implemented by a low monthly + charge and a high per-packet charge. + + + +Gigabit Working Group [Page 34] + +RFC 1077 November 1988 + + + Network administrators have a relatively small number of variables + with which to implement policy objectives. Nevertheless, these + variables can be combined in a number of innovative ways. Some of + the possibilities include: + + 1. Classes of users (e.g., large or small institutions, for- + profit or non-profit). + + 2. Classes of service. + + 3. Time varying (e.g., peak and off-peak). + + 4. Volume (e.g., volume discounts, or volume surcharges). + + 5. Access charges (e.g., per port, or port * [bandwidth of + port]). + + 6. Distance (e.g., circuit-miles, airline miles, number of hops). + + Generally, an accounting procedure can be developed to support + voluntary user cooperation with almost any single policy objective. + Difficulties most often arise when there are multiple competing + policy objectives, or when there is no clear policy at all. + + Another aspect of accounting and billing procedures which must be + carefully considered is the cost of accumulating and processing the + data on which billing is based. Of particular concern is collection + of detailed data on a per-packet basis. As network circuit data + rates increase, the number of instructions which must be executed on + a per-packet basis can become the limiting factor in system + throughput. Thus, it may be appropriate to prefer accounting and + billing policies and procedures which minimize the difficulty of + collecting data, even if this approach requires a compromise of other + objectives. Similarly, node memory required for data collection and + any network bandwidth required for transmission of the data to + administrative headquarters are factors which must be traded off + against the need to process user packets. + + + 3.4.9. Priority and Preemption + + + The GN should support multiple levels of priority for traffic and the + preemption of network resources for higher priority use. Network + control traffic should be given the highest priority to ensure that + it is able to pass through the network unimpeded by congestion caused + by user-level traffic. There may be additional military uses for + multiple levels of priority which correspond to rank or level of + + + +Gigabit Working Group [Page 35] + +RFC 1077 November 1988 + + + importance of a user or the mission criticality of some particular + data. + + The use of and existence of priority levels may be different for + different types of traffic. For example, datagram traffic may not + have multiple priority levels. Because the network's transmission + speed is so high and traffic bursts may be short, it may not make + sense to do any processing in the switches to deal with different + priority levels. Priority will be more important for flow- (or + soft-connection-) oriented data or hard connections in terms of + permitting higher priority connections to be set up ahead of lower + priority connections. Preemption will permit requests for high + priority connections to reclaim network resources currently in use by + lower priority traffic. + + Networks such as the Wideband Satellite Network, which supports + datagram and stream traffic, implement four priority levels for + traffic with the highest reserved for network control functions and + the other three for user traffic. The Wideband Network supports + preemption of lower priority stream allocations by higher priority + requests. An important component of the use of priority and + preemption is the ability to notify users when requests for service + have been denied, or allocations have been modified or disrupted. + Such mechanisms have been implemented in the Wideband Network for + streams and dynamic multicast groups. + + Priority and preemption mechanisms for a GN will have to be + implemented in an extremely simple way so that they can take effect + very quickly. It is likely that they will have to built into the + hardware of the switch fabric. + + + 3.5. User and Network Services + + + As discussed in Section 2 above, there will need to be certain + services provided as part of the network operation to the users + (people) themselves and to the machines that connect to the network. + These services, which include such capabilities as white and yellow + pages (allowing users to determine what the appropriate network + identification is for other users and for network-available computing + resources) and distributed fault identification and isolation, are + needed in current networks and will continue to be required in the + networks of the future. The speed of the GN will serve to accentuate + this requirement, but at the same time will allow for new + architectures to be put in place for such services. For example, + Ethernet speeds in the local environment have allowed for more usable + services to be provided. + + + +Gigabit Working Group [Page 36] + +RFC 1077 November 1988 + + + 3.5.1. Impact of High Bandwidth + + + One issue that will need to be addressed is the impact on the user of + such high-bandwidth capabilities. Users are already becoming + saturated by information in the modern information-rich environment. + (Many of us receive more than 50 electronic mail messages each day, + each requiring some degree of human attention.) Methods will be + needed to allow users to cope with this ever-expanding access to + data, or we will run the risk of users turning back to the relative + peace and quiet of the isolated office. + + + 3.5.2. Distributed Network Directory + + + A distributed network directory can support the user-level directory + services and the lower-level name-to-address mapping services + described elsewhere in this report. It can also support distributed + systems and network management facilities by storing additional + information about named objects. For example, the network directory + might store node configurations or security levels. + + Distributing the directory eases and decentralizes the administrative + burdens and provides a more robust and survivable implementation. + + One approach toward implementing a distributed network directory + would be to base it upon the CCITT X.500/ISO DIS 9594 standard. This + avoids starting from ground zero and has the advantage of + facilitating interoperability with other communications networks. + However, research and development will be required even if this path + is chosen. + + One area in which research and development are required is in the + services supplied by the distributed network directory. The X.500 + standard is very general and powerful, but so far specific provisions + have been made only for storing information about network users and + applications. As mentioned elsewhere, multilevel security is not + addressed by X.500, and the approach taken toward authentication must + be carefully considered in view of DoD requirements. Also, X.500 + assumes that administration of the directory will be done locally and + without the need for standardization; this may not be true of GN or + the larger national research network. + + The model and algorithms used by a distributed network directory + constitute a second area of research. The model specified by X.500 + must be extended into a framework that provides the necessary + flexibility in terms of services, responsiveness, data management + + + +Gigabit Working Group [Page 37] + +RFC 1077 November 1988 + + + policies, and protocol layer utilization. Furthermore, the internal + algorithms and mechanisms of X.500 must be extended in a number of + areas; for example, to support redundancy of the X.500 database, + internal consistency checking, fuller sharing of information about + the distribution of data, and defined access-control mechanisms. + + + 4. Avenues of Approach + + + Ongoing research and commercial activities provide an opportunity for + more rapidly attacking some of the above research issues. At the + same time, there needs to be attention paid to the overall technical + approach used to allow multiple potential solutions to be explored + and allow issues to be attacked in parallel. + + + 4.1. Small Prototype vs. Nationwide Network + + + The central question is how far to jump, and how far can the current + approaches get. That is, how far will connectionless network service + get us, how far will packet switching get us, and how far do we want + to go. If our goal is a Gbit/s net, then that is what we should + build. Building a 100 Mbit/s network to achieve a GN is analogous to + climbing a tree to get to the moon. It may get you closer, but it + will never get you there. + + There are currently some network designs which can serve as the basis + for a GN prototype. The next step is some work by experts in + photonics and possibly high-speed electronics to explore ease of + implementation. Developing a prototype 3-5 node network at a Gbit/s + data rate is realistic at this point and would demonstrate wide-area + (40 km or more) Gbit/s networking. + + DARPA should consider installing a Gbit/s cross-country set of + connected links analogous to the NSF backbone in 2 years. A Gbit/s + link between the east and west coasts would open up a whole new + generation of (C3I), distributed computing, and parallel computing + research possibilities and would reestablish DARPA as the premier + network research funding agency in the country. This will require + getting "dark" fiber from one or more of the common carriers and some + collaboration with these organizations on repeaters, etc. With this + collaboration, the time to a commercial network in the Gbit/s range + would be substantially reduced, and the resulting nationwide GN would + give the United States an enormous technical and economic advantage + over countries without it. + + + + +Gigabit Working Group [Page 38] + +RFC 1077 November 1988 + + + Demonstrating a high-bandwidth WAN is not enough, however. As one + can see from the many research issues identified above, it will be + necessary to pursue via study and experiment the issues involved in + interconnecting high-bandwidth networks into a high-bandwidth + internet. These experiments can be done through use of a new + generation of internet, even if it requires starting at lower speeds + (e.g., T1 through 100 Mbit/s). Appropriate care must be given, + however, to assure that the capabilities that are demonstrated are + applicable to the higher bandwidths (Gbit/s) as they emerge. + + + 4.2. Need for Parallel Efforts/Approaches + + + Parallel efforts will therefore be required for two major reasons. + First is the need to pursue alternative approaches (e.g., different + strategies for high-bandwidth switching, different addressing + techniques, etc). This is the case for most research programs, but + it is made more difficult here by the costs of prototyping. Thus, it + is necessary that appropriate review take place in the decisions as + to which efforts are supported through prototyping. + + In addition, it will be necessary to pursue the different aspects of + the program in parallel. It will not be possible to wait until the + high-bandwidth network is available before starting on prototyping + the high-bandwidth internet. Thus, a phased and evolutionary + approach will be needed. + + + 4.3. Collaboration with Common Carriers + + + Computer communication networks in the United States today + practically ignore the STN (the Switched Telephone Network), except + for buying raw bandwidth through it. However, advances in network + performance are based on improvements in the underlying communication + media, including satellite communication, fiber optics, and photonic + switching. + + In the past we used "their" transmission under "our" switching. An + alternative approach is to utilize the common-carrier switching + capabilities as an integral part of the networking architecture. We + must take an objective scientific and economic look and reevaluate + this question. + + Another place for cooperation with the common carriers is in the area + of network addressing. Their addressing scheme ("numbering plan") + has a few advantages such as proven service to 300 million users [4]. + + + +Gigabit Working Group [Page 39] + +RFC 1077 November 1988 + + + On the other hand, the common carriers have far fewer administrative + domains (area codes) than the current plethora of locally + administered local area networks in the internet system. + + It is likely that future networks will eventually be managed and + operated by commercial communications providers. A way to maximize + technology transfer from the research discussed here to the + marketplace is to involve the potential carriers from the start. + However, it is not clear that the goals of commercial communications + providers, who have typically been most interested in meeting the + needs of 90+ percent of the user base, will be compatible with the + goals of the research described here. Thus, while we recommend that + the research program involve an appropriate amalgam of academia and + industry, paying particular attention to involvement of the potential + system developers and operators, we also caution that the specific + and unique goals of the DARPA program must be retained. + + + 4.4. Technology Transfer + + + As we said above, it is our belief that future networks will + ultimately be managed and operated by commercial communications + providers. (Note that this may not be the common carriers as we know + them today, but may be value-added networks using common carrier + facilities.) The way to assure technology transfer, in our belief, is + to involve the potential system developers from the start. We + therefore believe that the research program would benefit from an + appropriate amalgam of university and industry, with provision for + close involvement of the potential system developers and operators. + + + 4.5. Standards + + + The Internet program was a tremendous success in influencing national + and international standards. While there were changes to the + protocols, the underlying technology and approaches used by CCITT and + ISO in the standardization of packet-switched networks clearly had + its roots in the DARPA internet. Nevertheless, this has had some + negative impact on the research program, as the evolution of the + standards led to pressure to adopt them in the research environment. + + Thus, it appears that there is a "catch-22" here. It is desirable + for the technology base developed in the research program to have + maximal impact on the standards activities. This is expedited by + doing the research in the context of the standards environment. + However, standards by their very nature will always lag behind the + + + +Gigabit Working Group [Page 40] + +RFC 1077 November 1988 + + + research environment. + + The only reasonable approach, therefore, appears to be an occasional + "checkpointing" of the research environment, where the required + conversions take place to allow a new plateau of standards to be used + for future evolution and research. A good example is conducting + future research in mail using X.400 and X.500 where possible. + + + 5. Conclusions + + + We hope that this document has provided a useful compendium of those + research issues critical to achieving the FCCSET phase III + recommendations. These problems interact in a complex way. If the + only goal of a new network architecture was high speed, reasonable + solutions would not be difficult to propose. But if one must achieve + higher speeds while supporting multiple services, and at the same + time support the establishment of these services across + administrative boundaries, so that policy concerns (e.g., access + control) must be enforced, the interactions become complex. + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +Gigabit Working Group [Page 41] + +RFC 1077 November 1988 + + + APPENDIX + +A. Current R and D Activities + + In this appendix, we provide pointers to some ongoing activities in + the research and development community of which the group was aware + relevant to the goal of achieving the GN. In some cases, a short + abstract is provided of the research. Neither the order of the + listing (which is random) nor the amount of detail provided is meant + to indicate in any way the significance of the activity. We hope + that this set of pointers will be useful to anyone who chooses to + pursue the research issues discussed in this report. + + 1. Grumman (at Bethpage) is working on a three-year DARPA + contract, started in January 1988 to develop a 1.6 Gbit/s LAN, + for use on a plane or ship, or as a "building block". It is + really raw transport capacity running on two fibers in a + token-ring like mode. First milestone (after one year?) is to + be a 100 Mbit/s demonstration. + + 2. BBN Laboratories, as part of its current three-year DARPA + Network-Oriented Systems contract, has proposed design + concepts for a 10-100 Gbit/s wide area network. Work under + this effort will include wavelength division multiplexing, + photonic switching, self-routing packets, and protocol design. + + 3. Cheriton (Stanford) research on Blazenet, a high-bandwidth + network using photonic switching. + + 4. Acampora (Bell Labs) research on the use of wavelength + division multiplexing for building a shared optical network. + + 5. Yeh is reserching a VLSI approach to building high-bandwidth + parallel processing packet switch. + + 6. Bell Labs is working on a Metropolitan Area Network called + "Manhattan Street Net." This work, under Dr. Maxemchuck, is + similar to Blazenet. It is in the prototype stage for a small + number of street intersections; ultimately it is meant to be + city-wide. Like Blazenet, is uses photonic switching 2 x 2 + lithium niobate block switches. + + 7. Ultra Network Technologies is a Silicon Valley company which + has a (prototype) Gbit/s fiber link which connects backplanes. + This is based on the ISO-TP4 transport protocol. + + 8. Jonathan Turner, Washington University, is working on a + Batcher-Banyan Multicast Net, based on the "SONET" concept, + + + +Gigabit Working Group [Page 42] + +RFC 1077 November 1988 + + + which provides 150 Mbit/s per pipe. + + 9. David Sincowskie, Bellcore, is working with Batcher-Banyan + design and has working 32x32 switches. + + 10. Stratacom has a commercial product which is really a T1 voice + switch implemented internally by a packet switch, where the + packet is 192 bits (T1 frame). This switch can pass 10,000 + packets per second. + + 11. Stanford NAB provides 30-50 Mbit/s throughput on 100 Mbit/s + connection using Versatile Message Transaction Protocol (VMTP) + [see RFC 1045] + + 12. The December issue of IEEE Journal on Selected Areas in + Communications, provides much detail concerning interconnects. + + 13. Ultranet Technology has a 480 Mbit/s connection using modified + ISO TP4. + + 14. At MIT, Dave Clark has an architecture proposal of interest. + + 15. At CMU, the work of Eric Cooper is relevant. + + 16. At Protocol Engines, Inc., Greg Chesson is working on an XTP- + based system. + + 17. Larry Landweber at Wisconsin University is doing relevant + work. + + 18. Honeywell is doing relevant work for NASA. + + 19. Kung at CMU is working on a system called "Nectar" based on a + STARLAN on fiber connecting dissimilar processors. + + 20. Burroughs (now Unisys) has some relevant work within the IEEE + 802.6 committee. + + 21. Bellcore work in "Switched Multimedia Datanet Service" (SMDS) + is relevant (see paper supplied by Dave Clark). + + 22. FDDI-2, a scheme for making TDMA channel allocations at 200 + Mbit/s. + + 23. NRI, Kahn-Farber Proposal to NSF, is a paper design for high- + bandwidth network. + + 24. Barry Goldstein work, IBM-Yorktown. + + + +Gigabit Working Group [Page 43] + +RFC 1077 November 1988 + + + 25. Bell Labs S-Net, 1280 Mbit/s prototype. + + 26. Fiber-LAN owned by Bell South and SECOR, a pre-prototype 575 + Mbit/s Metro Area Net. + + 27. Bellcore chip implementation of FASTNET (1.2 Gbit/s). + + 28. Scientific Computer Systems, San Diego, 1.4 Gbit/s prototype. + + 29. BBN Monarch Switch, Space Division pre-prototype, chips being + fabricated, 64 Mbit/s per path. + + 30. Proteon, 80 Mbit/s token ring. + + 31. Toronto University, 150 Mbit/s "tree"--- really a LAN. + + 32. NSC Hyperchannel II, reputedly available at 250 Mbit/s. + + 33. Tobagi at Stanford working on EXPRESSNET; not commercially + available. + + 34. Columbia MAGNET-- 150 Mbit/s. + + 35. Versatile Message Transaction Protocol (VMTP). + + 36. ST integrated with IP. + + 37. XTP (Chesson). + + 38. Stanford Transport Gateway. + + 39. X.25/X.75. + + 40. Work of the Internet Activities Board. + + + + + + + + + + + + + + + + + +Gigabit Working Group [Page 44] + +RFC 1077 November 1988 + + +B. Gigabit Working Group Members + +Member Affiliation + +Gordon Bell Ardent Computers +Steve Blumenthal BBN Laboratories +Vint Cerf Corporation for National Research Initiatives +David Cheriton Stanford University +David Clark Massachusetts Institute of Technology +Barry Leiner (Chairman) Research Institute for Advanced Computer Science +Robert Lyons Defense Communication Agency +Richard Metzger Rome Air Development Center +David Mills University of Delaware +Kevin Mills National Bureau of Standards +Chris Perry MITRE +Jon Postel USC Information Sciences Institute +Nachum Shacham SRI International +Fouad Tobagi Stanford University + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +Gigabit Working Group [Page 45] + +RFC 1077 November 1988 + + +End Notes + + [1] Workshop on Computer Networks, 17-19 February 1987, San Diego, + CA. + + [2] "A Report to the Congress on Computer Networks to Support + Research in the United States: A Study of Critical Problems and + Future Options", White House Office of Scientific and Technical + Policy (OSTP), November 1987. + + [3] We distinguish in the report between development of a backbone + network providing gigabit capacity, the GB, and an + interconnected set of high-speed networks providing high- + bandwidth service to the user, the Gigabit Network (GN). + + [4] Incidentally, they already manage to serve 150 million + subscribers in an 11-digit address-space (about 1:600 ratio). + We have a 9.6-digit address-space and are running into troubles + with much less than 100,000 users (less than 1:30,000 ratio). + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +Gigabit Working Group [Page 46] + |