diff options
Diffstat (limited to 'doc/rfc/rfc1152.txt')
| -rw-r--r-- | doc/rfc/rfc1152.txt | 1291 | 
1 files changed, 1291 insertions, 0 deletions
diff --git a/doc/rfc/rfc1152.txt b/doc/rfc/rfc1152.txt new file mode 100644 index 0000000..da6ab89 --- /dev/null +++ b/doc/rfc/rfc1152.txt @@ -0,0 +1,1291 @@ + + + + + + +Network Working Group                                       C. Partridge +Request for Comments: 1152                  BBN Systems and Technologies +                                                              April 1990 + + +                            Workshop Report +              Internet Research Steering Group Workshop on +                        Very-High-Speed Networks + +Status of this Memo + +   This memo is a report on a workshop sponsored by the Internet +   Research Steering Group.  This memo is for information only.  This +   RFC does not specify an Internet standard.  Distribution of this memo +   is unlimited. + +Introduction + +   The goal of the workshop was to gather together a small number of +   leading researchers on high-speed networks in an environment +   conducive to lively thinking.  The hope is that by having such a +   workshop the IRSG has helped to stimulate new or improved research in +   the area of high-speed networks. + +   Attendance at the workshop was limited to fifty people, and attendees +   had to apply to get in.  Applications were reviewed by a program +   committee, which accepted about half of them.  A few key individuals +   were invited directly by the program committee, without application. +   The workshop was organized by Dave Clark and Craig Partridge. + +   This workshop report is derived from session writeups by each of the +   session chairman, which were then reviewed by the workshop +   participants. + +Session 1: Protocol Implementation (David D. Clark, Chair) + +   This session was concerned with what changes might be required in +   protocols in order to achieve very high-speed operation. + +   The session was introduced by David Clark (MIT LCS), who claimed that +   existing protocols would be sufficient to go at a gigabit per second, +   if that were the only goal.  In fact, proposals for high-speed +   networks usually include other requirements as well, such as going +   long distances, supporting many users, supporting new services such +   as reserved bandwidth, and so on.  Only by examining the detailed +   requirements can one understand and compare various proposals for +   protocols.  A variety of techniques have been proposed to permit +   protocols to operate at high speeds, ranging from clever + + + +Partridge                                                       [Page 1] + +RFC 1152                  IRSG Workshop Report                April 1990 + + +   implementation to complete relayering of function.  Clark asserted +   that currently even the basic problem to be solved is not clear, let +   alone the proper approach to the solution. + +   Mats Bjorkman (Uppsala University) described a project that involved +   the use of an outboard protocol processor to support high-speed +   operation.  He asserted that his approach would permit accelerated +   processing of steady-state sequences of packets.  Van Jacobson (LBL) +   reported results that suggest that existing protocols can operate at +   high speeds without the need for outboard processors.  He also argued +   that resource reservation can be integrated into a connectionless +   protocol such as IP without losing the essence of the connectionless +   architecture.  This is in contrast to a more commonly held belief +   that full connection setup will be necessary in order to support +   resource reservation.  Jacobson said that he has an experimental IP +   gateway that supports resource reservation for specific packet +   sequences today. + +   Dave Borman (Cray Research) described high-speed execution of TCP on +   a Cray, where the overhead is most probably the system and I/O +   architecture rather than the protocol.  He believes that protocols +   such as TCP would be suitable for high-speed operation if the windows +   and sequence spaces were large enough. He reported that the current +   speed of a TCP transfer between the processors of a Cray Y-MP was +   over 500 Mbps.  Jon Crowcroft (University College London) described +   the current network projects at UCL.  He offered a speculation that +   congestion could be managed in very high-speed networks by returning +   to the sender any packets for which transmission capacity was not +   available. + +   Dave Feldmeier (Bellcore) reported on the Bellcore participation in +   the Aurora project, a joint experiment of Bellcore, IBM, MIT, and +   UPenn, which has the goal of installing and evaluating two sorts of +   switches at gigabit speeds between those four sites.  Bellcore is +   interested in switch and protocol design, and Feldmeier and his group +   are designing and implementing a 1 Gbps transport protocol and +   network interface.  The protocol processor will have special support +   for such things as forward error correction to deal with ATM cell +   loss in VLSI; a new FEC code and chip design have been developed to +   run at 1 Gbps. + +   Because of the large number of speakers, there was no general +   discussion after this session. + + + + + + + + +Partridge                                                       [Page 2] + +RFC 1152                  IRSG Workshop Report                April 1990 + + +Session 2: High-Speed Applications (Keith Lantz, Chair) + +   This session focused on applications and the requirements they impose +   on the underlying networks.  Keith Lantz (Olivetti Research +   California) opened by introducing the concept of the portable office +   - a world where a user is able to take her work with her wherever she +   goes.  In such an office a worker can access the same services and +   the same people regardless of whether she is in the same building +   with those services and people, at home, or at a distant site (such +   as a hotel) - or whether she is equipped with a highly portable, +   multi-media workstation, which she can literally carry with her +   wherever she goes.  Thus, portable should be interpreted as referring +   to portability of access to services rather than to portability of +   hardware.  Although not coordinated in advance, each of the +   presentations in this session can be viewed as a perspective on the +   portable office. + +   The bulk of Lantz's talk focused on desktop teleconferencing - the +   integration of traditional audio/video teleconferencing technologies +   with workstation-based network computing so as to enable +   geographically distributed individuals to collaborate, in real time, +   using multiple media (in particular, text, graphics, facsimile, +   audio, and video) and all available computer-based tools, from their +   respective locales (i.e., office, home, or hotel).  Such a facility +   places severe requirements on the underlying network.  Specifically, +   it requires support for several data streams with widely varying +   bandwidths (from a few Kbps to 1 Gbps) but generally low delay, some +   with minimal jitter (i.e., isochronous), and all synchronized with +   each other (i.e., multi-channel or media synchronization).  It +   appears that high-speed network researchers are paying insufficient +   attention to the last point, in particular.  For example, the bulk of +   the research on ATM has assumed that channels have independent +   connection request and burst statistics; this is clearly not the case +   in the context of desktop teleconferencing. + +   Lantz also stressed the need for adaptive protocols, to accommodate +   situations where the capacity of the network is exceeded, or where it +   is necessary to interoperate with low-speed networks, or where human +   factors suggest that the quality of service should change (e.g., +   increasing or decreasing the resolution of a video image).  Employing +   adaptive protocols suggests, first, that the interface to the network +   protocols must be hardware-independent and based only on quality of +   service.  Second, a variety of code conversion services should be +   available, for example, to convert from one audio encoding scheme to +   another.  Promising examples of adaptive protocols in the video +   domain include variable-rate constant-quality coding, layered or +   embedded coding, progressive transmission, and (most recently, at +   UC-Berkeley) the extension of the concepts of structured graphics to + + + +Partridge                                                       [Page 3] + +RFC 1152                  IRSG Workshop Report                April 1990 + + +   video, such that the component elements of the video image are kept +   logically separate throughout the production-to-presentation cycle. + +   Charlie Catlett (National Center for Supercomputing Applications) +   continued by analyzing a specific scientific application, simulation +   of a thunderstorm, with respect to its network requirements.  The +   application was analyzed from the standpoint of identifying data flow +   and the interrelationships between the computational algorithms, the +   supercomputer CPU throughput, the nature and size of the data set, +   and the available network services (throughput, delay, etc). + +   Simulation and the visualization of results typically involves +   several steps: + +      1.  Simulation + +      2.  Tessellation (transform simulation data into three-dimensional +          geometric volume descriptions, or polygons) + +      3.  Rendering (transform polygons into raster image) + +   For the thunderstorm simulation, the simulation and tessellation are +   currently done using a Cray supercomputer and the resulting polygons +   are sent to a Silicon Graphics workstation to be rendered and +   displayed.  The simulation creates data at a rate of between 32 and +   128 Mbps (depending on the number of Cray-2 processors working on the +   simulation) and the tessellation output data rate is in typically in +   the range of 10 to 100 Mbps, varying with the complexity of the +   visualization techniques.  The SGI workstation can display 100,000 +   polygons/sec which for this example translates to up to 10 +   frames/sec.  Analysis tools such as tracer particles and two- +   dimensional slices are used interactively at the workstation with +   pre-calculated polygon sets. + +   In the next two to three years, supercomputer speeds of 10-30 GFLOPS +   and workstation speeds of up to 1 GFLOPS and 1 million +   polygons/second display are projected to be available.  Increased +   supercomputer power will yield a simulation data creation rate of up +   to several Gbps for this application.  The increased workstation +   power will allow both tessellation and rendering to be done at the +   workstation.  The use of shared window systems will allow multiple +   researchers on the network to collaborate on a simulation, with the +   possibility of each scientist using his or her own visualization +   techniques with the tessellation process running on his or her +   workstation.  Further developments, such as network virtual memory, +   will allow the tessellation processes on the workstations to access +   variables directly in supercomputer memory. + + + + +Partridge                                                       [Page 4] + +RFC 1152                  IRSG Workshop Report                April 1990 + + +   Terry Crowley (BBN Systems and Technologies) continued the theme of +   collaboration, in the context of real-time video and audio, shared +   multimedia workspaces, multimedia and video mail, distributed file +   systems, scientific visualization, network access to video and image +   information, transaction processing systems, and transferring data +   and computational results between workstations and supercomputers. +   In general, such applications could help groups collaborate by +   directly providing communication channels (real-time video, shared +   multimedia workspaces), by improving and expanding on the kinds of +   information that can be shared (multimedia and video mail, +   supercomputer data and results), and by reducing replication and the +   complexity of sharing (distributed file systems, network access to +   video and image information). + +   Actual usage patterns for these applications are hard to predict in +   advance.  For example, real-time video might be used for group +   conferencing, for video phone calls, for walking down the hall, or +   for providing a long-term shared viewport between remote locations in +   order to help establish community ties.  Two characteristics of +   network traffic that we can expect are the need to provide multiple +   data streams to the end user and the need to synchronize these +   streams.  These data streams will include real-time video, access to +   stored video, shared multimedia workspaces, and access to other +   multimedia data.  A presentation involving multiple data streams must +   be synchronized in order to maintain cross-references between them +   (e.g., pointing actions within the shared multimedia workspace that +   are combined with a voice request to delete this and save that). +   While much traffic will be point-to-point, a significant amount of +   traffic will involve conferences between multiple sites.  A protocol +   providing a multicast capability is critical. + +   Finally, Greg Watson (HP) presented an overview of ongoing work at +   the Hewlett-Packard Bristol lab.  Their belief is that, while +   applications for high-speed networks employing supercomputers are the +   the technology drivers, the economic drivers will be applications +   requiring moderate bandwidth (say 10 Mbps) that are used by everyone +   on the network. + +   They are investigating how multimedia workstations can assist +   distributed research teams - small teams of people who are +   geographically dispersed and who need to work closely on some area of +   research.  Each workstation provides multiple video channels, +   together with some distributed applications running on personal +   computers.  The bandwidth requirements per workstation are about 40 +   Mbps, assuming a certain degree of compression of the video channels. +   Currently the video is distributed as an analog signal over CATV +   equipment.  Ideally it would all be carried over a single, unified +   wide-area network operating in the one-to-several Gbps range. + + + +Partridge                                                       [Page 5] + +RFC 1152                  IRSG Workshop Report                April 1990 + + +   They have constructed a gigabit network prototype and are currently +   experimenting with uncompressed video carried over the same network +   as normal data traffic. + +Session 3: Lightwave Technology and its Implications (Ira Richer, Chair) + +   Bob Kennedy (MIT) opened the session with a talk on network design in +   an era of excess bandwidth.  Kennedy's research is focused on multi- +   purpose networks in which bandwidth is not a scarce commodity, +   networks with bandwidths of tens of terahertz.  Kennedy points out +   that a key challenge in such networks is that electronics cannot keep +   up with fiber speeds.  He proposes that we consider all-optical +   networks (in which all signals are optical) with optoelectronic nodes +   or gateways capable of recognizing and capturing only traffic +   destined for them, using time, frequency, or code divisions of the +   huge bandwidth.  The routing algorithms in such networks would be +   extremely simple to avoid having to convert fiber-optics into slower +   electronic pathways to do switching. + +   Rich Gitlin (AT&T Bell Labs) gave a talk on issues and opportunities +   in broadband telecommunications networks, with emphasis on the role +   of fiber optic and photonic technology.  A three-level architecture +   for a broadband telecommunications network was presented.  The +   network is B-ISDN/ATM 150 (Mbps) based and consists of: customer +   premises equipment (PBXs, LANs, multimedia terminals) that access the +   network via a router/gateway, a Network Node (which is a high +   performance ATM packet switch) that serves both as a LAN-to-LAN +   interconnect and as a packet concentrator for traffic destined for +   CPE attached to other Network Nodes, and a backbone layer that +   interconnects the NODES via a Digital Cross-Connect System that +   provide reconfigurable SONET circuits between the NODES (the use of +   circuits minizes delay and avoids the need for implementation of +   peak-transmission-rate packet switching).  Within this framework, the +   most likely places for near-term application of photonics, apart from +   pure transport (ie, 150 Mbps channels in a 2.4 Gbps SONET system), +   are in the Cross-Connect (a Wavelength Division Multiplexed based +   structure was described) and in next-generation LANs that provide +   Gigabit per second throughputs by use of multiple fibers, concurrent +   transmission, and new access mechanisms (such as store and forward). + +   A planned interlocation Bell Labs multimedia gigabit/sec research +   network, LuckyNet, was described that attempts to extend many of the +   above concepts to achieve its principal goals: provision of a gigabit +   per second capability to a heterogeneous user community, the +   stimulation of applications that require Gpbs throughput (initial +   applications are video conferencing and LAN interconnect), and, to +   the extent possible, be based on standards so that interconnection +   with other Gigabit testbeds is possible. + + + +Partridge                                                       [Page 6] + +RFC 1152                  IRSG Workshop Report                April 1990 + + +Session 4: High Speed Networks and the Phone System +           (David Tennenhouse, Chair) + +   David Tennenhouse (MIT) reported on the ATM workshop he hosted the +   two days previous to this workshop.  His report will appear as part +   of the proceedings of his workshop. + +   Wally St. John (LANL) followed with a presentation on the Los Alamos +   gigabit testbed.  This testbed is based on the High Performance +   Parallel Interface (HPPI) and on crossbar switch technology.  LANL +   has designed its own 16x16 crossbar switch and has also evaluated the +   Network Systems 8x8 crossbar switch. Future plans for the network +   include expansion to the CASA gigabit testbed.  The remote sites (San +   Diego Supercomputer Center, Caltech, and JPL) are configured +   similarly to the LANL testbed.  The long-haul interface is from HPPI +   to/from SONET (using ATM if in time). + +   Wally also discussed some of the problems related to building a +   HPPI-SONET gateway: + +      a)  Flow control.  The HPPI, by itself, is only readily extensible +          to 64 km because of the READY-type flow control used in the +          physical layer.  The gateway will need to incorporate larger +          buffers and independent flow control. + +      b)  Error-rate expectations.  SONET is only specified to have a +          1E-10 BER on a per hop basis.  This is inadequate for long +          links.  Those in the know say that SONET will be much better +          but the designer is faced with the poor BER in the SONET spec. + +      c)  Frame mapping.  There are several interesting issues to be +          considered in finding a good mapping from the HPPI packet +          to the SONET frame.  Some are what SONET STS levels will be +          available in what time frame, the availability of concatenated +          service, and the error rate issue. + +   Dan Helman (UCSC) talked about work he has been doing with Darrell +   Long to examine the interconnection of Internet networks via an ATM +   B-ISDN network.  Since network interfaces and packet processing are +   the expensive parts of high-speed networks, they believe it doesn't +   make sense to use the ATM backbone only for transmission; it should +   be used for switching as well.  Therefore gateways (either shared by +   a subnet or integrated with fast hosts) are needed to encapsulate or +   convert conventional protocols to ATM format.  Gateways will be +   responsible for caching connections to recently accessed +   destinations.  Since many short-lived low-bandwidth connections as +   foreseen (e.g., for mail and ftp), routing in the ATM network (to set +   up connections) should not be complicated - a form of static routing + + + +Partridge                                                       [Page 7] + +RFC 1152                  IRSG Workshop Report                April 1990 + + +   should be adequate.  Connection performance can be monitored by the +   gateways.  Connections are reestablished if unacceptable.  All +   decision making can be done by gateways and route servers at low +   packet rates, rather than the high aggregate rate of the ATM network. +   One complicated issue to be addressed is how to transparently +   introduce an ATM backbone alongside the existing Internet. + +Session 5: Distributed Systems (David Farber, Chair) + +   Craig Partridge (BBN Systems and Technologies) started this session +   by arguing that classic RPC does not scale well to gigabit-speed +   networks.  The gist of his argument was that machines are getting +   faster and faster, while the round-trip delay of networks is staying +   relatively constant because we cannot send faster than the speed of +   light.  As a result, the effective cost of doing a simple RPC, +   measured in instruction cycles spent waiting at the sending machine, +   will become extremely high (millions of instruction cycles spent +   waiting for the reply to an RPC).  Furthermore, the methods currently +   used to improve RPC performance, such as futures and parallel RPC, do +   not adequately solve this problem.  Future requests will have to be +   made much much earlier if they are to complete by the time they are +   needed.  Parallel RPC allows multiple threads, but doesn't solve the +   fact that each individual sequence of RPCs still takes a very long +   time. + +   Craig went on to suggest that there are at least two possible ways +   out of the problem.  One approach is to try to do a lot of caching +   (to waste bandwidth to keep the CPU fed).  A limitation of this +   approach is that at some point the cache becomes so big that you have +   to keep in consistent with other systems' caches, and you suddenly +   find yourself doing synchronization RPCs to avoid doing normal RPCs +   (oops!).  A more promising approach is to try to consolidate RPCs +   being sent to the same machine into larger operations which can be +   sent as a single transaction, run on the remote machine, and the +   result returned.  (Craig noted that he is pursuing this approach in +   his doctoral dissertation at Harvard). + +   Ken Schroder (BBN Systems and Technologies) gave a talk on the +   challenges of combining gigabit networks with wide-area heterogeneous +   distributed operating systems.  Ken feels the key goals of wide area +   distributed systems will be to support large volume data transfers +   between users of conferencing and similar applications, and to +   deliver information to a large number of end users sharing services +   such as satellite image databases.  These distributed systems will be +   motivated by the natural distribution of users, of information and of +   expensive special purpose computer resources. + +   Ken pointed to three of the key problems that must be addressed at + + + +Partridge                                                       [Page 8] + +RFC 1152                  IRSG Workshop Report                April 1990 + + +   the system level in these environments: how to provide high +   utilization; how to manage consistency and synchronization in the +   presence of concurrency and non-determinism; and how to construct +   scalable system and application services.  Utilization is key only to +   high performance applications, where current systems would be limited +   by the cost of factors such as repeatedly copying messages, +   converting data representations and switching between application and +   operating system.  Concurrency can be used improve performance, but +   is also likely to occur in many programs inadvertently because of +   distribution.  Techniques are required both to exploit concurrency +   when it is needed, and to limit it when non-determinism can lead to +   incorrect results.  Extensive research on ensuring consistency and +   resolving resource conflicts has been done in the database area, +   however distributed scheduling and the need for high availability +   despite partial system failures introduce special problems that +   require additional research.  Service scalability will be required to +   support customer needs as the size of the user community grow.  It +   will require attention both ensuring that components do not break +   when they are subdivided across additional processors to support a +   larger user population, and to ensure that performance does to each +   user can be affordably maintained as new users are added. + +   In a bold presentation, Dave Cheriton (Stanford) made a sweeping +   argument that we are making a false dichotomy between distributed +   operating systems and networks.  In a gigabit world, he argued, the +   major resource in the system is the network, and in a normal +   operating system we would expect such a critical resource to be +   managed by the operating system.  Or, put another way, the gigabit +   network distributed operating system should manage the network. +   Cheriton went on to say that if a gigabit distributed operating +   system is managing the network, then it is perfectly reasonable to +   make the network very dumb (but fast) and put the system intelligence +   in the operating systems on the hosts that form the distributed +   system. + +   In another talk on interprocess communication, Jonathan Smith (UPenn) +   again raised the problem of network delay limiting RPC performance. +   In contrast to Partridge's earlier talk, Smith argued that the +   appropriate approach is anticipation or caching.  He justified his +   argument with a simple cost example.  If a system is doing a page +   fetch between two systems which have a five millisecond round-trip +   network delay between them, the cost of fetching n pages is: + +                         5 msec + (n-1) * 32 usec + +   Thus the cost of fetching an additional page is only 32 usec, but +   underfetching and having to make another request to get a page you +   missed costs 5000 usec.  Based on these arguments, Smith suggested + + + +Partridge                                                       [Page 9] + +RFC 1152                  IRSG Workshop Report                April 1990 + + +   that we re-examine work in virtual memory to see if there are +   comfortable ways to support distributed virtual memory with +   anticipation. + +   In the third talk on RPC in the session, Tommy Joseph (Olivetti), for +   reasons similar to those of Partridge and Smith, argued that we have +   to get rid of RPC and give programmers alternative programming +   paradigms.  He sketched out ideas for asynchronous paradigms using +   causal consistency, in which systems ensure that operations happen in +   the proper order, without synchronizing through a single system. + +Session 6: Hosts and Host Interfaces (Gary Delp, Chair) + +   Gary Delp (IBM Research) discussed several issues involved in the +   increase in speed of network attachment to hosts of increasing +   performance.  These issues included: + +      -  Media Access - There are aspects of media access that are +         best handled by dedicated silicon, but there are also aspects +         that are best left to a general-purpose processor. + +      -  Compression - Some forms of compression/expansion may belong +         on the network interface; most will be application-specific. + +      -  Forward Error Correction - The predicted major packet loss +         mode is packet drops due to internal network congestion, rather +         than bit errors, so forward error correction internal to a +         packet may not be useful.  On the other hand, the latency cost +         of not being able to recover from bit errors is very high. +         Some proposals were discussed which suggest that FEC among +         packet groups, with dedicated hardware support, is the way +         to go. + +      -  Encryption/Decryption - This is a computationally intensive +         task.  Most agree that if it is done with all traffic, some +         form of hardware support is helpful.  Where does it fit in the +         protocol stack? + +      -  Application Memory Mapping - How much of the host memory +         structure should be exposed to the network interface? +         Virtual memory and paging complicate this issue considerably. + +      -  Communication with Other Channel Controllers - Opinions were +         expressed that ranged from absolutely passive network +         interfaces to interfaces that run major portions of the +         operating system and bus arbitration codes. + +      -  Blocking/Segmentation - The consensus is that B/S should + + + +Partridge                                                      [Page 10] + +RFC 1152                  IRSG Workshop Report                April 1990 + + +         occur wherever the transport layer is processed. + +      -  Routing - This is related to communications with other +         controllers.  A routing-capable interface can reduce the bus +         requirements by a factor of two. + +      -  Intelligent participation in the host structure as a gateway, +         router, or bridge. + +      -  Presentation Layer issues - All of the other overheads can be +         completely overshadowed by this issue if it is not solved well +         and integrated into the overall host architecture.  This points +         out the need for some standardization of representation (IEEE +         floating point, etc.) + +   Eric Cooper (CMU) summarized some initial experience with Nectar, a +   high-speed fiber-optic LAN that has been built at Carnegie Mellon. +   Nectar consists of an arbitrary mesh of crossbar switches connected +   by means of 100 Mbps fiber-optic links.  Hosts are connected to +   crossbar switches via communication processor boards called CABs. +   The CAB presents a memory-mapped interface to user processes and +   off-loads all protocol processing from the host. + +   Preliminary performance figures show that latency is currently +   limited by the number of VME operations required by the host-to-CAB +   shared memory interface in the course of sending and receiving a +   message.  The bottleneck in throughput is the speed of the VME +   interface: although processes running on the CABs can communicate +   over Nectar at 70 Mbps, processes on the hosts are limited to +   approximately 25 Mbps. + +   Jeff Mogul (DEC Western Research Lab) made these observations: +   Although off-board protocol processors have been a popular means to +   connect a CPU to a network, they will be less useful in the future. +   In the hypothetical workstation of the late 1990s, with a 1000-MIPS +   CPU and a Gbps LAN, an off-board protocol processor will be of no +   use.  The bottleneck will not be the computation required to +   implement the protocol, but the cost of moving the packet data into +   the CPU's cache and the cost of notifying the user process that the +   data is available.  It will take far longer (hundreds of instruction +   cycles) to perform just the first cache miss (required to get the +   packet into the cache) than to perform all of the instructions +   necessary to implement IP and TCP (perhaps a hundred instructions). + +   A high-speed network interface for a reasonably-priced system must be +   designed with this cost structure in mind; it should also eliminate +   as many CPU interrupts as possible, since interrupts are also very +   expensive.  It makes more sense to let a user process busy-wait on a + + + +Partridge                                                      [Page 11] + +RFC 1152                  IRSG Workshop Report                April 1990 + + +   network-interface flag register than to suspend it and then take an +   interrupt; the normal CPU scheduling mechanism is more efficient than +   interrupts if the network interactions are rapid. + +   David Greaves (Olivetti Research Ltd.) briefly described the need for +   a total functionality interface architecture that would allow the +   complete elimination of communication interrupts.  He described the +   Cambridge high-speed ring as an ATM cell-like interconnect that +   currently runs at 500-1000 MBaud, and claims that ATM at that speed +   is a done deal.   Dave Tennenhouse also commented that ATM at high +   speeds with parallel processors is not the difficult thing that +   several others have been claiming. + +   Bob Beach (Ultra Technologies) started his talk with the observation +   that networking could be really fast if only we could just get rid of +   the hosts.   He then supported his argument with illustrations of +   80-MByte/second transfers to frame buffers from Crays that drop to +   half that speed when the transfer is host-to-host.  Using null +   network layers and proprietary MAC layers, the Ultra Net system can +   communicate application-to-application with ISO TP4 as the transport +   layer at impressive rates of speed.  The key to high-speed host +   interconnects has been found to be both large packets and large (on +   the order of one megabyte) channel transfer requests.  Direct DMA +   interfaces exhibit much smaller transfer latencies. + +   Derek McAuley (University Cambridge Computer Laboratory) described +   work of the Fairisle project which is producing an ATM network based +   on fast packet switches.  A RISC processor (12 MIPS) is used in the +   host interface to do segmentation/reassembly/demultiplexing.  Line +   rates of up to 150 Mbps are possible even with this modest processor. +   Derek has promised that performance and requirement results from this +   system will be published in the spring. + +   Bryan Lyles (XEROX PARC) volunteered to give an abbreviated talk in +   exchange for discussion rights.  He reported that Xerox PARC is +   interested in ATM technology and wants to install an ATM LAN at the +   earliest possible opportunity.  Uses will include such applications +   as video where guaranteed quality of service (QOS) is required.  ATM +   technology and the desire for guaranteed QOS places a number of new +   constraints on the host interface.  In particular, they believe that +   they will be forced towards rate-based congestion control.  Because +   of implementation issues and burst control in the ATM switches, the +   senders will be forced to do rate based control on a cell-by-cell +   basis. + +   Don Tolmie (Los Alamos National Laboratory) described the High- +   Performance Parallel Interface (HPPI) of ANSI task group X3T9.3.  The +   HPPI is a standardized basic building block for implementing, or + + + +Partridge                                                      [Page 12] + +RFC 1152                  IRSG Workshop Report                April 1990 + + +   connecting to, networks at the Gbps speeds, be they ring, hub, +   cross-bar, or long-haul based.  The HPPI physical layer operates at +   800 or 1600 Mbps over 25-meter twisted-pair copper cables in a +   point-to-point configuration.  The HPPI physical layer has almost +   completed the standards process, and a companion HPPI data framing +   standard is under way, and a Fiber Channel standard at comparable +   speeds is also being developed.  Major companies have completed, or +   are working on, HPPI interfaces for supercomputers, high-end +   workstations, fiber-optic extenders, and networking components. + +   The discussion at the end of the session covered a range of topics. +   The appropriateness of outboard protocol processing was questioned. +   Several people agreed that outboarding on a Cray (or similar +   cost/performance) machines makes economic sense.  Van Jacobson +   contended that for workstations, a simple memory-mapped network +   interface that provides packets visible to the host processor may +   well be the ideal solution. + +   Bryan Lyles reiterated several of his earlier points, asserting that +   when we talk about host interfaces and how to build them we should +   remember that we are really talking about process-to-process +   communication, not CPU-to-CPU communication.  Not all processes run +   on the central CPU, e.g., graphics processors and multimedia. +   Outboard protocol processing may be a much better choice for these +   architectures. + +   This is especially true when we consider that memory/bus bandwidth is +   often a bottleneck.  When our systems run out of bandwidth, we are +   forced towards a NUMA model and multiple buses to localize memory +   traffic. + +   Because of QOS issues, the receiver must be able to tell the sender +   how fast it can send.  Throwing away cells (packets) will not work +   because unwanted packets will still clog the receiver's switch +   interface, host interface, and requires processing to throw away. + +Session 7: Congestion Control (Scott Shenker, Chair) + +   The congestion control session had six talks.  The first two talks +   were rather general, discussing new approaches and old myths.  The +   other four talks discussed specific results on various aspects of +   packet (or cell) dropping: how to avoid drops, how to mitigate their +   impact on certain applications, a calculation of the end-to-end +   throughput in the presence of drops, and how rate-based flow control +   can reduce buffer usage.  Thumbnail sketches of the talks follow. + +   In the first of the general talks, Scott Shenker (XEROX PARC) +   discussed how ideas from economics can be applied to congestion + + + +Partridge                                                      [Page 13] + +RFC 1152                  IRSG Workshop Report                April 1990 + + +   control.  Using economics, one can articulate questions about the +   goals of congestion control, the minimal feedback necessary to +   achieve those goals, and the incentive structure of congestion +   control.  Raj Jain (DEC) then discussed eight myths related to +   congestion control in high-speed networks.  Among other points, Raj +   argued that (1) congestion problems will not become less important +   when memory, processors, and links become very fast and cheap, (2) +   window flow control is required along with rate flow control, and (3) +   source-based controls are required along with router-based control. + +   In the first of the more specific talks, Isidro Castineyra (BBN +   Communications Corporation) presented a back-of-the-envelope +   calculation on the effect of cell drops on end-to-end throughput. +   While at extremely low drop rates the retransmission strategies of +   go-back-n and selective retransmission produced similar end-to-end +   throughput, at higher drop rates selective retransmission achieved +   much higher throughput.  Next, Tony DeSimone (AT&T) told us why +   high-speed networks are not just fast low-speed networks.   If the +   buffer/window ratio is fixed, the drop rate decreases as the network +   speed increases.  Also, data was presented which showed that adaptive +   rate control can greatly decrease buffer utilization.  Jamal +   Golestani (Bellcore) then presented his work on stop-and-go queueing. +   This is a simple stalling algorithm implemented at the switches which +   guarantees no dropped packets and greatly reduces delay jitter.  The +   algorithm requires prior bandwidth reservation and some flow control +   on sources, and is compatible with basic FIFO queues.  In the last +   talk, Victor Frost (University of Kansas) discussed the impact of +   different dropping policies on the perceived quality of a voice +   connection.  When the source marks the drop priority of cells and the +   switch drops low priority cells first, the perceived quality of the +   connection is much higher than when cells are dropped randomly. + +Session 8: Switch Architectures (Dave Sincoskie, Chair) + +   Dave Mills (University of Delaware) presented work on a project now +   under way at the University of Delaware to study architectures and +   protocols for a high-speed network and packet switch capable of +   operation to the gigabit regime over distances spanning the country. +   It is intended for applications involving very large, very fast, very +   bursty traffic typical of supercomputing, remote sensing, and +   visualizing applications.  The network is assumed to be composed of +   fiber trunks, while the switch architecture is based on a VLSI +   baseband crossbar design which can be configured for speeds from 25 +   Mbps to 1 Gbps. + +   Mills' approach involves an externally switched architecture in which +   the timing and routing of flows between crossbar switches are +   determined by sequencing tables and counters in high-speed memory + + + +Partridge                                                      [Page 14] + +RFC 1152                  IRSG Workshop Report                April 1990 + + +   local to each crossbar.  The switch program is driven by a +   reservation-TDMA protocol and distributed scheduling algorithm +   running in a co-located, general-purpose processor.  The end-to-end +   customers are free to use any protocol or data format consistent with +   the timing of the network.  His primary interest in the initial +   phases of the project is the study of appropriate reservation and +   scheduling algorithms.  He expect these algorithms to have much in +   common with the PODA algorithm used in the SATNET and WIDEBAND +   satellite systems and to the algorithms being considered for the +   Multiple Satellite System (MSS). + +   John Robinson (JR, BBN Systems and Technologies) gave a talk called +   Beyond the Butterfly, which described work on a design for an ATM +   cell switch, known as MONET.  The talk described strategies for +   buffering at the input and output interfaces to a switch fabric +   (crossbar or butterfly).  The main idea was that cells should be +   introduced to the switch fabric in random sequence and to random +   fabric entry ports to avoid persistent traffic patterns having high +   cell loss in the switch fabric, where losses arise due to contention +   at output ports or within the switch fabric (in the case of a +   butterfly).  Next, the relationship of this work to an earlier design +   for a large-scale parallel processor, the Monarch, was described.  In +   closing, JR offered the claim that this class of switch is realizable +   in current technology (barely) for operation over SONET OC-48 2.4 +   Gbps links. + +   Dave Sincoskie (Bellcore) reported on two topics: recent switch +   construction at Bellcore, and high-speed processing of ATM cells +   carrying VC or DG information.  Recent switch design has resulted in +   a switch architecture named SUNSHINE, a Batcher-banyan switch which +   uses recirculation and multiple output banyans to resolve contention +   and increase throughput.  A paper on this switch will be published at +   ISS '90, and is available upon request from the author.  One of the +   interesting traffic results from simulations of SUNSHINE shows that +   per-port output queues of up to 1,000 cells (packets) may be +   necessary for bursty traffic patterns.  Also, Bill Marcus (at +   Bellcore) has recently produced Batcher-banyan (32x32) chips which +   test up to 170Mb/sec per port. + +   The second point in this talk was that there is little difference in +   the switching processing of Virtual Circuit (VC) and Datagram (DG) +   traffic that which has been previously broken into ATM cells at the +   network edge.  The switch needs to do a header translation operation +   followed by some queueing (not necessarily FIFO).  The header +   translation of the VC and DG cells differs mainly in the memory +   organization of the address translation tables (dense vs. sparse). + +   The discussion after the presentations seemed to wander off the topic + + + +Partridge                                                      [Page 15] + +RFC 1152                  IRSG Workshop Report                April 1990 + + +   of switching, back to some of the source-routing vs. network routing +   issues discussed earlier in the day. + +Session 9: Open Mike Night (Craig Partridge, Chair) + +   As an experiment, the workshop held an open mike session during the +   evening of the second day.  Participants were invited to speak for up +   to five minutes on any subject of their choice.  Minutes of this +   session are sketchy because the chair found himself pre-occupied by +   keeping speakers roughly within their time limits. + +   Charlie Catlett (NSCA) showed a film of the thunderstorm simulations +   he discussed earlier. + +   Dave Cheriton (Stanford) made a controversial suggestion that perhaps +   one could manage congestion in the network simply by using a steep +   price curve, in which sending large amounts of data cost +   exponentially more than sending small amounts of data (thus leading +   people only to ask for large bandwidth when they needed it, and +   having them pay so much, that we can afford to give it to them). + +   Guru Parulkar (Washington University, St. Louis) argued that the +   recent discussion on appropriateness of existing protocol and need +   for new protocols (protocol architecture) for gigabit networking +   lacks the right focus.  The emphasis of the discussion should be on +   what is the right functionality for gigabit speeds, which is simpler +   per packet processing, combination of rate and window based flow +   control, smart retransmission strategy, appropriate partitioning of +   work among host cpu+os, off board cpu, and custom hardware, and +   others.  It is not surprising that the existing protocols can be +   modified to include this functionality.  By the same token, it is not +   surprising that new protocols can be designed which take advantage of +   lessons of existing protocols and also include other features +   necessary for gigabit speeds. + +   Raj Jain (DEC) suggested we look at new ways to measure protocol +   performance, suggesting our current metrics are insufficiently +   informative. + +   Dan Helman (UCSC) asked the group to consider, more carefully, who +   exactly the users of the network will be.  Large consumers? or many +   small consumers? + + + + + + + + + +Partridge                                                      [Page 16] + +RFC 1152                  IRSG Workshop Report                April 1990 + + +Session 10: Miscellaneous Topics (Bob Braden, Chair) + +   As its title implies, this session covered a variety of different +   topics relating to high-speed networking. + +   Jim Kurose (University of Massachussetts) described his studies of +   scheduling and discard policies for real-time (constrained delay) +   traffic.  He showed that by enforcing local deadlines at switches +   along the path, it is possible to significantly reduce overall loss +   for such traffic.  Since his results depend upon the traffic model +   assumptions, he ended with a plea for work on traffic models, stating +   that Poisson models can sometimes lead to results that are wrong by +   many orders of magnitude. + +   Nachum Shacham (SRI International) discussed the importance of error +   correction schemes that can recover lost cells, and as an example +   presented a simple scheme based upon longitudinal parity.  He also +   showed a variant, diagonal parity, which allows a single missing cell +   to be recreated and its position in the stream determined. + +   Two talks concerned high-speed LANs.  Biswanath Muhkerjee (UC Davis) +   surveyed the various proposals for fair scheduling on unidirectional +   bus networks, especially those that are distance insensitive, i.e., +   that can achieve 100% channel utilization independent of the bus +   length and data rate.  He described in particular his own scheme, +   which he calls p-i persistant. + +   Howard Salwen (Proteon), speaking in place of Mehdi Massehi of IBM +   Zurich who was unable to attend, also discussed high-speed LAN +   technologies.  At 100 Mbps, a token ring has a clear advantage, but +   at 1 Gbps, the speed of light kills 802.6, for example.  He briefly +   described Massehi's reservation-based scheme, CRMA (Cyclic- +   Reservation Multiple-Access). + +   Finally, Yechiam Yemeni (YY, Columbia University) discussed his work +   on a protocol silicon compiler.  In order to exploit the potential +   parallelism, he is planning to use one processor per connection. + +   The session closed with a spirited discussion of about the relative +   merits of building an experimental network versus simulating it. +   Proponents of simulation pointed out the high cost of building a +   prototype and limitation on the solution space imposed by a +   particular hardware realization.  Proponents of building suggested +   that artificial traffic can never explore the state space of a +   network as well as real traffic can, and that an experimental +   prototype is important for validating simulations. + + + + + +Partridge                                                      [Page 17] + +RFC 1152                  IRSG Workshop Report                April 1990 + + +Session 11: Protocol Architectures (Vint Cerf, Chair) + +   Nick Maxemchuk (AT&T Bell Labs) summarized the distinctions between +   circuit switching, virtual circuits, and datagrams.  Circuits are +   good for (nearly) constant rate sources.  Circuit switching dedicates +   resources for the entire period of service.  You have to set up the +   resource allocation before using it.  In a 1.7 Gbps network, a 3000- +   mile diameter consumes 10**7 bytes during the circuit set-up round- +   trip time, and potentially the same for circuit teardown.  Some +   service requirements (file transfer, facsimile transmission) are far +   smaller than the wasted 2*10**7 bytes these circuit management delays +   impose.  (Of course, these costs are not as dramatic if the allocated +   bandwidth is less than the maximum possible.) + +   Virtual circuits allow shared use of bandwidth (multiplexing) when +   the primary source of traffic is idle (as in Voice Time Assigned +   Speech Interpolation).  The user notifies the network of planned +   usage. + +   Datagrams (DG) are appropriate when there is no prior knowledge of +   use statistics or usage is far less than the capacity wasted during +   circuit or virtual circuit set-up.  One can adaptively route traffic +   among equivalent resources. + +   In gigabit ATMs, the high service speed and decreased cell size +   increases the relative burstiness of service requests.  All of these +   characteristics combine to make DG service very attractive. + +   Maxemchuk then described a deflection routing notion in which traffic +   would be broken into units of fixed length and allowed into the +   network when capacity was available and routed out by any available +   channel, with preference being given to the channel on the better +   path.  This idea is similar to the hot potato routing of Paul Baran's +   1964 packet switching design.  With buffering (one buffer), Maxemchuk +   achieved a theoretical 90% utilization.  Large reassembly buffers +   provide for better throughput. + +   Maxemchuk did not have an answer to the question: how do you make +   sure empty "slots" are available where needed? This is rather like +   the problem encountered by D. Davies at the UK National Physical +   Laboratory in his isarythmic network design in which a finite number +   of crates are available for data transport throughout the network. + +   Guru Parulkar (Washington University, St. Louis) presented a broad +   view of an Internet architecture in which some portion of the system +   would operate at gigabit speeds.  In his model, internet, transport, +   and application protocols would operate end to end.  The internet +   functions would be reflected in gateways and in the host/net + + + +Partridge                                                      [Page 18] + +RFC 1152                  IRSG Workshop Report                April 1990 + + +   interface, as they are in the current Internet.  However, the +   internet would support a new type of service called a congram which +   aims at combining strengths of both soft connection and datagram. + +   In this architecture, a variable grade of service would be provided. +   Users could request congrams (UCON) or the system could set them up +   internally (Picons) to avoid end-to-end setup latency.  The various +   grades of service could be requested, conceptually, by asserting +   various required (desired) levels of error control, throughput, +   delay, interarrival jitter, and so on.  Gateways based on ATM +   switches, for example, would use packet processors at entry/exit to +   do internet specific per packet processing, which may include +   fragmentation and reassembly of packets (into and out of ATM cells). + +   At the transport level, Parulkar argued for protocols which can +   provide application-oriented flow and error control with simple per +   packet processing.  He also mentioned the notion of a generalized RPC +   (GRPC) in which code, data, and execution might be variously local or +   remote from the procedure initiator.  GRPC can be implemented using +   network level virtual storage mechanisms. + +   The basic premise of Raj Yavatkar's presentation (University of +   Kentucky) was that processes requiring communication service would +   specify their needs in terms of peak and average data rate as well as +   defining burst parameters (frequency and size).  Bandwidth for a +   given flow would be allocated at the effective data rate that is +   computed on the basis of flow parameters.  The effective data rate +   lies somewhere between the peak and average data rate based on the +   burst parameters.  Statistical multiplexing would take up the gap +   between peak and effective rate when a sudden burst of traffic +   arrives.  Bounds on packet loss rate can be computed for a given set +   of flow parameters and corresponding effective data rate. + +   This presentation led to a discussion about deliberate disciplining +   of inter-process communication demands to match the requested flow +   (service) profile.  This point was made in response to the +   observation that we often have little information about program +   behavior and might have trouble estimating the network service +   requirements of any particular program. + +Architectural Discussion + +   An attempt was made to conduct a high-level discussion on various +   architectural questions.  The discussion yielded a variety of +   opinions: + +      1.  The Internet would continue to exist in a form similar +          to its current incarnation, and gateways would be required, + + + +Partridge                                                      [Page 19] + +RFC 1152                  IRSG Workshop Report                April 1990 + + +          at least to interface the existing facilities to the high +          speed packet switching environment. + +      2.  Strong interest was expressed by some participants in access +          to raw (naked ATM) services.  This would permit users +          to construct their own gigabit nets, at the IP level, at any +          rate.  The extreme view of this was taken by David Cheriton +          who would prefer to have control over routing decisions and +          other behavior of the ATM network. + +      3.  The speed of light problem (latency, round-trip delay) +          is not going to go away and will have serious impact on +          control of the system.  The optimistic view was taken, +          for example, by Craig Partridge and Van Jacobson, who felt +          that many of the existing network and communications +          management mechanisms used in the present Internet protocols +          would suffice, if suitably implemented, at higher speeds. +          A less rosy view was taken by David Clark who observed +          (as did others) that many transactions would be serviced in +          much less than one round-trip time, so that any end-to-end +          controls would be largely useless. + +      4.  For applications requiring fixed, periodic service, +          reservation of resource seemed reasonably attractive to many +          participants, as long as the service period dominated the +          set-up time (round-trip delay) by an appreciable +          margin. + +      5.  There was much discussion throughout the workshop of +          congestion control and flow control.  Although these +          problems were not new, they took on somewhat newer +          character in the presence of much higher round-trip delays +          (measured in bits outstanding).  One view is that end-to-end +          flow control is needed, in any case, to moderate sources +          sending to limited bandwidth receivers.  End-to-end flow +          control may not, however, be sufficient to protect the +          interior of the network from congestion problems, so +          additional, intra-network means are needed to cope with +          congestion hot spots.   Eventually such conditions +          have to be reflected to the periphery of the network to +          moderate traffic sources. + +      6.  There was disagreement on the build or simulate +           question.  One view was in favor of building network +          components so as to collect and understand live application +          data.  Another view held that without some careful +          simulation, one might have little idea what to build +          (for example, Sincoskie's large buffer pool requirement was + + + +Partridge                                                      [Page 20] + +RFC 1152                  IRSG Workshop Report                April 1990 + + +          not apparent until the system was simulated). + +Comments from Workshop Evaluation Forms + +   At the end of the IRSG workshop, we asked attendees to fill out an +   evaluation form.  Of the fifty-one attendees, twenty-nine (56%) +   turned in a form. + +   The evaluation form asked attendees to answer two questions: + +      #1.  Do you feel that having attended this workshop will help you +           in your work on high speed networks during the next year? + +      #2.  What new ideas, questions, or issues, did you feel were +           brought up in the workshop? + +   In this section we discuss the answers we got to both questions. + +Question #1 + +   There was a satisfying unanimity of opinion on question #1.  Twenty- +   four attendees answered yes, often strongly (e.g., Absolutely and +   very much so).  Of the remaining five respondents, three said they +   expected it to have some effect on their research and only two said +   the workshop would have little or no effect. + +   Some forms had some additional notes about why the workshop helped +   them.  Several people mentioned that there was considerable benefit +   to simply meeting and talking with people they hadn't met before.  A +   few other people noted that the workshop had broadened their +   perspective, or improved their understanding of certain issues.  A +   couple of people noted that they'd heard ideas they thought they +   could use immediately in their research. + +Question #2 + +   Almost everyone listed ideas they'd seen presented at the conference +   which were new to them. + +   It is clear that which new ideas were important was a matter of +   perspective - the workshop membership was chosen to represent a broad +   spectrum of specialties, and people in different specialities were +   intrigued by different ideas.  However, there were some general +   themes raised in many questionnaires: + + +      (1)  Limitations of our traffic models.  This particular subject +           was mentioned, in some form, on many forms.  The attendees + + + +Partridge                                                      [Page 21] + +RFC 1152                  IRSG Workshop Report                April 1990 + + +           generally felt we didn't understand how network traffic would +           behave on a gigabit network, and were concerned that people +           might design (or worse, standardize) network protocols for +           high speed networks that would prove inadequate when used +           with real traffic.  Questions were raised about closed-loop +           vs. open-loop traffic models and the effects of varying types +           of service.  This concern led several people to encourage the +           construction of a high-speed testbed, so we can actually see +           some real traffic. + +      (2)  Congestion control.  Despite the limitations of our traffic +           models, respondents felt that congestion control at both +           switching elements and network wide was going to be even more +           important than today, due to the wider mix of traffic that +           will appear on gigabit networks.  Most forms mentioned at +           least one of the congestion control talks as a containing a +           new idea.  The talks by Victor Frost, Jamal Golestani and +           Scott Shenker received the most praise.  Some attendees were +           also interested in methods for keeping the lower-layer +           switching fabric from getting congested and mentioned the +           talks by Robinson and Maxemchuk as of interest. + +      (3)  Effects of fixed delay.  While the reviews were by no means +           unanimous, many people had come to the conclusion that the +           most serious problem in gigabit networking was not bandwidth, +           but delay.  The workshop looked at this issue in several +           guises, and most people listed at least one aspect of fixed +           delay as a challenging new problem.  Questions that people +           mentioned include: + + +    -    How to avoid a one round-trip set up delay, for less than one +         round-trip time's worth of data? + +    -    How to recover from error without retransmission (and thus +         additional network delays)?  Several people were intrigued by +         Nachum Shacham's work on error detection and recovery. + +    -    Should we use window flow-control or rate-based flow control +         when delays were long? + +    -    Can we modify the idea of remote procedure calls to deal with +         the fact that delays are relatively long? + +A couple of attendees noted that while some of these problems looked +similar to those of today, the subtle differences caused by operating a +network at gigabit speeds led them to believe the actual approaches to +solving these problems would have to be very different from those of + + + +Partridge                                                      [Page 22] + +RFC 1152                  IRSG Workshop Report                April 1990 + + +today. + +Security Considerations + +   Security issues are not discussed in this memo. + +Author's Address + +   Craig Partridge +   Bolt Beranek and Newman Inc. +   50 Moulton Street +   Cambridge, MA 02138 + +   Phone: (617) 873-2459 + +   EMail: craig@BBN.COM + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +Partridge                                                      [Page 23] +
\ No newline at end of file  |