summaryrefslogtreecommitdiff
path: root/doc/rfc/rfc1254.txt
diff options
context:
space:
mode:
Diffstat (limited to 'doc/rfc/rfc1254.txt')
-rw-r--r--doc/rfc/rfc1254.txt1403
1 files changed, 1403 insertions, 0 deletions
diff --git a/doc/rfc/rfc1254.txt b/doc/rfc/rfc1254.txt
new file mode 100644
index 0000000..8cc3b10
--- /dev/null
+++ b/doc/rfc/rfc1254.txt
@@ -0,0 +1,1403 @@
+
+
+
+
+
+
+Network Working Group A. Mankin
+Request for Comments: 1254 MITRE
+ K. Ramakrishnan
+ Digital Equipment Corporation
+ Editors
+ August 1991
+
+
+ Gateway Congestion Control Survey
+
+Status of this Memo
+
+ This memo provides information for the Internet community. It is a
+ survey of some of the major directions and issues. It does not
+ specify an Internet standard. Distribution of this memo is
+ unlimited.
+
+Abstract
+
+ The growth of network intensive Internet applications has made
+ gateway congestion control a high priority. The IETF Performance and
+ Congestion Control Working Group surveyed and reviewed gateway
+ congestion control and avoidance approaches. The purpose of this
+ paper is to present our review of the congestion control approaches,
+ as a way of encouraging new discussion and experimentation. Included
+ in the survey are Source Quench, Random Drop, Congestion Indication
+ (DEC Bit), and Fair Queueing. The task remains for Internet
+ implementors to determine and agree on the most effective mechanisms
+ for controlling gateway congestion.
+
+1. Introduction
+
+ Internet users regularly encounter congestion, often in mild forms.
+ However, severe congestion episodes have been reported also; and
+ gateway congestion remains an obstacle for Internet applications such
+ as scientific supercomputing data transfer. The need for Internet
+ congestion control originally became apparent during several periods
+ of 1986 and 1987, when the Internet experienced the "congestion
+ collapse" condition predicted by Nagle [Nag84]. A large number of
+ widely dispersed Internet sites experienced simultaneous slowdown or
+ cessation of networking services for prolonged periods. BBN, the
+ firm responsible for maintaining the then backbone of the Internet,
+ the ARPANET, responded to the collapse by adding link capacity
+ [Gar87].
+
+ Much of the Internet now uses as a transmission backbone the National
+ Science Foundation Network (NSFNET). Extensive monitoring and
+ capacity planning are being done for the NSFNET backbone; still, as
+
+
+
+Performance and Congestion Control Working Group [Page 1]
+
+RFC 1254 Gateway Congestion Control Survey August 1991
+
+
+ the demand for this capacity grows, and as resource-intensive
+ applications such as wide-area file system management [Sp89]
+ increasingly use the backbone, effective congestion control policies
+ will be a critical requirement.
+
+ Only a few mechanisms currently exist in Internet hosts and gateways
+ to avoid or control congestion. The mechanisms for handling
+ congestion set forth in the specifications for the DoD Internet
+ protocols are limited to:
+
+ Window flow control in TCP [Pos81b], intended primarily for
+ controlling the demand on the receiver's capacity, both in terms
+ of processing and buffers.
+
+ Source quench in ICMP, the message sent by IP to request that a
+ sender throttle back [Pos81a].
+
+ One approach to enhancing Internet congestion control has been to
+ overlay the simple existing mechanisms in TCP and ICMP with more
+ powerful ones. Since 1987, the TCP congestion control policy, Slow-
+ start, a collection of several algorithms developed by Van Jacobson
+ and Mike Karels [Jac88], has been widely adopted. Successful Internet
+ experiences with Slow-start led to the Host Requirements RFC [HREQ89]
+ classifying the algorithms as mandatory for TCP. Slow-start modifies
+ the user's demand when congestion reaches such a point that packets
+ are dropped at the gateway. By the time such overflows occur, the
+ gateway is congested. Jacobson writes that the Slow-start policy is
+ intended to function best with a complementary gateway policy
+ [Jac88].
+
+1.1 Definitions
+
+ The characteristics of the Internet that we are interested in include
+ that it is, in general, an arbitrary mesh-connected network. The
+ internetwork protocol is connectionless. The number of users that
+ place demands on the network is not limited by any explicit
+ mechanism; no reservation of resources occurs and transport layer
+ set-ups are not disallowed due to lack of resources. A path from a
+ source to destination host may have multiple hops, through several
+ gateways and links. Paths through the Internet may be heterogeneous
+ (though homogeneous paths also exist and experience congestion).
+ That is, links may be of different speeds. Also, the gateways and
+ hosts may be of different speeds or may be providing only a part of
+ their processing power to communication-related activity. The
+ buffers for storing information flowing through Internet gateways are
+ finite. The nature of the internet protocol is to drop packets when
+ these buffers overflow.
+
+
+
+
+Performance and Congestion Control Working Group [Page 2]
+
+RFC 1254 Gateway Congestion Control Survey August 1991
+
+
+ Gateway congestion arises when the demand for one or more of the
+ resources of the gateway exceeds the capacity of that resource. The
+ resources include transmission links, processing, and space used for
+ buffering. Operationally, uncongested gateways operate with little
+ queueing on average, where the queue is the waiting line for a
+ particular resource of the gateway. One commonly used quantitative
+ definition [Kle79] for when a resource is congested is when the
+ operating point is greater than the point at which resource power is
+ maximum, where resource power is defined as the ratio of throughput
+ to delay (See Section 2.2). At this operating point, the average
+ queue size is close to one, including the packet in service. Note
+ that this is a long-term average queue size. Several definitions
+ exist for the timescale of averaging for congestion detection and
+ control, such as dominant round-trip time and queue regeneration
+ cycle (see Section 2.1).
+
+ Mechanisms for handling congestion may be divided into two
+ categories, congestion recovery and congestion avoidance. Congestion
+ recovery tries to restore an operating state, when demand has already
+ exceeded capacity. Congestion avoidance is preventive in nature. It
+ tries to keep the demand on the network at or near the point of
+ maximum power, so that congestion never occurs. Without congestion
+ recovery, the network may cease to operate entirely (zero
+ throughput), whereas the Internet has been operating without
+ congestion avoidance for a long time. Overall performance may
+ improve with an effective congestion avoidance mechanism. Even if
+ effective congestion avoidance was in place, congestion recovery
+ schemes would still be required, to retain throughput in the face of
+ sudden changes (increase of demand, loss of resources) that can lead
+ to congestion.
+
+ In this paper, the term "user" refers to each individual transport
+ (TCP or another transport protocol) entity. For example, a TCP
+ connection is a "user" in this terminology. The terms "flow" and
+ "stream" are used by some authors in the same sense. Some of the
+ congestion control policies discussed in this paper, such as
+ Selective Feedback Congestion Indication and Fair Queueing aggregate
+ multiple TCP connections from a single host (or between a source
+ host-destination host pair) as a virtual user.
+
+ The term "cooperating transport entities" will be defined as a set of
+ TCP connections (for example) which follow an effective method of
+ adjusting their demand on the Internet in response to congestion.
+ The most restrictive interpretation of this term is that the
+ transport entities follow identical algorithms for congestion control
+ and avoidance. However, there may be some variation in these
+ algorithms. The extent to which heterogeneous end-system congestion
+ control and avoidance may be accommodated by gateway policies should
+
+
+
+Performance and Congestion Control Working Group [Page 3]
+
+RFC 1254 Gateway Congestion Control Survey August 1991
+
+
+ be a subject of future research. The role played in Internet
+ performance of non-cooperating transport entities is discussed in
+ Section 5.
+
+1.2 Goals and Scope of This Paper
+
+ The task remains for Internet implementors to determine effective
+ mechanisms for controlling gateway congestion. There has been
+ minimal common practice on which to base recommendations for Internet
+ gateway congestion control. In this survey, we describe the
+ characteristics of one experimental gateway congestion management
+ policy, Random Drop, and several that are better-known: Source
+ Quench, Congestion Indication, Selective Feedback Congestion
+ Indication, and Fair Queueing, both Bit-Round and Stochastic. A
+ motivation for documenting Random Drop is that it has as primary
+ goals low overhead and suitability for scaling up for Internets with
+ higher speed links. Both of these are important goals for future
+ gateway implementations that will have fast links, fast processors,
+ and will have to serve large numbers of interconnected hosts.
+
+ The structure of this paper is as follows. First, we discuss
+ performance goals, including timescale and fairness considerations.
+ Second, we discuss the gateway congestion control policies. Random
+ Drop is sketched out, with a recommendation for using it for
+ congestion recovery and a separate section on its use as congestion
+ avoidance. Third, since gateway congestion control in itself does
+ not change the end-systems' demand, we briefly present the effective
+ responses to these policies by two end-system congestion control
+ schemes, Slow-start and End-System Congestion Indication. Among our
+ conclusions, we address the issues of transport entities that do not
+ cooperate with gateway congestion control. As an appendix, because
+ of the potential interactions with gateway congestion policies, we
+ report on a scheme to help in controlling the performance of Internet
+ gateways to connection-oriented subnets (in particular, X.25).
+
+ Resources in the current Internet are not charged to users of them.
+ Congestion avoidance techniques cannot be expected to help when users
+ increase beyond the capacity of the underlying facilities. There are
+ two possible solutions for this, increase the facilities and
+ available bandwidth, or forcibly reduce the demand. When congestion
+ is persistent despite implemented congestion control mechanisms,
+ administrative responses are needed. These are naturally not within
+ the scope of this paper. Also outside the scope of this paper are
+ routing techniques that may be used to relocate demand away from
+ congested individual resources (e.g., path-splitting and load-
+ balancing).
+
+
+
+
+
+Performance and Congestion Control Working Group [Page 4]
+
+RFC 1254 Gateway Congestion Control Survey August 1991
+
+
+2. Performance Goals
+
+ To be able to discuss design and use of various mechanisms for
+ improving Internetwork performance, we need to have clear performance
+ goals for the operation of gateways and sets of end-systems.
+ Internet experience shows that congestion control should be based on
+ adaptive principles; this requires efficient computation of metrics
+ by algorithms for congestion control. The first issue is that of the
+ interval over which these metrics are estimated and/or measured.
+
+2.1 Interval for Measurement/Estimation of Performance Metrics
+
+ Network performance metrics may be distorted if they are computed
+ over intervals that are too short or too long relative to the dynamic
+ characteristics of the network. For instance, within a small
+ interval, two FTP users with equal paths may appear to have sharply
+ different demands, as an effect of brief, transient fluctuations in
+ their respective processing. An overly long averaging interval
+ results in distortions because of the changing number of users
+ sharing the resource measured during the time. It is similarly
+ important for congestion control mechanisms exerted at end systems to
+ find an appropriate interval for control.
+
+ The first approach to the monitoring, or averaging, interval for
+ congestion control is one based on round-trip times. The rationale
+ for it is as follows: control mechanisms must adapt to changes in
+ Internet congestion as quickly as possible. Even on an uncongested
+ path, changed conditions will not be detected by the sender faster
+ than a round-trip time. The effect of a sending end-system's control
+ will also not be seen in less than a round-trip time in the entire
+ path as well as at the end systems. For the control mechanism to be
+ adaptive, new information on the path is needed before making a
+ modification to the control exerted. The statistics and metrics used
+ in congestion control must be able to provide information to the
+ control mechanism so that it can make an informed decision.
+ Transient information which may be obsolete before a change is made
+ by the end-system should be avoided. This implies the
+ monitoring/estimating interval is one lasting one or more round
+ trips. The requirements described here give bounds on:
+
+ How short an interval: not small enough that obsolete information
+ is used for control;
+
+ How long: not more than the period at which the end-system makes
+ changes.
+
+ But, from the point of view of the gateway congestion control policy,
+ what is a round-trip time? If all the users of a given gateway have
+
+
+
+Performance and Congestion Control Working Group [Page 5]
+
+RFC 1254 Gateway Congestion Control Survey August 1991
+
+
+ the same path through the Internet, they also have the same round-
+ trip time through the gateway. But this is rarely the case.
+
+ A meaningful interval must be found for users with both short and
+ long paths. Two approaches have been suggested for estimating this
+ dynamically, queue regeneration cycle and frequency analysis.
+
+ Use of the queue regeneration cycle has been described as part of the
+ Congestion Indication policy. The time period used for averaging
+ here begins when a resource goes from the idle to busy state. The
+ basic interval for averaging is a "regeneration cycle" which is in
+ the form of busy and idle intervals. Because an average based on a
+ single previous regeneration may become old information, the
+ recommendation in [JRC87] is to average over the sum of two
+ intervals, that is, the previous (busy and idle) period, and the time
+ since the beginning of the current busy period.
+
+ If the gateway users are window-based transport entities, it is
+ possible to see how the regeneration interval responds to their
+ round-trip times. If a user with a long round-trip time has the
+ dominant traffic, the queue length may be zero only when that user is
+ awaiting acknowledgements. Then the users with short paths will
+ receive gateway congestion information that is averaged over several
+ of their round-trip times. If the short path traffic dominates the
+ activity in the gateway, i.e., the connections with shorter round-
+ trip times are the dominant users of the gateway resources, then the
+ regeneration interval is shorter and the information communicated to
+ them can be more timely. In this case, users with longer paths
+ receive, in one of their round-trip times, multiple samples of the
+ dominant traffic; the end system averaging is based on individual
+ user's intervals, so that these multiple samples are integrated
+ appropriately for these connections with longer paths.
+
+ The use of frequency analysis has been described by [Jac89]. In this
+ approach, the gateway congestion control is done at intervals based
+ on spectral analysis of the traffic arrivals. It is possible for
+ users to have round-trip times close to each other, but be out of
+ phase from each other. A spectral analysis algorithm detects this.
+ Otherwise, if multiple round-trip times are significant, multiple
+ intervals will be identified. Either one of these will be
+ predominant, or several will be comparable. An as yet difficult
+ problem for the design of algorithms accomplishing this technique is
+ the likelihood of "locking" to the frequency of periodic traffic of
+ low intensity, such as routing updates.
+
+
+
+
+
+
+
+Performance and Congestion Control Working Group [Page 6]
+
+RFC 1254 Gateway Congestion Control Survey August 1991
+
+
+2.2 Power and its Relationship to the Operating Point
+
+ Performance goals for a congestion control/avoidance strategy embody
+ a conflict in that they call for as high a throughput as possible,
+ with as little delay as possible. A measure that is often used to
+ reflect the tradeoff between these goals is power, the ratio of
+ throughput to delay. We would like to maximize the value of power
+ for a given resource. In the standard expression for power,
+
+ Power = (Throughput^alpha)/Delay
+
+ the exponent alpha is chosen for throughput, based on the relative
+ emphasis placed on throughput versus delay: if throughput is more
+ important, then a value of A alpha greater than one is chosen. If
+ throughput and delay are equally important (e.g., both bulk transfer
+ traffic and interactive traffic are equally important), then alpha
+ equal to one is chosen. The operating point where power is maximized
+ is the "knee" in the throughput and delay curves. It is desirable
+ that the operating point of the resource be driven towards the knee,
+ where power is maximized. A useful property of power is that it is
+ decreasing whether the resource is under- or over-utilized relative
+ to the knee.
+
+ In an internetwork comprising nodes and links of diverse speeds and
+ utilization, bottlenecks or concentrations of demand may form. Any
+ particular user can see a single bottleneck, which is the slowest or
+ busiest link or gateway in the path (or possibly identical "balanced"
+ bottlenecks). The throughput that the path can sustain is limited by
+ the bottleneck. The delay for packets through a particular path is
+ determined by the service times and queueing at each individual hop.
+ The queueing delay is dominated by the queueing at the bottleneck
+ resource(s). The contribution to the delay over other hops is
+ primarily the service time, although the propagation delay over
+ certain hops, such as a satellite link, can be significant. We would
+ like to operate all shared resources at their knee and maximize the
+ power of every user's bottleneck.
+
+ The above goal underscores the significance of gateway congestion
+ control. If techniques can be found to operate gateways at their
+ resource knee, it can improve Internet performance broadly.
+
+2.3 Fairness
+
+ We would like gateways to allocate resources fairly to users. A
+ concept of fairness is only relevant when multiple users share a
+ gateway and their total demand is greater than its capacity. If
+ demands were equal, a fair allocation of the resource would be to
+ provide an equal share to each user. But even over short intervals,
+
+
+
+Performance and Congestion Control Working Group [Page 7]
+
+RFC 1254 Gateway Congestion Control Survey August 1991
+
+
+ demands are not equal. Identifying the fair share of the resource
+ for the user becomes hard. Having identified it, it is desirable to
+ allocate at least this fair share to each user. However, not all
+ users may take advantage of this allocation. The unused capacity can
+ be given to other users. The resulting final allocation is termed a
+ maximally fair allocation. [RJC87] gives a quantitative method for
+ comparing the allocation by a given policy to the maximally fair
+ allocation.
+
+ It is known that the Internet environment has heterogeneous transport
+ entities, which do not follow the same congestion control policies
+ (our definition of cooperating transports). Then, the controls given
+ by a gateway may not affect all users and the congestion control
+ policy may have unequal effects. Is "fairness" obtainable in such a
+ heterogeneous community? In Fair Queueing, transport entities with
+ differing congestion control policies can be insulated from each
+ other and each given a set share of gateway bandwidth.
+
+ It is important to realize that since Internet gateways cannot refuse
+ new users, fairness in gateway congestion control can lead to all
+ users receiving small (sub-divided) amounts of the gateway resources
+ inadequate to meet their performance requirements. None of the
+ policies described in this paper currently addresses this. Then,
+ there may be policy reasons for unequal allocation of the gateway
+ resources. This has been addressed by Bit-Round Fair Queueing.
+
+2.4 Network Management
+
+ Network performance goals may be assessed by measurements in either
+ the end-system or gateway frame of reference. Performance goals are
+ often resource-centered and the measurement of the global performance
+ of "the network," is not only difficult to measure but is also
+ difficult to define. Resource-centered metrics are more easily
+ obtained, and do not require synchronization. That resource-centered
+ metrics are appropriate ones for use in optimization of power is
+ shown by [Jaf81].
+
+ It would be valuable for the goal of developing effective gateway
+ congestion handling if Management Information Base (MIB) objects
+ useful for evaluating gateway congestion were developed. The
+ reflections on the control interval described above should be applied
+ when network management applications are designed for this purpose.
+ In particular, obtaining an instantaneous queue length from the
+ managed gateway is not meaningful for the purposes of congestion
+ management.
+
+
+
+
+
+
+Performance and Congestion Control Working Group [Page 8]
+
+RFC 1254 Gateway Congestion Control Survey August 1991
+
+
+3. Gateway Congestion Control Policies
+
+ There have been proposed a handful of approaches to dealing with
+ congestion in the gateway. Some of these are Source Quench, Random
+ Drop, Congestion Indication, Selective Feedback Congestion
+ Indication, Fair Queueing, and Bit-Round Fair Queueing. They differ
+ in whether they use a control message, and indeed, whether they view
+ control of the end-systems as necessary, but none of them in itself
+ lowers the demand of users and consequent load on the network. End-
+ system policies that reduce demand in conjunction with gateway
+ congestion control are described in Section 4.
+
+3.1 Source Quench
+
+ The method of gateway congestion control currently used in the
+ Internet is the Source Quench message of the RFC-792 [Pos81a]
+ Internet Control Message Protocol (ICMP). When a gateway responds to
+ congestion by dropping datagrams, it may send an ICMP Source Quench
+ message to the source of the dropped datagram. This is a congestion
+ recovery policy.
+
+ The Gateway Requirements RFC, RFC-1009 [GREQ87], specifies that
+ gateways should only send Source Quench messages with a limited
+ frequency, to conserve CPU resources during the time of heavy load.
+ We note that operating the gateway for long periods under such loaded
+ conditions should be averted by a gateway congestion control policy.
+ A revised Gateway Requirements RFC is being prepared by the IETF.
+
+ Another significant drawback of the Source Quench policy is that its
+ details are discretionary, or, alternatively, that the policy is
+ really a family of varied policies. Major Internet gateway
+ manufacturers have implemented a variety of source quench
+ frequencies. It is impossible for the end-system user on receiving a
+ Source Quench to be certain of the circumstances in which it was
+ issued. This makes the needed end-system response problematic: is
+ the Source Quench an indication of heavy congestion, approaching
+ congestion, a burst causing massive overload, or a burst slightly
+ exceeding reasonable load?
+
+ To the extent that gateways drop the last arrived datagram on
+ overload, Source Quench messages may be distributed unfairly. This
+ is because the position at the end of the queue may be unfairly often
+ occupied by the packets of low demand, intermittent users, since
+ these do not send regular bursts of packets that can preempt most of
+ the queue space.
+
+ [Fin89] developed algorithms for when to issue Source Quench and for
+ responding to it with a rate-reduction in the IP layer on the sending
+
+
+
+Performance and Congestion Control Working Group [Page 9]
+
+RFC 1254 Gateway Congestion Control Survey August 1991
+
+
+ host. The system controls end-to-end performance of connections in a
+ manner similar to the congestion avoidance portion of Slow-start TCP
+ [Jac88].
+
+3.2 Random Drop
+
+ Random Drop is a gateway congestion control policy intended to give
+ feedback to users whose traffic congests the gateway by dropping
+ packets on a statistical basis. The key to this policy is the
+ hypothesis that a packet randomly selected from all incoming traffic
+ will belong to a particular user with a probability proportional to
+ the average rate of transmission of that user. Dropping a randomly
+ selected packet results in users which generate much traffic having
+ a greater number of packets dropped compared with those generating
+ little traffic. The selection of packets to be dropped is completely
+ uniform. Therefore, a user who generates traffic of an amount below
+ the "fair share" (as defined in Section 2.3) may also experience a
+ small amount of packet loss at a congested gateway. This character of
+ uniformity is in fact a primary goal that Random Drop attempts to
+ achieve.
+
+ The other primary goal that Random Drop attempts to achieve is a
+ theoretical overhead which is scaled to the number of shared
+ resources in the gateway rather than the number of its users. If a
+ gateway congestion algorithm has more computation the more users
+ there are, this can lead to processing demands that are higher as
+ congestion increases. Also the low-overhead goal of Random Drop
+ addresses concerns about the scale of gateway processing that will be
+ required in the mid-term Internet as gateways with fast processors
+ and links are shared by very large active sets of users.
+
+3.2.1 For Congestion Recovery
+
+ Random Drop has been proposed as an improvement to packet dropping at
+ the operating point where the gateway's packet buffers overflow.
+ This is using Random Drop strictly as a congestion recovery
+ mechanism.
+
+ In Random Drop congestion recovery, instead of dropping the last
+ packet to arrive at the queue, a packet is selected randomly from the
+ queue. Measurements of an implementation of Random Drop Congestion
+ Recovery [Man90] showed that a user with a low demand, due to a
+ longer round-trip time path than other users of the gateway, had a
+ higher drop rate with RDCR than without. The throughput accorded to
+ users with the same round-trip time paths was nearly equal with RDCR
+ as compared to without it. These results suggest that RDCR should be
+ avoided unless it is used within a scheme that groups traffic more or
+ less by round-trip time.
+
+
+
+Performance and Congestion Control Working Group [Page 10]
+
+RFC 1254 Gateway Congestion Control Survey August 1991
+
+
+3.2.2 For Congestion Avoidance
+
+ Random Drop is also proposed as a congestion avoidance policy
+ [Jac89]. The intent is to initiate dropping packets when the gateway
+ is anticipated to become congested and remain so unless there is some
+ control exercised. This implies selection of incoming packets to
+ be randomly dropped at a rate derived from identifying the level of
+ congestion at the gateway. The rate is the number of arrivals
+ allowed between drops. It depends on the current operating point and
+ the prediction of congestion.
+
+ A part of the policy is to determine that congestion will soon occur
+ and that the gateway is beginning to operate beyond the knee of the
+ power curve. With a suitably chosen interval (Section 2.1), the
+ number of packets from each individual user in a sample over that
+ interval is proportional to each user's demand on the gateway. Then,
+ dropping one or more random packets indicates to some user(s) the
+ need to reduce the level of demand that is driving the gateway beyond
+ the desired operating point. This is the goal that a policy of
+ Random Drop for congestion avoidance attempts to achieve.
+
+ There are several parameters to be determined for a Random Drop
+ congestion avoidance policy. The first is an interval, in terms of
+ number of packet arrivals, over which packets are dropped with
+ uniform probability. For instance, in a sample implementation, if
+ this interval spanned 2000 packet arrivals, and a suitable
+ probability of drop was 0.001, then two random variables would be
+ drawn in a uniform distribution in the range of 1 to 2,000. The
+ values drawn would be used by counting to select the packets dropped
+ at arrival. The second parameter is the value for the probability of
+ drop. This parameter would be a function of an estimate of the
+ number of users, their appropriate control intervals, and possibly
+ the length of time that congestion has persisted. [Jac89] has
+ suggested successively increasing the probability of drop when
+ congestion persists over multiple control intervals. The motivation
+ for increasing the packet drop probability is that the implicit
+ estimate of the number of users and random selection of their packets
+ to drop does not guarantee that all users have received enough
+ signals to decrease demand. Increasing the probability of drop
+ increases the probability that enough feedback is provided.
+ Congestion detection is also needed in Random Drop congestion
+ avoidance, and could be implemented in a variety of ways. The
+ simplest is a static threshold, but dynamically averaged measures of
+ demand or utilization are suggested.
+
+ The packets dropped in Random Drop congestion avoidance would not be
+ from the initial inputs to the gateway. We suggest that they would
+ be selected only from packets destined for the resource which is
+
+
+
+Performance and Congestion Control Working Group [Page 11]
+
+RFC 1254 Gateway Congestion Control Survey August 1991
+
+
+ predicted to be approaching congestion. For example, in the case of
+ a gateway with multiple outbound links, access to each individual
+ link is treated as a separate resource, the Random Drop policy is
+ applied at each link independently. Random Drop congestion avoidance
+ would provide uniform treatment of all cooperating transport users,
+ even over individual patterns of traffic multiplexed within one
+ user's stream. There is no aggregation of users.
+
+ Simulation studies [Zha89], [Has90] have presented evidence that
+ Random Drop is not fair across cooperating and non-cooperating
+ transport users. A transport user whose sending policies included
+ Go-Back-N retransmissions and did not include Slow-start received an
+ excessive share of bandwidth from a simple implementation of Random
+ Drop. The simultaneously active Slow-start users received unfairly
+ low shares. Considering this, it can be seen that when users do not
+ respond to control, over a prolonged period, the Random Drop
+ congestion avoidance mechanism would have an increased probability of
+ penalizing users with lower demand. Their packets dropped, these
+ users exert the controls leading to their giving up bandwidth.
+
+ Another problem can be seen to arise in Random Drop [She89] across
+ users whose communication paths are of different lengths. If the
+ path spans congested resources at multiple gateways, then the user's
+ probability of receiving an unfair drop and subsequent control (if
+ cooperating) is exponentially increased. This is a significant
+ scaling problem.
+
+ Unequal paths cause problems for other congestion avoidance policies
+ as well. Selective Feedback Congestion Indication was devised to
+ enhance Congestion Indication specifically because of the problem of
+ unequal paths. In Fair Queueing by source-destination pairs, each
+ path gets its own queue in all the gateways.
+
+3.3 Congestion Indication
+
+ The Congestion Indication policy is often referred to as the DEC Bit
+ policy. It was developed at DEC [JRC87], originally for the Digital
+ Network Architecture (DNA). It has also been specified for the
+ congestion avoidance of the ISO protocols TP4 and CLNP [NIST88].
+ Like Source Quench, it uses explicit communications from the
+ congested gateway to the user. However, to use the lowest possible
+ network resources for indicating congestion, the information is
+ communicated in a single bit, the Congestion Experienced Bit, set in
+ the network header of the packets already being forwarded by a
+ gateway. Based on the condition of this bit, the end-system user
+ makes an adjustment to the sending window. In the NSP transport
+ protocol of DECNET, the source makes an adjustment to its window; in
+ the ISO transport protocol, TP4, the destination makes this
+
+
+
+Performance and Congestion Control Working Group [Page 12]
+
+RFC 1254 Gateway Congestion Control Survey August 1991
+
+
+ adjustment in the window offered to the sender.
+
+ This policy attempts to avoid congestion by setting the bit whenever
+ the average queue length over the previous queue regeneration cycle
+ plus part of the current cycle is one or more. The feedback is
+ determined independently at each resource.
+
+3.4 Selective Feedback Congestion Indication
+
+ The simple Congestion Indication policy works based upon the total
+ demand on the gateway. The total number of users or the fact that
+ only a few of the users might be causing congestion is not
+ considered. For fairness, only those users who are sending more than
+ their fair share should be asked to reduce their load, while others
+ could attempt to increase where possible. In Selective Feedback
+ Congestion Indication, the Congestion Experienced Bit is used to
+ carry out this goal.
+
+ Selective Feedback works by keeping a count of the number of packets
+ sent by different users since the beginning of the queue averaging
+ interval. This is equivalent to monitoring their throughputs. Based
+ on the total throughput, a fair share for each user is determined and
+ the congestion bit is set, when congestion approaches, for the users
+ whose demand is higher than their fair share. If the gateway is
+ operating below the throughput-delay knee, congestion indications are
+ not set.
+
+ A min-max algorithm used to determine the fair share of capacity and
+ other details of this policy are described in [RJC87]. One parameter
+ to be computed is the capacity of each resource to be divided among
+ the users. This metric depends on the distribution of inter-arrival
+ times and packet sizes of the users. Attempting to determine these
+ in real time in the gateway is unacceptable. The capacity is instead
+ estimated from on the throughput seen when the gateway is operating
+ in congestion, as indicated by the average queue length. In
+ congestion (above the knee), the service rate of the gateway limits
+ its throughput. Multiplying the throughput obtained at this
+ operating point by a capacity factor (between 0.5 and 0.9) to adjust
+ for the distributions yields an acceptable capacity estimate in
+ simulations.
+
+ Selective Feedback Congestion Indication takes as input a vector of
+ the number of packets sent by each source-destination pair of end-
+ systems. Other alternatives include 1) destination address, 2)
+ input/output link, and 3) transport connection (source/destination
+ addresses and ports).
+
+ These alternatives give different granularities for fairness. In the
+
+
+
+Performance and Congestion Control Working Group [Page 13]
+
+RFC 1254 Gateway Congestion Control Survey August 1991
+
+
+ case where paths are the same or round-trip times of users are close
+ together, throughputs are equalized similarly by basing the selective
+ feedback on source or destination address. In fact, if the RTTs are
+ close enough, the simple congestion indication policy would result in
+ a fair allocation. Counts based on source/destination pairs ensure
+ that paths with different lengths and network conditions get a fair
+ throughput at the individual gateways. Counting packets based on
+ link pairs has a low overhead, but may result in unfairness to users
+ whose demand is below the fair share and are using a long path.
+ Counts based on transport layer connection identifiers, if this
+ information was available to Internet gateways, would make good
+ distinctions, since the differences of demand of different
+ applications and instances of applications would be separately
+ monitored.
+
+ Problems with Selective Feedback Congestion Indication include that
+ the gateway has significant processing to do. With the feasible
+ choice of aggregation at the gateway, unfairness across users within
+ the group is likely. For example, an interactive connection
+ aggregated with one or more bulk transfer connections will receive
+ congestion indications though its own use of the gateway resources is
+ very low.
+
+3.5 Fair Queueing
+
+ Fair Queueing is the policy of maintaining separate gateway output
+ queues for individual end-systems by source-destination pair. In the
+ policy as proposed by [Nag85], the gateway's processing and link
+ resources are distributed to the end-systems on a round-robin basis.
+ On congestion, packets are dropped from the longest queue. This
+ policy leads to equal allocations of resources to each source-
+ destination pair. A source-destination pair that demands more than a
+ fair share simply increases its own queueing delay and congestion
+ drops.
+
+3.5.1 Bit-Round Fair Queueing
+
+ An enhancement of Nagle Fair Queueing, the Bit-Round Fair Queuing
+ algorithm described and simulated by [DKS89] addresses several
+ shortcomings of Nagle's scheme. It computes the order of service to
+ packets using their lengths, with a technique that emulates a bit-
+ by-bit round-robin discipline, so that long packets do not get an
+ advantage over short ones. Otherwise the round-robin would be
+ unfair, for example, giving more bandwidth to hosts whose traffic is
+ mainly long packets than to hosts sourcing short packets.
+
+ The aggregation of users of a source-destination pair by Fair
+ Queueing has the property of grouping the users whose round-trips are
+
+
+
+Performance and Congestion Control Working Group [Page 14]
+
+RFC 1254 Gateway Congestion Control Survey August 1991
+
+
+ similar. This may be one reason that the combination of Bit-Round
+ Fair Queueing with Congestion Indication had particularly good
+ simulated performance in [DKS89].
+
+ The aggregation of users has the expected drawbacks, as well. A
+ TELNET user in a queue with an FTP user does not get delay benefits;
+ and host pairs with high volume of connections get treated the same
+ as a host pair with small number of connections and as a result gets
+ unfair services.
+
+ A problem can be mentioned about Fair Queueing, though it is related
+ to implementation, and perhaps not properly part of a policy
+ discussion. This is a concern that the resources (processing) used
+ for determining where to queue incoming packets would themselves be
+ subject to congestion, but not to the benefits of the Fair Queuing
+ discipline. In a situation where the gateway processor was not
+ adequate to the demands on it, the gateway would need an alternative
+ policy for congestion control of the queue awaiting processing.
+ Clever implementation can probably find an efficient way to route
+ packets to the queues they belong in before other input processing is
+ done, so that processing resources can be controlled, too. There is
+ in addition, the concern that bit-by-bit round FQ requires non-FCFS
+ queueing even within the same source destination pairs to allow for
+ fairness to different connections between these end systems.
+
+ Another potential concern about Fair Queueing is whether it can scale
+ up to gateways with very large source-destination populations. For
+ example, the state in a Fair Queueing implementation scales with the
+ number of active end-to-end paths, which will be high in backbone
+ gateways.
+
+3.5.2 Stochastic Fairness Queuing
+
+ Stochastic Fairness Queueing (SFQ) has been suggested as a technique
+ [McK90] to address the implementation issues relating to Fair
+ Queueing. The first overhead that is reduced is that of looking up
+ the source-destination address pair in an incoming packet and
+ determining which queue that packet will have to be placed in. SFQ
+ does not require as many memory accesses as Fair Queueing to place
+ the packet in the appropriate queue. SFQ is thus claimed to be more
+ amenable to implementation for high-speed networks [McK90].
+
+ SFQ uses a simple hash function to map from the source-destination
+ address pair to a fixed set of queues. Since the assignment of an
+ address pair to a queue is probabilistic, there is the likelihood of
+ multiple address pairs colliding and mapping to the same queue. This
+ would potentially degrade the additional fairness that is gained with
+ Fairness Queueing. If two or more address pairs collide, they would
+
+
+
+Performance and Congestion Control Working Group [Page 15]
+
+RFC 1254 Gateway Congestion Control Survey August 1991
+
+
+ continue to do so. To deal with the situation when such a collision
+ occurs, SFQ periodically perturbs the hash function so that these
+ address pairs will be unlikely to collide subsequently.
+
+ The price paid for achieving this implementation efficiency is that
+ SFQ requires a potentially large number of queues (we must note
+ however, that these queues may be organized orthogonally from the
+ buffers in which packets are stored. The buffers themselves may be
+ drawn from a common pool, and buffers in each queue organized as a
+ linked list pointed to from each queue header). For example, [McK90]
+ indicates that to get good, consistent performance, we may need to
+ have up to 5 to 10 times the number of active source-destination
+ pairs. In a typical gateway, this may require around 1000 to 2000
+ queues.
+
+ [McK90] reports simulation results with SFQ. The particular hash
+ function that is studied is using the HDLC's cyclic redundancy check
+ (CRC). The hash function is perturbed by multiplying each byte by a
+ sequence number in the range 1 to 255 before applying the CRC. The
+ metric considered is the standard deviation of the number of packets
+ output per source-destination pair. A perfectly fair policy would
+ have a standard deviation of zero and an unfair policy would have a
+ large standard deviation. In the example studied (which has up to 20
+ source-destination (s-d) pairs going through a single overloaded
+ gateway), SFQ with 1280 queues (i.e., 64 times the number of source-
+ destination pairs), approaches about 3 times the standard deviation
+ of Fairness Queueing. This must be compared to a FCFS queueing
+ discipline having a standard deviation which is almost 175 times the
+ standard deviation of Fairness Queueing.
+
+ It is conjectured in [McK90] that a good value for the interval in
+ between perturbations of the hash function appears to be in the area
+ between twice the queue flush time of the stochastic fairness queue
+ and one-tenth the average conversation time between a source-
+ destination pair.
+
+ SFQ also may alleviate the anticipated scaling problems that may be
+ an issue with Fair Queueing by not strictly requiring the number of
+ queues equal to the possible source-destination pairs that may be
+ presenting a load on a particular gateway. However, SFQ achieves this
+ property by trading off some of the fairness for implementation
+ efficiency.
+
+ [McK90] examines alternative strategies for implementation of Fair
+ Queueing and SFQ and estimates their complexity on common hardware
+ platforms (e.g., a Motorola 68020). It is suggested that mapping an
+ IP address pair may require around 24 instructions per packet for
+ Fair Queueing in the best case; in contrast SFQ requires 10
+
+
+
+Performance and Congestion Control Working Group [Page 16]
+
+RFC 1254 Gateway Congestion Control Survey August 1991
+
+
+ instructions in the worst case. The primary source of this gain is
+ that SFQ avoids a comparison of the s-d address pair on the packet to
+ the identity of the queue header. The relative benefit of SFQ over
+ Fair Queueing is anticipated to be greater when the addresses are
+ longer.
+
+ SFQ offers promising implemenatation benefits. However, the price to
+ be paid in terms of having a significantly larger number of queues
+ (and the consequent data structures and their management) than the
+ number of s-d pairs placing a load on the gateway is a concern. SFQ
+ is also attractive in that it may be used in concert with the DEC-bit
+ scheme for Selective Feedback Congestion Indication to provide
+ fairness as well as congestion avoidance.
+
+4. END-SYSTEM CONGESTION CONTROL POLICIES
+
+ Ideally in gateway congestion control, the end-system transport
+ entities adjust (decrease) their demand in response to control
+ exerted by the gateway. Schemes have been put in practice for
+ transport entities to adjust their demand dynamically in response to
+ congestion feedback. To accomplish this, in general, they call for
+ the user to gradually increase demand until information is received
+ that the load on the gateway is too high. In response to this
+ information, the user decreases load, then begins an exploratory
+ increases again. This cycle is repeated continuously, with the goal
+ that the total demand will oscillate around the optimal level.
+
+ We have already noted that a Slow-start connection may give up
+ considerable bandwidth when sharing a gateway with aggressive
+ transport entities. There is currently no way to enforce that end-
+ systems use a congestion avoidance policy, though the Host
+ Requirements RFC [HR89] has defined Slow-start as mandatory for TCP.
+ This adverse effect on Slow-start connections is mitigated by the
+ Fair Queueing policy. Our conclusions discuss further the
+ coexistence of different end-system strategies.
+
+ This section briefly presents two fielded transport congestion
+ control and avoidance schemes, Slow-start and End-System Congestion
+ Indication, and the responses by means of which they cooperate with
+ gateway policies. They both use the control paradigm described
+ above. Slow-start, as mentioned in Section 1, was developed by
+ [Jac88], and widely fielded in the BSD TCP implementation. End-
+ system Congestion Indication was developed by [JRC87]. It is fielded
+ in DEC's Digital Network Architecture, and has been specified as well
+ for ISO TP4 [NIST88].
+
+ Both Slow-start and End-system Congestion Indication view the
+ relationship between users and gateways as a control system. They
+
+
+
+Performance and Congestion Control Working Group [Page 17]
+
+RFC 1254 Gateway Congestion Control Survey August 1991
+
+
+ have feedback and control components, the latter further broken down
+ into a procedure bringing a new connection to equilibrium, and a
+ procedure to maintain demand at the proper level. They make use of
+ policies for increasing and decreasing the transport sender's window
+ size. These require the sender to follow a set of self-restraining
+ rules which dynamically adjust the send window whenever congestion is
+ sensed.
+
+ A predecessor of these, CUTE, developed by [Jai86], introduced the
+ use of retransmission timeouts as congestion feedback. The Slow-
+ start scheme was also designed to use timeouts in the same way. The
+ End-System policies for Congestion Indication use the Congestion
+ Experienced Bit delivered in the network header as the primary
+ feedback, but retransmission timeouts also provoke an end-system
+ congestion response.
+
+ In reliable transport protocols like TCP and TP4, the retransmission
+ timer must do its best to satisfy two conflicting goals. On one hand,
+ the timer must rapidly detect lost packets. And, on the other hand,
+ the timer must minimize false alarms. Since the retransmit timer is
+ used as a congestion signal in these end-system policies, it is all
+ the more important that timeouts reliably correspond to packet drops.
+ One important rule for retransmission is to avoid bad sampling in the
+ measurements that will be used in estimating the round-trip delay.
+ [KP87] describes techniques to ensure accurate sampling. The Host
+ Requirements RFC [HR89] makes these techniques mandatory for TCP.
+
+ The utilization of a resource can be defined as the ratio of its
+ average arrival rate to its mean service rate. This metric varies
+ between 0 and 1.0. In a state of congestion, one or more resources
+ (link, gateway buffer, gateway CPU) has a utilization approaching
+ 1.0. The average delay (round trip time) and its variance approach
+ infinity. Therefore, as the utilization of a network increases, it
+ becomes increasingly important to take into account the variance of
+ the round trip time in estimating it for the retransmission timeout.
+
+ The TCP retransmission timer is based on an estimate of the round
+ trip time. [Jac88] calls the round trip time estimator the single
+ most important feature of any protocol implementation that expects to
+ survive a heavy load. The retransmit timeout procedure in RFC-793
+ [Pos81b] includes a fixed parameter, beta, to account for the
+ variance in the delay. An estimate of round trip time using the
+ suggested values for beta is valid for a network which is at most 30%
+ utilized. Thus, the RFC-793 retransmission timeout estimator will
+ fail under heavy congestion, causing unnecessary retransmissions that
+ add to the load, and making congestive loss detection impossible.
+
+ Slow-start TCP uses the mean deviation as an estimate of the variance
+
+
+
+Performance and Congestion Control Working Group [Page 18]
+
+RFC 1254 Gateway Congestion Control Survey August 1991
+
+
+ to improve the estimate. As a rough view of what happens with the
+ Slow-start retransmission calculation, delays can change by
+ approximately two standard deviations without the timer going off in
+ a false alarm. The same method of estimation may be applicable to
+ TP4. The procedure is:
+
+ Error = Measured - Estimated
+ Estimated = Estimated + Gain_1 * Error
+ Deviation = Deviation + Gain_2 * (|Error| - Deviation)
+ Timeout = Estimated + 2 * Deviation
+
+ Where: Gain_1, Gain_2 are gain factors.
+
+4.1 Response to No Policy in Gateway
+
+ Since packets must be dropped during congestion because of the finite
+ buffer space, feedback of congestion to the users exists even when
+ there is no gateway congestion policy. Dropping the packets is an
+ attempt to recover from congestion, though it needs to be noted that
+ congestion collapse is not prevented by packet drops if end-systems
+ accelerate their sending rate in these conditions. The accurate
+ detection of congestive loss by all retransmission timers in the
+ end-systems is extremely important for gateway congestion recovery.
+
+4.2 Response to Source Quench
+
+ Given that a Source Quench message has ambiguous meaning, but usually
+ is issued for congestion recovery, the response of Slow-start to a
+ Source Quench is to return to the beginning of the cycle of increase.
+ This is an early response, since the Source Quench on overflow will
+ also lead to a retransmission timeout later.
+
+4.3 Response to Random Drop
+
+ The end-system's view of Random Drop is the same as its view of loss
+ caused by overflow at the gateway. This is a drawback of the use of
+ packet drops as congestion feedback for congestion avoidance: the
+ decrease policy on congestion feedback cannot be made more drastic
+ for overflows than for the drops the gateway does for congestion
+ avoidance. Slow-start responds rapidly to gateway feedback. In a
+ case where the users are cooperating (all Slow-start), a desired
+ outcome would be that this responsiveness would lead quickly to a
+ decreased probability of drop. There would be, as an ideal, a self-
+ adjusting overall amount of control.
+
+
+
+
+
+
+
+Performance and Congestion Control Working Group [Page 19]
+
+RFC 1254 Gateway Congestion Control Survey August 1991
+
+
+4.4 Response to Congestion Indication
+
+ Since the Congestion Indication mechanism attempts to keep gateways
+ uncongested, cooperating end-system congestion control policies need
+ not reduce demand as much as with other gateway policies. The
+ difference between the Slow-start response to packet drops and the
+ End-System Congestion Indication response to Congestion Experienced
+ Bits is primarily in the decrease policy. Slow-start decreases the
+ window to one packet on a retransmission timeout. End-System
+ Congestion Indication decreases the window by a fraction (for
+ instance, to 7/8 of the original value), when the Congestion
+ Experienced Bit is set in half of the packets in a sample spanning
+ two round-trip times (two windows full).
+
+4.5 Response to Fair Queuing
+
+ A Fair Queueing policy may issue control indications, as in the
+ simulated Bit-Round Fair Queueing with DEC Bit, or it may depend only
+ on the passive effects of the queueing. When the passive control is
+ the main effect (perhaps because most users are not responsive to
+ control indications), the behavior of retransmission timers will be
+ very important. If retransmitting users send more packets in
+ response to increases in their delay and drops, Fair Queueing will be
+ prone to degraded performance, though collapse (zero throughput for
+ all users) may be prevented for a longer period of time.
+
+5. Conclusions
+
+ The impact of users with excessive demand is a driving force as
+ proposed gateway policies come closer to implementation. Random Drop
+ and Congestion Indication can be fair only if the transport entities
+ sharing the gateway are all cooperative and reduce demand as needed.
+ Within some portions of the Internet, good behavior of end-systems
+ eventually may be enforced through administration. But for various
+ reasons, we can expect non-cooperating transports to be a persistent
+ population in the Internet. Congestion avoidance mechanisms will not
+ be deployed all at once, even if they are adopted as standards.
+ Without enforcement, or even with penalties for excessive demand,
+ some end-systems will never implement congestion avoidance
+ mechanisms.
+
+ Since it is outside the context of any of the gateway policies, we
+ will mention here a suggestion for a relatively small-scale response
+ to users which implement especially aggressive policies. This has
+ been made experimentally by [Jac89]. It would implement a low
+ priority queue, to which the majority of traffic is not routed. The
+ candidate traffic to be queued there would be identified by a cache
+ of recent recipients of whatever control indications the gateway
+
+
+
+Performance and Congestion Control Working Group [Page 20]
+
+RFC 1254 Gateway Congestion Control Survey August 1991
+
+
+ policy makes. Remaining in the cache over multiple control intervals
+ is the criterion for being routed to the low priority queue. In
+ approaching or established congestion, the bandwidth given to the
+ low-service queue is decreased.
+
+ The goal of end-system cooperation itself has been questioned. As
+ [She89] points out, it is difficult to define. A TCP implementation
+ that retransmits before it determines that has been loss indicated
+ and in a Go-Back-N manner is clearly non-cooperating. On the other
+ hand, a transport entity with selective acknowledgement may demand
+ more from the gateways than TCP, even while responding to congestion
+ in a cooperative way.
+
+ Fair Queueing maintains its control of allocations regardless of the
+ end-system congestion avoidance policies. [Nag85] and [DKS89] argue
+ that the extra delays and congestion drops that result from
+ misbehavior could work to enforce good end-system policies. Are the
+ rewards and penalties less sharply defined when one or more
+ misbehaving systems cause the whole gateway's performance to be less?
+ While the tax on all users imposed by the "over-users" is much less
+ than in gateways with other policies, how can it be made sufficiently
+ low?
+
+ In the sections on Selective Feedback Congestion Indication and Bit-
+ Round Fair Queueing we have pointed out that more needs to be done on
+ two particular fronts:
+
+ How can increased computational overhead be avoided?
+
+ The allocation of resources to source-destination pairs is, in
+ many scenarios, unfair to individual users. Bit-Round Fair
+ Queueing offers a potential administrative remedy, but even if it
+ is applied, how should the unequal allocations be propagated
+ through multiple gateways?
+
+ The first question has been taken up by [McK90].
+
+ Since Selective Feedback Congestion Indication (or Congestion
+ Indication used with Fair Queueing) uses a network bit, its use in
+ the Internet requires protocol support; the transition of some
+ portions of the Internet to OSI protocols may make such a change
+ attractive in the future. The interactions between heterogeneous
+ congestion control policies in the Internet will need to be explored.
+
+ The goals of Random Drop Congestion Avoidance are presented in this
+ survey, but without any claim that the problems of this policy can be
+ solved. These goals themselves, of uniform, dynamic treatment of
+ users (streams/flows), of low overhead, and of good scaling
+
+
+
+Performance and Congestion Control Working Group [Page 21]
+
+RFC 1254 Gateway Congestion Control Survey August 1991
+
+
+ characteristics in large and loaded networks, are significant.
+
+Appendix: Congestion and Connection-oriented Subnets
+
+ This section presents a recommendation for gateway implementation in
+ an areas that unavoidably interacts with gateway congestion control,
+ the impact of connection-oriented subnets, such as those based on
+ X.25.
+
+ The need to manage a connection oriented service (X.25) in order to
+ transport datagram traffic (IP) can cause problems for gateway
+ congestion control. Being a pure datagram protocol, IP provides no
+ information delimiting when a pair of IP entities need to establish a
+ session between themselves. The solution involves compromise among
+ delay, cost, and resources. Delay is introduced by call
+ establishment when a new X.25 SVC (Switched Virtual Circuit) needs to
+ be established, and also by queueing delays for the physical line.
+ Cost includes any charges by the X.25 network service provider.
+ Besides the resources most gateways have (CPU, memory, links), a
+ maximum supported or permitted number of virtual circuits may be in
+ contest.
+
+ SVCs are established on demand when an IP packet needs to be sent and
+ there is no SVC established or pending establishment to the
+ destination IP entity. Optionally, when cost considerations, and
+ sufficient numbers of unused virtual circuits allow, redundant SVCs
+ may be established between the same pair of IP entities. Redundant
+ SVCs can have the effect of improving performance when coping with
+ large end-to-end delay, small maximum packet sizes and small maximum
+ packet windows. It is generally preferred though, to negotiate large
+ packet sizes and packet windows on a single SVC. Redundant SVCs must
+ especially be discouraged when virtual circuit resources are small
+ compared with the number of destination IP entities among the active
+ users of the gateway.
+
+ SVCs are closed after some period of inactivity indicates that
+ communication may have been suspended between the IP entities. This
+ timeout is significant in the operation of the interface. Setting
+ the value too low can result in closing of the SVC even though the
+ traffic has not stopped. This results in potentially large delays
+ for the packets which reopen the SVC and may also incur charges
+ associated with SVC calls. Also, clearing of SVCs is, by definition,
+ nongraceful. If an SVC is closed before the last packets are
+ acknowledged, there is no guarantee of delivery. Packet losses are
+ introduced by this destructive close independent of gateway traffic
+ and congestion.
+
+ When a new circuit is needed and all available circuits are currently
+
+
+
+Performance and Congestion Control Working Group [Page 22]
+
+RFC 1254 Gateway Congestion Control Survey August 1991
+
+
+ in use, there is a temptation to pick one to close (possibly using
+ some Least Recently Used criterion). But if connectivity demands are
+ larger than available circuit resources, this strategy can lead to a
+ type of thrashing where circuits are constantly being closed and
+ reopened. In the worst case, a circuit is opened, a single packet
+ sent and the circuit closed (without a guarantee of packet delivery).
+ To counteract this, each circuit should be allowed to remain open a
+ minimum amount of time.
+
+ One possible SVC strategy is to dynamically change the time a circuit
+ will be allowed to remain open based on the number of circuits in
+ use. Three administratively controlled variables are used, a minimum
+ time, a maximum time and an adaptation factor in seconds per
+ available circuit. In this scheme, a circuit is closed after it has
+ been idle for a time period equal to the minimum plus the adaptation
+ factor times the number of available circuits, limited by the maximum
+ time. By administratively adjusting these variables, one has
+ considerable flexibility in adjusting the SVC utilization to meet the
+ needs of a specific gateway.
+
+Acknowledgements
+
+ This paper is the outcome of discussions in the Performance and
+ Congestion Control Working Group between April 1988 and July 1989.
+ Both PCC WG and plenary IETF members gave us helpful reviews of
+ earlier drafts. Several of the ideas described here were contributed
+ by the members of the PCC WG. The Appendix was written by Art
+ Berggreen. We are particularly thankful also to Van Jacobson, Scott
+ Shenker, Bruce Schofield, Paul McKenney, Matt Mathis, Geof Stone, and
+ Lixia Zhang for participation and reviews.
+
+References
+
+ [DKS89] Demers, A., Keshav, S., and S. Shenker, "Analysis and
+ Simulation of a Fair Queueing Algorithm", Proceedings of SIGCOMM '89.
+
+ [Fin89] Finn, G., "A Connectionless Congestion Control Algorithm",
+ Computer Communications Review, Vol. 19, No. 5, October 1989.
+
+ [Gar87] Gardner, M., "BBN Report on the ARPANET", Proceedings of the
+ McLean IETF, SRI-NIC IETF-87/3P, July 1987.
+
+ [GREQ87] Braden R., and J. Postel, "Requirements for Internet
+ Gateways", RFC 1009, USC/Information Sciences Institute, June 1987.
+
+ [HREQ89] Braden R., Editor, "Requirements for Internet Hosts --
+ Communications Layers", RFC 1122, Internet Engineering Task Force,
+ October 1989.
+
+
+
+Performance and Congestion Control Working Group [Page 23]
+
+RFC 1254 Gateway Congestion Control Survey August 1991
+
+
+ [Has90] Hashem, E., "Random Drop Congestion Control", M.S. Thesis,
+ Massachusetts Institute of Technology, Department of Computer
+ Science, 1990.
+
+ [Jac88] Jacobson, V., "Congestion Avoidance and Control", Proceedings
+ of SIGCOMM '88.
+
+ [Jac89] Jacobson, V., "Presentations to the IETF Performance and
+ Congestion Control Working Group".
+
+ [Jaf81] Jaffe, J., "Bottleneck Flow Control", IEEE Transactions on
+ Communications, COM-29(7), July, 1981.
+
+ [Jai86] Jain, R., "A Timeout-based Congestion Control Scheme for
+ Window Flow-controlled Networks", IEEE Journal on Selected Areas in
+ Communications, SAC-4(7), October 1986.
+
+ [JRC87] Jain, R., Ramakrishnan, K., and D. Chiu, "Congestion
+ Avoidance in Computer Networks With a Connectionless Network Layer",
+ Technical Report DEC-TR-506, Digital Equipment Corporation.
+
+ [Kle79] Kleinrock, L., "Power and Deterministic Rules of Thumb for
+ Probabilistic Problems in Computer Communications", Proceedings of
+ the ICC '79.
+
+ [KP87] Karn, P., and C. Partridge, "Improving Round Trip Estimates in
+ Reliable Transport Protocols", Proceedings of SIGCOMM '87.
+
+ [Man90] Mankin, A., "Random Drop Congestion Control", Proceedings of
+ SIGCOMM '90.
+
+ [McK90] McKenney, P., "Stochastic Fairness Queueing", Proceedings of
+ INFOCOM '90.
+
+ [Nag84] Nagle, J., "Congestion Control in IP/TCP Internetworks", RFC
+ 896, FACC Palo Alto, 6 January 1984.
+
+ [Nag85] Nagle, J., "On Packet Switches With Infinite Storage", RFC
+ 970, FACC Palo Alto, December 1985.
+
+ [NIST88] NIST, "Stable Implementation Agreements for OSI Protocols,
+ Version 2, Edition 1", National Institute of Standards and Technology
+ Special Publication 500-162, December 1988.
+
+ [Pos81a] Postel, J., "Internet Control Message Protocol - DARPA
+ Internet Program Protocol Specification", RFC-792, USC/Information
+ Sciences Institute, September 1981.
+
+
+
+
+Performance and Congestion Control Working Group [Page 24]
+
+RFC 1254 Gateway Congestion Control Survey August 1991
+
+
+ [Pos81b] Postel, J., "Transmission Control Protocol - DARPA Internet
+ Program Protocol Specification", RFC-793, DARPA, September 1981.
+
+ [RJC87] Ramakrishnan, K., Jain, R., and D. Chiu, "A Selective Binary
+ Feedback Scheme for General Topologies", Technical Report DEC-TR-510,
+ Digital Equipment Corporation.
+
+ [She89] Shenker, S., "Correspondence with the IETF Performance and
+ Congestion Control Working Group".
+
+ [Sp89] Spector, A., and M. Kazar, "Uniting File Systems", Unix
+ Review, Vol. 7, No. 3, March 1989.
+
+ [Zha89] Zhang, L., "A New Architecture for Packet Switching Network
+ Protocols", Ph.D Thesis, Massachusetts Institute of Technology,
+ Department of Computer Science, 1989.
+
+Security Considerations
+
+ Security issues are not discussed in this memo.
+
+Authors' Addresses
+
+ Allison Mankin
+ The MITRE Corporation
+ M/S W425
+ 7525 Colshire Drive
+ McLean, VA 22102
+
+ Email: mankin@gateway.mitre.org
+
+
+ K.K. Ramakrishnan
+ Digital Equipment Corporation
+ M/S LKG1-2/A19
+ 550 King Street
+ Littleton, MA 01754
+
+ Email: rama@kalvi.enet.dec.com
+
+
+
+
+
+
+
+
+
+
+
+
+Performance and Congestion Control Working Group [Page 25]
+ \ No newline at end of file