1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
|
Network Working Group H-W. Braun
Request for Comments: 1222 San Diego Supercomputer Center
Y. Rekhter
IBM T.J. Watson Research Center
May 1991
Advancing the NSFNET Routing Architecture
Status of this Memo
This RFC suggests improvements in the NSFNET routing architecture to
accommodate a more flexible interface to the Backbone clients. This
memo provides information for the Internet community. It does not
specify an Internet standard. Distribution of this memo is
unlimited.
Introduction
This memo describes the history of NSFNET Backbone routing and
outlines two suggested phases for further evolution of the Backbone's
routing interface. The intent is to provide a more flexible
interface for NSFNET Backbone service subscribers, by providing an
attachment option that is simpler and lower-cost than the current
one.
Acknowledgements
The authors would like to thank Scott Brim (Cornell University),
Bilal Chinoy (Merit), Elise Gerich (Merit), Paul Love (SDSC), Steve
Wolff (NSF), Bob Braden (ISI), and Joyce K. Reynolds (ISI) for their
review and constructive comments.
1. NSFNET Phase 1 Routing Architecture
In the first phase of the NSFNET Backbone, a 56Kbps infrastructure
utilized routers based on Fuzzball software [2]. The Phase 1
Backbone used the Hello Protocol for interior routing. At the
periphery of the Backbone, the client networks were typically
connected by using a gatedaemon ("gated") interface to translate
between the Backbone's Hello Protocol and the interior gateway
protocol (IGP) of the mid-level network.
Mid-level networks primarily used the Routing Information Protocol
(RIP) [3] for their IGP. The gatedaemon system acted as an interface
between the Hello and RIP environments. The overall appearance was
that the Backbone, mid-level networks, and the campus networks formed
a single routing system in which information was freely exchanged.
Braun & Rekhter [Page 1]
^L
RFC 1222 Advancing the NSFNET Routing Architecture May 1991
Network metrics were translated among the three network levels
(backbone, mid-level networks, and campuses).
With the development of the gatedaemon, sites were able to introduce
filtering based on IP network numbers. This process was controlled
by the staff at each individual site.
Once specific network routes were learned, the infrastructure
forwarded metric changes throughout the interconnected network. The
end-result was that a metric fluctuation on one end of the
interconnected network could permeate all the way to the other end,
crossing multiple network administrations. The frequency of metric
fluctuations within the Backbone itself was further increased when
event-driven updates (e.g., metric changes) were introduced. Later,
damping of the event driven updates lessened their frequency, but the
overall routing environment still appeared to be quite unstable.
Given that only limited tools and protocols were available to
engineer the flow of dynamic routing information, it was fairly easy
for routing loops to form. This was amplified as the topology became
more fully connected without insulation of routing components from
each other.
All six nodes of the Phase 1 Backbone were located at client sites,
specifically NSF funded supercomputer centers.
2. NSFNET Phase 2 Routing Architecture
The routing architecture for the second phase of the NSFNET Backbone,
implemented on T1 (1.5Mbps) lines, focused on the lessons learned in
the first NSFNET phase. This resulted in a strong decoupling of the
IGP environments of the backbone network and its attached clients
[5]. Specifically, each of the administrative entities was able to
use its own IGP in any way appropriate for the specific network. The
interface between the backbone network and its attached client was
built by means of exterior routing, initially via the Exterior
Gateway Protocol (EGP) [1,4].
EGP improved provided routing isolation in two ways. First, EGP
signals only up/down transitions for individual network numbers, not
the fluctuations of metrics (with the exception of metric acceptance
of local relevance to a single Nodal Switching System (NSS) only for
inbound routing information, in the case of multiple EGP peers at a
NSS). Second, it allowed engineering of the dynamic distribution of
routing information. That is, primary, secondary, etc., paths can be
determined, as long as dynamic externally learned routing information
is available. This allows creation of a spanning tree routing
Braun & Rekhter [Page 2]
^L
RFC 1222 Advancing the NSFNET Routing Architecture May 1991
topology, satisfying the constraints of EGP.
The pre-engineering of routes is accomplished by means of a routing
configuration database that is centrally controlled and created, with
a subsequent distribution of individual configuration information to
all the NSFNET Backbone nodes. A computer controlled central system
ensures the correctness of the database prior to its distribution to
the nodes.
All nodes of the 1.5Mbps NSFNET Backbone (currently fourteen) are
located at client sites, such as NSF funded supercomputer centers and
mid-level network attachment points.
3. T3 Phase of the NSFNET Backbone
The T3 (45Mbps) phase of the NSFNET Backbone is implemented by means
of a new architectural model, in which the principal communication
nodes (core nodes) are co-located with major phone company switching
facilities. Those co-located nodes then form a two-dimensional
networking infrastructure "cloud". Individual sites are connected
via exterior nodes (E-NSS) and typically have a single T3 access line
to a core node (C-NSS). That is, an exterior node is physically at
the service subscriber site.
With respect to routing, this structure is invisible to client sites,
as the routing interface uses the same techniques as the T1 NSFNET
Backbone. The two backbones will remain independent infrastructures,
overlaying each other and interconnected by exterior routing, and the
T1 Backbone will eventually be phased out as a separate network.
4. A Near-term Routing Alternative
The experience with the T1/T3 NSFNET routing demonstrated clear
advantages of this routing architecture in which the whole
infrastructure is strongly compartmentalized. Previous experience
also showed that the architecture imposes certain obligations upon
the attached client networks. Among them is the requirement that a
service subscriber must deploy its own routing protocol peer,
participating in the IGP of the service subscriber and connected via
a common subnet to the subscriber-site NSFNET node. The router and
the NSFNET Backbone exchange routing information via an EGP or BGP
[7] session.
The drawbacks imposed by this requirement will become more obvious
with the transition to the new architecture that is employed by the
T3 phase of the NSFNET Backbone. This will allow rapid expansion to
many and smaller sites for which a very simple routing interface may
be needed.
Braun & Rekhter [Page 3]
^L
RFC 1222 Advancing the NSFNET Routing Architecture May 1991
We strongly believe that separating the routing of the service
subscriber from the NSFNET Backbone routing via some kind of EGP is
the correct routing architecture. However, it should not be
necessary to translate this architecture into a requirement for each
service subscriber to install and maintain additional equipment, or
for the subscriber to deal with more complicated routing
environments. In other words, while maintaining that the concept of
routing isolation is correct, we view the present implementation of
the concept as more restrictive than necessary.
An alternative implementation of this concept may be realized by
separating the requirement for an EGP/BGP session, as the mechanism
for exchanging routing information between the service subscriber
network and the backbone, from the actual equipment that has to be
deployed and maintained to support such a requirement. The only
essential requirement for routing isolation is the presence of two
logical routing entities. The first logical entity participates in
the service subscriber's IGP, the second logical entity participates
in the NSFNET Backbone IGP, and the two logical entities exchange
information with each other by means of inter-domain mechanisms. We
suggest that these two logical entities could exist within a single
physical entity.
In terms of implementation, this would be no different from a
gatedaemon system interfacing with the previous 56Kbps NSFNET
Backbone from the regional clients, except that we want to continue
the strong routing and administrative control that decouple the two
IGP domains. Retaining an inter-domain mechanism (e.g., BGP) to
connect the two IGP domains within the single physical entity allows
the use of a well defined and understood interface. At the same
time, care must be taken in the implementation that the two daemons
will not simultaneously interact with the system kernel in unwanted
ways.
The possibility of interfacing two IGP domains within a single router
has also been noted in [8]. For the NSFNET Backbone case, we propose
in addition to retain strong firewalls between the IGP domains. The
IGP information would need to be tagged with exterior domain
information at its entry into the other IGP. It would also be
important to allow distributed control of the configuration. The
NSFNET Backbone organization and the provider of the attached client
network are each responsible for the integrity of their own routing
information.
An example implementation might be a single routing engine that
executed two instances of routing daemons. In the NSFNET Backbone
case, one of the daemons would participate in the service
subscriber's IGP, and the other would participate in the NSFNET
Braun & Rekhter [Page 4]
^L
RFC 1222 Advancing the NSFNET Routing Architecture May 1991
Backbone IGP. These two instances could converse with each other by
running EGP/BGP via a local loopback mechanism or internal IPC. In
the NSFNET Backbone implementation, the NSFNET T1 E-PSP or T3 E-NSS
are UNIX machines, so the local loopback interface (lo0) of the UNIX
operating system may be used.
Putting both entities into the same physical machine means that the
E-PSP/E-NSS would participate in the regional IGP on its exterior
interface. We would still envision the Ethernet attachment to be the
demarcation point for the administrative control and operational
responsibility. However, the regional client could provide the
configuration information for the routing daemon that interfaced to
the regional IGP, allowing the regional to continue to exercise
control over the introduction of routing information into its IGP.
5. Long-Term Alternatives
As technology employed by the NSFNET Backbone evolves, one may
envision the demarcation line between the Backbone and the service
subscribers moving in the direction of the "C-NSS cloud", so that the
NSFNET IGP will be confined to the C-NSS, while the E-NSS will be a
full participant in the IGP of the service subscriber.
Clearly, one of the major prerequisites for such an evolution is the
ability for operational management of the physical medium connecting
a C-NSS with an E-NSS by two different administrative entities (i.e.,
the NSFNET Backbone provider as well as the service subscriber). It
will also have to be manageable enough to be comparable in ease of
use to an Ethernet interface, as a well-defined demarcation point.
The evolution of the Point-to-Point Protocol, as well as a
significantly enhanced capability for managing serial lines via
standard network management protocols, will clearly help. This may
not be the complete answer, as a variety of equipment is used on
serial lines, making it difficult to isolate a hardware problem.
Similar issues may arise for future demarcation interfaces to
Internet infrastructure (e.g., SMDS interfaces).
In summary, there is an opportunity to simplify the management,
administration, and exchange of routing information by collapsing the
number of physical entities involved.
6. References
[1] Mills, D., "Exterior Gateway Protocol Formal Specification", RFC
904, BBN, April 1984.
[2] Mills, D., and H-W. Braun, "The NSFNET Backbone Network", SIGCOMM
Braun & Rekhter [Page 5]
^L
RFC 1222 Advancing the NSFNET Routing Architecture May 1991
1987, August 1987.
[3] Hedrick, C., "Routing Information Protocol", RFC 1058, Rutgers
University, June 1988.
[4] Rekhter, Y., "EGP and Policy Based Routing in the New NSFNET
Backbone", RFC 1092, IBM T.J. Watson Research Center, February
1989.
[5] Braun, H-W., "The NSFNET Routing Architecture", RFC 1093,
Merit/NSFNET, February 1989.
[6] Braun, H-W., "Models of Policy Based Routing", RFC 1104,
Merit/NSFNET, June 1989.
[7] Lougheed, K., and Y. Rekhter, "A Border Gateway Protocol (BGP)",
RFC 1163, cisco Systems, IBM T.J. Watson Research Center, June
1990.
[8] Almquist, P., "Requirements for Internet IP Routers", to be
published as a RFC.
7. Security Considerations
Security issues are not discussed in this memo.
8. Authors' Addresses
Hans-Werner Braun
San Diego Supercomputer Center
P.O. Box 85608
La Jolla, CA 92186-9784
Phone: (619) 534-5035
Fax: (619) 534-5113
EMail: HWB@SDSC.EDU
Yakov Rekhter
T.J. Watson Research Center
IBM Corporation
P.O. Box 218
Yorktown Heights, NY 10598
Phone: (914) 945-3896
EMail: Yakov@Watson.IBM.COM
Braun & Rekhter [Page 6]
^L
|