[comp.protocols.tcp-ip] THE INTERNET CRUCIBLE - Volume 1, Issue 1

geoff@FERNWOOD.MPK.CA.US (the terminal of Geoff Goodfellow) (09/01/89)

THE CRUCIBLE                                                  INTERNET EDITION
an eleemosynary publication of the                                August, 1989
Anterior Technology IN MODERATION NETWORK(tm)               Volume 1 : Issue 1

			     Geoff Goodfellow
				 Moderator

   In this issue:
	A Critical Analysis of the Internet Management Situation

   THE CRUCIBLE is an irregularly published, refereed periodical on the
   Internet.  The charter of the Anterior Technology IN MODERATION NETWORK 
   is to provide the Internet and Usenet community with useful, instructive
   and entertaining information which satisfies commonly accepted standards
   of good taste and principled discourse.  All contributions and editorial
   comments to THE CRUCIBLE are reviewed and published without attribution.
   Cogent, cohesive, objective, frank, polemic submissions are welcomed.

Mail contributions/editorial comments to:	crucible@fernwood.mpk.ca.us

------------------------------------------------------------------------------

	 A Critical Analysis of the Internet Management Situation:
		       The Internet Lacks Governance


				  ABSTRACT

At its July 1989 meeting, the Internet Activities Board made some
modifications in the management structure for the Internet.  An outline of
the new IAB structure was distributed to the Internet engineering community
by Dr. Robert Braden, Executive Director.  In part, the open letter stated:

   "These changes resulted from an appreciation of our successes, especially
    as reflected in the growth and vigor of the IETF, and in rueful
    acknowledgment of our failures (which I will not enumerate).  Many on
    these lists are concerned with making the Internet architecture work in
    the real world."

In this first issue of THE INTERNET CRUCIBLE we will focus on the failures
and shortcomings in the Internet.  Failures contain the lessons one often
needs to achieve success.  Success rarely leads to a search for new
solutions.  Recommendations are made for short and long term improvements 
to the Internet.


			A Brief History of Networking

The Internet grew out of the early pioneering work on the ARPANET.  This
influence was more than technological, the Internet has also been
significantly influenced by the economic basis of the ARPANET.

The network resources of the ARPANET (and now Internet) are "free".  There
are no charges based on usage (unless your Internet connection is via an
X.25 Public Data Network (PDN) in which case you're well endowed, or better
be).  Whether a site's Internet connection transfers 1 packet/day or a 1M
packets/day, the "cost" is the same.  Obviously, someone pays for the
leased lines, router hardware, and the like, but this "someone" is, by and
large, not the same "someone" who is sending the packets.

In the context of the Research ARPANET, the "free use" paradigm was an
appropriate strategy, and it has paid handsome dividends in the form of
developing leading edge packet switching technologies.  Unfortunately,
there is a significant side-effect with both the management and technical
ramifications of the current Internet paradigm: there is no 
accountability, in the formal sense of the word.

In terms of management, it is difficult to determine who exactly is
responsible for a particular component of the Internet.  From a technical
side, responsible engineering and efficiency has been replaced by the
purchase of T1 links.

Without an economic basis, further development of short-term Internet
technology has been skewed.  The most interesting innovations in
Internet engineering over the last five years have occurred in resource
poor, not resource rich, environments.

Some of the best known examples of innovative Internet efficiency
engineering are John Nagle's tiny-gram avoidance and ICMP source-quench
mechanisms documented in RFC896, Van Jacobsen's slow-start algorithms and
Phil Karn's retransmission timer method.

In the Nagle, Jacobsen and Karn environments, it was not possible or cost
effective to solve the performance and resource problems by simply adding
more bandwidth -- some innovative engineering had to be done.
Interestingly enough, their engineering had a dramatic impact on our
understanding of core Internet technology.

It should be noted that highly efficient networks are important when
dealing with technologies such as radio where there is a finite amount of
bandwidth/spectrum to be had.  As in the Nagle, Jacobsen and Karn cases,
there are many environments where adding another T1 link can not be used
to solve the problem.  Unless innovation continues in Internet
technology, our less than optimal protocols will perform poorly in
bandwidth or resource constrained environments.

Developing at roughly the same time as Internet technology have been the
"cost-sensitive" technologies and services, such as the various X.25-based
PDNs, the UUCP and CSNET dial-up networks.  These technologies are all
based on the notion that bandwidth costs money and the subscriber pays for
the resources used.  This has the notable effect of focusing innovation to
control costs and maximize efficiency of available resources and bandwidth.
Higher efficiency is achieved by concentrating on sending the most amount
of information through the pipe in the most efficient manner thereby making
the best use of available bandwidth/cost ratio.

For example, bandwidth conservation in the UUCP dial-up network has
multiplied by leaps and bounds in the modem market with the innovation of
Paul Baran's (the grandfather of packet switching technology) company,
Telebit, which manufactures a 19.2KB dial-up modem especially optimized for
UUCP and other well known transfer protocols.  For another example,
although strictly line-at-a-time terminal sessions are less "user friendly"
than character-oriented sessions, they make for highly efficient use of
X.25 PDN network resources with echoing and editing performed locally on
the PAD.

While few would argue the superiority of X.25 and dial-up CSNET and UUCP,
these technologies have proved themselves both to spur innovation and to be
accountable.  The subscribers to such services appreciate the cost of the
services they use, and often such costs form a well-known "line item" in
the subscriber's annual budget.

Nevertheless, the Internet suite of protocols are eminently successful,
based solely on the sheer size and rate of growth of both the Internet and
the numerous private internets, both domestically and internationally.  You
can purchase internet technology with a major credit card from a mail order
catalog.  Internet technology has achieved the promise of Open Systems,
probably a decade before OSI will be able to do so.


			   Failures of the Internet

The evolution and growth of Internet technology have provided the basis for
several failures.  We think it is important to examine failures in detail,
so as to learn from them.  History often tends to repeat itself.


Failure 1:- Network Nonmanagement

The question of responsibility in todays proliferated Internet is completely
open.  For the last three years, the Internet has been suffering from
non-management.  While few would argue that a centralized czar is necessary
(or possible) for the Internet, the fact remains there is little to be done
today besides finger-pointing when a problem arises.
 
In the NSFNET, MERIT is incharge of the backbone and each regional network
provider is responsible for its respective area.  However, trying to debug
a networking problem across lines of responsibility, such as intermittent
connectivity, is problematic at best.  Consider three all too true refrains
actually heard from NOC personal at the helm:

	"You can't ftp from x to y?  Try again tomorrow, it will
	 probably work then."

	"If you are not satisfied with the level of [network] 
         service you are receiving you may have it disconnected."

	"The routers for network x are out of table space for routes,
	 which is why hosts on that network can't reach your new
	 (three-month old) network.  We don't know when the routers will
	 be upgraded, but it probably won't be for another year." 

One might argue that the recent restructuring of the IAB may work towards
bringing the Internet under control and Dr. Vinton G. Cerf's recent
involvement is a step in the right direction.  Unfortunately, from a
historical perspective, the new IAB structure is not likely to be
successful in achieving a solution.  Now the IAB has two task forces, the
Internet Research Task Force (IRTF) and the Internet Engineering Task Force
(IETF).  The IRTF, responsible for long-term Internet research, is largely
composed of the various task forces which used to sit at the IAB level.
The IETF, responsible for the solution of short-term Internet problems, has
retained its composition.

The IETF is a voluntary organization and its members participate out of
self interest only.  The IETF has had past difficulties in solving some of
the Internet's problems (i.e., it has taken the IETF well over a year to
not yet produce RFCs for either a Point-To-Point Serial Line IP or Network
Management enhancements).  It is unlikely that the IETF has the resources
to mount a concerted attack against the problems of today's ever expanding
Internet.  As one IETF old-timer put it: "No one's paid to go do these
things, I don't see why they (the IETF management) think they can tell us
what to do" and "No one is paying me, why should I be thinking about the
these things?"
 
Even if the IETF had the technical resources, many of the Internet's
problems are also due to lack of "hands on" management.  The IETF

	o  Bites off more than it can chew;
	o  Sometimes fails to understand a problem before making a solution;
	o  Attempts to solve political/marketing problems with technical
	     solutions;
	o  Has very little actual power.

The IETF has repeatedly demonstrated the lack of focus necessary to
complete engineering tasks in a timely fashion.  Further, the IRTF is
chartered to look at problems on the five-year horizon, so they are out of
the line of responsibility.  Finally, the IAB, per se, is not situated to
resolve these problems as they are inherent to the current structure of
nonaccountability.

During this crisis of non-management, the Internet has evolved into a
patch quilt of interconnected networks that depend on lots of
seat-of-the-pants flying to keep interoperating.  It is not an unusual
occurrence for an entire partition of the Internet to remain disconnected
for a week because the person responsible for a key connection went on
vacation and no one else knew how to fix it.  This situation is but one
example of an endemic problem of the global Internet.


Failure 2:- Network Management

The current fury over network management protocols for TCP/IP is but a
microcosm of the greater Internet vs. OSI debate going on in the
marketplace.  While everyone in the market says they want OSI, anyone
planning on getting any work done today buys Internet technology.  So it
is with network management, the old IAB made the CMOT an Internet
standard despite the lack of a single implementation, while the only
non-proprietary network management protocol in use in the Internet is the
SNMP.  The dual network management standardization blessings will no
doubt have the effect of confusing end-users of Internet
technology--making it appear there are two choices for network
management, although only one choice, the SNMP has been implemented.  The
CMOT choice isn't implemented, doesn't work, or isn't interoperable.

To compound matters, after spending a year trying to achieve consensus on
the successor to the current Internet standard SMI/MIB, the MIB working
group was disbanded without ever producing anything: the political climate
prevented them from resolving the matter.  (Many congratulatory notes were
sent to the chair of the group thanking him for his time.  This is an
interesting new trend for the Internet--congratulating ourselves on our
failures.)

Since a common SMI/MIB could not be advanced, an attempt was made to
de-couple the SNMP and the CMOT (RFC1109).  The likely result of RFC1109
will be that the SNMP camp will continue to refine their experience
towards workable network management systems, whilst the CMOT camp will
continue the never-ending journey of tracking OSI while producing demo
systems for trade shows exhibitions.  Unfortunately the end-user will
remain ever confused because of the IAB's controversial (and technically
questionable) decision to elevate the CMOT prior to implementation.

While the network management problem is probably too large for the SNMP
camp to solve by themselves they seem to be the only people who are
making any forward progress.


Failure 3:- Bandwidth Waste

Both the national and regional backbone providers are fascinated with T1
(and now T3) as the solution towards resource problems.  T1/T3 seems to
have become the Internet panacea of the late 80's.  You never hear anything
from the backbone providers about work being done to get hosts to implement
the latest performance/congestion refinements to IP, TCP, or above.
Instead, you hear about additional T1 links and plans for T3 links.  While
T1 links certainly have more "sex and sizzle" than efficient technology
developments like slow-start, tiny gram avoidance and line mode telnet, the
majority of users on the Internet will probably get much more benefit from
properly behaving hosts running over a stable backbone than the current
situation of misbehaving and semi-behaved hosts over an intermittent catenet.


Failure 4:- Routing

The biggest problem with routing today is that we are still using phase I
(ARPANET) technology, namely EGP.  The EGP is playing the role of routing
glue in providing the coupling between the regional IGP and the backbone
routing information.  It was designed to only accommodate a single point of
attachment to the catenet (which was all DCA could afford with the PSNs).
However with lower line costs, one can build a reasonably inexpensive
network using redundant links.  However the EGP does not provide enough
information nor does the model it is based upon support multiple
connections between autonomous systems.  Work is progressing in the
Interconnectivity WG of the IETF to replace EGP.  They are in the process
of redefining the model to solve some of the current needs.  BGP or the
Border Gateway Protocol (RFC1105) is an attempt to codify some of the ideas
the group is working on.

Other problems with routing are caused by regionals wanting a backdoor
connection to another regional directly.  These connections require some
sort of interface between the two routing systems.  These interfaces are
built by hand to avoid routing loops.  Loops can be caused when information
sent into one regional network is sent back towards the source.  If the
source doesn't recognize the information as its own, packets can flow until
their time to live field expires.

Routing problems are caused by the interior routing protocol or IGP.  This
is the routing protocol which is used by the regionals to pass information
to and from its users.  The users themselves can use a different IGP than
the regional.  Depending on the number of connections a user has to the
regional network, routing loops can be an issue.  Some regionals pass
around information about all known networks in the entire catenet to their
users.  This information deluge is a problem with some IGPs.  Newer IGPs
such as the new OSPF from the IETF and IGRP from cisco attempt to provide
some information hiding by adding hierarchy.  OSPF is the internets first
attempt at using a Dykstra type algorithm as an IGP.  BBN uses it to route
between their packet switch nodes below the 1822 or X.25 layer.

Unstable routing is caused by hardware or hosts software.  Older BSD
software sets the TTL field in the IP header to a small number.  The
Internet today is growing and its diameter has exceed the software's
ability to reach the other side.  This problem is easily fixed by
knowledgeable systems people, but one must be aware of the problem before
they can fix it.

Routing problems are also perceived when in fact a serial line problem or
hardware problem is the real cause.  If a serial line is intermittent or
quickly cycles from the up state into the down state and back again,
routing information will not be supplied in a uniform or smooth manner.
Most current IGPs are Bellman-Ford based and employ some stabilizing
techniques to stem the flow of routing oscillations due to "flapping"
lines.  Often when a route to a network disappears, it may take several
seconds for it to reappear.  This can occur at the source router who waits
for the route to "decay" from the system.  This pause should be short
enough so that active connections persist but long enough that all routers
in the routing system "forget" about routes to that network.  Older host
software with over-active TCP retransmission timers will time out
connections instead of persevering in the face of this problem.  Also
routers, according to RFC1009, must be able to send ICMP unreachables when
a packet is sent to a route which is not present in its routing database.
Some host products on the market close down connections when a single ICMP
reachable is received.  This bug flies in the face of the Internet parable
"be generous in what you accept and rigorous in what you send".

Many of the perceived routing problems are really complex multiple
interactions of differing products.


			    Causes of the Failures

The Internet failures and shortcomings can be traced to several sources:
			
First and foremost, there is little or no incentive for efficiency and/or
economy in the current Internet.  As a direct result, the resources of the
Internet and its components are limited by factors other than economics.
When resources wear thin, congestion and poor performance result.  There is
little to no incentive to make things better, if 1 packet out of 10 gets
through things "sort of work".  It would appear that Internet technology
has found a loophole in the "Tragedy of The Commons" allegory--things get
progressively worse and worse, but eventually something does get through.

The research community is interested in technology and not economics,
efficiency or free-markets.  While this tack has produced the Internet
suite of protocols, the de facto International Standard for Open Systems,
it has also created an atmosphere of intense in-breeding which is overly
sensitive to criticism and quite hardened against outside influence.
Meanwhile, the outside world goes on about developing economically viable
and efficient networking technology without the benefit of direct
participation on the part of the Internet.

The research community also appears to be spending a lot of its time
trying to hang onto the diminishing number of research dollars available
to it (one problem of being a successful researcher is eventually your
sponsors want you to be successful in other things).  Despite this, the
research community actively shuns foreign technology (e.g., OSI), but,
inexplicably has not recently produced much innovation in new Internet
technology.  There is also a dearth of new and nifty innovative
applications on the Internet.  Business as usual on the Internet is mostly
FTP, SMTP and Telnet or Rlogin as it has been for many years.  The most
interesting example of a distributed application on the Internet today is
the Domain Name System, which is essentially an administrative facility,
not an end-user service.

The engineering community must receive equal blame in these matters.
While there have been some successes on the part of the engineering
community, such as those by Nagel, Jacobsen and Karn mentioned above, the
output of the IETF, namely RFCs and corresponding implementations, has
been surprisingly low over its lifetime.

Finally, the Internet has become increasingly dependent on vendors for
providing implementations of Internet technology.  While this is no doubt
beneficial in the long-term, the vendor community, rather than investing
"real" resources when building these products, do little more than
shrink-wrap code written primarily by research assistants at universities.
This has lead to cataclysmic consequences (e.g., the Internet worm
incident, where Sendmail with "debug" command and all was packaged and
delivered to customers without proper consideration).  Of course, when
problems are found and fixed (either by the vendor's customers or software
sources), the time to market with these fixes is commonly a year or longer.
Thus, while vendors are vital to the long-term success of Internet
technology, they certainly don't receive high marks in the short-term.


			       Recommendations

Short-term solutions (should happen by year's end):

In terms of hardware, the vendor community has advanced to the point where
the existing special-purpose technologies (Butterfly, NSSs) can be replaced
by off-the-shelf routers at far less cost and with superior throughput and
reliability.  Obvious candidates for upgrade are both the NSFNET and
ARPANET backbones.  Given the extended unreliability of the mailbridges,
the ARPA core is an immediate candidate (even though the days of net 10 are
numbered).

In terms of software, ALL devices in the Internet must be network
manageable.  This is becoming ever more critical when problems must be
resolved.  Since SNMP is the only open network management protocol
functioning in the Internet, all devices must support SNMP and the Internet
standard SMI and MIB.

Host implementations must be made to support the not-so-recent TCP
enhancements (e.g., those by Nagle, Jacobsen and Karn) and the more
recent linemode TELNET.

The national and regional providers must coordinate to share network
management information and tools so that user problems can be dealt with in
a predictable and timely fashion.  Network management tools are a big help,
but without the proper personnel support above this, the benefits can not
be fully leveraged.

The Internet needs leadership and hands-on guidance.  No one is seemingly
in charge today, and the people who actually care about the net are pressed
into continually fighting the small, immediate problems.  

Long-term solutions:

To promote network efficiency and a free-market system for the delivery of
Internet services, it is proposed to switch the method by which the network
itself is supported.  Rather than a top-down approach where the money goes
from funding agencies to the national backbone or regional providers, it is
suggested the money go directly to end-users (campuses) who can then select
from among the network service providers which among them best satisfies
their needs and costs.

This is a strict economic model: by playing with the full set of the laws of
economics, a lot of the second-order problems of the Internet, both present
and on the horizon, can be brought to heel.  The Internet is no longer a
research vehicle, it is a vibrant production facility.  It is time to
acknowledge this by using a realistic economic model in the delivery of
Internet services to the community (member base).

When Internet sites can vote with their pocketbooks, some new regionals
will be formed; some, those which are non-performant or uncompetitive, will
go away; and, the existing successful ones will grow.  The existing
regionals will then be able to use their economic power, as any consumer
would, to ensure that the service providers (e.g., the national backbone
providers) offer responsive service at reasonable prices.  "The Market" is
a powerful forcing function: it will be in the best interests of the
national and regional providers to innovate, so as to be more competitive.
Further, such a scheme would also allow the traditional telecommunications
providers a means for becoming more involved in the Internet, thus allowing
cross-leverage of technologies and experience.

The transition from top-down to economic model must be handled carefully,
but this is exactly the kind of statesmanship that the Internet should
expect from its leadership.
-------

henry@utzoo.uucp (Henry Spencer) (09/07/89)

In article <8909010312.AA03885@fernwood.MPK.CA.US> geoff@FERNWOOD.MPK.CA.US (the terminal of Geoff Goodfellow) writes:
>While few would argue the superiority of X.25 and dial-up CSNET and UUCP,
>these technologies have proved themselves both to spur innovation and to be
>accountable.  The subscribers to such services appreciate the cost of the
>services they use, and often such costs form a well-known "line item" in
>the subscriber's annual budget.

This is undoubtedly true for X.25 and dialup CSNET, but for UUCP the exact
reverse is often the case:  a network connection exists precisely because,
unlike the alternatives, a UUCP connection does *not* require a line item
in the budget.  All it requires is suitable software (present in most
versions of Unix already), dialup modems on both ends (usually present
anyway), and a bit of setup work.  If no obscure problems intervene, such
a connection can be up and running in fifteen minutes, with *no* paperwork.
If traffic remains modest -- where the exact semantics of "modest" depend
on modem type and the cost (if any) of the calls -- often no formal
justification of the connection is ever required.  When communication is
a means to an end rather than a research topic, this can be a major asset.
-- 
V7 /bin/mail source: 554 lines.|     Henry Spencer at U of Toronto Zoology
1989 X.400 specs: 2200+ pages. | uunet!attcan!utzoo!henry henry@zoo.toronto.edu