[sci.math] Weizenbaum keynote address at U of Waterloo

brewster@watdcsu.UUCP (dave brewer, SD Eng, PAMI ) (11/02/86)

The Hagey Lectures at the University of Waterloo provide an
opportunity for a distinguished researcher to address the
community at large every year.  This year, Dr. Weizenbaum of
MIT was the chosen speaker, and he has just delivered two
key note addresses entitled; "Prospects for AI" and "The Arms
Race, Without Us".

The important points of the first talk can be summarized as :
  1) AI has good prospects from an investment prospective since
     a strong commitment to marketing something called AI has
     been made.
  2) the early researchers did not understand how difficult
     the problems they addressed were and so the early claims
     of the possibilities were greatly exaggerated.  The trend
     still continues but on a reduced scale. 
  3) AI has been a handle for some portion of the US military
     to hang SDI on, since whenever a "difficult" problem
     arises it is always possible to say , " Well, we don't
     understand that now, but we can use AI techniques to
     solve that problem later."
  4) the actual achievements of AI are small.
  5) the ability of expert systems to continuously monitor
     stock values and react has led to increased volatility
     and crisis situations in the stock markets of the world
     recently.  What happens if machine induced technical trading
     drops the stock market by 20 % in one day , 50 % in one day ?

The important points of the second talk can be summarized as :
   1) not all problems can be reduced to computation, for
      example how could you conceive of coding the human
      emotion loneliness.
   2) AI will never duplicate or replace human intelligence
      since every organism is a function of its history.
   3) research can be divided into performance mode or theory
      mode research.  An increasing percentage of research is
      now conducted in performance mode, despite possible
      desires to do theory mode research, since funds (mainly
      military), are available for performance mode research. 
   4) research on "mass murder machines" is possible because
      the researchers (he addressed computer scientists
      directly although extension to any technical or
      scientific discipline was implied), are able to
      psychologically distance themselves from the end use
      of their work. 
   5) technical education that neglects language, culture,
      and history, may need to be rethought.
   6) courage is infectious, and while it may not seem to be
      a possibility to some, the arms race could be stopped cold
      if an entire group of professions, (ie computer scientists),
      refused to participate. 
   7) the search for funds has led to an increased rate of
      performance mode research, and has even induced many
      institutions to prostitute themselves to the highest bidder.
      Specific situations within MIT were used for examples. 
      Weizenbaum had the graciousness to ignore related (albeit
      proportionally smaller), circumstances at this 
      university.
   8) every researcher should assess the possible end use of
      their own research, and if they are not morally comfortable
      with this end use, they should stop their research.  Weizenbaum
      did not believe that this would be the end of all research,
      but if that was the case then he would except this result.
      He specifically referred to research in machine vision, which he
      felt would be used directly and immediately by the military for 
      improving their killing machines.  While not saying so, he implied
      that this line of AI should be stopped dead in its tracks. 

Posters comments :
  1) Weizenbaum seemed to be technically out of date in some areas,
     and admitted as much at one point.  Some of his opinions 
     regarding state of the art were suspect.
  2) His background, technical and otherwise, seems to predispose
     him to dismissing some technical issues a priori. i.e. a machine
     can never duplicate a human, why ?, because !.  
  3) His most telling point, and one often ignored, is that 
     researchers have to be responsible for their work, and should
     consider its possible end uses.  
  4) He did not appear to have thought through all the consequences
     of a sudden end to research, and indeed many of his solutions
     appear overly simplistic, in light of the complicated
     world we live in. 
  5) You have never seen an audience squirm, as they did for the
     second lecture.  A once premier researcher, addresses his
     contemporaries, and tells them they are ethically and morally
     bankrupt, and every member of the audience has at least some
     small buried doubt that maybe he is right. 
  6) Weizenbaum intended the talks to be "controversial and 
     provocative" and has achieved his goal within the U of W
     community.   While not agreeing with many of his points, I
     believe that there are issues raised which are relevant to
     the entire world-wide scientific community, and have posted 
     for this reason. 

The main question that I see arising from the talks is : is it time
to consider banning, halting, slowing, or otherwise rethinking 
certain AI or technical adventures, such as machine vision, as was
done in the area of recombinant DNA.

Disclaimer : The opinions above are mine and may not accurately
	     reflect those of U of Waterloo, Dr.Weizenbaum, or
	     anyone else for that matter.  I make no claims as
	     to the accuracy of the above summarization and advise
	     that transcripts of the talks are available from some
	     place within U of W, but expect to pay for them because
	     thats the recent trend. 


UUCP  : {decvax|ihnp4}!watmath!watdcsu!brewster
Else  : Dave Brewer, (519) 886-6657

brad@looking.UUCP (Brad Templeton) (11/02/86)

In article <2689@watdcsu.UUCP> brewster@watdcsu.UUCP (dave brewer,  SD Eng, PAMI ) writes:
>  5) You have never seen an audience squirm, as they did for the
>     second lecture.  A once premier researcher, addresses his
>     contemporaries, and tells them they are ethically and morally
>     bankrupt, and every member of the audience has at least some
>     small buried doubt that maybe he is right. 

The audienced "squirmed" as Dr. Weizenbaum accused, not so much because
he told us we were morally bankrupt, but because he told us we might be
so.

His contstant message, or so he said, throughout the evening was,
"Computer Scientists, examine the morality of what you are doing."

I don't want to belittle this message, because it's important, but it doesn't
need a full talk.  It was obvious that he had more to say than that, even
though he denied he was saying more, but he waffled about this further ground.

An audience does not expect hints at immorality, they expect to hear a speaker's
real opinions and solid arguments.  That the audience wanted an answer to
the question, "what sort of work do you think people should abandon" is not
surprising.  The questions that people like myself, Kelly & Ian! asked all
touched on serious aspects of this issue.  Dr. Weizenbaum's answers spoke
of gray areas and the need for miracles.  The only concrete things discussed
were the areas of computer vision & self programming natural language
interfaces.

Perhaps most people in this field really are moral blank slates who have to
be reminded that they should consider the risks in what they do.  Those who
asked the questions certainly were not, and perhaps came seeking more.
-- 
Brad Templeton, Looking Glass Software Ltd. - Waterloo, Ontario 519/884-7473

rggoebel@watdragon.UUCP (Randy Goebel LPAIG) (11/02/86)

Both reviews of Weizenbaum's lectures were quite polite; while it is important
to consider both the means and ends of one's work, it is unrealistic to believe
that some loosely defined community like ``computer scientists'' are all morally
bankrupt, and should collectively rethink their position.  Weizenbaum's motiva-
tion is well taken...but he has no suggestion of what to do about it.

As far as science goes, Weizenbaum believes that certain things about human
intelligence are currently unexplainable, and should remain that way.  There is
nothing scientific about that attitude.  I don't believe that Weizenbaum created
ANY worthwhile controversy with his lectures.

trost@reed.UUCP (Bill Trost) (11/03/86)

In article <2689@watdcsu.UUCP> brewster@watdcsu.UUCP (dave brewer,  SD Eng, PAMI ) writes:
>
>The main question that I see arising from the talks is : is it time
>to consider banning, halting, slowing, or otherwise rethinking 
>certain AI or technical adventures, such as machine vision, as was
>done in the area of recombinant DNA.

Somehow, I don't think that banning machine vision makes much sense.  It
seems that it would be similar to banning automatic transmissions because
you can use them to make tanks.  The device itself is not the hazard (as it
is in genetic research) -- it is the application.
-- 
Bill Trost, tektronix!reed!trost
	"ACK!"
	(quoted, without permission, from Bloom County)