[comp.research.japan] Kahaner Report: NIPT Symposium

rick@cs.arizona.edu (Rick Schlichting) (04/09/91)

  [Dr. David Kahaner is a numerical analyst visiting Japan for two-years
   under the auspices of the Office of Naval Research-Asia (ONR/Asia).  
   The following is the professional opinion of David Kahaner and in no 
   way has the blessing of the US Government or any agency of it.  All 
   information is dated and of limited life time.  This disclaimer should 
   be noted on ANY attribution.]

  [Copies of previous reports written by Kahaner can be obtained from
   host cs.arizona.edu using anonymous FTP.]

To: Distribution
From: David Kahaner ONR Asia [kahaner@xroads.cc.u-tokyo.ac.jp]
Re: New Information Processing Technology Symposium (6th Generation Proj)
7 April 1991 

ABSTRACT. An International Symposium on New Information Processing
Technologies (NIPT) '91 13-14 March 1991, Tokyo Japan is described. NIPT
is to be the successor to the 5th Generation Project. 

INTRODUCTION.
On 13-14 March 1991, an open International Symposium on New Information
Processing Technologies was held in Tokyo, Japan, attended by almost 400
people, including about 30 from 12 overseas countries.  The Symposium
program is attached at the end of this report.  Three months ago, a much
smaller workshop was held in Hakone on the same topic, and was reported
on in "nipt.90, 28 Dec 1990". Except when required for clarity, I will
omit material that duplicates information given there. Consequently the
earlier report should be thought of as an important supplement to this
one. 

WHAT IS NIPT?
There are several parts to NIPT, but the main thread is that some things
are easy for humans and difficult (presently) for computers, such as
reading and interpreting comics. The reverse is also true, computers
seem much better at floating point arithmetic than we are. NIPT seeks to
concentrate on those areas where computers are currently weak. These
include friendliness, flexibility, processing diverse and parallel
information such as speech and image, adaptation, and integration of
ambiguous information. Humans can perform logical symbol manipulation,
intuitive thinking based on pattern dynamics, integration of multimodal
incomplete information, learning from examples, fault tolerance, self
organization, etc.  The Japanese believe that computers should be
made more human-like. Until now computer programming has emphasized
logical sequential processing, corresponding to activities in the human
left-brain, but they feel it will be important to build computers that 
mimic right brain capabilities.  In this report I am deeply greatful to
many helpful comments and corrections from Mr. Satoshi Ishihara (ETL),
Dr. Paul Refenes (DTI and UCL), Prof Bruce MacLennan (UTenn), and others. 
In addition, several Japanese participants had an opportunity to review a 
draft of this report and voiced no specific objections to it.  

The most succinct and complete description of NIPT was provided by 
Professor Takemochi Ishii (U-Tokyo, Department of Mechanical 
Engineering), who chairs the research committee studying the project. I 
could not do better than quoting his remarks, which follow. My comments 
[bracketed] are inserted when appropriate.  

"Since the 1989 fiscal year, the Research Committee on NIPT has been 
conducting a two-year preliminary study for MITI to explore the 
possibility of a new international R&D program to realize advanced 
information processing technology which can break through the limitations 
of conventional computers and must be a basis of the Information Network 
Society of the 21st century. [A one-year, feasibility study begins now.  
The program per-se, if approved, will begin in 1992. An official budget 
has not been announced, but MITI claims this will around $30million US.  
There are also likely to be industrial contributions that may be 
difficult to measure.] 

The Committee has three sub-committees, Fundamental Theory, Computer 
Science, and Social Impact, consisting of more than 100 researchers 
participating from various fields such as neuro-physiology, cognitive 
science, economics, philosophy as well as computer science and 
engineering. Here is a perspective of the outcome of the preliminary 
study as an introduction to the detailed reports by chairs of the sub-
committees. [Each of the three sub-committee chairs presented detailed 
papers at this Symposium. The computer science sub-committee was the 
largest of these, and was broken into two working groups, Integrated 
Computing, and Optical Computer/Devices. These were in turn broken into 
several smaller sub-working groups.]

1. Objective
   The goals of the NIPT program should be:
  * Establish humanized flexible advanced information processing 
    technologies which will be the basis of the information network 
    society of the 21st century; [Since the earlier meeting there is now 
    a more definite statement that the focus and goals of the program are 
    diverse, multi-directional and pluralistic. It was admitted that some 
    research groups will fail, or get lost. I felt that the program has 
    been enunciated much more clearly than previously, although parts are 
    still vague. On the other hand, some Western attendees who were not at 
    the Hakone workshop were confused by its imprecision. One called it 
    "alice in wonderland", but I think that progress is definitely being 
    made.] 
  * Encourage the international cooperative researches to make some 
    contribution for the development of the fundamental generic 
    technologies in high-tech fields and construct the international co-
    prosperous relationship. [The lofty nature of the goals and limited 
    resources available require cooperation and coordination. Overseas 
    participation was welcomed, but admittedly difficult to manage.  The 
    5th Generation Project was thought of entirely Japanese, whereas NIPT 
    has international cooperation as a major stated goal.] 


2. Strategy
   These goals will be accomplished through promoting an international 
    cooperative research program to: 
  * Realize humanized advanced flexible computers through integration of 
    logical information processing and intuitive information processing by 
      ** Development of massively parallel and distributed hardware with 
          new device technologies such as optical devices and 
         [At the earlier meeting, it was apparent that building a 
          massively parallel computer was to be a part of this project, 
          but this is the first time I have seen it explicitly stated. In 
          later sessions this was spelled out in more detail, but also
          specifically proposed as something more challenging than that 
          selected by industry, such as a one million processor parallel 
          computer, a fully optical computer, or a neural computer.
          It was left undecided how a neural computer will be 
          integrated into this program, and if the focus will be on a general
          purpose computer using a neural engine as one part, or on a
          more specialized neural-optical computer.] 
      ** Controlling the hardware in the flexible and adaptive way based 
          on the theory of flexible information processing.  [Later, 
          Amari made clear than bold new theories are needed, and what is 
          envisioned is not an extension of current ideas.] 


3. Background and Needs for the NIPT
   In the 21st century, because of the development of sensor technologies 
and telecommunication service, trend for multi-media information 
processing, popularization of HDTV, and weight shift from material 
production to information production, quantity and variety of information 
which flow through the society will be expected to increase explosively. 
Computational ability of conventional computers seem to be insufficient 
to process such massive and multi-modal information.  [It is extremely
important to note the role that networking plays in Japan's view 
of the future. In virtually every plan, application, and product, the 
existence of high performance networks connecting substantial segments 
of the society is a given, something to be folded in and used, rather 
than something for which a case needs to be made.]

   To cope with the crisis of information explosion of the 21st century, 
new information processing technologies enabling us to realize 
revolutionary advanced information processing devices which will have 
some features as the following: 
  * Learning and self-organizing ability for software reduction; 
  * Adaptability for individual users or situations; 
  * Information integration ability which enables not only analyzing 
    information but also synthetic decision and use of many kinds of 
    knowledge cooperatively for problem solving; 
  * Sufficient complexity of systems compared with the complexity of the 
    phenomena which the system treat; and 
  * Affinity for optical communication technologies for enabling the 
    integration of communication and information processing will be 
    needed.  

4. Fundamental Policy
   In carrying out this program, to keep the following fundamental policy 
   is important; 

  * Construction of international cooperative and co-prosperous 
    relationship in high-tech fields by keeping the door open to foreign 
    companies and universities as well as domestic ones, and guaranteeing 
    impartial distribution of the results; [Several Western governments 
    are already involved in discussions as are a few private companies.  
    Ishii admitted that handling the results was crucial, could 
    eventually become a model for international R&D in generic high-tech 
    fields, and that the Japanese people felt the need to become more 
    global and share world intellectual assets.  In particular he noted 
    that advanced research needed to be used by the developing countries, 
    and that cooperation with their researchers was important.  Although 
    several private meetings occurred there still is no public statement 
    of how cooperation and intellectual property are to be handled.  
    However, at about the time of the Symposium some proposed changes 
    were announced in the way MITI treats patent rights arising out of 
    R&D projects that are funded either by MITI or by NEDO (New Energy 
    and Industrial Technology Development Organization).  The proposals 
    have been submitted to the Diet as a revision of the "NEDO Law". The 
    claim is that the measure is intended to promote international 
    research cooperation among the private sectors in order to develop 
    new technology, and is a first step toward realizing the government's 
    "technoglobalism" policy. Until now, all patent rights arising from 
    projects funded by MITI have been held by the government.  
    Participating companies have had to pay license fees to use these 
    patent rights. (In the case of projects funded by NEDO, companies can 
    hold up to 50% of the patent rights, but they still have to pay 
    license fees.) Under the proposed new measure, companies can hold up 
    to 50% of patent rights, even if those patent rights arise from 
    government-funded R&D projects. In addition, companies can use patent 
    rights arising from R&D projects funded by MITI or NEDO free of 
    charge or for a small fee. The law is expected to be passed and go 
    into effect this summer. I assume that the proposed legislation will 
    also apply to research collaborations with universities.  

    Because of MITI's support for Japanese industry, as contrasted to 
    MOMBUSHO which supports Japanese universities, some Western 
    participants were wary about assuring themselves that cooperation 
    will mean that benefits flow in both directions. Perhaps it would 
    make sense to offer up parts of the program directly through 
    MOMBUSHO. I hope as the next study begins we will get even more 
    details of who is going to do what, and with what resources. If this 
    information is open and complete it will go a long way to allieviate 
    a number of uneasy comments that I heard. See also remarks from 
    Refenes below.] 

  * Challenge for creating fundamental and revolutionary generic 
    technologies by encouraging the participation of the universities and 
    national laboratories; and 
  * Shift from R&D for technology itself to R&D for the human beings by 
    encouraging the participation of users of the technologies and 
    researchers from related basic research fields such as brain science 
    and cognitive science. " [End of Ishii's remarks. Western 
    participants in the Symposium who also attended the workshop were 
    impressed with the increased role assigned to biology, a topic whose 
    absence was criticized at the earlier workshop. See also the summary of  
    Suzuki's lecture below.]

REMARKS.
Both formal and informal discussions suggested that at the moment four 
major ideas are active. These were also mentioned by Amari in his lecture 
describing the computer science research subcommittee.
(1) Develop a theory of flexible information processing, parallel and 
    distributed computation, learning and self organization, neural and 
    optical computation, etc.  
(2) Develop a theory and implement an integrated information processing 
    model of the brain that can be useful for recognition, 
    understanding, and control in a real world of information, such as 
    control by complex speech input.  
(3) Build a (probably) general purpose, object-oriented, data-driven massively 
    parallel computer system with 10^6 to 10^9 processors, using optical 
    or neural chips.
(4) Continue research and development on optical computing, including 
    optical interconnects, devices, optical neural computers, and fully 
    optical computers. 

With respect to a massively parallel system, Amari described what could 
be built using technology that would be working within the next ten 
years. A 32bit processing element (PE) would contain 2K words of memory, 
and require 600K transistors. Eight processors would be on one chip, 1K 
processors would be on a one board module, and 1000 boards would compose 
a one million processor module, or system. He emphasized that the NIPT 
program would only go so far as to develop proof of concept or do the 
research necessary to overcome bottlenecks. Further work would be left 
for industry.
                Professor Shun-ichi Amari
                Faculty of Engineering
                University of Tokyo
                Bunkyo-ku, Hongo 7-3-1
                Tokyo 113 Japan
                Phone: 03 812-2111, ext 6910, Fax: 03 816-7805
                AMARI@SAT.T.U-TOKYO.AC.JP



PANEL DISCUSSION. 
One of the most interesting parts of the Symposium was a panel composed 
of eight well known researchers.
    M.A. Arbib Univ Southern Cal, USA
    R. Eckmiller Univ of Dusseldorf, Germany
    A. Hartmann MCC USA
    B. MacLennan Univ of Tenn, USA
    P.N. Refenes Dept Trade and Industry and Univ College London, UK
    J. Shimada ETL, Japan
    K. Toyama Kyoto Pref Univ of Medicine, Japan
    T. Yuba ETL, Japan

The topic was "Towards Computing in the 21st Century", and each panelist 
was given an opportunity to make some formal remarks and then there was 
to be a dialog. One or two questions from the floor were allowed to each 
speaker.  Although all of the presentations were exceptionally well 
thought out and extremely valuable, most of the non-Japanese panelists 
seemed unable to keep to the schedule and an important opportunity for a 
real discussion was lost, save for Arbib, who spoke last and was able to 
make a few comments about earlier presentations.  As readers will 
discover scanning the summaries below, panelists dealt with very different 
topics and an interchange among them would have been most useful.  My own 
feeling is that the Japanese should market an overhead projector which 
displays a large clock on its face showing the time remaining (to be set 
via remote control by the chair), blinks to alert the speaker when time 
is almost up, and after an appropriate grace interval turns the projector 
light off.  This would also eliminate the embarrassing need for the 
secretary to try and catch the speaker's attention with a sign reading 
"no more time".  

A brief summary of the significant remarks are given below.  Eckmiller, 
Hartmann, Refenes, Shimada, and Toyama submitted papers which are 
included in the proceedings. Arbib and Yuba gave other lectures which are 
also in the proceedings, and in those two cases I have used their papers 
as reference material.  

Toyama: Neurobiologist, discussed aspects of cortical machinery that 
might help in future design of parallel computers. (Another biology paper 
was presented by Hideo Sakata of Nippon University. I am not qualified to 
comment on either of these.)

Eckmiller: A very general appeal to use the international cooperation 
aspect of this project to work on solutions to the Earth's major 
problems, population, environmental protection, etc. Also pointed out 
that the European Esprit model was probably not the right one for MITI to 
follow for cooperation.  MITI's Nishikawa responded to a question that as 
yet there was no contract with any other country, or any real model for 
how cooperation should be done. On the other hand, I have learned that 
are some negotiations already underway with Western universities for the 
exchange of scholars, setting up of research institutes, etc.  
                Prof Rolf Eckmiller
                Dept of Biophysics
                Heinrich-Heine-Universitat Dusseldorf
                D-4000 Dusseldorf, FR Germany
                Tel: (211) 311-4540

MacLennan: Current "hot" computer science areas, AI, expert systems, 
fuzzy logic, etc., cannot cope well with flexibility (too brittle--nice 
term). Urged the NIPT planners to consider studying how discrete objects 
come from continuous ones (neural processes and subsymbolic cognition).
His paper was not included in the proceedings, but copies can obtained by 
contacting the author.     
                Professor Bruce J. MacLennan
                University of Tennessee at Knoxville
                Computer Science Department
                107 Ayres Hall
                Knoxville, TN 37996-1301
                Tel: (615) 974-5067
                Email: MACLENNAN@CS.UTK.EDU
After the meeting, MacLennan forwarded to me some remarks that are
quoted below. I believe that they agree with my own observations as well
as those other participants. 

"One of the most important characteristics of the Japanese initiative,
as I understand it, is that it is not a narrowly technological project.
First, it is based on a comprehensive vision of "the information
network society of the 21st century."*  This is seen to imply a need
for: (1) flexible computing ("intuitive information processing");
(2) adaptive computing; and (3) massively parallel computing, including
optical computers.  Second, the Japanese are aware that much basic
research remains to be done before this vision can be fully realized,
and so they are including researchers in "neuro-physiology, cognitive
science, economics, philosophy as well as computer science and
engineering."  Furthermore, they apparently realize that the same 
fundamental understanding of the cooperative/competitive dynamics
of complex systems that informs our understanding of neural networks
will also inform our understanding of the multinational projects,
societies and economies that produce them.  Thus they are abandoning
the top-down, unidirectional organization of the Fifth Generation
project in favor of a bottom-up, pluralistic, cooperative/competitive
strategy.  All of this shows, I think, a breadth of vision that will
carry the Japanese project much further than would the view that
the goal is no more than a new computer technology."

Refenes: The most focused of the panel presentations in my opinion. 
Claimed that the NIPT program was ambitious. Pointed out that soft 
information processing is not yet a theory, and in fact that there are 
several competing theories, no accepted model nor evidence of the 
emergence of any unifying theory. Developing the technologies necessary 
is a considerable task requiring large resources and long lead times. 
Claimed it was unclear if a massively parallel computer should be MIMD, 
SIMD, etc., and that problems of overhead explosion are not solved. 
Finally, that basic system software development is significant task. 
Suggests that as much as possible various problems should be treated 
independently, and that objectives should be narrowed. In particular to 
decouple the architecture from neurocomputing. Also claimed that high 
speed was not really necessary for neurocomputing except during training, 
and that neural networks will typically be trained once, off-site, using 
general purpose computers. (Arbib disagreed with this; I think I do too.)
Pointed out that there was an important need to provide network 
development tools, and network compilation tools. Finally noted that 
technology transfer was an essential element in international 
cooperation. 

Subsequent to the Symposium, Refenes sent me some additonal comments
specifically related to cooperation. He agrees with most other observers
that NIPT is more a Program than a Project, and that a central research
facility, such as ICOT (5th Generation) is inappropriate. He notes that
"From the Japanese point of view, there  are  two  main reasons  for
seeking International  collaboration in NIPT.  Firstly,  to  establish
a strong  presence  in  the  world research  community  and  thus
"legitimize" their Industrial exploitation of world R&D results. The aim
is to  counteract long  standing  criticism  of  the  Japanese
contribution to global R&D efforts, and  is  in  line  with  the  policy
of establishing research Laboratories in the west (e.g. NEC and Sharp in
the USA and Europe  respectively).  Secondly,  NIPT requires significant
numbers  of  high caliber researchers which, for many but not all areas
of NIPT, are  not  readily available in Japan."

He also points out that in the past there was not much Western interest
in collaboration because it was felt that Western R&D was significantly
ahead of Japan's and that collaboration would have too many one sided
benefits. In many areas, especially those related to device technology,
this is no longer true.  Collaboration on this current project could
provide Western access to Japanese markets (as much by exposure as for
any other reason), and also because of the possibility of new markets
developing from NIPT technology. This might be particularly true in
consumer electronics, as these are likely candidates for intelligence.
(I agree emphatically, and have repeatedly emphasized the role this part
of the industry plays here in Japan.) Finally, Refenes notes that
horizontal collaboration, between Japanese companies and Western
universities is likely to happen anyway, especially if NIPT becomes a
major funding source. He notes that intellectual property rights is the
key to be worked out, that the Eureka model is not a bad one to copy,
and that there need not be any central funding source, except to cover
pure collaboration costs, allowing participating governments to apply
their normal rules. 
                Dr. Paul N. Refenes
                Information Technology Division
                Department of Trade and Industry (DTI)
                Kingsgate House 
                66-74 Victoria Street
                London SW1E 6SW
                United Kingdom
                Tel: +44 (0)71-215-5000, Fax: 071-215-8318
                Email: P.REFENES@CS.UCL.AC.UK

Yuba: Explained that NIPT is thinking about a "super parallel" machine, 
which he defined as one with more than one million processors. A starting 
point for this is the ETL EM-4 parallel data-flow machine. Also explained 
that they have proposed an EM-5, to be built by 1995, with 16K 
processors, 1.3T-Ops, 655GIPS, using an object-oriented model, a 
universal network, as well as a robust architecture that uses both 
hardware and software to adapt to load scheduling. 
                Dr. Toshitsugu Yuba
                Director of Intelligent Systems Division
                Electrotechnical Laboratory
                Tsukuba Science City
                305 Japan
                Fax:    0298-58-5176,  Tel:  0298-54-5412
                Email:  YUBA@ETL.GO.JP


Hartmann: Showed an interesting slide giving the potential payoff of each 
of three technologies in processing, communication, and storage.
(Read this table across.)
                 Processing     Communication    Storage
    Photonics     Third            First           Second
    Electronics/
      Semicond    Second           Third           First
    Electronics/
      Supercond   First            Second          Third

Thought that 21st century computing will be characterized by an 
additional dimension, processing planar data through volumetric 
processors. "Dense data planes can be communicated photonically, while a 
compatible ultrafast ultradense processing capability could be achieved 
in superconductor electronics, using a storage hierarchy of semiconductor 
and photonic storage devices. 
                Dr. Alfred Hartmann
                MCC Computer Physics Lab
                3500 W. Balcones Center Drive
                Austin, Texas 78759-6509

Shimada: Discussed the pros and cons of optical computers. Clear 
description of technical problems, but conclusion "Optical 
interconnection is strongly advocated as a basis of optical computers" 
was conservative. 

Arbib: Began by saying he didn't like the term "soft" and suggested the 
use of "flexible" instead. Felt that the de-emphasis on programming was 
wrong, and that all computers will need to be programmed. Rather than 
reducing the need for programmers we will be making programming easier, 
and also easier to describe more complex issues.  Also (similar to 
Refenes) suggested that the program set more modest goals, establish 
benchmarks, and develop specific applications. He felt that 10 years was 
a very short time, and wondered if time lines of 100 or 1000 years would 
be necessary to computerize "wisdom".  His vision of 6th generation 
computer is one of "cooperative computation; the computer will be a 
problem solving network, rather than a single serial machine. The average 
user will use an expert system to configure a network of standard 
components with established network protocols, whereas the advanced user 
will "program" new networks for new applications, using environments for 
distributed programming, including design of new components (silicon 
compilers' mechatronics) and network protocols. Each 6th generation 
computer will thus be a network of subsystems, including general-purpose 
engines and special-purpose machines some of which (such as the front 
ends for perceptual processors, and devices for matrix manipulation) will 
be highly parallel machines.  Some subsystems will use optical computing; 
more speculatively, some may employ biomaterials. Another key aspect ...  
is the use of learning in artificial neural networks, which can adapt 
automatically to new tasks in a manner based on the learning principles 
of the brain. We will also see devices and computers more tightly 
integrated so the perceptual robotics will be an integral part of the 6th 
generation design, with computers including robotic actuators and multi-
modal intelligent interfaces among their subsystems.  
                 Professor Michael A. Arbib
                 Director, Center for Neural Engineering
                 University of Southern California
                 Room 03, Nedco Neurosciences Building
                 Los Angeles, CA 90089-2520
                 Tel: (213) 740-9220, Fax: (213) 746-2863
                 Email: ARBIB@POLLUX.USC.EDU
Subsequently, Arbib read a draft of this report and agreed that it
correctly summarized the content and spirit of the meeting.

OPTICAL COMPUTING.
Last year I wrote a summary of optical computing activities in Japan, see 
"optical, 17 August 1990, but other remarks are also given in the 
"nipt.90" report cited earlier.  An excellent survey of Japanese research 
in optical computing is given in "Optical Computing in Japan", S. 
Ishihara (ed) 1990, Nova Science Publishers Inc., 283 Commack Road, Suite 
300, Commack NY 11725. For additional information contact the editor 
             Mr. Satoshi Ishihara
             Senior Researcher, Optical Information Section
             Electrotechnical Laboratory
             Tsukuba Science City, 305 Japan
             Tel: (0298) 58-5625, Fax: (0298) 58-5627
             Email: ISHIHARA@ETL.GO.JP

At this meeting, two lectures were presented on this topic, by
             Prof. Takanori Okoshi
             Research Center for Advanced Science and Technology (RCAST)
             University of Tokyo
             4-6-1 Komaba, Meguro-ku, Tokyo 153 Japan
             Tel: (03) 3481-4436
and
             Dr. Alan Huang
             Head, Optical Computing Research Dept
             AT&T Bell Labs
             Room 4G514
             Crawfords Corner Road
             Holmdel, NJ 07733
             Tel: (201) 949-7631

Okoshi explained that about 35 university professors have just launched a 
Grant-in-Aid Special Research Project, to end March 1994, entitled 
"Ultrafast Ultra-Parallel Optoelectronics", and his talk centered on 
three examples of the work associated with that project. He displayed a 
figure showing operating time versus operating power on which various 
devices (silicon transistors, GaAs, Josephson Junction, etc) were 
plotted, along with boundaries associated with cooling, numbers of 
photons, and uncertainty, showing what kinds of devices will make sense 
in different regions.  For example, with 1 pico watt power, the 
uncertainty limit requires no less than 20 pico second operations, 
whereas with 1 milliwatt this can be reduced to 0.001 pico second. 
However, the photon limit, that is point below which not enough photons 
are being delivered to reliably make decisions about "0" or "1" is much 
more restrictive, forcing operation times of more than 1 micro second
with one pico watt of power. Details of this work is cited as T. Kamiya,
"private communication" in Okoshi's paper, but it was (firstly in
English) published in the Nova book mentioned above (T. Kamiya;
pp.407-417). [Thanks to Mr. Ishihara for pointing this out to me.]
Finally Okoshi concluded that "if an ultrafast optical computer is to
realized in future, we will be obliged to take advantage of the parallel
computation capability of the optical approach, because its advantage
cannot be emphasized too much on the power-speed trade-off graph."

Okoshi then went on to describe a parallel logic system OPALS, two 
dimensional surface-emitting laser arrays, and an experiment in 
fabricating both AND and EOR units using a semi-insulating GaAs-wafer-
based high-mobility epitaxial GaAs layer as active medium. As these have 
already been published in English their descriptions are omitted here.  

He ended with the general remarks that optical transmission is 
still ahead of optical computing. With respect to the latter, premature 
research is still working toward prototypes, and that many years will be 
required with lots of room for new ideas. However some technology is 
already practical, such as optical memory (CD-ROM), although 
opportunities are open for innovation. In the near future, optical 
interconnections will become more important via fiber optics, free space 
optics, and wave guides. In the former (optical interconnections) we 
already (1990) can transmit at 10Gbits/second. Okoshi also showed a table 
of the rate of improvement in communication capability, which is 
summarized below.  

 Year    M-bits/second          #Telephone Lines     #TV Channels
 1981           32                      460                1
 1987         1600                    23000               48
 1990        10000                   150000              200

Finally, he pointed out the improvements in long distance transmission 
capability in 1990, about 10km using coaxial cable, vs (experimentally 
realized) 364km using optical lines, between repeaters. 

My impression of this lecture was that it was very conservative, with a 
great deal of hesitancy to commit as to whether optical computing will 
really work, and if so how long it will take. The tone was entirely 
different from that of Huang's lecture. Huang  gave the last, and one of 
the most up-beat presentations of the work that his group is engaged in.  
He began his lecture by noting that today's supercomputer has a clock in 
the range of a few nanoseconds, while the transistor runs at a few pico 
seconds, a factor of one thousand difference, which he feels can be made 
up by use of optical connections.  Again, much of this has been published 
so I only summarize his conclusions, i.e., he expects that using optical 
output pads and various architecture modifications will allow 100GBit/sec 
output. Using more speculative weak nonlinear optical materials he also
believes that femtosecond reaction times might be achievable. He also 
stated that "optics can easily achieve over 50 times more connectivity" 
(parallelism).  


SPECIAL LECTURE.
A dazzlingly elegant lecture on symmetry was presented by Professor 
T.D.Lee, Columbia University Physics Dept, and winner of the 1957 Nobel 
Prize in Physics. Nevertheless, as far as I could tell, the only 
connection with this Symposium were his remarks about quantum 
chromodynamics (QCD) calculations requiring very fast parallel computers.  
He showed a graph on which various special purpose QCD computers' speed 
were plotted against time. Early machines (mid 1980s) were capable of 
about 100MFlops, current machines are in the range of 10GFlops, including 
one (GF11) at the speaker's institution (Columbia), and another (QCDPAX) 
at Tsukuba University.  These are still orders of magnitude below the 
performance that is required. At exactly the same time as this meeting, a 
collection of high energy physicists were also conferring at Tsukuba to 
discuss the same problem of Computing in High Energy Physics, and one of 
the speakers there showed essentially the identical slide as Lee did.  

OTHER LECTURES.
A few other lectures are worth noting briefly. 

H. Tanaka (U-Tokyo) described the Expectations and Problems in the World 
of New Info Processing in a blitz talk loaded with facts and figures 
going far too fast for me to take in or get much out of. A few details. 
He pointed out that ICOT's PIM/P machine will be generating 8GIPS next 
year. He also mentioned (***check this**) development of Micro 2000, 
using 0.1um(micro meter) CMOS, 25x25mm chips (with FPU, 5*10^7 ? transistors), 64bit 
words, 4 PE per chip, capable of 2000 MIPS.  Neural chips 125x125 
(connections?) or more, from Hitachi and Mitsubishi.  

W. Goloi (GMD-Berlin) discussed two research topics that his group has 
been working on related to methods of programming massively parallel 
systems. These are virtual shared memory, in which a distributed memory 
machine can be programmed as if it had shared memory, and virtual 
processor model in which the user can pretend that there are as many 
synchronized processors as are appropriate for the application. He 
claimed that such ideas are very well suited for real applications such 
as lattice guage (QCD), and finite difference computations. As Giloi 
publishes in English it is not necessary to detail this further, except 
to say that he made a very persuasive case (to me) about his activities, 
and seemed deeply involved in system building and testing of these ideas. 
                Dr. W.K. Giloi
                GMD Research Center for Innovative Computer
                        Systems and Technology
                Technical University of Berlin
                Hardenbergplatz 2, Berlin 12, Germany                   
                Email: GILOI@GMDTUB.UUCP


Ryoji Suzuki (U-Tokyo) discussed general principles that biology can 
teach us about computing. Suzuki is the chair of the fundamental theory 
subcommittee. These principles are 
(1) Highly parallel distributed processing, including the architecture of 
the brain, the role of efferent signals, how information is represented 
in the brain (including the possibility that chaotic behavior of a 
network could be a candidate for information coding), and understanding 
the neuron as a processing unit.  
(2) Learning and self organization, including the multilevel organization 
of the brain (molecular, network, behavioral).
(3) Integrated processing of patterned and symbolic information (this 
includes unconscious parallel processing, and later conscious serial 
processing in the recognition system, integrated processing in the motor 
control system, and mutual interaction between these).  


International Symposium on New Information Processing Technologies '91
13-14 March 1991, Tokyo Japan

Program:
 Greetings.
  Eiji Kageyama, President JIPDEC (Japan Information Processing
      Development Center)
  Kohsuke Yamamoto Director-General of Machinery and Information
      Industries Bureau, MITI

 Keynote Speech
  Toward New Information Processing Technologies
   Takemochi Ishii, Prof Univ of Tokyo & Chairperson, Research Committee
     on The New Information Processing Technology

 Special Lecture
  Symmetry Principles in Physics
   T.D. Lee, Prof Columbia Univ, Winner 1957 Nobel Prize in Physics

 Research Reports
  Information Processing Age in the 21st Century--an Impact of the New
   Information Processing Technologies in Society
   Tadashi Sasaki, Senior Advisor, Sharp Corp & Chairperson Social
     Impact Subcommittee

  What the Brain Tells Us Towards a New Computational Principle
   Ryoji Suzuki, Prof Univ of Tokyo & Chairperson Fundamental Theory
    Subcommittee 

  Perspectives of New Information Processing Technologies
   Shun-ichi Amari, Prof Univ of Tokyo & Chairperson Computer Science
    Subcommittee

 Panel Discussion
  Towards Computing in the 21st Century
   Coordinator: Shun-ishi Amari
   Panelists: 
    M.A. Arbib Univ Southern Cal, USA
    R. Eckmiller Univ of Dusseldorf, Germany
    A. Hartmann MCC USA
    B. MacLennan Univ of Tenn, USA
    P.N. Refenes Univ College London, UK
    J. Shimada ETL, Japan
    K. Toyama Kyoto Pref Univ of Medicine, Japan
    T. Yuba ETL, Japan

 Fundamental Theory Session
  Implementation of Learning Computational Principles in the Cerebellar
   Neuronal Circuity
   Masao Ito, Inst of Physical and Chemical Research, Japan
  (because of illness, this was replaced at short notice by)
  Neural Mechanisms in Association Cortex
   Hideo Sakata, Nippon Univ

  Schemas and Neural Networks: From Brain Models to Cooperative Computation
   M.A. Arbib, Univ of Southern Cal, USA

 Integrated Computing Session
  The World of New Information Processing-The Expectations and Problems
   H. Tanaka, Univ of Tokyo, Japan

  Programming Models for Massively Parallel Systems
   W.K. Giloi, Prof GMD, Germany

 Optical Computer and Devices
  Ultrafast Ultra-Parallel Optical Information Processing and Transmission
   T. Okoshi, Univ of Tokyo Japan

  The Evolving Relationship Between Optics and Electronics
   A. Huang, AT&T Bell Labs, USA

 Closing Remarks
   Masao Teruyama, Executive Director JIPDEC
              
-----------------END OF REPORT----------------------------------------