[net.works] ICONS: Passing Fad or New Found Wisdom?

Vongehren%OFFICE-12@sri-unix.UUCP (06/22/83)

My current work has brought me to the question as stated in the subject.  Rather
than elaborate on it in this message, I would like to make contact with others 
who would be able to contribute to the discussion.  I will, however, briefly 
state a few of the questions which fall out of this:

A - Do you think that the current interest in the use of icons on terminals and 
computer displays is just a passing fad?

B - Aren't some of the current 'graphics' just a little too 'cute' e.g. IBM 
upper-case 'lock'.  Is this just a sign of an immature field?  Will the 
marketplace tolerate this long enough for growth and maturity to occur?

C - What must happen for this field to mature?

D - Standardization of Signs and Symbols has occurred in other fields, e.g. 
Traffic.  Is there any effort to standardize within the computer field?  Should 
this be done?

E - Are there any obvious indicators for when it is inappropriate to use an icon
in place of a word?

F - What would you offer as guiding principles for the use of icons in computer 
displays?  Will these differ for icon use on keyboards?

I'll be glad to hear from you if you are willing to do some (or have done some) 
thinking on these issues.

                    Ed vonGehren,
                    Bell-Northern Research

-------

laura@utcsstat.UUCP (06/26/83)

I am posting this because I dont know how to reach you (Vogehren@OFFICE-12).
If you want to send mail to me, please find out how mail from the
usenet is supposed to be addressed so that it will get to you rather
than come back to me with a nasty message about "attempted break-in
of ARPA security". I have had rotten luck lately with mail, and
if anyone is keeping track I have probably already been labelled as
"dangerous foreign spy with multiple attempts to crack security on
multiple machines" -- and all I want to do is *TALK* to five different
people on the Arpa-net.

Tonight I am in a cynical mood; tomorrow everything may look rosier
--- but i doubt it. The problem with "cute" things is that they
catch on among the illiterate and/or stupid users. If computers
are going out to the masses then I think we are stuck with "cute".
There are a lot of people who cant read; there are even more who
dont want to read. I read that a 1975 survey reported that only
19% of Americans reported reading even *one* non-work related
book a year. If you forget the people whose one book was the
novelisation of the movie hit of the year or a Harlequin romance
you are left with the staggering figure that 9% of the population
of the USA is buying all the books. These figures did not include
magazines or school textbooks, but nonetheless the very magnitude
of the problem is enough to make your blood run cold.

Canada is no better, I am sure. My mother teaches grade 8 and
the single worst thing she can say to her class is "Today we
are going to spend the next 2 hours reading." For the past five
years she has kept a record, and to date she has found only
three or four students who actually admitted to READING for
PLEASURE.

I dont know what one can do about this. When I recognise how
much of everything I do is based on things that I have read I
begin to conclude that anyone who doesnt read just isnt quite
human. Then I remember that it is I who is in the minority; it
is I who is not quite human.

Be that as it may, the conclusion that can be drawn is that people
who do not read do not apreciate "cleanness, conciseness and elegance in
ideas".  Most ideas are found in books -- unless you read or travel
you are imprisoned with the local ideas of your area with only the 
television or the movie theater to bring you other ideas. Most of
tv is dreck, and most watchers of tv arent watching the documentaries
on PBS. TV and movies are both visual media also, and do little or
nothing to present abstract ideas. TV and movies both exploit the
themes of adventure and so-called "human interest" -- it is 
impossible to show on a screen the precise mental glow you get when
your compiler starts generating correct code, or you understand some
of the significance of E=mc**2. You can present the information and
hope that the lights go on in somebody's head, but you cannot do it
for him. How much easier it is to show a shark eating a swimmer...

But what, you may be asking, has this to do with ICONS? Everything.
If you want to sell to the functional illiterates then you are
going to have to be understood by them. I go down to the local
shopping centre and see that most people have difficulty with the
simplest menu-driven "show me where the shoe stores are" type of
computer aid. The have trouble understanding simple commands
like "push the panel beside the type of store you want". What
do they like? The pictures. The silly sleezy drawings that crunch
the CPU so much to draw that those of us who know what we are doing
wish there was some way the touch-panels had "touch-ahead" because
we dont need a picture of a book to know that we want a book store,
and we could do without the wait for the drawing of the various pictures.

ICONS, I am afraid, are here to stay -- at least in devices that
will be made available to John Q. Public. They have to be.
Old JQ thhks that E.T. is a much better movie than "Das Boot".
Cute friendly monsters are "in" now, but "cute" is always in.
Old JQ needs picures because JQ doesnt read -- and any manufacturer
interested in selling to the public understands this and will
continue to sell "cute".

Laura Creighton
utzoo!utcsstat!laura

ravi@hcr.UUCP (Ravi Pandya) (06/26/83)

Whew! You certainly were feeling cynical -- but your article had a lot of
unpleasant (but true) points that set my thoughts on a path they haven't
been along in a while.  It is quite chilling to see how few people use their
intelligence at all (let alone to its full capacity).  Following the rise of
anti-science movements like the Creationists, the more rabid of the ecology
groups (although some of them have many valid and intelligent points), the
mystic and pseudo-scientific cults (like the Scientologists that our windows
look down upon), over the last few years gives the impression of an almost
palpable fear of rational and scientific thought, a kind of "look where
thinking has brought us -- let's stop thinking" attitude.  If anything is
going to destroy us, I think this is the prime candidate, for people who
don't think can be swayed with ease by any skilful demagogue with a bit of
media coverage.  Look at Ronald Reagan, the president of the United States,
a "world superpower", in command of enough power to destroy the earth many
times over -- a B-grade movie actor.  (Pierre Trudeau, despite his faults,
is a well-read, well-educated, well-travelled, intelligent man -- he toured
through the Third World on a motorcycle in his thirties to get first-hand
experience with the conditions there).  

So what is there to be done?  It would be impossible to give 180 million
Americans (I'm still amazed by that figure!) an idea of what that sense of
wonder, of discovery, of enlightenment is really like, and (cynically) if
you try something of that massive a scope, you're simply going to be
disappointed.  I content myself with designing tools that make those who do
want to think and create more effective at the task, and perhaps to make the
activity seem less foreign to those unaccustomed to it.  Despite their
"cuteness", the iconic referents and similar graphical user-interface
techniques are the best way that I can see to do that.  If you don't have to
spend your time fighting your way through the Sargasso Seas of the abysmal
user interface facilities of Unix (which is one of the better traditional
systems), then you can spend more time solving the problem you're working
on.

CPU power requirement isn't the issue -- mind power requirement is.
Computer science suffers greatly from the archaic idea that the computer is
still the expensive part of the system, and many people waste much time
adapting themselves to computer systems that are well-designed, but with the
wrong set of design parameters; the designers tried to minimize demand on
the computer rather than demand on the user.

	"You can lead a whore to culture, but you can't make her think"
		[I think either George Bernard Shaw or Oscar Wilde]

	--ravi

ravi@hcr.UUCP (Ravi Pandya) (06/26/83)

I have some more detailed comments on the use of icons that didn't get into
my previous article because I was more concerned with the issues that Laura
raised.  So, for what they're worth, here are my thoughts on the issue.

A - I am quite pleased to see the use of icons and other visual techniques
in computer systems, and I certainly hope (and think) it will continue, for
a couple of reasons.  To begin, I think that because computers have been
communicating using a medium that is highly abstract and symbolized (print)
they have been limited in their usage to those who are willing to take the
time required to master the many inconsistent, illogical, and complex symbol
systems that software designers have seen fit to inflict upon their "lusers"
(that term itself holds a wealth of information about why things are as they
are).  Thus, many people have been unable to take advantage of the
"augmentation of the intellect" (to use Engelbart's term) that a
well-designed computer system can bestow upon its user, and have been
wasting much time with less efficient media; on the other hand, those who
have taken the time to learn such systems have been wasting much time
fighting with systems that were, in many cases, actively preventing them
from doing what they wanted to do.  I also think that computer science has
suffered greatly from its incestuous nature (e.g.  computer systems are
designed by computer scientists for the use of other computer scientists who
are designing computer systems) in the development of some rather gross
mutations (for an excellent example, take a close and critical look at C
some day and ask yourself if it can be anything else but a rather virulent
form of conceptual cancer), and the rise in the use of graphical and visual
techniques may provide some cross-fertilization from the very different,
lively, and creative field of graphic design (as well as other related
areas).

B, C - The cuteness of which you speak is frequently a result of
poorly-designed iconic systems -- the concept of "shift lock" has in fact
little to do with a picture of a lock, and so the process of establishing a
connection is closer to figuring out the punch line of a joke than
understanding a piece of visual communication.  Effective visual
communication is not "cute", it simply gets the message across cleanly and
quickly.

D - Standards are a rather moot point.  I would inevitably pick a
creatively, effectively designed system over one that rigidly follows a
standard that is a mediocre compromise (as standards tend to be).  I think
that encouraging good design is more important than establishing a standard,
but the latter at least avoids having to learn seventeen different mediocre
systems -- with a standard, you only have to learn one mediocre system (and
with a good system, you hardly need to learn it at all -- it meshes with the
way you think so well that the boundary between you and the system
disappears entirely).

E - For guidelines on the use of icons, a good source is the
well-established body of knowledge in the graphic arts.  Icons, because they
operate below the symbolic (linguistic) level, are appropriate when they
represent a package of concepts that can be grasped almost intuitively
without having to really "think" about them at all.  So using an icon in the
middle of a sentence of words (like those oh-so-cute children's books) is a
very ineffective way to communicate, as you have to make a connection
between two widely separated modes of understanding.  Icons *can* be arranged
in structural relationships as words can, but only when the relationships
can also be expressed in a simple, graphic, and visual manner -- a
dependency network, for example.

F - My idea of guiding principles for the use of icons, other than those
above, are the same as for the design of any interactive system: solve the
most general problem in a way that gives the user the most power with the
least conceptual clutter; if you do this, and don't let your design be
limited by short-sighted concerns about "efficiency", you can usually create
something for which there exists an elegant, general, and very efficient
algorithm.  You thus end up with a much, much better system and pay a small
price in "efficiency" for a large gain in effectiveness.

Well, I guess it's time to get off the soapbox now.  Good luck.
	--ravi

					...!decvax!hcr!hcrvax!ravi

brucec@orca.UUCP (07/07/83)

IN-REPLY-TO: Vongehren@OFFICE-12 Wed Jun 22 00:50:00 1983
Subject: ICONS: Passing Fad or New Found Wisdom?

I meant to write this article and get it out to the net some time ago,
but got caught up in work, and completely forgot that I intended to do
it.  Since this is a subject about which I have strong feelings, and
since it relates to the work that I do, I decided "better late than
never," and went ahead to write this.

	A - Do you think that the current interest in the use of icons
	on terminals and computer displays is just a passing fad?

No.  The purpose of icons is to increase the bandwidth of the
information channel from a computer to a user.  It is a
well-established principle that people can recognize relatively
abstract geometric symbols faster than clusters of words in a language
they can read.  That's not to say that the recognition of an icon
can't be aided, especially for new users of a system, by one well
chosen caption word.  It seems to be the time in cmprehending multiple
words as phrases that slows down the recognition of written symbols.
The experienced user can easily ignore a caption, or even recognize it
as part of the abstract shape of the icon rather than a word (I doubt
you perceive the writing on a stop sign as a word after years of
driving).  I make no claim that icons can or should be "instantly
recognizable" in any sense without a period of training in recognition
of the icons in use on a system.  I do claim that once trained, a user
can operate faster using familiar icons than text.


	B - Aren't some of the current 'graphics' just a little too
	'cute' e.g. IBM upper-case 'lock'.  Is this just a sign of
	an immature field?  Will the marketplace tolerate this long
	enough for growth and maturity to occur?

I agree that there seems to be more marketing than cognitive
psychology in the design of a lot of icons.  Oh, well, it doesn't
really hurt anything but the feelings of people who want this field to
be taken seriously.  The real questions in the design of icons are
distinguishability between different icons on a given system, and size
and complexity of shape.  Even the cutest icon becomes just a shape to
be recognized after enough use (as an experiment, try writing the word
"thrash" down on paper and staring at it for a while.  Pretty soon, it
stops looking like a word.).


	C - What must happen for this field to mature?

I don't see anything that will stop it from maturing.  For the market
to take thsi form of interface seriously, it has to become easily
available (i.e. cheap), and easily usable (i.e. fast).  The economics
of the microcomuter industry guarantee that this will happen, as more
powerful hardware gets cheaper, and as more powerful operatin systems
become more common.  As long as there is a paradigm held in front of
the vendors in the marketplace, giving the less inovative of them a
model to copy, then progress will continue.  As flawed as I think the
Xerox Star and Appple Lisa are, they have provided such paradigms.
Enough vendors are working on copies or extensions of these models
that I think the field has the critical mass to keep going.


	D - Standardization of Signs and Symbols has occurred in other
	fields, e.g. Traffic.  Is there any effort to standardize within
	the computer field?  Should this be done?

Having been involved in the standards process in the computer field, I
can truthfully say that everyone wants a standard - his own.  The
field of iconic interfaces is too new for standards to be desirable,
since we don't yet know what we want.  It's probably best to leave
things alone until something like a de facto standard arises.


	E - Are there any obvious indicators for when it is inappropriate
	to use an icon in place of a word?

Clearly there is amaximum number of distinguishable icons in a system,
depending on the criteria used for desired speed of recognition.  The
less often an icon is used, the less recognizable it is.  So sometimes
it's better to use a word or two than an icon that the user has never
seen before.  Icons can be ambiguous in some circumstances, so for
communication requiring great precision (error messages from the
computer fro instance) icons are not desirable.  And I for one would
rather see a text string saying "the building is on fire" than a
little picture of a house with flames coming out.


	F - What would you offer as guiding principles for the use of
	icons in computer displays?  Will these differ for icon use on
	keyboards?

I can't offer any guiding principles since I haven't yet gone through
the cycle of designing and testing the effect of a set of icons.  I
will point out that the legends on a keyboard necessarily have a
different look from the icons on a display screen, due to the
difference betweem reflective and luminant shades, the texture of the
surfaces, and the lower resolution of the key legend.  This means that
key legends must be more cartoon-like (hard edges, high contrast,
etc.) than screen icons.  To date, most of the icons I've seen on
screens have been cartoon-like, possibly with some use of gray scale
half-toning.   I suspect that this is more a lack of imagination than a
basic limitation of the technique.


				Bruce Cohen

				UUCP:	...!teklabs!tekecs!brucec
				CSNET:	tekecs!brucec@tektronix
				ARPA:	tekecs!brucec.tektronix@rand-relay