[wordup] Have We Entered a Post-Literate Technological Age?
Adam Shand
adam at shand.net
Sat Aug 22 19:38:41 EDT 2009
It's an interesting article but something doesn't sit right with me
about it and I'm going to need to think about it more.
As Adam Engst says at the beginning, the video from Google asking
"What is a browser?" is flawed as anything another other then a lark.
In my experience even asking "What program do you use to access the
Web?" would have produced much better results, and if you'd given them
a multi-choice selection I bet most of them would have been able to
tell you the name of the browser they use.
Also I think I disagree with the premise that we're getting less and
less technically literate. Perhaps I'm in an usual situation but I'm
constantly running into people who are surprisingly knowledgeable
about computers, who aren't "geeks" but have learned as part of their
job or hobbies. For example lots of digital musicians end up knowing
quite a lot about how computers work in order to do what they do. I
also think that there are huge numbers of exceptionally bright people
who know absurd quantities of information about computers, and there
are more and more of them every year. However is the rate of adoption
among relatively technically illiterate people is much higher.
</musing>,
Adam.
Via: David Nicol <davidnicol at gmail...>
Source: http://db.tidbits.com/article/10493
18 Aug 2009
Have We Entered a Post-Literate Technological Age?
by Adam C. Engst
Not long ago, Google produced a video that's making the rounds on the
Internet. In it, a Google employee asks people in Times Square in New
York City a series of questions, such as "What is a browser?", "What
browser do you use?", and "Have you heard of Google Chrome?" (Chrome
is Google's new Web browser; it's available for Windows and in pre-
release test versions for the Mac.)
Among the geek set, the video has gotten a lot of play because most of
the people in the video - who appear to be functional adults and who
use the Internet regularly - come off as highly clueless. According to
the video, only 8 percent of people queried that day knew what a
browser is.
The video is clearly not a scientific study, and suffers from horrible
methodology. It's likely, for instance, that simply asking "What is
aWeb browser?" would have produced better results, and the middle of
Times Square is undoubtedly not where most people are thinking about
the names of programs on their computers. But let's leave aside such
criticisms for the moment.
What's Your Browser?
Instead, let's take the results on face value and consider their
implications. What does it say about the technological world in which
we live that 92 percent of the people asked could not identify the
name of the program they use to access the Web? If other statistics
are to be believed, browsing the Web is the primary use of computers
today, so that's saying these people couldn't name the program they
use more than any other.
Worse, some of the answers on the video reveal that they don't even
know what a program is. A number of them identified their browser as
"a search engine" and "Google." When asked which browser he used, one
guy said "the big E," undoubtedly meaning Microsoft Internet Explorer,
which has a stylized lowercase letter E as its icon.
When the best someone can come up with is a vague recollection of a
program's icon, it says to me that we've entered a "post-literate"
technological society, one in which people have lost not just the
ability to read and write about a topic, but also the ability to speak
about it, all while retaining the ability to use it.
As someone who earns a living crafting text to help people learn how
to use technology, I found myself profoundly troubled by Google's
video. After all, if someone doesn't know what browser they use, or
even that a browser is a program on their computer, how could I
possibly expect them to be interested in buying my company's "Take
Control of Safari 4" book (written, with infinite care, by the
estimable Sharon Zardetto)? How could they even learn of its
existence, if they had no idea that Safari is a Web browser or that
they were using Safari?
(One concern that I don't explore further in this article are the
implications of a post-literate technological society for marketing
technology itself - will even technology marketing be forced to rely
solely on pretty pictures and emotional appeals? In fact, are we
already there? Apple's "I'm a Mac" ads help customers identify with
the actor playing the Mac but give little solid information, and Apple
conceals many technical specifications about the iPhone.)
But perhaps I'm barking up the wrong tree, and Google's video in fact
shows that we've taken great technological strides. TidBITS editor
Glenn Fleishman, when we were discussing the video, suggested that
it's a good thing that the Web browser has become so ubiquitous that
people need not know what it's called to use it effectively.
(Linguistically, this same devolution has happened with the Web
itself. Although it's TidBITS house style to capitalize "Web" - a
proper noun that's a shortening of "World Wide Web" - it's commonplace
to see even professionally edited publications lowercase the word,
thus de-emphasizing the fact that it's a unique thing. I think they're
wrong: "Web" should always be capitalized, as should "Internet.")
From a usability stance, I think I agree with Glenn - it's a good
thing that using the Web has become so easy that a myriad of people
can do so without even knowing the name of the tool they use to access
it. Most people just use the browser that comes bundled with their
computer, and despite the issues with Microsoft Internet Explorer over
the years, Firefox has garnered only a bit over 20 percent of the
browser market since 2004 - largely from the small subset of people
who know what a browser is.
On a platform like the iPhone, it's even easier to see this trend
toward obscuring the identity of the browser. Although Safari is the
iPhone's Web browser, and its icon is clearly named, applications like
Twitterrific can display Web content internally, and others, like
Mail, can open a Web link in Safari without ever informing you that
Safari is displaying your page. It would be difficult to quibble with
someone who didn't realize that their iPhone browser was Safari, when
in fact, much of the time they would be viewing the Web via some other
app that piggybacks on top of OS X's WebKit core.
Tied up in all of this is the fact that if what's bundled with your
computer or phone just works, you don't need to learn much more.
Dissatisfaction is the mother of exploration - only if Safari or
Internet Explorer isn't meeting your needs do you have much impetus to
learn about and switch to Firefox. So the better technology works, the
less we'll learn about how it works. I can't say that's entirely a bad
thing.
When the Thing Breaks
But I remain troubled by this post-literate inability to talk about
everyday activities and the tools used to perform them, using the
proper nouns that are not only generally agreed-upon by those in the
know, but with which the graphical representations of those tools are
clearly labeled. What happens when something goes wrong, and such a
person can't connect to the Internet at all? Can you imagine the tech
support call?
"Hi, this is tech support. How may I help you?"
"I can't get on the Google."
"OK, what browser are you using?"
"I told you - Google."
"Let's step back for a second. What program are you running on your
computer to access the Web?"
"I don't know - I just Google when I want to find something."
"Perhaps we should go a bit further back. What icon do you click on
when you want to use Google?"
"The picture? It's blue and kind of round, I think."
"OK, that's probably Internet Explorer. Can you load any Web sites
other than Google?"
"If I can't get on Google, how can I load any other Web sites?!"
I could draw this out further, but it's not far-fetched (TidBITS
staffer Doug McLean confirmed that my contrived dialog was painfully
reminiscent of tech support calls he took in a previous job). In
essence, the caller and the support rep don't share a common language.
They may both be speaking English, but that's as far as it goes, and
as soon as domain-specific words like "browser" come into play,
communication breaks down. A good support rep would undoubtedly adjust
his questions upon realizing that there's a terminology barrier, and
like Captain Kirk meeting an alien, would attempt to build up some
shared terminology based on visual appearance before attempting to
solve the problem.
Generational Problem Solving
If I asked you to tell me something about the caller in my fabricated
script above, you might fall back on stereotypes and describe the
caller as being elderly, or at least as someone who didn't grow up
with technology and therefore has come to it, perhaps grudgingly,
later in life. But what if I told you it could be a college student?
My neighbor Peter Rothbart teaches music at Ithaca College, and he's
been noticing a disturbing trend among his students. Although they're
capable of using the digital music software necessary for his courses,
he says that many of them have trouble with the most basic of computer
tasks, like saving files in a particular location on the hard disk.
Worse, if something does go wrong, he finds, they have absolutely no
idea how to solve the problem.
These aren't the sort of kids who are befuddled by high school -
they're students at a well-respected institution of higher education.
(It's the alma mater of Disney CEO Robert Iger, for instance.) No,
they're not computer science majors, but they're not being asked to
program, just to use off-the-shelf music software and perform
commonplace tasks. And now those commonplace tasks are not only
something that they apparently have never had to do, but lack the
skills to figure out on their own.
Could this inability to solve a problem with a device with which they
are otherwise familiar be a result of losing some ability to talk
about it? I wouldn't go so far as to say it's impossible to
troubleshoot without terminology, but it's less radical to suggest
that troubleshooting will become more difficult without being able to
communicate effectively with people who are experts in the field.
Not all that long ago, when adults had trouble getting something
working on a computer, they would sarcastically say that they needed a
teenager to explain it to them. That was largely true of those of us
who were teenagers in the 1980s and 1990s, but if Peter Rothbart's
experience is at all representative, today you'd be better off finding
a 30- or 40-year-old geek to help.
Don't get me wrong - I'm not saying that all young people are
incapable of solving technical problems or going beyond the basics. My
friend Dave Burbank, whose full-time job is as a fireman in the City
of Ithaca, is also a serious geek known for taking hundreds of photos
on his kids' class trips, posting constant updates via Twitter, and
updating a photo Web site for the trip before turning in each night.
His 15-year-old son Istvan is currently a 3D animator at Moving Box
Studios in Ithaca and is perfectly capable of maintaining a technical
discussion on the evolution of backup media and other such geeky topics.
In other words, there will always be geeks, and in my mind, that's a
darn good thing. The technological sophistication of those people of
my generation (I'm 41 now) who were interested in technology created
the meme that young people were fluid with technology. But what we all
missed was that being fluid with technology doesn't mean you
understand how it works or can fix it when it breaks. Being able to
dash off text messages on a mobile phone demonstrates fluidity; being
able to troubleshoot a dead Internet connection down to a corrupted
preference file or flaky cable demonstrates understanding.
So what will most members of society do when something on their
computers or smartphones fails to work? Let's not pretend that
problems won't happen - technology may have become more reliable over
time, but the rate at which things go wrong even for undemanding users
is still shamefully high.
Just recently, my father called because his iPod wouldn't show up in
iTunes. After some back and forth, I suggested that he reset the iPod,
and when he went to use it, he realized it was indeed entirely frozen.
A hard reset brought it back to life and resolved his problem, but had
he been on his own, it's possible that he - or at least someone less
experienced than he is - would have concluded it was broken and bought
another one.
This isn't a new concern. In 1909, E.M. Forster wrote a piece of early
science fiction, "The Machine Stops," in which he imagined a future in
which face-to-face contact was considered bizarre, humanity lived
underground, and the "Machine" fed all our needs. Of course, one
day...the machine stopped. More recently and amusingly, consider the
Pixar movie "Wall-E."
Cars and Computers
The obvious analogy in today's world, and one that several people have
suggested in response to our discussions, is the car. At one time,
knowledge of keeping a car running was a kind of patriarchal rite of
passage. Failure to monitor oil levels, radiator fluids, and other
factors could lead to a dead horseless carriage.
Few people know how cars work these days, and even those of us who do
have a basic understanding of them can't really work on a modern car.
If the car stutters when accelerating, or sometimes won't start, most
of us simply take it in to the repair shop and get it fixed. Problem
solved with the application of money, and of course, since cars work
relatively well these days, much less monitoring is needed. When was
the last time you checked your car's fluids?
Like so many automotive analogies, this one sounds good, but suffers
under scrutiny. In part, repairing cars has become a specialty not so
much because intelligent people couldn't understand what's wrong or
figure out how to troubleshoot it, but because the training and
equipment necessary to diagnose problems and effect repairs have
themselves become highly specialized. Gone are the days when you could
fix a car with a few screwdrivers and a set of wrenches. The shops all
download data from the car computer for diagnosis.
But the more serious problem with the analogy is that cars are single-
purpose machines - they do one thing, and they do it moderately well.
Thus, the type of problems they can suffer, while troubling,
frustrating, and sometimes seemingly inexplicable, are still
relatively limited in scope, more like a household appliance. How
often do you have to check the inner workings of your washing machine
or refrigerator?
In contrast, computers are general purpose machines that can perform a
vast number of wildly different tasks, such as browsing the Web,
reading email, writing a book, developing a company budget, tracking a
database of customers, composing music, editing video, and so on.
We have up-and-coming geeks like Istvan Burbank, but even bright young
men like Istvan have their limits. While I'd happily ask him to fix a
Mac that's not booting, I'm not sure he'd have any idea how to help if
I showed him a PDF where the text on some pages appeared darker and
bitmapped when viewed in certain PDF readers (even Adobe hasn't been
able to fix that problem reliably for me). There's a limit to how much
any one of us can learn, but there's no limit to what a computer can do.
In a way, this is an odd situation for those of us who grew up with
the personal computer. Before Apple, before the IBM PC, we had
mainframes and minicomputers that we interacted with via dumb
terminals. You couldn't do all that much, and you were sharing
resources with many other people, but you also didn't have to worry
about things going wrong as much, because when they did, the computer
operators would fix them.
They were the gatekeepers, the wizards who controlled access and could
say who was allowed to do what. Personal computers were supposed to
democratize computing so anyone and everyone could do their own work.
While that's come to pass in some ways, it seems to me that we've
returned to the days when you need a wizard to solve problems or do
anything beyond the norm. It's a somewhat uncomfortable situation,
since those of us who grew up with personal computers are finding that
we're the new wizards.
Technological Illiteracy
So how did we get here? I'd argue that Apple - and we Macintosh users
- are perhaps more to blame for this state of affairs than any other
group. After all, no one has championed usability like Apple, with the
Mac's vaunted ease-of-use. For years, many Mac users scoffed at
manuals. "Why would anyone need a manual when the program is so easy
to use?" they'd ask. It was a fair point, for the users of the time,
who were highly interested in the technology, well versed in how it
actually worked under the hood, and amenable to poking and prodding
when things didn't go right.
But then we got our wish, and ever more companies started writing
software that was easy enough for most people to use without reading a
manual, at least at some level. That was the death of documentation, a
phrase I first coined more than 10 years ago (see "The Death of
Documentation," 1998-05-04). Of course, it was really the death of the
manual, and technical books have remained popular, in part because of
the lack of the manual (how else could David Pogue have made a mint on
his Missing Manual series?).
Even still, back when I started writing technical books in the early-
to-mid 1990s, the average computer book would sell about 12,000
copies. Today, despite a vastly larger audience (though with much more
competition), 5,000 copies is considered acceptable.
I'd argue there was a more insidious effect from the loss of manuals -
it caused an entire class of users to become technologically
functional while remaining technologically illiterate. When I asked my
mother-in-law, Linda Byard, what browser she used, she became somewhat
flustered and guessed at Outlook. This is a woman who uses the Web
fluidly and for all sorts of tasks far more sophisticated than simply
browsing static Web pages. And yet, the fact that she used Internet
Explorer to do so escaped her.
As the conversation proceeded (and keep in mind that my father-in-law,
Cory Byard, helped design personal computers for NCR back in the 1980s
and now consults on massive database projects for Teradata - Tonya
didn't grow up in a technologically backward household), it came out
that Linda had stopped reading about how to use technology when
manuals gave way to inferior online help.
She didn't stop learning how to use various programs, but without any
sort of formalized instruction or written reference, she lost the
terminology necessary to talk about the technology she was using. Of
course, she had Cory around to fix anything that went wrong, and she
said that the same was true of all her peers too - there was always
someone technologically adept in the family to deal with troubles.
Although it's harder to pin this loss of technological literacy on the
lack of manuals when looking at schoolkids, the problem isn't
necessarily being addressed there either. When my son Tristan was in
second and third grade in the public schools in Ithaca, NY, the
closest he was taught to computer skills were typing (not a terrible
idea, but tricky for kids whose hands aren't large enough to touch-
type properly) and PowerPoint.
Although some level of presentation skills are certainly worthwhile,
why would you have second graders focus on something that's guaranteed
to be different (if not entirely obsolete) by the time they're in
college?
I'd argue that some of the basics of technology - the concept of a
program as a set of instructions and the essentials of networking -
would be both more compelling for kids and more useful for
understanding the way the world works later in life.
When TidBITS contributing editor Matt Neuburg tried to teach a group
of his friends' kids REALbasic one summer, he found himself frustrated
at almost every turn - they lacked the conceptual underpinning that
they could make the computer do something. And more important, they
didn't care, since they were accustomed to technology just working. It
wasn't until he got them to draw a stick figure and, by changing the
location of its parts repeatedly, make it walk across the screen, that
one of them said, "Hey, this must be how my video games are made."
And networking? No, you don't need to know it works to use the
Internet, but isn't it wondrous that an email message sent to a friend
on the other side of the globe in Australia is broken up into many
small pieces, shuttled from computer to computer at nearly the speed
of light, and reassembled at its destination, no more than seconds
later? Wouldn't it be fun to act out a packet-switched network with an
entire class of second graders and the pieces of a floor puzzle? Or at
least more fun than PowerPoint?
Luckily, this lack in the public education system isn't uniform. Glenn
Fleishman's son Ben is about to enter a public elementary school in
Seattle, where the beginning curriculum teaches kids about opening,
saving, and printing files; later, it moves to task-based - not
program-oriented - computer projects. That's much better.
But I digress.
Illiteracy Stifling Innovation?
My more serious concern with our society's odd fluency with a
technology that we cannot easily communicate about is that it might
slowly stifle innovation. Already we're in a situation where browser
innovation is almost the sole province of Apple and Microsoft, with
contributions from Mozilla, Google, and maybe Opera.
Iterative changes from the incumbents can be worked in, since everyone
will be forced to accept them, but does it become harder to convince
most people to try a ground-breaking new technology because it's
different, because it's talked about using strange new terminology,
and perhaps because no paradigm-shifting new technology can by
definition be so easy to use that it doesn't require some level of
training? I fear that might be the case.
In the dawn of the computer age, the stakes weren't as high and the
market wasn't as large, so I'd suggest that companies were more likely
to take risks on innovative technologies that might appeal to only a
small subset of the population. Today, with everyone using technology,
I suspect that business plans and funding proposals all assume a large
potential audience, which in turn causes the ideas to be vetted more
on their business chances than their technological innovation.
Put another way, there have always been technological haves and have
nots, but since there was no chance of selling technology to the have
nots, technology of the past was less limited by the literacy of the
audience. Since the technologically illiterate are not just buying
technology now, but are the primary market for it, that has to be
affecting the kind of ideas that get funding and are being developed
in a real way.
Plus, think back to the point about dissatisfaction being the mother
of exploration. We geeks may be willing to belly up to the new
technology feeding trough since we're never satisfied. But once
technology reaches a certain plateau of working well enough, if this
lack of technological literacy is indeed a more general concern,
spreading technological successes into the population as a whole may
become all the more difficult.
I'm fully aware that my musings here are largely hypothetical and
based on anecdotal evidence. But I think there's a new technology on
the horizon that could serve as a test of my theory that anything
sufficiently innovative will face an uphill battle due to the
technological illiteracy of the user base: Google Wave.
For those who didn't see Google's announcement of Google Wave (we
didn't cover it in TidBITS at the time because it was a technology
announcement, not a service that people could use), it's a personal
communication and collaboration tool that's designed to merge the
strengths of email, instant messaging, wikis, and social networking
services. (You can read more about it at Wikipedia.)
On the plus side, Google Wave has the power of Google behind it, and
Google could potentially merge it into Gmail, thus introducing it to
146 million users nearly instantaneously. But Google Wave will
undoubtedly be quite different from Gmail, and will require a learning
curve. Will that hamper its adoption, since email and instant
messaging and other service work well enough that people aren't
sufficiently dissatisfied to learn about and try Google Wave? Only
time will tell.
-------------- next part --------------
An HTML attachment was scrubbed...
URL: http://lists.spack.org/pipermail/wordup/attachments/20090823/80b7a673/attachment.htm
More information about the wordup
mailing list