Imagine I told you in 1970 that I was going to invent a
wondrous tool. This new tool would make it possible for anyone with access —
and most of humanity would have access — to quickly communicate and collaborate
with anyone else. It would store nearly the sum of human knowledge and thought up
to that point, and all of it would be searchable, sortable, and portable. Text
could be instantly translated from one language to another, news would be
immediately available from all over the world, and it would take no longer for
a scientist to download a journal paper from 15 years ago than to flip to an
entry in the latest issue.
اضافة اعلان
What would you have predicted this leap in information and
communication and collaboration would do for humanity? How much faster would
our economies grow?
Now imagine I told you that I was going to invent a sinister
tool (perhaps, while telling you this, I would cackle). As people used it,
their attention spans would degrade, as the tool would constantly shift their
focus, weakening their powers of concentration and contemplation. This tool
would show people whatever it is they found most difficult to look away from —
which would often be what was most threatening about the world, from the worst
ideas of their political opponents to the deep injustices of their society. It
would fit in their pockets and glow on their nightstands and never truly be
quiet; there would never be a moment when people could be free of the sense
that the pile of messages and warnings and tasks needed to be checked.
Our society wide obsession with speed and efficiency has given us a flawed model of human cognition that I have come to think of as the Matrix theory of knowledge. Many of us wish we could use the little jack from “The Matrix” to download the knowledge of a book … into our heads, and then we would have it, instantly. But that misses much of what is really happening when we spend nine hours reading a biography.
What would you have thought this engine of distraction,
division and cognitive fracture would do to humanity?
Thinking of the internet in these terms helps solve an
economic mystery. The embarrassing truth is that productivity growth — how much
more we can make with the same number of people and factories and land — was
far faster for much of the 20th century than it is now. We average about half
the productivity growth rate today that we saw in the 1950s and ’60s. That
means stagnating incomes, sluggish economies and a political culture that’s
more about fighting over what we have than distributing the riches and wonders
we have gained. So what went wrong?
You can think of two ways the internet could have sped up
productivity growth. The first way was obvious: by allowing us to do what we
were already doing and do it more easily and quickly. And that happened. You
can see a bump in productivity growth from roughly 1995 to 2005 as companies
digitized their operations. But it’s the second way that was always more
important: By connecting humanity to itself and to nearly its entire storehouse
of information, the internet could have made us smarter and more capable as a
collective.
I do not think that promise proved false, exactly. Even in
working on this article, it was true for me: The speed with which I could find
information, sort through research, contact experts — it’s marvelous. Even so,
I doubt I wrote this faster than I would have in 1970. Much of my mind was
preoccupied by the constant effort needed just to hold a train of thought in a
digital environment designed to distract, agitate, and entertain me.
And I am not alone.
Gloria Mark, a professor of information science at the
University of California, Irvine, and the author of “Attention Span,” started
researching the way people used computers in 2004. The average time people
spent on a single screen was 2.5 minutes. “I was astounded,” she told me. “That
was so much worse than I’d thought it would be.” But that was just the
beginning. By 2012, Mark and her colleagues found the average time on a single
task was 75 seconds. Now it’s down to about 47.
This is an acid bath for human cognition. Multitasking is
mostly a myth. We can focus on one thing at a time. “It’s like we have an
internal whiteboard in our minds,” Mark said. “If I’m working on one task, I
have all the info I need on that mental whiteboard. Then I switch to email. I
have to mentally erase that whiteboard and write all the information I need to
do email. And just like on a real whiteboard, there can be a residue in our
minds. We may still be thinking of something from three tasks ago.”
A question to ask about large language models, then, is where does trustworthiness not matter? Those are the areas where adoption will be fastest.
The cost is in more than just performance. Mark and others
in her field have hooked people to blood pressure machines and heart rate
monitors and measured chemicals in the blood. The constant switching makes us
stressed and irritable. I didn’t exactly need experiments to prove that — I
live that, and you probably do, too — but it was depressing to hear it
confirmed.
Which brings me to artificial intelligence. Here I’m talking
about the systems we are seeing now: large language models like OpenAI’s GPT-4
and Google’s Bard. What these systems do, for the most part, is summarize
information they have been shown and create content that resembles it. I
recognize that sentence can sound a bit dismissive, but it shouldn’t: That’s a
huge amount of what human beings do, too.
Already, we are being told that AI is making coders and
customer service representatives and writers more productive. At least one
chief executive plans to add ChatGPT use in employee performance evaluations.
But I am skeptical of this early hype. It is measuring AI’s potential benefits
without considering its likely costs — the same mistake we made with the
internet.
I worry we are headed in the wrong direction in at least
three ways.
One is that these systems will do more to distract and
entertain than to focus. Right now, the large language models tend to
hallucinate information: Ask them to answer a complex question, and you will
receive a convincing, erudite response in which key facts and citations are
often made up. I suspect this will slow their widespread use in important
industries much more than is being admitted, akin to the way driverless cars
have been tough to roll out because they need to be perfectly reliable rather
than just pretty good.
A question to ask about large language models, then, is
where does trustworthiness not matter? Those are the areas where adoption will
be fastest. An example from media is telling, I think. CNET, the technology
website, quietly started using these models to write articles, with humans
editing the pieces. But the process failed. Forty-one of the 77 AI-generated
articles proved to have errors the editors missed, and CNET, embarrassed,
paused the program. BuzzFeed, which recently shuttered its news division, is
racing ahead with using AI to generate quizzes and travel guides. Many of the
results have been shoddy, but it doesn’t really matter. A BuzzFeed quiz doesn’t
have to be reliable.
Marcela Martin, BuzzFeed’s president, encapsulated my next
worry nicely when she told investors, “Instead of generating 10 ideas in a
minute, AI can generate hundreds of ideas in a second.” She meant that as a
good thing, but is it? Imagine that multiplied across the economy. Someone
somewhere will have to process all that information. What will this do to
productivity?
One lesson of the digital age is that more is not always
better. More emails and more reports and more Slacks and more tweets and more
videos and more news articles and more slide decks and more Zoom calls have not
led, it seems, to more great ideas. “We can produce more information,” Mark
said. “But that means there’s more information for us to process. Our
processing capability is the bottleneck.”
The embarrassing truth is that productivity growth — how much more we can make with the same number of people and factories and land — was far faster for much of the 20th century than it is now.
Email and chat systems like Slack offer useful analogies
here. Both are widely used across the economy. Both were initially sold as
productivity boosters, allowing more communication to take place faster. And as
anyone who uses them knows, the productivity gains — though real — are more
than matched by the cost of being buried under vastly more communication, much
of it junk and nonsense.
The magic of a large language model is that it can produce a
document of almost any length in almost any style, with a minimum of user
effort. Few have thought through the costs that will impose on those who are
supposed to respond to all this new text. One of my favorite examples of this
comes from The Economist, which imagined NIMBYs — but really, pick your
interest group — using GPT-4 to rapidly produce a 1,000-page complaint opposing
a new development. Someone, of course, will then have to respond to that
complaint. Will that really speed up our ability to build housing?
My third concern is related to that use of AI: Even if those
summaries and drafts are pretty good, something is lost in the outsourcing.
Part of my job is reading 100-page Supreme Court documents and composing crummy
first drafts of columns. It would certainly be faster for me to have AI do that
work. But the increased efficiency would come at the cost of new ideas and
deeper insights.
Our society wide obsession with speed and efficiency has
given us a flawed model of human cognition that I have come to think of as the
Matrix theory of knowledge. Many of us wish we could use the little jack from
“The Matrix” to download the knowledge of a book (or, to use the movie’s
example, a kung fu master) into our heads, and then we would have it,
instantly. But that misses much of what is really happening when we spend nine
hours reading a biography. It is the time inside that book spent drawing
connections to what we know and having thoughts we would not otherwise have had
that matters.
“Nobody likes to write reports or do emails, but we want to
stay in touch with information,” Mark said. “We learn when we deeply process
information. If we’re removed from that and we’re delegating everything to GPT
— having it summarize and write reports for us — we’re not connecting to that
information.”
We understand this intuitively when it is applied to
students. No one thinks that reading the SparkNotes summary of a great piece of
literature is akin to actually reading the book. And no one thinks that if
students have ChatGPT write their essays, they have cleverly boosted their
productivity rather than lost the opportunity to learn. The analogy to office
work is not perfect — there are many dull tasks worth automating so people can
spend their time on more creative pursuits — but the dangers of overautomating
cognitive and creative processes are real.
These are old concerns, of course. Socrates questioned the
use of writing (recorded, ironically, by Plato), worrying that “if men learn
this, it will implant forgetfulness in their souls; they will cease to exercise
memory because they rely on that which is written, calling things to
remembrance no longer from within themselves but by means of external marks.” I
think the trade-off here was worth it — I am, after all, a writer — but it was
a trade-off. Human beings really did lose faculties of memory we once had.
To make good on its promise, artificial intelligence needs
to deepen human intelligence. And that means human beings need to build AI, and
build the workflows and office environments around it, in ways that don’t
overwhelm and distract and diminish us. We failed that test with the internet.
Let’s not fail it with AI.
Read more Opinion and Analysis
Jordan News