[...]
Underpinning much of the buzz over artificial intelligence in London and elsewhere is the implicit premise that AI is thetransformative
technology of the moment, or maybe of the decade, or even of the
century or, well, just about ever. Promises like the AI Summit’s claim
that the technology goes “beyond the hype” to “deliver real value in
business” only drives the corporate feeding frenzy among executives
desperate not to be left behind.
But there is something else going
on “beyond the hype,” something that ought to be disconcerting for AI
boosters: Among those closest to the cutting edge of machine learning,
there is a sense–perhaps a faint but creeping suspicion—that deep
learning, the techniques which underpin most of what people think of as
ground-breaking AI, may not deliver on its promise.
For one thing,
there’s a growing consensus among AI researchers that deep learning
alone will probably not get us to artificial general intelligence (the
idea of a single piece of software that is more intelligent than humans
at a wide variety of tasks). But there’s also a growing fear that AI may
not create systems that are reliably useful for even a narrower set of
real-world challenges, like autonomous driving or making investment
decisions.
Filip Piekniewski, an expert on computer vision in San
Diego who is currently the principal AI scientist for Koh Young
Technology Inc, a company the builds 3D measurement devices, recently
kicked off a debate with a viral blog post predicting a looming “AI Winter,”
a period of disillusionment and evaporating funding for AI-related
research. “(The field has experienced several such periods over the past
half-century.)
Piekniewski’s evidence? The pace of AI
breakthroughs seems to be slowing and those breakthroughs that are
occurring seem to require ever-larger amounts of data and computer
power. Several top AI researchers who had been hired by big tech firms
to head in-house AI labs have left or moved into slightly less prominent
roles. And most importantly, Piekniewski argues that the recent crashes of
self-driving cars point to fundamental issues with the ability of deep
learning to handle the complexity of the real world. Even more notable
than the crashes, he says, is how often machines lose confidence in
their ability to make safe decisions and cede control back to human
drivers.
Piekniewski also references the work of New York
University’s Gary Marcus, who earlier this year published a
much-discussed paper critiquing the failings of today’s deep learning
systems. This software, Marcus argues, can identify objects in images,
but lacks any model of the real world. As a result, they often can’t
handle new situations, even if they are very similar to the ones they’ve
been trained to perform. For instance, the DeepMind algorithm that
performs so well at the Atari game “Breakout”—and which the company
often highlights in public presentations—does terribly if it is suddenly
presented with a different-sized paddle, whereas a top human player
would likely find the larger paddle wasn’t much of a handicap.
[..]
Fonte: aqui
Assinar:
Postar comentários (Atom)
Nenhum comentário:
Postar um comentário