Syndetics cover image
Image from Syndetics

The myth of artificial intelligence : why computers can't think the way we do / Erik J. Larson.

By: Material type: TextTextPublisher: Cambridge, Massachusetts : The Belknap Press of Harvard University Press, 2021Copyright date: ©2021Description: viii, 312 pages ; 22 cmContent type:
  • text
Media type:
  • unmediated
Carrier type:
  • volume
ISBN:
  • 9780674983519
  • 0674983513
Subject(s):
Contents:
Part One. The simplified world: The intelligence errors -- Turing at Bletchley -- The superintelligence error -- The singularity, then and now -- Natural language understanding -- AI as technological kitsch -- Simplifications and mysteries -- Part Two. The problem of inference: Don't calculate, analyze -- The puzzle of Peirce (and Peirce's Puzzle) -- Problems with deduction and induction -- Machine learning and big data -- Abductive inference -- Inference and language I -- Inference and language II -- Part Three. The future of the myth: Myths and heroes -- AI mythology invades neuroscience -- Neocortical theories of human intelligence -- The end of science?
Summary: "Futurists are certain that humanlike AI is on the horizon, but in fact engineers have no idea how to program human reasoning. AI reasons from statistical correlations across data sets, while common sense is based heavily on conjecture. Erik Larson argues that hyping existing methods will only hold us back from developing truly humanlike AI"-- Provided by publisher.Summary: Futurists insist that AI will soon eclipse the capacities of the most gifted human mind. What hope do we have against superintelligent machines? But we aren't really on the path to developing intelligent machines. In fact, we don't even know where that path might be. A tech entrepreneur and pioneering research scientist working at the forefront of natural language processing, Erik Larson takes us on a tour of the landscape of AI to show how far we are from superintelligence, and what it would take to get there. Ever since Alan Turing, AI enthusiasts have equated artificial intelligence with human intelligence. This is a profound mistake. AI works on inductive reasoning, crunching data sets to predict outcomes. But humans don't correlate data sets: we make conjectures informed by context and experience. Human intelligence is a web of best guesses, given what we know about the world. We haven't a clue how to program this kind of intuitive reasoning, known as abduction. Yet it is the heart of common sense. That's why Alexa can't understand what you are asking, and why AI can only take us so far. Larson argues that AI hype is both bad science and bad for science. A culture of invention thrives on exploring unknowns, not overselling existing methods. Inductive AI will continue to improve at narrow tasks, but if we want to make real progress, we will need to start by more fully appreciating the only true intelligence we know--our own."-- Provided by publisher.
Holdings
Item type Home library Collection Call number Materials specified Status Date due Barcode Item holds
Adult Book Adult Book Main Library NonFiction 006.3 L334 Checked out 06/19/2024 33111010513618
Total holds: 0

Enhanced descriptions from Syndetics:

"If you want to know about AI, read this book It shows how a supposedly futuristic reverence for Artificial Intelligence retards progress when it denigrates our most irreplaceable resource for any future progress: our own human intelligence."-Peter Thiel

A cutting-edge AI researcher and tech entrepreneur debunks the fantasy that superintelligence is just a few clicks away-and argues that this myth is not just wrong, it's actively blocking innovation and distorting our ability to make the crucial next leap.

Futurists insist that AI will soon eclipse the capacities of the most gifted human mind. What hope do we have against superintelligent machines? But we aren't really on the path to developing intelligent machines. In fact, we don't even know where that path might be.

A tech entrepreneur and pioneering research scientist working at the forefront of natural language processing, Erik Larson takes us on a tour of the landscape of AI to show how far we are from superintelligence, and what it would take to get there. Ever since Alan Turing, AI enthusiasts have equated artificial intelligence with human intelligence. This is a profound mistake. AI works on inductive reasoning, crunching data sets to predict outcomes. But humans don't correlate data sets: we make conjectures informed by context and experience. Human intelligence is a web of best guesses, given what we know about the world. We haven't a clue how to program this kind of intuitive reasoning, known as abduction. Yet it is the heart of common sense. That's why Alexa can't understand what you are asking, and why AI can only take us so far.

Larson argues that AI hype is both bad science and bad for science. A culture of invention thrives on exploring unknowns, not overselling existing methods. Inductive AI will continue to improve at narrow tasks, but if we want to make real progress, we will need to start by more fully appreciating the only true intelligence we know-our own.

Includes bibliographical references and index.

Part One. The simplified world: The intelligence errors -- Turing at Bletchley -- The superintelligence error -- The singularity, then and now -- Natural language understanding -- AI as technological kitsch -- Simplifications and mysteries -- Part Two. The problem of inference: Don't calculate, analyze -- The puzzle of Peirce (and Peirce's Puzzle) -- Problems with deduction and induction -- Machine learning and big data -- Abductive inference -- Inference and language I -- Inference and language II -- Part Three. The future of the myth: Myths and heroes -- AI mythology invades neuroscience -- Neocortical theories of human intelligence -- The end of science?

"Futurists are certain that humanlike AI is on the horizon, but in fact engineers have no idea how to program human reasoning. AI reasons from statistical correlations across data sets, while common sense is based heavily on conjecture. Erik Larson argues that hyping existing methods will only hold us back from developing truly humanlike AI"-- Provided by publisher.

Futurists insist that AI will soon eclipse the capacities of the most gifted human mind. What hope do we have against superintelligent machines? But we aren't really on the path to developing intelligent machines. In fact, we don't even know where that path might be. A tech entrepreneur and pioneering research scientist working at the forefront of natural language processing, Erik Larson takes us on a tour of the landscape of AI to show how far we are from superintelligence, and what it would take to get there. Ever since Alan Turing, AI enthusiasts have equated artificial intelligence with human intelligence. This is a profound mistake. AI works on inductive reasoning, crunching data sets to predict outcomes. But humans don't correlate data sets: we make conjectures informed by context and experience. Human intelligence is a web of best guesses, given what we know about the world. We haven't a clue how to program this kind of intuitive reasoning, known as abduction. Yet it is the heart of common sense. That's why Alexa can't understand what you are asking, and why AI can only take us so far. Larson argues that AI hype is both bad science and bad for science. A culture of invention thrives on exploring unknowns, not overselling existing methods. Inductive AI will continue to improve at narrow tasks, but if we want to make real progress, we will need to start by more fully appreciating the only true intelligence we know--our own."-- Provided by publisher.

Powered by Koha