Amazon cover image
Image from Amazon.com
Custom cover image
Custom cover image

The Alignment problem : machine learning and human values / Brian Christian.

By: Publisher: New York, NY : W.W. Norton & Company, 2020Description: xii, 476 pages ; 25 cmISBN:
  • 9780393868333
  • 9780393635829
Subject(s): LOC classification:
  • Q334.7 .C488A 2020
Summary: "A jaw-dropping exploration of everything that goes wrong when we build AI systems-and the movement to fix them. Today's "machine-learning" systems, trained by data, are so effective that we've invited them to see and hear for us-and to make decisions on our behalf. But alarm bells are ringing. Systems cull résumés until, years later, we discover that they have inherent gender biases. Algorithms decide bail and parole-and appear to assess black and white defendants differently. We can no longer assume that our mortgage application, or even our medical tests, will be seen by human eyes. And autonomous vehicles on our streets can injure or kill. When systems we attempt to teach will not, in the end, do what we want or what we expect, ethical and potentially existential risks emerge. Researchers call this the alignment problem. In best-selling author Brian Christian's riveting account, we meet the alignment problem's "first-responders," and learn their ambitious plan to solve it before our hands are completely off the wheel"-- Provided by publisher.
Tags from this library: No tags from this library for this title. Log in to add tags.
Star ratings
    Average rating: 0.0 (0 votes)
Holdings
Item type Current library Collection Shelving location Call number Status Date due Barcode Item holds
หนังสือ หนังสือ PIM Creative Learning Space Chaengwattana หนังสือภาษาจีน English Book Shelves Q334.7 .C488A 2020 (Browse shelf(Opens below)) Checked out 25/11/2025 32550000521990
Total holds: 0

Includes bibliographical references (pages [401]-451) and index.

"A jaw-dropping exploration of everything that goes wrong when we build AI systems-and the movement to fix them. Today's "machine-learning" systems, trained by data, are so effective that we've invited them to see and hear for us-and to make decisions on our behalf. But alarm bells are ringing. Systems cull résumés until, years later, we discover that they have inherent gender biases. Algorithms decide bail and parole-and appear to assess black and white defendants differently. We can no longer assume that our mortgage application, or even our medical tests, will be seen by human eyes. And autonomous vehicles on our streets can injure or kill. When systems we attempt to teach will not, in the end, do what we want or what we expect, ethical and potentially existential risks emerge. Researchers call this the alignment problem. In best-selling author Brian Christian's riveting account, we meet the alignment problem's "first-responders," and learn their ambitious plan to solve it before our hands are completely off the wheel"-- Provided by publisher.

There are no comments on this title.

to post a comment.