AI Snake Oil 09月25日
《AI的骗局》:揭示AI真相,辨别真伪
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

《AI的骗局》一书现已发布首章,作者深入剖析了人工智能的复杂性,指出AI是一个涵盖多种技术的广泛术语。书中详细阐述了预测性AI的局限性,探讨了AI预测未来的可能性,并追溯了生成式AI的发展历程。此外,本书还审慎评估了AI的生存风险,分析了AI在社交媒体内容审核中的不足,并揭示了AI神话持续存在的原因。作者旨在帮助读者理性看待AI,辨别AI技术中的炒作与欺骗,并引导大家共同塑造AI的未来。

📚 **AI概念的复杂性与多面性**:文章指出,AI并非单一技术,而是一个包含预测性AI、生成式AI等多种技术的广泛术语。这种模糊性常常被用来掩盖技术本身的局限性,因此理解AI的不同分支及其运作方式至关重要,以避免被“AI骗局”所误导。

🔮 **预测性AI的局限与误导**:书中深入探讨了预测性AI在实际应用中的不足,例如在招聘、信贷、教育等领域,尽管它能发现统计模式,但其预测能力常被夸大,导致实际应用中的重大失误。文章强调,预测未来的难度,无论是否有AI辅助,都应被审慎看待。

💡 **生成式AI的历史与未来展望**:文章追溯了生成式AI近十年来取得的巨大进步,并将其置于计算技术长达七十年的发展背景下。理解其发展历程有助于更现实地预测其未来潜力,避免过度神化。

⚠️ **审慎评估AI的生存风险与社会影响**:书中对AI可能带来的生存风险进行了批判性评估,认为许多流行的讨论存在缺陷。同时,文章也关注了AI在内容审核等领域的实际应用问题,并分析了AI神话持续存在的原因,旨在提升读者的批判性思维能力,识别AI炒作。

The first chapter of the AI snake oil book is available online. It is 30 pages long and summarizes the book’s main arguments. If you haven't ordered the book yet, we hope that reading the introductory chapter will convince you to get yourself a copy.

Update (September 2025): It has been a year since the release of AI Snake Oil. In the time since its release, the two of us have given talks, appeared on podcasts, published exercises to accompany the book, and written a new preface and epilogue for the paperback edition of the book. The book was included in Nature’s list of the 10 best books of 2024, Bloomberg’s 49 best books of 2024, and Forbes’s 10 must-read tech books of 2024. It has received many positive reviews, including in the New Yorker. We are grateful to readers of the book for engaging deeply with its ideas.

We have now started working on our next project together, AI as Normal Technology. The project picks up where AI Snake Oil left off: whereas AI Snake Oil was an attempt to understand the present and near-term impacts of AI, AI as Normal Technology is a framework to think about its future impacts. The new name of this newsletter reflects this change. We hope you will follow along.

The single most confusing thing about AI

Our book is about demystifying AI, so right out of the gate we address what we think is the single most confusing thing about it: 

AI is an umbrella term for a set of loosely related technologies

Because AI is an umbrella term, we treat each type of AI differently. We have chapters on predictive AI, generative AI, as well as AI used for social media content moderation. We also have a chapter on whether AI is an existential risk. We conclude with a discussion of why AI snake oil persists and what the future might hold. By AI snake oil we mean AI applications that do not (and perhaps cannot) work. Our book is a guide to identifying AI snake oil and AI hype. We also look at AI that is harmful even if it works well — such as face recognition used for mass surveillance. 

While the book is meant for a broad audience, it does not simply rehash the arguments we have made in our papers or on this newsletter. We make scholarly contributions and we wrote the book to be suitable for adoption in courses. We will soon release exercises and class discussion questions to accompany the book.

What's in the book

Chapter 1: Introduction. We begin with a summary of our main arguments in the book. We discuss the definition of AI (and more importantly, why it is hard to come up with one), how AI is an umbrella term, what we mean by AI Snake Oil, and who the book is for. 

Generative AI has made huge strides in the last decade. On the other hand, predictive AI is used for predicting outcomes to make consequential decisions in hiring, banking, insurance, education, and more. While predictive AI can find broad statistical patterns in data, it is marketed as far more than that, leading to major real-world misfires. Finally, we discuss the benefits and limitations of AI for content moderation on social media.

We also tell the story of what led the two of us to write the book. The entire first chapter is now available online.

Chapter 2: How predictive AI goes wrong. Predictive AI is used to make predictions about people—will a defendant fail to show up for trial? Is a patient at high risk of negative health outcomes? Will a student drop out of college? These predictions are then used to make consequential decisions. Developers claim predictive AI is groundbreaking, but in reality it suffers from a number of shortcomings that are hard to fix. 

We have discussed the failures of predictive AI in this blog. But in the book, we go much deeper through case studies to show how predictive AI fails to live up to the promises made by its developers.

Chapter 3: Can AI predict the future? Are the shortcomings of predictive AI inherent, or can they be resolved? In this chapter, we look at why predicting the future is hard — with or without AI. While we have made consistent progress in some domains such as weather prediction, we argue that this progress cannot translate to other settings, such as individuals' life outcomes, the success of cultural products like books and movies, or pandemics. 

Since much of our newsletter is focused on topics of current interest, this is a topic that we have never written about here. Yet, it is foundational knowledge that can help you build intuition around when we should expect predictions to be accurate.

Chapter 4: The long road to generative AI. Recent advances in generative AI can seem sudden, but they build on a series of improvements over seven decades. In this chapter, we retrace the history of computing advances that led to generative AI. While we have written a lot about current trends in generative AI, in the book, we look at its past. This is crucial for understanding what to expect in the future. 

Chapter 5: Is advanced AI an existential threat? Claims about AI wiping out humanity are common. Here, we critically evaluate claims about AI's existential risk and find several shortcomings and fallacies in popular discussion of x-risk. We discuss approaches to defending against AI risks that improve societal resilience regardless of the threat of advanced AI.

Chapter 6: Why can't AI fix social media? One area where AI is heavily used is content moderation on social media platforms. We discuss the current state of AI use on social media, and highlight seven reasons why improvements in AI alone are unlikely to solve platforms' content moderation woes. We haven't written about content moderation in this newsletter.

Chapter 7: Why do myths about AI persist? Companies, researchers, and journalists all contribute to AI hype. We discuss how myths about AI are created and how they persist. In the process, we hope to give you the tools to read AI news with the appropriate skepticism and identify attempts to sell you snake oil.

Chapter 8: Where do we go from here? While the previous chapter focuses on the supply of snake oil, in the last chapter, we look at where the demand for AI snake oil comes from. We also look at the impact of AI on the future of work, the role and limitations of regulation, and conclude with vignettes of the many possible futures ahead of us. We have the agency to determine which path we end up on, and each of us can play a role.

We hope you will find the book useful and look forward to hearing what you think. 

Early reviews

Book launch events

Podcasts and interviews

We’ve been on many other podcasts that will air around the time of the book’s release, and we will keep this list updated.

Purchase links

The book is available to preorder internationally on Amazon.

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

AI 人工智能 AI Snake Oil AI Hype Predictive AI Generative AI AI Risk AI Regulation Artificial Intelligence AI Hype Tech Books Demystifying AI
相关文章