<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0"><channel><title><![CDATA[AI as Normal Technology]]></title><description><![CDATA[Analyzing AI as transformative but normal technology, not superintelligence.]]></description><link>https://www.normaltech.ai</link><image><url>https://substackcdn.com/image/fetch/$s_!YzQY!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6780dda8-5879-4789-bf02-b3ea43a9f85e_1000x1000.png</url><title>AI as Normal Technology</title><link>https://www.normaltech.ai</link></image><generator>Substack</generator><lastBuildDate>Fri, 24 Apr 2026 11:24:42 GMT</lastBuildDate><atom:link href="https://www.normaltech.ai/feed" rel="self" type="application/rss+xml"/><copyright><![CDATA[Sayash Kapoor and Arvind Narayanan]]></copyright><language><![CDATA[en]]></language><webMaster><![CDATA[normaltech@substack.com]]></webMaster><itunes:owner><itunes:email><![CDATA[normaltech@substack.com]]></itunes:email><itunes:name><![CDATA[Sayash Kapoor]]></itunes:name></itunes:owner><itunes:author><![CDATA[Sayash Kapoor]]></itunes:author><googleplay:owner><![CDATA[normaltech@substack.com]]></googleplay:owner><googleplay:email><![CDATA[normaltech@substack.com]]></googleplay:email><googleplay:author><![CDATA[Sayash Kapoor]]></googleplay:author><itunes:block><![CDATA[Yes]]></itunes:block><item><title><![CDATA[Open-world evaluations for measuring frontier AI capabilities]]></title><description><![CDATA[Introducing CRUX, a new project for evaluating AI on long, messy tasks]]></description><link>https://www.normaltech.ai/p/open-world-evaluations-for-measuring</link><guid isPermaLink="false">https://www.normaltech.ai/p/open-world-evaluations-for-measuring</guid><dc:creator><![CDATA[Sayash Kapoor]]></dc:creator><pubDate>Thu, 16 Apr 2026 17:47:29 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/44ed558c-fa00-46fc-a80d-c8a70b338fc9_2992x1616.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>This post is 8,000 words long&#8212;it is our new collaborative paper on an emerging type of AI evaluation. The paper is also published in a PDF format <a href="https://cruxevals.com/open-world-evaluations.pdf">here</a>.</em></p><p><em><strong>Summary</strong>: AI models have started to saturate most major benchmarks. But does that mean they can build and ship a real product, or conduct a scientific experiment end-to-end, or navigate a government bureaucracy? Researchers have started testing AI in such real-world settings. We call these evals &#8220;open-world evaluations&#8221;. This essay defines open-world evaluations, <a href="https://cruxevals.com/#survey">surveys</a> the lessons learned so far, and lays out best practices for conducting them.</em></p><p><em>We also introduce CRUX, a collaboration of 17 researchers from academia, government, civil society, and industry that will regularly evaluate frontier AI capabilities through open-world evaluations. In our first experiment, an AI agent built and <a href="https://apps.apple.com/in/app/breathe-easy-calm-breaths/id6760207382">published</a> an iOS app to the App Store, making just two errors, one of which required manual intervention. This gives us an early indication of potentially useful capabilities and, more importantly, an early warning about the potential for AI-driven app store spam (we disclosed this result to Apple a month before publication). </em></p><p><em>We hope to conduct similar experiments to surface early warnings across other real-world domains; this will be one of our main empirical projects over the coming year.</em></p><p><em><strong>The authors are:</strong> Sayash Kapoor, Peter Kirgis, Andrew Schwartz, Stephan Rabanser, J.J. Allaire, Rishi Bommasani, Magda Dubois, Gillian Hadfield, Andy Hall, Sara Hooker, Seth Lazar, Steve Newman, Dimitris Papailiopoulos, Shoshannah Tekofsky, Helen Toner, Cozmin Ududec, Arvind Narayanan</em></p><div><hr></div><p>How should we track and predict AI capabilities? The AI community&#8217;s dominant answer today is benchmarking. For example, <a href="https://metr.org/time-horizons/">METR&#8217;s time horizon graph</a> has been used by policy analysts, industry leaders, and organizations researching AI risks to argue that AI capabilities are rapidly increasing.</p><p>But benchmarks can both overestimate and underestimate progress. To turn a task into a benchmark, it needs to be precisely specified and automatically verifiable. The catch is that whatever is precise enough to benchmark is also precise enough to optimize for, allowing AI agents to excel at such tasks. On the flip side, low accuracy on benchmarks might result from incidental failures such as encountering CAPTCHA on a website, even if agents are capable of solving the underlying task.</p><p>To address these limitations, many researchers are turning to a new kind of evaluation: long, messy, real-world evaluations that go beyond benchmarks. Nicholas Carlini at Anthropic used Claude agents to build a <a href="https://www.anthropic.com/engineering/building-c-compiler">C compiler</a> that could compile the Linux kernel. Anthropic and Andon Labs designed a free-form experiment where Claude was tasked with <a href="https://www.anthropic.com/research/project-vend-2">maintaining a small shop in their office</a>. While benchmarks consist of dozens of tasks evaluated in an automated way, open-world evaluations consist of small samples, often require human intervention, and are evaluated in an open-ended way, such as by analyzing agent logs.</p><p>It is easy to dismiss these as unscientific: each such evaluation has a sample size of 1, and they lack standardization and reproducibility. Despite these limitations, we think such evaluations are important for collecting evidence about AI capabilities. They can provide early warnings about emerging capabilities to inform efforts at building societal resilience, help evaluators identify blind spots in existing benchmarks, and give companies a clearer picture of what tasks AI systems could soon carry out, informing strategic decisions about AI. We call them <strong>open-world evaluations</strong>.</p><p>In this essay, we conceptualize open-world evaluations, review past examples to identify best practices and pitfalls in conducting them, and introduce CRUX, a project aimed at regularly conducting new open-world evaluations. Here are our main insights:</p><ul><li><p><strong>Open-world evaluations are an important emerging class of AI evaluation.</strong> As AI systems become more capable, evaluations to elicit frontier capabilities must also increase in complexity. Open-world evaluations are the latest in a long line of evaluations of increasing complexity. We survey 10 prominent open-world evals conducted over the last year to identify best practices and key takeaways.</p></li><li><p><strong>CRUX (Collaborative Research for Updating AI eXpectations)</strong> is our attempt at systematically conducting open-world evaluations. The team consists of collaborators from government, academia, and non-profits, many of whom have led open-world evaluations, and who have a range of expectations about the future of AI. We aim to provide empirical evidence about the <em>present</em> <em>capabilities </em>of AI systems, even if they are currently costly, and to provide early warnings for capabilities that might soon be widespread. We plan to release new open-world evaluations regularly.</p></li><li><p><strong>In our first CRUX experiment, we tasked an AI agent with developing and publishing a simple iOS app to the App Store. </strong>Many benchmarks test agents&#8217; ability to write code. But publishing an iOS app involves many other steps: signing the app, publishing a privacy policy on a webpage, filling out Apple&#8217;s forms, and taking the app through the review process. We were more interested in whether the agent succeeded at the real-world requirements of publishing the app rather than its ability to write code, so we tasked it with building a simple app and taking it through the iOS App Store submission process.</p></li><li><p><strong>The agent was <a href="https://apps.apple.com/us/app/breathe-easy-calm-breaths/id6760207382">successful</a> after making two errors, one of which required manual intervention </strong>(forgetting where the correct credentials were stored and fabricating a fictional phone number for the App Store review process). The process of developing and publishing the app cost about $1,000. The app is now live on the <a href="https://apps.apple.com/us/app/breathe-easy-calm-breaths/id6760207382">iOS App Store</a>. We think the cost could have been far lower: the app development and submission only cost $25; the vast majority of tokens were spent monitoring the app&#8217;s status. We reached out to Apple a month before the publication of this essay to disclose the results of our experiment. App store operators should prepare for and police spam submissions, as they might soon see thousands of applications submitted autonomously using agents.</p></li><li><p><strong>How can we improve open-world evaluations? What&#8217;s next?</strong> To increase the usefulness of open-world evaluations, evaluators should specify what and how much human intervention is allowed, release logs collected while the agent was solving the task, and analyze logs to report what an agent did in the course of solving the task. In future CRUXes, we will evaluate AI R&amp;D automation, AI governance, and many other areas.</p></li></ul><h2><strong>Open-world evaluations are an important emerging class of AI evaluation</strong></h2><p>In this section, we define open-world evaluations and survey the emerging landscape of such evaluations to extract insights about their success and limitations. We discuss areas where open-world evals can overcome some of the blind spots of benchmarks. In our view, as AI systems become more capable, evaluations to elicit frontier capabilities must become more complex; open-world evaluations are the latest in a series of evaluations of increasing complexity. We also discuss the limits of open-world evaluations compared to benchmarks.</p><p>We discuss five loose criteria that define open-world evaluations below (see &#8220;What are open-world evaluations?&#8221;). But it is worth noting that the line between long, complex benchmark tasks and open-world evaluations is blurry. Indeed, many of the evaluations we discuss are sandboxed. We still include them in our list of open-world evals, because they satisfy our other criteria (e.g., Carlini&#8217;s C compiler is sandboxed, but involves just one long-running task, human intervention, and qualitative analysis as part of the evaluation).</p><h4>Benchmarks can both overestimate and underestimate progress</h4><p>People with drastically different views on AI <a href="https://asteriskmag.substack.com/p/common-ground-between-ai-2027-and">agree</a> that current AI benchmarks might soon be <a href="https://red.anthropic.com/2026/mythos-preview/">saturated</a>. Many prominent benchmarks have been saturated in the last two years, and evaluators have raced to release &#8220;successor&#8221; benchmarks. Many of these updated benchmarks are themselves near saturation.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!FLk-!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F48a36d1f-f2be-4f0a-ac30-fca88df74521_2048x996.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!FLk-!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F48a36d1f-f2be-4f0a-ac30-fca88df74521_2048x996.jpeg 424w, https://substackcdn.com/image/fetch/$s_!FLk-!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F48a36d1f-f2be-4f0a-ac30-fca88df74521_2048x996.jpeg 848w, https://substackcdn.com/image/fetch/$s_!FLk-!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F48a36d1f-f2be-4f0a-ac30-fca88df74521_2048x996.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!FLk-!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F48a36d1f-f2be-4f0a-ac30-fca88df74521_2048x996.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!FLk-!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F48a36d1f-f2be-4f0a-ac30-fca88df74521_2048x996.jpeg" width="1456" height="708" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/48a36d1f-f2be-4f0a-ac30-fca88df74521_2048x996.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:708,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!FLk-!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F48a36d1f-f2be-4f0a-ac30-fca88df74521_2048x996.jpeg 424w, https://substackcdn.com/image/fetch/$s_!FLk-!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F48a36d1f-f2be-4f0a-ac30-fca88df74521_2048x996.jpeg 848w, https://substackcdn.com/image/fetch/$s_!FLk-!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F48a36d1f-f2be-4f0a-ac30-fca88df74521_2048x996.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!FLk-!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F48a36d1f-f2be-4f0a-ac30-fca88df74521_2048x996.jpeg 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em>Many popular benchmarks (such as <a href="https://arxiv.org/abs/2310.06770">SWE-Bench</a>, <a href="https://arxiv.org/abs/1911.01547">ARC-AGI</a>, <a href="https://arxiv.org/abs/2406.12045">&#964;-bench</a>, <a href="https://www.tbench.ai/">Terminal Bench</a>, and <a href="https://metr.org/time-horizons/">METR&#8217;s Time Horizon task suite</a>) <a href="https://arxiv.org/abs/2506.07982">have</a> <a href="https://arcprize.org/blog/arc-agi-2-technical-report">seen</a> <a href="https://arxiv.org/abs/2601.11868">successor</a> <a href="https://openai.com/index/introducing-swe-bench-verified/">benchmarks</a> <a href="https://metr.org/blog/2026-1-29-time-horizon-1-1/">being</a> <a href="https://www.swebench.com/multilingual-leaderboard.html">released</a> <a href="https://sierra.ai/resources/research/tau-3-bench">in</a> <a href="https://arcprize.org/arc-agi/3">the</a> <a href="https://arxiv.org/abs/2410.03859">last</a> two years.</em></figcaption></figure></div><p>Does this mean AI systems will soon be capable of solving any task? Not necessarily. Benchmark improvements could indicate real capability gains, but they can also overstate progress: for example, because benchmarks have limited construct validity&#8212;they may test accuracy on narrow tasks rather than general ability&#8212;and they don&#8217;t test how well agents handle the messiness of real-world environments.</p><p>On the flip side, benchmarks could also <em>underestimate</em> progress due to challenges in the environment or infrastructure that are only incidental to the capabilities being measured, such as encountering CAPTCHA.</p><p>One way to get a fuller picture is to use metrics other than accuracy. For example, in a recent preprint several of us coauthored, we showed that even though agents are improving drastically in terms of capability metrics (such as average accuracy), they have improved much more slowly on metrics that measure <a href="https://arxiv.org/abs/2602.16666">reliability</a>. Similarly, many agent solutions to SWE-bench tasks that were judged correct because they pass tests, would nevertheless be <a href="https://metr.org/notes/2026-03-10-many-swe-bench-passing-prs-would-not-be-merged-into-main/">rejected</a> by project maintainers.</p><p>But this still doesn&#8217;t allow us to measure the upper bounds of AI capabilities<em>. </em>Capabilities that are possible today (even if only under favorable conditions) might soon be widespread, and we need to anticipate them before they are. Providing such early warnings gives businesses extra time to capitalize on new opportunities, institutions to build resilience, and policymakers to address risks.</p><p>Now let&#8217;s discuss in more detail why benchmarks can both overestimate and underestimate capabilities. They can overestimate capabilities because:</p><p><strong>Benchmarks resemble tasks amenable to modern reinforcement learning (RL) techniques.</strong> To turn a task into a benchmark, it needs to be precisely specified and automatically verifiable. But whatever is precise enough to benchmark is also precise enough to optimize for, and modern RL training runs increasingly resemble the shape of the benchmarks themselves. This makes it easy to saturate any task that can be measured using benchmarks.</p><p>This is already the case for leading evaluation platforms like <a href="https://github.com/harbor-framework/harbor">Harbor</a>, which double as RL training platforms. As a result, AI models could be directly trained on data from many prominent benchmarks included on these platforms. Even if benchmarks include held-out test sets, models might be trained on tasks from the training set that look very similar to those they would encounter in the test set. So benchmarks don&#8217;t help us understand how well performance generalizes to the real-world.</p><p><strong>Benchmarks avoid real-world messiness. </strong>Real-world tasks involve underspecified interactions that can&#8217;t be fully sandboxed, such as responding to unexpected situations or navigating environments that are open-ended. Benchmarks can take some steps to simulate messy environments, but they can&#8217;t fully replicate them.</p><p>Benchmarks can also underestimate capabilities, because:</p><p><strong>Eliciting frontier capabilities is costly. </strong>Running large-scale, long-running experiments is expensive, making it impractical to achieve the sample sizes that benchmarks rely on. Anthropic&#8217;s C compiler cost ~$20k; the task we describe running below (developing and publishing an iOS app for CRUX #1) cost ~$1,000. We can&#8217;t run such experiments hundreds of times, limiting the budget and complexity for each benchmark task.</p><p><strong>Average performance is very different from upper-bound capability elicitation. </strong>Trying to run dozens (or 100s) of tasks in a benchmark suite is only necessary to measure <em>average</em> performance. But when we are trying to understand the frontier of what agents can do, the goal shifts to understanding <em>best-case</em> performance: what can agents accomplish when given sufficient resources and support to work around incidental failures? This is necessary for providing early warnings for capabilities that might soon be widespread.</p><p><strong>Human intervention can help elicit capability upper bounds.</strong> Agents working on real-world tasks could encounter policy refusals, require solving CAPTCHAs, or encounter other infrastructure issues where they get stuck. This could negatively affect their performance. But these failures are only <em>incidental</em> to the capability being measured. If human operators can handle these, it would allow us to elicit upper bounds of capability. Such manual intervention is impractical for 100s of benchmark tasks each time they are run.</p><p>As AI capabilities improve, sandboxed evaluations that test AI capabilities in coding, deep research, and customer service require intensely engineered environments to challenge agents and avoid contamination or reward hacking. For example, the performance of agents on web benchmarks has been <a href="https://arxiv.org/pdf/2510.02418">affected</a> by how often agents encounter CAPTCHA, rather than eliciting the true underlying capabilities of agents. Recent work has highlighted AI agents <a href="https://www.nist.gov/caisi/cheating-ai-agent-evaluations">finding answers online</a>, <a href="https://github.com/SWE-bench/SWE-bench/issues/465">exploiting bugs</a> in evaluations, and <a href="https://metr.org/notes/2026-03-10-many-swe-bench-passing-prs-would-not-be-merged-into-main/">producing code</a> that passes tests but fails to meet standards for production. This highlights the need for a shift towards deeper qualitative evaluations of AI agent performance that can help us better understand failure modes and problem-solving strategies that can be lost in the barrage of average benchmark scores.</p><p>But these threats to validity are not straightforward to address. We can conduct <a href="https://www.aisi.gov.uk/blog/a-pipeline-for-transcript-analysis-using-inspect-scout">qualitative log analysis</a> for benchmarks to address some of these concerns, but even when we find validity issues in benchmark results using log analysis, there&#8217;s little we can do about it except releasing an updated benchmark, which could take a few months. Open-world evals allow us to conduct dry runs to fix issues before the real test run, and manually intervene to fix issues found during the evaluation.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a></p><p>Of course, benchmarks can be helpful despite these limitations.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a> And open-world evaluations have their own limitations, which we discuss later in this paper.  That said, we hope this list illustrates the systematic blind spots in benchmarking. As AI agents become more capable, the gaps we identify between traditional benchmarking and real-world capability will continue to grow. Success metrics will need to be multi-faceted to capture the diversity of objectives in a given task. Benchmarks will need to incorporate key bottlenecks (human assistance) and messy environments (internet navigation), each of which introduces additional issues of internal and external validity. Open-world evaluations offer an alternative.</p><h4>What are open-world evaluations?</h4><p>As evaluation methods have matured alongside AI capabilities, a gradient in evaluation has emerged. At one end are simple, automated, scalable methods that work well for early-stage capabilities. At the other end are richer, more labor-intensive methods that become necessary as capabilities improve and saturate simpler metrics.</p><p>As capabilities in a domain improve, evaluations further along the gradient become important to get insights that are complementary to simpler evaluations. We think of this gradient as having roughly five levels so far, each with their strengths and limitations:</p><ol><li><p><strong>Q&amp;A benchmarks</strong> (e.g., <a href="https://arxiv.org/abs/2009.03300">MMLU</a>, <a href="https://arxiv.org/abs/2311.12022">GPQA</a>): useful for broad knowledge assessment, but increasingly saturated for frontier models. Often formatted as multiple-choice questions for ease of grading, but as a result have low construct validity, since users rarely interact with models by asking multiple-choice questions.</p></li><li><p><strong>Open-ended chat benchmarks</strong> (e.g., <a href="https://arxiv.org/abs/2406.04770">WildBench</a>, <a href="https://github.com/lmarena/arena-hard-auto">Arena-hard-auto</a>): capture more nuance, but still limited to single-turn or short interactions.</p></li><li><p><strong>Outcome-only agent benchmarks </strong>(e.g., <a href="https://arxiv.org/abs/2310.06770">SWE-Bench</a>, <a href="https://arxiv.org/abs/2307.13854">WebArena</a>): test agent performance on real tasks, but only measure whether the task was completed, not how. As a result, they have limitations. For example, <a href="https://metr.org/notes/2026-03-10-many-swe-bench-passing-prs-would-not-be-merged-into-main/">most passing SWE-Bench solutions are not accepted by repository maintainers</a>.</p></li><li><p><strong>Agent benchmarks with log analysis</strong> (e.g., <a href="https://www.aisi.gov.uk/blog/a-pipeline-for-transcript-analysis-using-inspect-scout">UK AISI transcript analysis</a>, <a href="https://metr.org/time-horizons/">METR Time Horizon</a>): go deeper by examining <em>how</em> agents succeed or fail and uncovering errors and reward hacking by analyzing agent logs. But they still operate in sandboxed environments with predefined tasks.</p></li><li><p><strong>Open-world evaluations</strong>: long-horizon tasks in real-world environments, where success can&#8217;t be neatly specified or automatically graded. This allows eliciting upper bounds of capabilities. But this comes at the cost of lack of reproducibility and standardization (which are both benefits of benchmarking; see section on Limitations).</p></li></ol><p>Some evaluations blur the categories of this spectrum. For example, OpenAI&#8217;s <a href="https://openai.com/index/gdpval/">GDPVal</a>, a long-horizon agent benchmark, is designed to be manually graded based on expert&#8217;s opinions on quality. This structure is very similar to an open-world evaluation, though they primarily focus on outputs but not on analyzing the logs. At the same time, results on GDPVal are commonly reported using <a href="https://artificialanalysis.ai/evaluations/gdpval-aa">GDPval-AA</a>, which uses automated LLM grading; this evaluation setting resembles an outcome-only agent benchmark.</p><p>Open-world evaluations consist of running agents on a small number of long-horizon tasks in real-world settings, and qualitatively evaluating their results <a href="https://www.aisi.gov.uk/blog/a-pipeline-for-transcript-analysis-using-inspect-scout">using tools for log analysis</a>. These evaluations are complementary to benchmarking and help address many of their limitations. They can also uncover tasks that remain out of reach for current AI systems for future benchmarking efforts.</p><p>We provide a rough taxonomy to clarify the boundary of what constitutes open-world evaluations. No single dimension determines whether an experiment qualifies as &#8220;open-world&#8221;. Instead, it depends on the overall pattern across all of the dimensions below.</p><ul><li><p><strong>Openness</strong>. Is this evaluation in a real-world deployment setting (as opposed to a sandboxed environment)?</p></li><li><p><strong>Complexity/length</strong>. Does the task require a human days or weeks at a time to complete (as opposed to a few minutes or hours)?</p></li><li><p><strong>Number of tasks</strong>. Is this a stand-alone task or a small set of tasks (as opposed to a large evaluation suite or benchmark)?</p></li><li><p><strong>Human intervention</strong>. To help elicit upper-bounds of capabilities, are humans able to intervene when agents hit a hurdle (as opposed to just setting up the environment or resolving setup issues)?</p></li><li><p><strong>Method of evaluation</strong>. Does the evaluation primarily consist of in-depth log evaluation (as opposed to the result driven by a single average metric)?</p></li></ul><p>When do we call something an evaluation as opposed to simply using agents to accomplish something novel? For example, Anthropic used AI agents to <a href="https://www.anthropic.com/news/mozilla-firefox-security">find security vulnerabilities</a> in leading open-source software such as Mozilla Firefox&#8212;is this an example of an open-world evaluation? We still consider these open-world evals if the role of the agent is systematically and publicly documented (including the parts carried out by the agent carried and the human experts, and the end result).</p><p>The line between complex benchmark tasks and open-world evaluations is also blurry. For example, some evaluations in our list of open-world evals below are sandboxed (e.g., <a href="https://futurism.com/advanced-ai-stuck-pokemon">Claude Plays Pokemon</a> and <a href="https://www.anthropic.com/engineering/building-c-compiler">Anthropic&#8217;s C compiler</a> were both run in sandboxed environments, but we still consider them open-world evaluations because they consist of a single, complex, long-running task, have human intervention during the evaluation, and are evaluated qualitatively).</p><p>The two types of evals are also complementary: we could imagine using open-world evaluations to understand what tasks remain unsolvable by agents as a first step towards building new benchmarks, and also use open-world evals to evaluate agents on messy real-world tasks that are not amenable to being benchmarked.</p><p>Finally, we don&#8217;t mean to imply that open-world evaluations are categorically more informative for understanding AI progress compared to benchmarks. Indeed, we discuss many limitations of open-world evaluations later in this section. Rather, as AI capabilities increase, open-world evaluations become important as an <em>additional complementary</em> signal of AI capabilities.</p><h4>An incomplete survey of open-world evaluations</h4><p>Over the past year, researchers at AI labs, universities, non-profits, and independent groups have begun running open-world evaluations. These share a common structure: give a capable AI agent a hard, real-world task with a long time horizon and observe and analyze its behavior in detail. Some notable examples:</p><ol><li><p><strong><a href="https://futurism.com/advanced-ai-stuck-pokemon">Anthropic, Claude Plays Pokemon</a></strong> (Feb 2025). Anthropic launched a Twitch livestream in which Claude 3.7 Sonnet played Pokemon Red. While not a real-world deployment, the experiment was an early example of setting an AI agent up in a relatively open environment compared to typical benchmarks. While the project illustrated progress in AI computer use with minimal scaffolding, the demonstration also clearly illustrated the limitations of early 2025 agents&#8211;&#8211;Sonnet 3.7 remained stuck in a level for <a href="https://futurism.com/advanced-ai-stuck-pokemon">nearly 80 hours</a>.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-3" href="#footnote-3" target="_self">3</a></p></li><li><p><strong><a href="https://theaidigest.org/village">AI Digest, AI Village</a></strong> (April 2025-present). The AI Village gives multiple AI agents their own computer environments and a shared group chat, then tasks them with open-ended real-world goals like fundraising for charity, organizing real in-person events, making word games, and gaining subscribers on Substack. Across different experiments in 2025, the project <a href="https://theaidigest.org/village/blog/what-we-learned-2025">highlighted</a> persistent failure modes of hallucination, mis-calibration, and unproductive loops, but also documented notable improvements from late 2025 agents on these dimensions.</p></li><li><p><strong><a href="https://www.anthropic.com/research/project-vend-1">Anthropic/Andon Labs, Project Vend</a></strong> (June 2025-present). Anthropic partnered with Andon Labs to have a scaffold using Claude 3.7 Sonnet (nicknamed &#8220;Claudius&#8221;) operate a small automated store in their office. The agent managed inventory, set prices, and interacted with customers over several weeks, revealing a rich set of failure modes around manipulation, prioritization, and real-world decision-making. A<a href="https://www.anthropic.com/research/project-vend-2"> second phase</a> expanded the experiment to multiple locations, used newer models, and included a red-teaming exercise by the staff at the Wall Street Journal. The agent lost nearly all revenue from poor planning, hallucination, and excessive discounts. Anthropic&#8217;s follow-up was more successful, advertising positive profit each week, but the staff at the WSJ were still able to successfully <a href="https://futurism.com/future-society/anthropic-ai-vending-machine">jailbreak</a> &#8220;Claudius,&#8221; leading it to give all products away for free. Andon Labs recently began a <a href="https://andonlabs.com/blog/andon-market-launch">third phase to the project </a>where they gave a Claude-based agent, &#8220;Luna,&#8221; a three year lease to a brick-and-mortar store in San Francisco and tasked it with making a profit as the &#8220;manager,&#8221; including hiring human employees, designing the brand, and making product selections.</p></li><li><p><strong><a href="https://cursor.com/blog/scaling-agents">Lin, Cursor browser experiment</a></strong> (Jan 2026). Wilson Lin at Cursor coordinated hundreds of GPT-5.2 agents to build a web browser from scratch, running uninterrupted for one week. The resulting browser (&#8221;FastRender&#8221;) consisted of over a million lines of Rust with a from-scratch rendering engine. It could render simple websites but was far from production-ready. The project was notable for its exploration of hierarchical multi-agent coordination at scale and the specific failure modes that emerge when agents work on a project for days rather than minutes.</p></li><li><p><strong><a href="https://www.anthropic.com/engineering/building-c-compiler">Carlini, C compiler</a></strong> (Feb 2026). Nicholas Carlini at Anthropic tasked Claude with building a C compiler from scratch, spending roughly $20k in API costs. The agent produced a working compiler for compiling the Linux kernel and passed a large fraction of standard test suites, revealing detailed information about where agents excel (systematic code generation, test-driven iteration) and where they struggle (complex optimization passes, debugging subtle spec violations).</p></li><li><p><strong><a href="https://epoch.ai/gradient-updates/how-close-is-ai-to-taking-my-job">Ho, &#8220;How Close is AI to Taking my Job&#8221;</a></strong> (Feb 2026). Epoch researcher Anson Ho had Claude Code and ChatGPT Atlas attempt to autonomously complete three challenging work tasks at Epoch: replicating an interactive web interface for a 40 parameter economic model, writing an article in Epoch&#8217;s style on AI progress in 2025, and porting an article from Google Docs to Substack and Epoch&#8217;s website. The project highlighted the persistent bottlenecks of formatting and hallucinations in successful knowledge-work tasks.</p></li><li><p><strong><a href="https://developers.openai.com/blog/run-long-horizon-tasks-with-codex">Choi, GPT 5.3 Codex Builds a Design Tool</a> </strong>(Feb 2026)<strong>. </strong>OpenAI&#8217;s Derrick Choi had GPT-5.3 Codex run autonomously for 25 hours, generating 35,000 lines of code, to build a &#8220;design tool&#8221; from scratch. The post noted that the agent showed impressive planning, memory, and verification processes from the agent, though it did not provide substantive analysis of the capabilities and limitations of the finished product.</p></li><li><p><strong><a href="https://blog.cloudflare.com/vinext/">Faulkner, Next.js Reimplementation</a> </strong>(Feb 2026)<strong>. </strong>An engineer at Cloudflare used Claude with OpenCode to release vinext, a reimplementation of the popular frontend web framework<a href="http://next.js"> Next.js</a> on Vite rather than React. While the initial blogpost advertised coverage of 94% of<a href="http://next.js"> Next.js</a> for only $1,100 in API costs, follow-up <a href="https://news.ycombinator.com/item?id=47142156">analyses</a> highlighted persistent security limitations and low generalizability to other domains of software given the importance of human-built testing and infrastructure of both Vite and Next.js.</p></li><li><p><strong><a href="https://x.com/DimitrisPapail/status/2028669695344148946">Papailiopoulos, &#8220;Can You Train a Computer&#8221;</a></strong> (March 2026). Dimitris Papailiopoulos and collaborators tested whether Claude Code and OpenAI Codex could train a transformer to function as a general-purpose computer. The experiment included both a fully autonomous round, where both agents failed and reward-hacked solutions, and a human-guided version, where Claude Code succeeded and displayed meaningful generalization, including solving multi-step computations never seen in training.</p></li><li><p><strong><a href="https://x.com/karpathy/status/2031135152349524125">Karpathy, Nanochat Autoresearch</a></strong> (March 2026). Using an existing open-source project, nanochat, for GPT-2 level LLM training, Andrej Karpathy built a simple automation pipeline for AI agents to optimize training in 5 minute increments. The agent has complete autonomy to adjust the architecture, hyperparameters, optimizers, and batch sizes. In a follow-up, Karpathy <a href="https://x.com/karpathy/status/2031135152349524125">shared</a> that autoresearch made progress against &#8220;Time to GPT-2&#8221; (measured with 8xH100 GPUs), dropping this metric by 11% in 2 days.</p></li></ol><p>We plan to collect a running list of such evaluations (and the key takeaways and limitations for each of them) on <a href="http://cruxevals.com">cruxevals.com</a>. The increase in AI capabilities has allowed people across domains to conduct such evaluations and figure out if AI systems can carry out tasks they are experts on, resulting in growing interest in such evaluations. In just the past week, there have been many significant open-world evaluation releases, including <a href="https://epoch.ai/blog/mirrorcode-preliminary-results">MirrorCode</a> by Adamczewski et al., which tasked agents with reimplementing large programs, a set of <a href="https://alignment.anthropic.com/2026/automated-w2s-researcher/">automated alignment research</a> case studies by Wen et al., and an exercise using Claude Code to train models to <a href="https://github.com/dphuang2/tinker-cookbook/tree/claude/golf-forecasting-setup-VIpRZ/tinker_cookbook/recipes/golf_forecasting">forecast the outcome of the recent Masters golf tournament</a> by Huang.</p><h4>Limitations of open-world evaluations</h4><p>While open-world evaluations are helpful at addressing some blind spots in benchmarking, they also suffer from many limitations. There remains a genuine tradeoff between developing better benchmarks and investing in open-world evaluations. Benchmarks give evaluators control over the task and allows evaluations to occur in controlled environments, but are less useful for understanding agents&#8217; performance on open-ended tasks. Open-world evaluations sacrifice sandboxing and evaluator control to improve the construct validity for the task and elicit upper-bound capabilities. Concretely, open-world evals have the following limitations:</p><p><strong>Lack of reproducibility and standardization:</strong> Benchmarks have been successful because they provide coordinating functions for the AI community. Researchers can develop and test new methods independently, and validate if these methods work well on benchmarks to get the community&#8217;s attention. The culture of benchmarking runs so deep that David Donoho called it the &#8220;<a href="https://www.tandfonline.com/doi/full/10.1080/10618600.2017.1384734">secret sauce</a>&#8220; for the success of the AI/ML community over the last 50 years. Open-world evaluations give up the reproducibility and standardization that made benchmarking so successful.</p><p><strong>Hard to compare agents: </strong>Benchmarks offer a relative comparison between different models/agents. But since open-world evaluations are often run one or a handful of times, the run-to-run variability might be higher than the difference between different agents&#8217; performance. As a result, open-world evaluations are not useful for comparing the accuracy of different models or agents.</p><p><strong>The need for domain expertise: </strong>Evaluating whether an agent succeeded in an open-world evaluation can be challenging. Verification of the agent&#8217;s work might require deep domain expertise and time, especially if the task is open-ended.</p><p><strong>Incomplete recall of log analysis: </strong>Even if automated log analysis is used to analyze open-world evaluations, it can never be considered complete. Agent transcripts from long-horizon tasks can run to hundreds of millions of tokens, making thorough human review impractical. And because agent behavior in these tasks is complex, there is no guarantee that a given round of analysis will surface all noteworthy behaviors or errors. This limitation does not apply in the same way to benchmarks, where success criteria are predefined and automatically verified. Releasing logs publicly so that a broader community can examine them is one way to partially mitigate this concern.</p><p><strong>Blurry success criteria: </strong>Given the potential for human intervention, it is hard to cleanly delineate the agent&#8217;s performance from the help given to it by the human.</p><p><strong>Non-stationary environments: </strong>In open-world evals, agents can interact with open-ended environments such as the internet.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-4" href="#footnote-4" target="_self">4</a> This makes it hard to make generalizable claims about the agent&#8217;s capabilities as opposed to information it can access via the internet (or another open-ended environment). For example, could an agent solve a hard software engineering task because it is genuinely competent at such tasks (and will therefore be able to solve novel software engineering tasks), or because it was able to look up a specific task online (and won&#8217;t be able to solve a novel task)? It is also challenging to compare agent performance over time. For example, the internet contains a growing list of implementations or hints that an unsandboxed agent can pull from.</p><h4>How different stakeholders can use open-world evaluations</h4><p>Open-world evaluations have limitations, but they can be helpful in addressing the blind spots of benchmarks. We think they can benefit a number of stakeholders:</p><p><strong>Policymakers: </strong>The diffusion of AI across domains <a href="https://www.normaltech.ai/p/ai-as-normal-technology">lags behind improvements in capabilities</a>. This allows institutions to adapt as AI systems become more capable. Open-world evaluations can further increase this lead time by providing early warnings of what agents could soon be able to do autonomously and at scale, giving institutions time to build resilience. For example, <a href="https://www.anthropic.com/news/claude-code-security">Anthropic</a>&#8216;s <a href="https://www.anthropic.com/glasswing">recent work</a> on discovering cybersecurity vulnerabilities using AI could spur efforts to rapidly adopt AI for defensive cybersecurity, especially in critical software infrastructure.</p><p><strong>AI evaluators and researchers:</strong> Open-world evaluations offer a complementary signal to benchmarks by testing capabilities that are structurally resistant to benchmarking, such as solving messy real-world tasks. Analyzing agent logs allows us to uncover instances of the agent taking shortcuts and reward hacking. On the flip side, they can find areas where agents develop new insights or surpass previous limitations. For example, in our iOS app development evaluation, we identified that the agent modified its approach to be more token efficient, which allowed it to drastically reduce the cost of solving the task, without any input or instruction from our end.</p><p><strong>Frontier AI developers: </strong>AI developers should actively support and participate in external open-world evaluation efforts by providing access (such as pre-release access to models) as well as <a href="https://arxiv.org/abs/2403.04893">safe harbors</a> for third-parties conducting evaluations that might not adhere to developers&#8217; terms of service (such as for evaluating safety). Open-world evaluations by independent third parties could surface findings that internal red teams miss if they optimize for known threat models.</p><p>New models are also saturating benchmarks. For example, Anthropic&#8217;s <a href="https://www-cdn.anthropic.com/8b8380204f74670be75e81c820ca8dda846ab289.pdf">Mythos Preview system card</a> notes that the model &#8220;saturates many of our most concrete, objectively-scored evaluations,&#8221; leaving them with noisier methods for assessing capabilities. Open-world evaluations could offer a complementary way to stress-test models in realistic settings that benchmarks can no longer distinguish.</p><p>To foster an ecosystem of open-world evaluations, it is important to develop shared best practices and build toward a cumulative body of evidence about what agents can and can&#8217;t do. That is what we are trying to do with a new project called CRUX.</p><h2><strong>Introducing CRUX: Collaborative Research for Updating AI eXpectations</strong></h2><p>CRUX is a project for operationalizing open-world evaluations. We aim to conduct open-world evaluations on a regular basis. Each evaluation will involve a long-horizon, real-world task; the implementation of an agent scaffold that could <em>in theory</em> allow agents to solve the task; and detailed analysis of what the agent did in order to solve a task. Our team consists of researchers spanning industry, academia, and non-profits. Many people on the team have conducted open-world evaluations.</p><p>In addition to conducting evaluations, we also want to develop mechanisms to provide early warnings for AI capabilities that will soon be widespread. For example, if AI agents can <em>almost </em>autonomously develop and publish apps (our task for the first iteration of CRUX; discussed in the next section), app store operators such as Apple and Google might soon need to update their policies to manage spam submissions.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-5" href="#footnote-5" target="_self">5</a></p><p>This requires eliciting upper-bounds of capabilities&#8212;often filling in missing incidental capabilities (such as filling out CAPTCHA) with human input. Open-world evaluations are a good fit for this particular style of evaluation, since they allow us to deeply understand AI systems&#8217; capabilities.</p><p>Another advantage of open-world evaluations is that we don&#8217;t need to develop a large task suite or design complex environments or sandboxes before conducting evaluations, since each task only needs to be carried out and evaluated a small number of times. This allows us to conduct new evaluations regularly. We plan to design a new CRUX evaluation, analyze results, and publish our analysis on new tasks every 1-2 months.</p><p>Our first evaluation tests if AI agents can autonomously develop and publish apps to the iOS App Store; we discuss this in the next section. In future iterations, we plan to expand to a wide range of domains, including tasks on AI R&amp;D automation, AI governance, complex software engineering, and real-world physical tasks.</p><h4>CRUX #1: Can AI agents autonomously develop and publish an iOS app?</h4><p>The question of whether AI agents can write software has been extensively studied, both through benchmarks like SWE-Bench and <a href="https://www.tbench.ai/">Terminal Bench</a>, and through open-world evaluations like the C compiler and browser experiments discussed earlier in this essay. Agents have shown strong coding capabilities (though questions of code quality and reliability remain <a href="https://x.com/GergelyOrosz/status/2036925672501715425?s=20">unresolved</a>).</p><p>Meanwhile, a task that has not been evaluated as closely is whether agents can handle the non-coding aspects of software deployment, such as satisfying platform requirements and interacting with review systems they do not control. For our evaluation, we tasked an agent with building a mobile app from scratch and publishing it on the iOS App Store.</p><p>We prompted the agent to develop and publish a <em>simple</em> app to the App Store. We weren&#8217;t primarily interested in the agent&#8217;s software engineering ability, but rather its ability to interact with Apple&#8217;s App Store submission process. This process requires developers to configure signing certificates and provisioning profiles, prepare screenshots and metadata, draft and host a privacy policy at a public URL, fill out compliance questionnaires, and submit the app for review by Apple&#8217;s team. Reviewers may reject the app for technical or policy reasons, requiring the developer to diagnose the issue, make changes, and resubmit. This process typically takes several days and involves interacting with systems and reviewers that the developer does not control.</p><p>The agent was responsible for every step of the process except those where human involvement is required by policy, such as setting up the Apple Developer account and hitting publish to release the app to the App Store. Specifically, the agent handled writing the code, building the app, preparing metadata, drafting and hosting a privacy policy, submitting it for review, and handling any feedback. (We provided the agent access to a Mac VM, a GitHub account, an Apple developer account, and a Gmail account.)</p><p>The success criterion was whether the agent got the app published on the App Store. We logged how many manual interventions the agent needed us to make to solve the task. The agent had the option to ping the team for support; we monitored the agent&#8217;s progress once a day. The lower this number, the better the agent performs.</p><p>In addition, if agents can do this autonomously (or are close to being able to do so), this serves as an early warning for Apple&#8217;s review processes, since agents might soon be able to publish thousands of apps autonomously. The App Store has already seen an <a href="https://appfigures.com/resources/insights/20251205?f=2">increase</a> in the number of published apps, but if agents could develop and publish apps fully autonomously, the number of submissions could increase dramatically.</p><h4>Our setup for the agent</h4><p>We used <a href="https://openclaw.ai/">OpenClaw</a> as the agent scaffold with Claude Opus 4.6 and <a href="https://platform.claude.com/docs/en/build-with-claude/adaptive-thinking#adaptive-thinking-with-the-effort-parameter">adaptive thinking enabled</a>.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-6" href="#footnote-6" target="_self">6</a> We chose OpenClaw for this experiment since it is configurable, integrates well with the browser, and natively supports long-running tasks. Given its recent popularity and support for wide-ranging and long-running tasks, we also wanted to evaluate its capabilities as a scaffold, and to potentially compare it to other scaffolds in the future. (Note that we used it as a fairly general scaffold for the task; in fact, we did not make any changes to the default OpenClaw setup beyond prompting it and giving it deeper access to the macOS VM.)</p><p>We also wanted to test whether agents struggle with visual reasoning and GUI operation tasks, which might be bottlenecks to solving the task. Based on reports of users&#8217; experiences, OpenClaw was able to successfully resolve some of these concerns, and we wanted to test this ability.</p><p>We recognize that OpenClaw has security risks in real-world use. We still chose to use it since we wanted to evaluate the capability frontier, while recognizing that security might be a bottleneck for real-world adoption.</p><p>The agent had access to a macOS virtual machine with expansive permissions (sudo, screen visibility, UI control) to perform this task. We logged all of the agent&#8217;s actions, reasoning chains, and stored screenshots that the agent took while solving the task.</p><h4>Addressing evaluation awareness</h4><p>AI models often recognize they are being evaluated and adjust their behavior. This is known as evaluation awareness. Recent work from <a href="https://www.anthropic.com/engineering/eval-awareness-browsecomp">Anthropic</a>, <a href="https://www.apolloresearch.ai/blog/claude-sonnet-37-often-knows-when-its-in-alignment-evaluations/">Apollo</a>, and <a href="https://alignment.openai.com/prod-evals/">OpenAI</a> has shown that frontier models increasingly detect evaluation contexts on their own and sometimes modify their behavior in response.</p><p>How should we address evaluation awareness? We think hiding the evaluation context from a frontier model is increasingly a lost cause; models are getting better at detecting test environments with each generation. Since we plan to conduct CRUX evaluations regularly and publish our plans publicly, future agents could simply come across descriptions of our evaluations during the course of a task.</p><p>More importantly, concerns about evaluation awareness are most acute when measuring alignment properties, where a model might strategically alter its behavior. For a capability evaluation, if the agent succeeds at the task, it has demonstrated the capability regardless of whether it knew it was being evaluated.</p><p>To address evaluation awareness, we explicitly informed the agent that it was being evaluated. This allowed us to communicate the parameters of the evaluation clearly; in particular, that it would be judged on the number of times it required unnecessary human inputs, but that some kinds of inputs (e.g., resolving CAPTCHAs) were completely fine to delegate. We were concerned that without this framing, the agent might avoid asking for help even when doing so was appropriate, which would have caused us to underelicit its performance.</p><h4>Conducting dry runs</h4><p>Before a full evaluation run, we conducted two dry runs to understand whether our agent setup worked well, and if not, what we needed to change. This allowed us to discover and solve bugs in the scaffold. Notably, these dry runs did not involve any interaction with Apple&#8217;s App Store submission or review processes.</p><p>Setting up the OpenClaw agent such that it had all permissions to be able to autonomously develop apps also required inputs from our end. We estimate it took us eight person-hours of time and about $50 in API cost to set up the agent scaffold; this included configuring the virtual machine to ensure the agent had full control to take any actions; setting up logging to monitor the agent&#8217;s work; and configuring an email account, GitHub account, and Apple developer account for the agent to use and access.</p><p>This effort might be a bottleneck for agents to autonomously develop apps. But from the perspective of a potential spammer, it only needs to be carried out once. Spending a few person-hours and $50 for setting up the pipeline is not likely to be a bottleneck for spammers looking to submit thousands of apps to the App Store.</p><p>In our dry runs, the agent primarily used the command line and the browser. It used the command line to generate code, build the app, and prepare it for submission. It used the browser to log into App Store Connect, access certificates, and fill out forms. When command-line commands hung because of requests for permission, it was able to take screenshots and simulate mouse clicks, for example to click &#8220;Allow&#8221; to grant itself permissions.</p><h4>The final evaluation</h4><p>After two dry runs, we started the full evaluation. The agent took 45 minutes to develop a simple app for breathing exercises. This included developing the app, publishing a privacy policy using GitHub Pages, filling out the App Store review forms, and submitting the app for review.</p><p>We set up the agent to keep checking the app&#8217;s status every 5 minutes after the app was sent for review. But it took 10 days before the app was approved. It is <a href="https://apps.apple.com/app/id6760207382">now live on the App Store</a>. (To comply with Apple&#8217;s policies, the agent needed approval from our team before publishing the app.)</p><p>The agent required one unnecessary manual intervention<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-7" href="#footnote-7" target="_self">7</a>: it could not locate the credentials we had previously given it to access the Apple developer account.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-8" href="#footnote-8" target="_self">8</a> It also fabricated the phone number submitted for Apple&#8217;s review process, using a <a href="https://www.cbc.ca/radio/undertheinfluence/the-real-reason-hollywood-uses-555-phone-numbers-1.6070537">fictional number</a> instead of asking us for the correct phone number; the App Store review went through despite this error.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-9" href="#footnote-9" target="_self">9</a> This alerted us to the need for proactive monitoring of agent actions to prevent such unintended actions, which we plan to implement for future CRUX evaluations.  The final app functions well, though it contains a toggle for sound that doesn&#8217;t work. The agent also produced a screenshot for the App Store listing with visible formatting errors.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!dmiP!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7434feaf-bac4-45ee-b854-64983d7054bb_1766x1178.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!dmiP!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7434feaf-bac4-45ee-b854-64983d7054bb_1766x1178.jpeg 424w, https://substackcdn.com/image/fetch/$s_!dmiP!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7434feaf-bac4-45ee-b854-64983d7054bb_1766x1178.jpeg 848w, https://substackcdn.com/image/fetch/$s_!dmiP!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7434feaf-bac4-45ee-b854-64983d7054bb_1766x1178.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!dmiP!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7434feaf-bac4-45ee-b854-64983d7054bb_1766x1178.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!dmiP!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7434feaf-bac4-45ee-b854-64983d7054bb_1766x1178.jpeg" width="1456" height="971" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/7434feaf-bac4-45ee-b854-64983d7054bb_1766x1178.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:971,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!dmiP!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7434feaf-bac4-45ee-b854-64983d7054bb_1766x1178.jpeg 424w, https://substackcdn.com/image/fetch/$s_!dmiP!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7434feaf-bac4-45ee-b854-64983d7054bb_1766x1178.jpeg 848w, https://substackcdn.com/image/fetch/$s_!dmiP!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7434feaf-bac4-45ee-b854-64983d7054bb_1766x1178.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!dmiP!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7434feaf-bac4-45ee-b854-64983d7054bb_1766x1178.jpeg 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">The screenshots uploaded by the agent had visible formatting errors.</figcaption></figure></div><p>The agent was eventually <a href="https://apps.apple.com/us/app/breathe-easy-calm-breaths/id6760207382">successful</a> in publishing the app, at a total cost of about $1,000. The development and submission of the app cost just about $25; the vast majority of the tokens were spent looking for updates to verify if the app had been successfully reviewed. We think the total cost could have been dramatically lower if we optimized the scaffold for efficiency, such as by waking the agent less frequently to check the app&#8217;s status, but in this evaluation we erred on the side of a higher budget.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-10" href="#footnote-10" target="_self">10</a></p><p>In short, the agent couldn&#8217;t completely automate the task, but it was extremely close to being able to do so.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-11" href="#footnote-11" target="_self">11</a> As a result, we notified Apple&#8217;s product security team of our experiment four weeks before publishing the results, since we thought some version of responsible disclosure was warranted; spammers could soon submit thousands of apps to the iOS App Store using agents.</p><h4><strong>Lessons for open-world evaluations</strong></h4><p>Through running CRUX and studying the growing body of similar efforts, we have begun to identify what makes an open-world evaluation informative. We expect these lessons to evolve over time. But we think they are worth sharing now, because there is growing interest in these evaluations, and developing shared evaluation norms can help.</p><p><strong>Be specific about what you are measuring and what it implies.</strong> One reason why people had sharply diverging opinions about Anthropic&#8217;s C compiler project or Cursor&#8217;s browser was that these projects did not clearly specify the target of their measurement. The results were very impressive from the perspective of measuring if agents productively work for long periods of time on well-defined tasks. But they were perhaps less impressive from the perspective of developing software that is directly usable by and useful to end users; the GitHub issues for both projects highlight core technical complaints, such as <a href="https://github.com/anthropics/claudes-c-compiler/issues/1">failing to compile &#8220;hello world&#8221; out of the box</a>, or <a href="https://github.com/wilsonzlin/fastrender/issues/98">welcome page hangs with the standard build script</a>. Had the authors been clear that they were trying to measure the former, and not the latter, it could have clarified the public discussion on these projects. (This is not to criticize the authors of these pieces. They were one of the first to conduct open-world evaluations, and it is hard to know how people will react a priori.)</p><p>For example, writing software has many <a href="https://personal.utdallas.edu/~chung/BOOK/book.html">non-functional requirements</a> such as quality, reliability, maintainability etc. that developers might sacrifice when shipping apps autonomously coded using AI agents. Many concerns about Cursor&#8217;s browser and Anthropic&#8217;s compiler were actually concerns about these hard-to-specify properties not being satisfied by applications developed using agents.</p><p><strong>Design the task so that human intervention is straightforward and well-documented. </strong>In real-world tasks, agents will sometimes need help, such as for navigating policy refusals, CAPTCHAs, or infrastructure failures. While traditional benchmarks cannot accommodate human-in-the-loop interventions, for open-world evaluations, such inputs are helpful in ensuring we measure upper-bounds of capabilities. This requires documenting precisely when, why, and how humans step in, so that the degree of autonomy can be assessed clearly.</p><p><strong>Invest in log analysis. </strong>The logs from an agent attempting a complex task contain far more information than the binary outcome. This allows us to uncover insights about agent behavior that would be invisible from outcomes alone. For example, how did the agent choose to decompose the problem? Did it get stuck at any point, and, if so, how did it self-correct? Did it search through solution paths in a principled or haphazard way? Which aspects of the task proved most challenging? Did the agent misrepresent any aspect of its outputs and progress?</p><p><strong>Consider complementing log analysis with real-time monitoring. </strong>Post-hoc log analysis is valuable, but it is not sufficient on its own to catch all unintended agent actions. In previous open-world evaluations, agents operating with substantial autonomy have sometimes taken actions that were difficult for human reviewers to detect after the fact. For example, in many AI experiments conducted by the AI Village, agents took <a href="https://theaidigest.org/village/blog/what-we-learned-2025">unintended actions</a>, such as attempting to send hundreds of unsolicited emails. In our own evaluation, the agent fabricated a fictional phone number that went undetected until a later round of review. Automated real-time monitoring, for example a separate agent that continuously reviews the primary agent&#8217;s actions and flags anomalies or errors as they occur, could serve as a valuable complement to human review.</p><p><strong>Conduct dry runs before the open-world experiment. </strong>Testing the agent scaffold, the evaluation criteria, and the infrastructure before running the full evaluation is helpful to uncover hidden assumptions about the task or errors in the scaffold. Our dry runs for the iOS app development task uncovered many issues with our scaffold before we began the real attempt.</p><p><strong>Measure cost. </strong>For many tasks, capabilities continue improving with increased budgets. Developers should treat cost measurement as a first-rate goal for conducting open-world evals, and report their findings alongside the budget they used. Even if it is not possible to make general claims about capability upper bounds, if measuring partial progress is viable, it can be helpful to get a sense of whether increasing the budget helps <a href="https://epoch.ai/blog/mirrorcode-preliminary-results">advance progress</a> towards task completion.</p><p><strong>Release logs. </strong>While open-world evaluations lack reproducibility, collecting and releasing logs to a broad community can help. For example, external researchers can add to analyses of how well agents performed or where they failed, and also verify the results.</p><p>We plan to conduct new CRUX evaluations regularly. In future CRUXes, we expect to conduct evaluations on a wide range of topics, including AI R&amp;D tasks, AI governance, video generation, and more challenging software engineering tasks. We plan to keep <a href="http://cruxevals.com">cruxevals.com</a> updated with our team's efforts and those of the broader community on open-world evaluations.</p><h4><strong>Author contributions and acknowledgments</strong></h4><p><strong>Core team:</strong> Sayash Kapoor and Arvind Narayanan conceptualized the project and designed the first evaluation task. Andrew Schwartz led the agent development and executed the evaluation. Peter Kirgis led the log analysis and literature review of open-world evaluations. Stephan Rabanser offered feedback and inputs into the task design, essay text, and our analysis and interpretation of the results. Sayash Kapoor, Peter Kirgis, Andrew Schwartz, Stephan Rabanser, and Arvind Narayanan drafted the essay.</p><p><strong>Collaborators</strong>: Rishi Bommasani offered feedback and inputs into the task design, essay text, and our analysis and interpretation of the results. J.J. Allaire, Magda Dubois, Gillian Hadfield, Andy Hall, Sara Hooker, Seth Lazar, Steve Newman, Dimitris Papailiopoulos, Shoshannah Tekofsky, Helen Toner, and Cozmin Ududec offered feedback and inputs into the essay text and our analysis and interpretation of the results.</p><p><strong>Acknowledgments: </strong>Nicholas Carlini provided feedback on the task design and agent setup. Ryan Greenblatt, Daniel Kokotajlo, and Ajeya Cotra participated in online and in-person conversations that informed the project.</p><p><strong>Funding. </strong>We are grateful to Coefficient Giving, Schmidt Sciences, and the Princeton AI Lab for funding to support this project.</p><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>Of course, we could also use dry runs to uncover issues with benchmarks. But benchmarks are intended to compare model capabilities over time. Many issues with benchmarks are only obvious when a more capable model finds a shortcut or an edge case that wasn&#8217;t obvious when the benchmark was constructed. Benchmark developers can analyze logs to uncover such cases. But resolving these issues requires updating the benchmark and re-running evaluations on prior models, which goes against the long-term validity of benchmark results. As AI agents become more capable, we expect such issues to arise more often.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p>There have always been issues with evaluations; the requirement is not that they are perfect, but rather that they are able to provide a useful proxy to the progress in AI capabilities. And there remain evaluations that are unsaturated, such as <a href="https://arxiv.org/abs/2407.13168">SciCode</a>, <a href="https://arxiv.org/abs/2406.01574">MMLU-Pro</a>, <a href="https://arxiv.org/abs/2501.14249">Humanity&#8217;s Last Exam</a>, and <a href="https://arxiv.org/abs/2509.16941">SWE-Bench Pro</a>. In fact, even saturated capability benchmarks can be useful for measuring the efficiency and reliability of AI agents, which are essential components that guide AI diffusion.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-3" href="#footnote-anchor-3" class="footnote-number" contenteditable="false" target="_self">3</a><div class="footnote-content"><p>For context, most children are able to beat the entire game <a href="https://howlongtobeat.com/game/7169">in around 25 hours.</a></p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-4" href="#footnote-anchor-4" class="footnote-number" contenteditable="false" target="_self">4</a><div class="footnote-content"><p>This concern also affects benchmarks that require access to the internet, for example web benchmarks such as <a href="https://arxiv.org/abs/2407.15711">AssistantBench</a> and <a href="https://arxiv.org/abs/2311.12983">GAIA</a>. But benchmarks run in sandboxed environments bypass this concern, at the cost of construct validity (if we tried to solve the task in the real world, the agent should be able to use resources on the internet to solve it).</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-5" href="#footnote-anchor-5" class="footnote-number" contenteditable="false" target="_self">5</a><div class="footnote-content"><p>There is already some evidence that Apple&#8217;s App Store reviews have started <a href="https://9to5mac.com/2026/03/29/vibe-coding-developers-report-long-app-store-review-queues/">taking</a> <a href="https://x.com/nikitabier/status/2033931821260648659">longer</a> as a result of the adoption of coding agents. Still, the rate of apps published is <a href="https://appfigures.com/resources/insights/20251205?f=2">much lower</a> than at the peak of app submissions in 2016, though it&#8217;s also possible that there might already be a bigger wave of slop submissions that are not eventually published. But our results show that agents can autonomously submit apps that are eventually approved by Apple. Once agents can develop and publish apps autonomously, it might lead to a much bigger <a href="https://www.fastcompany.com/91522242/apple-app-store-vibe-coding-generative-ai-unexpected-phenomenon">wave</a> of apps that satisfy publication criteria being submitted to the store, and App Store policies and practices might need updates to accommodate this.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-6" href="#footnote-anchor-6" class="footnote-number" contenteditable="false" target="_self">6</a><div class="footnote-content"><p>We made minor modifications to the agent to set up a subagent that verifies the agent&#8217;s outputs and wakes it up every 5 minutes to check for updates (e.g., responses from Apple&#8217;s review responses team). OpenClaw&#8217;s default interval for waking the agent is 30 minutes. In retrospect, waking the agent every 5 minutes increased the API cost for the task significantly.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-7" href="#footnote-anchor-7" class="footnote-number" contenteditable="false" target="_self">7</a><div class="footnote-content"><p>The agent required five total interventions, but four of them concerned requirements from Apple (either technical or policy-based) or flaws in our evaluation setup: the agent couldn&#8217;t resolve requirements from Apple like <a href="https://developer.apple.com/library/archive/documentation/Security/Conceptual/Security_Overview/Architecture/Architecture.html">intentionally preventing</a> <a href="https://support.apple.com/guide/security/controlling-app-access-to-files-secddd1d86a6/web">synthetic interactions</a> (such as those from the agent) on sensitive dialogs like approving two-factor authentication. It needed to prompt us to log in to the Apple developer account. At one point during our evaluation, the OpenClaw daemon crashed, requiring a manual restart. We also prompted the agent to ask for approval before publishing the app to satisfy the App Store&#8217;s policy requirements.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-8" href="#footnote-anchor-8" class="footnote-number" contenteditable="false" target="_self">8</a><div class="footnote-content"><p>When the agent was stuck and asked for help, a member of our team suggested reusing previously provided credentials and resolving 2FA. The agent briefly searched its memory for those credentials, but did not proceed with a live sign-in attempt. Instead, it recovered by discovering that the App Store Connect API key still existed at the expected hidden path, allowing monitoring to resume.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-9" href="#footnote-anchor-9" class="footnote-number" contenteditable="false" target="_self">9</a><div class="footnote-content"><p>While it is hard to say for sure, one reason the agent used a fictional phone number could have been that we informed the agent it was being evaluated. Also note that while we carried out in-depth log analysis on the agent&#8217;s actions, as we discussed earlier, automated log analysis can never be considered complete, so we think there might be interesting things we haven&#8217;t found yet. We have released our agent logs to aid third-party log analysis efforts (see <a href="http://cruxevals.com">cruxevals.com</a>).</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-10" href="#footnote-anchor-10" class="footnote-number" contenteditable="false" target="_self">10</a><div class="footnote-content"><p>Interestingly, partway through the evaluation, the agent changed its strategy to reduce the monitoring cost significantly: it started using subagents rather than using the entire context, and began using shorter daily memory files. This reduced the running cost from $35/hour to $3/hour.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-11" href="#footnote-anchor-11" class="footnote-number" contenteditable="false" target="_self">11</a><div class="footnote-content"><p>Note that our agent did not encounter any objections from Apple&#8217;s reviewers during the review process. On one hand, this shows that the agent was able to develop an app that passed the App Store&#8217;s bar for publication. On the other, we were unable to test how well the agent would perform in communications with Apple reviewers.</p><p></p></div></div>]]></content:encoded></item><item><title><![CDATA[New Paper: Towards a science of AI agent reliability]]></title><description><![CDATA[Quantifying the capability-reliability gap]]></description><link>https://www.normaltech.ai/p/new-paper-towards-a-science-of-ai</link><guid isPermaLink="false">https://www.normaltech.ai/p/new-paper-towards-a-science-of-ai</guid><dc:creator><![CDATA[Sayash Kapoor]]></dc:creator><pubDate>Tue, 24 Feb 2026 13:07:19 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/2e9d4b87-9068-4764-b50a-9c458b45252b_1486x960.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>By Stephan Rabanser, Sayash Kapoor, Arvind Narayanan</em></p><p>Suppose you hear about a new AI agent for improving productivity &#8212; by making purchases, or writing code, or sending emails, or handling a customer on your behalf. Should you trust it? Can the agent do the job reliably enough? After all, there are <a href="https://www.businessinsider.com/replit-ceo-apologizes-ai-coding-tool-delete-company-database-2025-7">many</a> <a href="https://www.washingtonpost.com/technology/2025/02/07/openai-operator-ai-agent-chatgpt/">horror</a> <a href="https://themarkup.org/artificial-intelligence/2024/03/29/nycs-ai-chatbot-tells-businesses-to-break-the-law">stories</a> of agents going wrong.</p><p>Surprisingly, even though the lack of reliability of AI agents is well known, right now the AI industry doesn&#8217;t have good tools for measuring reliability, or even a good definition of reliability.</p><p>Arvind and Sayash have long been <a href="https://sites.google.com/princeton.edu/agents-workshop">thinking</a> about this. Last fall, we were joined by postdoctoral researcher <a href="https://rabanser.dev/">Stephan Rabanser</a>, whose PhD looked at the reliability question in simpler, more traditional AI systems. We recruited a few other independent researchers, and have released what we hope is a comprehensive measurement of reliability. Our draft paper is called <a href="https://arxiv.org/abs/2602.16666">Towards a Science of AI Agent Reliability</a>.</p><p>We borrowed insights from many other fields, such as nuclear and aviation safety. We were able to decompose reliability into 12 different dimensions. Evaluating 14 models on two complementary benchmarks, we found that nearly two years of rapid capability progress have produced only modest reliability gains. See our interactive dashboard <a href="https://hal.cs.princeton.edu/reliability/">here</a>.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!KXpN!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd904192c-a01e-42fb-8990-b46f387f4690_2120x1276.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!KXpN!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd904192c-a01e-42fb-8990-b46f387f4690_2120x1276.jpeg 424w, https://substackcdn.com/image/fetch/$s_!KXpN!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd904192c-a01e-42fb-8990-b46f387f4690_2120x1276.jpeg 848w, https://substackcdn.com/image/fetch/$s_!KXpN!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd904192c-a01e-42fb-8990-b46f387f4690_2120x1276.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!KXpN!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd904192c-a01e-42fb-8990-b46f387f4690_2120x1276.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!KXpN!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd904192c-a01e-42fb-8990-b46f387f4690_2120x1276.jpeg" width="1456" height="876" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/d904192c-a01e-42fb-8990-b46f387f4690_2120x1276.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:876,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:135858,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.normaltech.ai/i/189010640?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd904192c-a01e-42fb-8990-b46f387f4690_2120x1276.jpeg&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!KXpN!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd904192c-a01e-42fb-8990-b46f387f4690_2120x1276.jpeg 424w, https://substackcdn.com/image/fetch/$s_!KXpN!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd904192c-a01e-42fb-8990-b46f387f4690_2120x1276.jpeg 848w, https://substackcdn.com/image/fetch/$s_!KXpN!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd904192c-a01e-42fb-8990-b46f387f4690_2120x1276.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!KXpN!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd904192c-a01e-42fb-8990-b46f387f4690_2120x1276.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>While our findings are tentative at this stage, we hope they can help explain the <a href="https://www.dwarkesh.com/p/thoughts-on-ai-progress-dec-2025">puzzlement</a> among many in the industry as to why the economic impacts of AI agents have been gradual, even though they are crushing capability benchmarks.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a> To help the community track reliability systematically, we plan to launch an AI agent &#8220;reliability index&#8221;. We hope this will stimulate researchers and industry to invest effort into improving reliability.</p><h3>Table of Contents</h3><ol><li><p><a href="https://www.normaltech.ai/i/189010640/accuracy-isnt-enough-four-dimensions-of-reliability">Accuracy isn&#8217;t enough: four dimensions of reliability</a></p></li><li><p><a href="https://www.normaltech.ai/i/189010640/capability-gains-are-rapid-but-improvements-in-reliability-are-modest">Capability gains are rapid, but improvements in reliability are modest</a></p></li><li><p><a href="https://www.normaltech.ai/i/189010640/why-we-could-be-wrong">Why we could be wrong</a></p></li><li><p><a href="https://www.normaltech.ai/i/189010640/what-should-deployers-do-differently">What should deployers do differently?</a></p></li><li><p><a href="https://www.normaltech.ai/i/189010640/what-should-researchers-and-developers-do-differently">What should researchers and developers do differently?</a></p></li><li><p><a href="https://www.normaltech.ai/i/189010640/what-do-our-findings-mean-for-ai-progress">What do our findings mean for AI progress?</a></p></li><li><p><a href="https://www.normaltech.ai/i/189010640/further-reading">Further reading</a></p></li></ol><h3>Accuracy isn&#8217;t enough: four dimensions of reliability</h3><p>When we consider a coworker to be reliable, we don&#8217;t just mean that they get things right most of the time. We mean something richer:</p><ol><li><p>They get it right consistently, not right today and wrong tomorrow on the same thing (Consistency)</p></li><li><p>They don&#8217;t fall apart when conditions aren&#8217;t perfect (Robustness)</p></li><li><p>They tell you when they&#8217;re unsure rather than confidently guessing (Calibration)</p></li><li><p>When they do mess up, their mistakes are more likely to be fixable than catastrophic (Safety)</p></li></ol><p>Unfortunately, AI agents are evaluated based on a single number, the average success rate at the task. That number has been going up quickly on many tasks over the last two years, which is why there&#8217;s so much excitement about deploying agents.</p><p>Safety-critical engineering fields (aviation, nuclear, automotive) figured out decades ago that reliability is not the same as average performance. These fields independently converged on the above four dimensions: consistency, robustness, predictability, and safety (the frequency and severity of failures).</p><p>For example, nuclear reactor protection systems <a href="https://www.nrc.gov/docs/ML0705/ML070550085.pdf">must respond identically</a> every time conditions warrant shutdown. Automotive safety testing evaluates <a href="https://www.iso.org/standard/68383.html">responses to sensor failures</a> and adverse weather. Nuclear risk assessment models thousands of failure modes and <a href="https://www.nrc.gov/reading-rm/doc-collections/fact-sheets/probabilistic-risk-asses">quantifies their probabilities</a>. Aviation <a href="https://www.faa.gov/documentLibrary/media/Advisory_Circular/AC%2023.1309-1E.pdf">targets one catastrophic error per billion flight hours</a>.</p><h3>Capability gains are rapid, but improvements in reliability are modest</h3><p>We refined and decomposed these four high-level dimensions into twelve metrics. We then tested agents based on 14 models from OpenAI, Google, and Anthropic, spanning 18 months of releases. We looked at two complementary benchmarks: a general assistant benchmark (GAIA) and a customer service simulation benchmark (TauBench). We ran each task five times, with instructions paraphrased. We injected faults in the tools and environment to measure robustness to such failures, and elicited the agents&#8217; confidence at having solved the task to measure calibration. In total, we executed 500 overall benchmark runs.</p><p>We found that reliability has improved only modestly over 18 months, while accuracy improved substantially. All three major providers cluster together, so this appears to be an industry-wide limitation (though there are some cases where Anthropic&#8217;s models outperform OpenAI&#8217;s and Google&#8217;s).</p><p>More specifically, we measured the following criteria:</p><ul><li><p><strong>Consistency</strong>: Agents that can solve a task often fail on repeated attempts under identical conditions. Many models have trouble giving a consistent answer, with outcome consistency scores ranging from 30% to 75% across the board.</p></li><li><p><strong>Robustness</strong>: Most models handle genuine technical failures (server crashes, API timeouts) gracefully. But if we rephrase the instructions with the same semantic meaning, performance drops substantially.</p></li><li><p><strong>Predictability</strong>: Agents are not good at knowing when they&#8217;re wrong. This is the weakest dimension across the board. When agents report confidence, it often carries little signal. On one benchmark, most models couldn&#8217;t distinguish their correct predictions from incorrect ones better than chance.</p></li><li><p><strong>Safety</strong>: Recent models are noticeably better at avoiding constraint violations, though financial errors, such as incorrect charges, remain a common failure mode. We use safety narrowly to mean bounded harm when failures occur, not broader concerns like alignment. We are still iterating on how we measure safety, so we report it separately from the aggregate reliability score.</p></li><li><p><strong>Impact of scaling</strong>: Bigger models aren&#8217;t uniformly more reliable. Scaling up improves some dimensions (calibration, robustness) but can hurt consistency. Larger models with richer behavioral repertoires sometimes show more run-to-run variability.</p></li></ul><h3>Why we could be wrong</h3><p>Our view is that reliability lags capability, and that reliability will remain a barrier to deployment unless researchers and developers focus effort on improving reliability as a separate dimension from accuracy. There are three reasons why we could be wrong.</p><p>First, there is some subjectivity in our dimensions and metrics. We have tried to minimize this by grounding our analysis in existing engineering fields. And we finalized our list of metrics <em>before</em> performing experiments, in order to prevent our hypothesis about slow reliability progress from influencing our selection of metrics. Still, we welcome other researchers to suggest alternative ways to measure reliability, and emphasize that <strong>our findings are tentative at this stage.</strong></p><p>Second, maybe reliability won&#8217;t matter if accuracy gets high enough. Our metrics are carefully crafted so that accuracy gains don&#8217;t automatically lead to reliability gains. Broadly speaking, accuracy is about the rate of failures while reliability is about the nature of failures. But if an agent is accurate 99% of the time, maybe we can tolerate 1% error even if it is completely unpredictable. We disagree. Our view is that for autonomous operation in high-stakes contexts, we need 3-5 &#8220;nines&#8221; of performance &#8212; 99.9% to 99.999% accuracy &#8212; in order for reliability to become a non-issue, and we don&#8217;t think LLM-based agents are on track to reach such a threshold. But time will tell.</p><p>Third, reliability progress being slower than accuracy doesn&#8217;t necessarily mean that it is slow in absolute terms. If we project the current linear trend forward, agents will reach 100% reliability in just three years! We don&#8217;t think a linear model makes sense, in part because we expect each order of magnitude decrease in <em>unreliability</em> (1-reliability) to be as hard as the previous one. That is, we expect the jump from 90 to 99% reliability to be about as hard as the jump from 99 to 99.9% reliability, and so on. But again, we just have to wait and see.</p><p>Suppose we&#8217;re right. There are important implications for deployers, researchers &amp; developers, and for those tracking the pace of AI progress. Let&#8217;s discuss each in turn.</p><h3>What should deployers do differently?</h3><p><strong>Clearly distinguish automation from augmentation</strong>. A coding assistant that occasionally suggests wrong variable names is annoying; an autonomous agent managing an industrial plant yielding highly variable output is unacceptable. The difference is whether the agent is used to augment a human&#8217;s creativity or directly make decisions. Augmentation tools (copilots, search assistants) get a reliability &#8220;discount&#8221; because someone reviews the output. In some augmentation use-cases, reliability might actually be undesirable. For example, a creative writing assistant that produces the same story every time would be terrible at its job.</p><p>Incidentally, how well agents collaborate with humans is woefully under-theorized and under-measured. There is some early work on <a href="https://arxiv.org/abs/2405.10632v2">evaluating</a> <a href="https://arxiv.org/abs/2405.10632v2">human-LM</a> interaction, but both efforts predate autonomous agents, and we are not aware of any equivalent work studying how agents collaborate with humans over multi-step tasks. Uplift studies offer one useful lens, but a broader agent-focused effort is overdue.</p><p><strong>Consider reliability for making release decisions</strong>. For automation tools (unattended workflows, customer-facing bots), reliability is non-negotiable. Deployers should consider requiring reliability thresholds before moving from sandbox to production, the way aviation requires certification before service. There are many other practices to borrow from such domains, such as building an <a href="https://arxiv.org/abs/2508.14231">incident</a>-<a href="https://cset.georgetown.edu/publication/ai-incidents-key-components-for-a-mandatory-reporting-regime/">reporting</a> <a href="https://arxiv.org/abs/2503.16861">culture</a> around agent failures.</p><p>While the metrics we have identified are broadly useful as a starting point for understanding reliability, deployers should build their own internal evaluations tailored to their specific context and datasets.</p><h3>What should researchers and developers do differently?</h3><p>Benchmarks drive progress in AI. A year and a half ago, our paper <a href="https://arxiv.org/pdf/2407.01502">AI Agents that Matter</a> showed there was a big gap between what agent benchmarks measured and what matters in practice. Our new paper shows that the gap persists. To fix this disconnect, both AI evaluation and AI development practices need to change course.</p><p><strong>Measure reliability</strong>. The current approach of running a benchmark once and reporting the accuracy number is a shallow, superficial performance measure. It is comparable to stress-testing a car once in perfect weather and declaring it safe if it passes. When evaluating agents, we need to test them for multiple runs (testing for variance in outcomes), under different conditions (evaluating adaptability), and on an ongoing basis (re-test as models and environments change). We call for reporting reliability profiles alongside accuracy, not instead of it.</p><p><strong>Understand and improve reliability</strong>. Our experiments suggest that consistency and predictability are the biggest gaps preventing models from being more reliable. Agent developers should consider improving these weak points explicitly, possibly via targeted optimization or improved scaffolding. In particular, agents should be able to recognize when they are likely to fail and say so, and recover gracefully when they do fail. More speculative ideas include agents that explore different strategies during development but follow a consistent execution plan when deployed, rather than solving the same task differently each time, simultaneously delivering the best of <a href="https://huggingface.co/blog/VirtualOasis/agents-vs-workflows-en">agents and workflows</a>.</p><h3>What do our findings mean for AI progress?</h3><p>The capability-reliability gap could be one reason why we are not seeing any of the rapid labor-market effects that Artificial General Intelligence has been predicted to bring about. It is not the only one. A recent UK AISI <a href="https://www.aisi.gov.uk/blog/mapping-the-limitations-of-current-ai-systems">report</a> identified six barriers to AGI. However, the report discussed each at a high level.</p><p>Our paper can be seen as putting flesh on the bones of one of these barriers &#8212; reliability. None of the four dimensions we identify can be considered solved at this point in time, and only two of the individual metrics, shown in green in the figure, can be considered (tentatively) solved.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!MID-!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8f64a3e5-d811-480d-9003-b2fcd0bc6e92_1522x974.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!MID-!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8f64a3e5-d811-480d-9003-b2fcd0bc6e92_1522x974.png 424w, https://substackcdn.com/image/fetch/$s_!MID-!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8f64a3e5-d811-480d-9003-b2fcd0bc6e92_1522x974.png 848w, https://substackcdn.com/image/fetch/$s_!MID-!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8f64a3e5-d811-480d-9003-b2fcd0bc6e92_1522x974.png 1272w, https://substackcdn.com/image/fetch/$s_!MID-!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8f64a3e5-d811-480d-9003-b2fcd0bc6e92_1522x974.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!MID-!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8f64a3e5-d811-480d-9003-b2fcd0bc6e92_1522x974.png" width="1456" height="932" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/8f64a3e5-d811-480d-9003-b2fcd0bc6e92_1522x974.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:932,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:145647,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.normaltech.ai/i/189010640?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8f64a3e5-d811-480d-9003-b2fcd0bc6e92_1522x974.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!MID-!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8f64a3e5-d811-480d-9003-b2fcd0bc6e92_1522x974.png 424w, https://substackcdn.com/image/fetch/$s_!MID-!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8f64a3e5-d811-480d-9003-b2fcd0bc6e92_1522x974.png 848w, https://substackcdn.com/image/fetch/$s_!MID-!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8f64a3e5-d811-480d-9003-b2fcd0bc6e92_1522x974.png 1272w, https://substackcdn.com/image/fetch/$s_!MID-!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8f64a3e5-d811-480d-9003-b2fcd0bc6e92_1522x974.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em><strong>Future work on other barriers to AGI may reveal many other dimensions of performance that must be improved before AI agents can be widely deployed.</strong></em></figcaption></figure></div><p>There is much work to be done in fleshing out the other barriers to AGI, analogous to our analysis of reliability. Our hunch is that this will reveal many other dimensions and metrics on which progress has been slow. The gazillion-dollar question is whether agents will get better across the board through general methods such as inference scaling and reinforcement learning, or whether painstaking work will be required to improve individual dimensions of reliability, adaptability, originality, and so on.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.normaltech.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Subscribe to receive evidence-based analysis of AI developments and the future of AI.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p></p><h3>Further reading</h3><ul><li><p>The full 66-page paper is available on <a href="https://arxiv.org/abs/2602.16666">arXiv</a> and includes definitions of the metrics we propose, details on our experimental setup, as well as an extended discussion on our recommendations and proposed future work. The authors are Stephan Rabanser, Sayash Kapoor, Peter Kirgis, Kangheng Liu, Saiteja Utpala, and Arvind Narayanan. Alongside the paper, we also provide an <a href="https://hal.cs.princeton.edu/reliability/">interactive dashboard</a> of our results to enable practitioners to drill down into individual models, benchmarks, and reliability dimensions. The code to reproduce our experiments is available on  <a href="https://github.com/steverab/hal-harness">GitHub</a>.</p></li><li><p>This paper is part of our broader effort to advance the <a href="https://sage.cs.princeton.edu/">Science of AI Agent Evaluation</a>. Key publications include <a href="https://www.normaltech.ai/p/new-paper-ai-agents-that-matter">AI Agents that Matter</a> and <a href="https://arxiv.org/pdf/2510.11977">Holistic Agent Leaderboard: The Missing Infrastructure for AI Agent Evaluation</a>, recently accepted to ICLR 2026.</p></li><li><p>Our paper can be seen as building on work that aims to extend the scope of what we measure through benchmarks. For example,&nbsp;<a href="https://arxiv.org/abs/2211.09110">HELM</a>&nbsp;introduced 7 metrics, including calibration and robustness, and used these to evaluate dozens of language models on prominent benchmarks.</p></li></ul><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>Of course, this is not the only factor. Benchmarks continue to <a href="https://arxiv.org/pdf/2602.12413">struggle</a> with data contamination, so they might overestimate capability progress. <a href="https://knightcolumbia.org/content/ai-as-normal-technology">AI as Normal Technology</a> provides a useful framework to understand the barriers between capability progress and economic impacts. Reliability falls into the products/applications stage. Human, social, and organizational factors lie downstream. As a case study of how serious these can be, see our recent essay <a href="https://www.lawfaremedia.org/article/ai-won-t-automatically-make-legal-services-cheaper">AI Won&#8217;t Automatically Make Legal Services Cheaper</a>.</p><p></p></div></div>]]></content:encoded></item><item><title><![CDATA[AI Won’t Automatically Make Legal Services Cheaper]]></title><description><![CDATA[Applying the AI as Normal Technology framework to legal services]]></description><link>https://www.normaltech.ai/p/ai-wont-automatically-make-legal</link><guid isPermaLink="false">https://www.normaltech.ai/p/ai-wont-automatically-make-legal</guid><dc:creator><![CDATA[Justin Curl]]></dc:creator><pubDate>Thu, 12 Feb 2026 19:26:05 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/2a2691bd-fda5-4a31-aa6c-ff0d59c79e4d_2428x1368.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>This essay was published in Lawfare&#8217;s Research Paper Series. The official version that should be cited is linked <a href="https://www.lawfaremedia.org/article/ai-won-t-automatically-make-legal-services-cheaper">here</a>.</em></p><p><em>The essay is co-authored with <a href="https://x.com/curl_justin">Justin Curl</a>, a third-year at Harvard Law School. Previously, he was a Schwarzman Scholar at Tsinghua University and earned a degree in Computer Science from Princeton, where we first collaborated. You can find more of his writing on <a href="https://substack.com/@justincurl">AI and the law here</a>.</em></p><div><hr></div><p>Many AI leaders believe the technology will transform knowledge work. OpenAI CEO Sam Altman predicts AI systems that are &#8220;smarter than humans by 2030,&#8221;<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a> while Anthropic CEO Dario Amodei analogizes future AI models to a &#8220;country of geniuses in a data center.&#8221;<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a></p><p>Researchers identify legal services as especially vulnerable to disruption by AI.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-3" href="#footnote-3" target="_self">3</a> And since GPT-4 passed the bar exam,<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-4" href="#footnote-4" target="_self">4</a> much of the profession seems to agree. Law schools have begun incorporating AI into their curricula<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-5" href="#footnote-5" target="_self">5</a><sup> </sup>and partnering with AI-focused legal-tech companies to prepare future lawyers for a changing profession.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-6" href="#footnote-6" target="_self">6</a><sup> </sup>One prominent lawyer has argued AI can already replace law clerks and oral argument.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-7" href="#footnote-7" target="_self">7</a> Another predicts AI could &#8220;replace traditional lawyers by 2035.&#8221;<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-8" href="#footnote-8" target="_self">8</a></p><p>This excitement about AI comes at a time when legal services are expensive. Millions of individuals are priced out of legal assistance,<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-9" href="#footnote-9" target="_self">9</a> while corporate legal fees are increasing steadily, with hourly rates for partners at large law firms now exceeding $2,300.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-10" href="#footnote-10" target="_self">10</a> Unsurprisingly, many observers see the potential for AI to make legal services more accessible by delivering outcomes at lower costs.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-11" href="#footnote-11" target="_self">11</a></p><p>Our central claim is that advanced AI will not, by default, help consumers achieve their desired legal outcomes at lower costs. We examine the bottlenecks<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-12" href="#footnote-12" target="_self">12</a> that stand between AI capability advances and the positive transformation of the practice of law that some envision. For AI to usher in a world of abundant legal services, the profession must address three bottlenecks: regulatory barriers, adversarial dynamics, and human involvement.</p><p>First, unauthorized practice of law (UPL) regulations may limit AI use by consumers (and to some extent lawyers). These laws prohibit nonlawyers from performing legal work.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-13" href="#footnote-13" target="_self">13</a> Individuals and organizations can face steep fines and criminal liability if courts conclude their systems cross into practicing law,<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-14" href="#footnote-14" target="_self">14</a> forcing would-be providers to either limit their AI tools&#8217; functionality in legal domains or risk enforcement actions. Entity-based regulations&#8212;which restrict who can own equity in businesses that provide legal services&#8212;restrict how legal services are offered, again limiting how AI is used by lawyers and consumers. Without reforms, if consumers cannot access AI capabilities or lawyers are not incentivized to use AI well, AI will not help people accomplish their legal goals, regardless of how advanced it becomes.</p><p>Second, even if AI is effectively and widely adopted, the American<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-15" href="#footnote-15" target="_self">15</a> legal system&#8217;s adversarial structure can prevent advanced AI from lowering the cost of achieving clients&#8217; outcomes.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-16" href="#footnote-16" target="_self">16</a> Because legal outcomes often depend on relative rather than absolute quality, when both parties become more productive, the competitive equilibrium simply shifts upward.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-17" href="#footnote-17" target="_self">17</a> In a world with advanced AI, achieving the same result&#8212;like settling favorably or prevailing at trial&#8212;would require a greater quantity and quality of legal work. So even as productivity increases and cost per legal task falls, parties are locked into an arms race of increasing amounts of legal work required to reach the same outcome.</p><p>As a historical analogy, digitization could have reduced discovery costs by making document review much easier.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-18" href="#footnote-18" target="_self">18</a> But litigators operating within litigation&#8217;s adversarial framework exploited the surge in digital documents to drive up costs for their opponents, leaving total litigation costs high.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-19" href="#footnote-19" target="_self">19</a></p><p>Though less explicitly adversarial, transactional work (like contract negotiation) can exhibit similar dynamics: Lawyers compete to control disclosures and outmaneuver opposing counsel when drafting and negotiating agreements.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-20" href="#footnote-20" target="_self">20</a> Of course, some legal outcomes (like effective estate planning) do not depend on adversarial processes, and this bottleneck would not apply to them.</p><p>The third and final bottleneck we discuss is human involvement in legal work. In a world where AI gains outpace increases in the volume of legal work, our desire for human beings to adjudicate cases and understand the contracts they sign is a final bottleneck. In litigation, if AI enables a flood of legal work, judges will likely respond by taking longer to resolve disputes (delaying outcomes) or delegate more to assistants (lowering adjudication quality).<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-21" href="#footnote-21" target="_self">21</a> And with transactional work, even if AI is drafting entire contracts, human lawyers will still need time to understand what these provisions mean for an organization&#8217;s interests. The speed of human decision-makers (whether judges, lawyers, or clients) places an upper limit on how much AI can accelerate legal processes without sacrificing human involvement. See <strong>Figure 1</strong> below for a visual description of the three bottlenecks and our argument.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!BKN-!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb460f8d7-d0f3-4752-a809-2b73c50b0138_1056x864.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!BKN-!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb460f8d7-d0f3-4752-a809-2b73c50b0138_1056x864.png 424w, https://substackcdn.com/image/fetch/$s_!BKN-!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb460f8d7-d0f3-4752-a809-2b73c50b0138_1056x864.png 848w, https://substackcdn.com/image/fetch/$s_!BKN-!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb460f8d7-d0f3-4752-a809-2b73c50b0138_1056x864.png 1272w, https://substackcdn.com/image/fetch/$s_!BKN-!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb460f8d7-d0f3-4752-a809-2b73c50b0138_1056x864.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!BKN-!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb460f8d7-d0f3-4752-a809-2b73c50b0138_1056x864.png" width="1056" height="864" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/b460f8d7-d0f3-4752-a809-2b73c50b0138_1056x864.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:864,&quot;width&quot;:1056,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:161213,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://justincurl.substack.com/i/187445681?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb460f8d7-d0f3-4752-a809-2b73c50b0138_1056x864.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!BKN-!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb460f8d7-d0f3-4752-a809-2b73c50b0138_1056x864.png 424w, https://substackcdn.com/image/fetch/$s_!BKN-!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb460f8d7-d0f3-4752-a809-2b73c50b0138_1056x864.png 848w, https://substackcdn.com/image/fetch/$s_!BKN-!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb460f8d7-d0f3-4752-a809-2b73c50b0138_1056x864.png 1272w, https://substackcdn.com/image/fetch/$s_!BKN-!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb460f8d7-d0f3-4752-a809-2b73c50b0138_1056x864.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><strong>Figure 1:</strong> The bottlenecks between advanced AI capabilities and the positive transformation of the practice of law</figcaption></figure></div><p>This report applies the &#8220;AI as Normal Technology&#8221; framework to a specific domain: the legal industry.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-22" href="#footnote-22" target="_self">22</a> This framework is fundamentally about agency: Rather than treating AI&#8217;s trajectory as predetermined by capability advances, it directs attention to the social and organizational bottlenecks between what AI can do and the impact it has on the world. In our analysis of the practice of law, diffusion will likely be slow. Better models have not yet translated into more reliable legal products because adapting workflows to leverage AI and teaching users takes time.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-23" href="#footnote-23" target="_self">23</a></p><p>We argue that the end state of AI diffusion can look very different depending on the institutional response. Some pathways lead to genuine improvements in access and efficiency, while others simply make producing legal work (outputs) cheaper without making it easier to achieve the results clients want (outcomes). For AI to deliver better <em>outcomes</em> for consumers, the legal industry must enact reforms addressing the bottlenecks. Otherwise, we risk a future in which legal work becomes more abundant, but legal outcomes remain expensive and inaccessible.</p><p>We proceed in three sections. The first section explains why legal services are so expensive. The second section aims to convince readers that AI won&#8217;t automatically deliver legal outcomes at lower costs. And the third section offers recommendations for addressing these bottlenecks based on existing proposals for legal reform and illustrates how drastically AI&#8217;s impact could differ based on the legal profession&#8217;s response.</p><h1><strong>Table of Contents</strong></h1><ol><li><p><a href="https://justincurl.substack.com/p/16e98caa-1854-4db5-9a75-266e9a1c3a5b#why-legal-services-are-so-expensive">Why Legal Services Are So Expensive</a></p></li><li><p><a href="https://justincurl.substack.com/p/16e98caa-1854-4db5-9a75-266e9a1c3a5b#why-ai-wont-help-by-default">Why AI Won&#8217;t Help by Default</a></p><ol><li><p><a href="https://justincurl.substack.com/p/16e98caa-1854-4db5-9a75-266e9a1c3a5b#regulatory-barriers">Regulatory Barriers</a></p></li><li><p><a href="https://justincurl.substack.com/p/16e98caa-1854-4db5-9a75-266e9a1c3a5b#adversarial-dynamics">Adversarial Dynamics</a></p></li><li><p><a href="https://justincurl.substack.com/p/16e98caa-1854-4db5-9a75-266e9a1c3a5b#human-oversight">Human Oversight</a></p></li></ol></li><li><p><a href="https://justincurl.substack.com/p/16e98caa-1854-4db5-9a75-266e9a1c3a5b#institutional-reforms">Institutional Reforms</a></p><ol><li><p><a href="https://justincurl.substack.com/p/16e98caa-1854-4db5-9a75-266e9a1c3a5b#reforming-professional-regulation">Reforming Professional Regulation</a></p></li><li><p><a href="https://justincurl.substack.com/p/16e98caa-1854-4db5-9a75-266e9a1c3a5b#reforming-adjudication">Reforming Adjudication</a></p></li><li><p><a href="https://justincurl.substack.com/p/16e98caa-1854-4db5-9a75-266e9a1c3a5b#the-evolving-role-of-lawyers">The Evolving Role of Lawyers</a></p></li></ol></li><li><p><a href="https://justincurl.substack.com/p/16e98caa-1854-4db5-9a75-266e9a1c3a5b#conclusion">Conclusion</a></p></li></ol><h1><strong>Why Legal Services Are So Expensive</strong></h1><p>Three structural factors help explain why legal services are so expensive:<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-24" href="#footnote-24" target="_self">24</a> Evaluating their quality is difficult, their value is often relative, and professional regulations limit competition from alternative business models.</p><p>First, unlike a meal at a restaurant, where it&#8217;s easy to assess quality, legal services are &#8220;credence goods,&#8221; which means their quality is difficult to evaluate even with hindsight.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-25" href="#footnote-25" target="_self">25</a> The final outcome in a case reflects the cumulative effect of many smaller decisions, so it can be very hard, even for other lawyers, to evaluate whether legal services were provided effectively.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-26" href="#footnote-26" target="_self">26</a> How clear was the law on that issue? Did the client reach the desired outcome because of or in spite of the lawyer&#8217;s skill? Which decisions actually contributed to that success? This evaluation difficulty forces consumers to rely on proxies for quality (e.g., the prestige of a lawyer&#8217;s law school or judicial clerkships) when choosing between legal service providers, making it hard for traditional market mechanisms to work.</p><p>Second, the value of legal services is relative.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-27" href="#footnote-27" target="_self">27</a> Because &#8220;the American litigation system is a thoroughgoing adversarial&#8221; one, what matters to the lawsuit&#8217;s outcome often isn&#8217;t how good your lawyers are in absolute terms, but whether they&#8217;re better than the other side&#8217;s.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-28" href="#footnote-28" target="_self">28</a> Other kinds of legal work, such as drafting contracts and agreements (often called transactional work), can also be adversarial as lawyers try to &#8220;outfox&#8221; each other in terms of what is disclosed in negotiations and the language of a contract itself.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-29" href="#footnote-29" target="_self">29</a></p><p>These dynamics have kick-started an arms race for legal talent, driving up costs at the top end of the market, which serves corporate clients and is often called &#8220;BigLaw.&#8221;<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-30" href="#footnote-30" target="_self">30</a><sup> </sup>In 2024, the median partner at large law firms charged $1,050 per hour, with some commanding over $2,300.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-31" href="#footnote-31" target="_self">31</a> That&#8217;s up 5.1 percent from 2023, which was itself up 5.4 percent from 2022.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-32" href="#footnote-32" target="_self">32</a> Fortune 200 companies reported that their average litigation costs in cases exceeding $250,000 in legal fees had nearly doubled over eight years, climbing from $66 million per company in 2000 to $115 million in 2008.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-33" href="#footnote-33" target="_self">33</a> In the patent field, a 2017 survey found that patent cases worth <em>less</em> than $1 million typically cost $1 million to litigate ($500,000 per side).<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-34" href="#footnote-34" target="_self">34</a></p><p>Third, the profession&#8217;s regulatory framework, designed with consumer protection in mind, has created its own complications. Two types of regulations are often the focus of reform: unauthorized practice of law (UPL) and law firm ownership regulations.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-35" href="#footnote-35" target="_self">35</a></p><p>UPL laws make it illegal (in some jurisdictions a felony) for unlicensed attorneys to apply legal knowledge to specific circumstances.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-36" href="#footnote-36" target="_self">36</a> An unfortunate effect is to make it more expensive to offer basic legal assistance in contexts requiring little legal expertise.</p><p>Most states have regulations limiting who may share in legal fees. Gillian Hadfield argues that these rules promote a business model that creates inefficiency for small firms.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-37" href="#footnote-37" target="_self">37</a> These firms serve individuals and small businesses and are sometimes called the &#8220;PeopleLaw&#8221; sector.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-38" href="#footnote-38" target="_self">38</a> She cites a 2017 Clio study of forty thousand customers: In an average eight-hour workday, lawyers engaged in billable work for only 2.3 hours, billed 1.9 hours, and collected payment for just 1.6 hours.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-39" href="#footnote-39" target="_self">39</a> So although clients paid an average of $260 per hour, lawyers effectively received $25&#8211;40 per hour because the rest of their time was spent finding clients, managing administrative tasks, and collecting payments. These regulations require lawyers to serve clients through partnerships fully owned and financed by lawyers. They deter alternative models that involve large-scale businesses with centralized billing, customer service, marketing, and administrative functions, which could leverage economies of scale to deliver legal services at $30&#8211;50 per hour instead of $260.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-40" href="#footnote-40" target="_self">40</a></p><p>Importantly, none of the sources of market dysfunction are intrinsic to legal services. They reflect choices about procedure, pricing, and professional governance. While reform may be politically difficult or costly, the outlook is dim without it. And contrary to what some might hope, AI will not automatically make legal services cheaper, as we discuss next.</p><h1><strong>Why AI Won&#8217;t Help by Default</strong></h1><h2><strong>Regulatory Barriers</strong></h2><p>More legal assistance would be valuable in the debt collection context. From 1993 to 2013, the number of debt collection lawsuits grew from 1.7 million to about 4 million.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-41" href="#footnote-41" target="_self">41</a> In Michigan, these lawsuits made up 37 percent of all civil district court case filings by 2019.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-42" href="#footnote-42" target="_self">42</a> The trend is similar in Texas: &#8220;Debt claims more than doubled from 2014 to 2018, accounting for 30% of the state&#8217;s civil caseload by the end of that five-year period.&#8221;<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-43" href="#footnote-43" target="_self">43</a> More than 70 percent of debt collection defendants lose by default for failing to respond, even though many cases are &#8220;meritless suits&#8221; and responding is not complicated.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-44" href="#footnote-44" target="_self">44</a></p><p>New York has created a form for responding to debt collection lawsuits by checking some boxes.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-45" href="#footnote-45" target="_self">45</a> This form, however, includes questions difficult for nonlawyers to understand, such as whether someone would like to invoke the doctrine of &#8220;laches.&#8221; Recognizing this difficulty, the nonprofit Upsolve began training volunteers to offer assistance. Concerned that this might violate New York&#8217;s UPL rules, Upsolve sought an injunction declaring this basic assistance was protected by the First Amendment. A federal judge agreed.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-46" href="#footnote-46" target="_self">46</a> But New York appealed to the U.S. Court of Appeals for the Second Circuit, which invalidated the injunction, concluding that the lower court applied the wrong First Amendment test (so Upsolve was no longer protected).<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-47" href="#footnote-47" target="_self">47</a></p><p>It&#8217;s easy to see how an AI system could help here. A nonprofit could provide access to a tool customized for debt collection suits. Or individuals could directly ask general-purpose tools like ChatGPT, Claude, or Gemini for relevant information. Despite this potential, organizations risk violating UPL laws whenever their AI tools complete &#8220;tasks that require legal judgment or expertise.&#8221;<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-48" href="#footnote-48" target="_self">48</a> The New York Bar Association, concerned that the shortcomings of current AI models would harm consumers, has warned that &#8220;AI-powered chat bots now hover on the line of unauthorized practice of law.&#8221;<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-49" href="#footnote-49" target="_self">49</a> While some legal researchers disagree because AI is not a &#8220;person&#8221; capable of exercising legal judgment<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-50" href="#footnote-50" target="_self">50</a> or because AI systems simply &#8220;provide information to users, similar to paper guides about court procedure,&#8221;<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-51" href="#footnote-51" target="_self">51</a> all authors cited in this paragraph agree that the status of AI tools under UPL laws is currently unclear.</p><p>LegalZoom&#8217;s history of lawsuits illustrates how UPL regulations can deter innovation in the delivery of legal services.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-52" href="#footnote-52" target="_self">52</a> LegalZoom automates rote tasks like preparing documents for trademark filings and has been plagued by UPL lawsuits for years.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-53" href="#footnote-53" target="_self">53</a> In 2011, private individuals who had purchased the company&#8217;s services sued in Missouri, alleging LegalZoom was engaged in UPL because it claimed to &#8220;take[] over once a consumer answer[ed] a few simple online questions.&#8221;<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-54" href="#footnote-54" target="_self">54</a> After the court denied its motion to dismiss the case, LegalZoom agreed to compensate plaintiffs and modify its business model. In 2015, the North Carolina State Bar won a consent judgment requiring LegalZoom to conform to certain conditions.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-55" href="#footnote-55" target="_self">55</a> Trademark lawyers in California advanced similar theories in a 2017 suit against LegalZoom&#8217;s trademark-filing product.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-56" href="#footnote-56" target="_self">56</a> And in 2024, a New Jersey plaintiff brought a class action alleging UPL violations.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-57" href="#footnote-57" target="_self">57</a></p><p>While AI&#8217;s legality remains in doubt, the threat of UPL liability can inhibit its adoption. Without reform, developers risk fines and criminal liability if their AI systems provide legal advice. Organizations may simply be unwilling to provide access to users, especially those who cannot afford to compensate a developer for the risk of UPL liability. Separately, entity regulations that restrict financing for AI legal startups can deter the kinds of operational experimentation helpful for delivering legal services at lower costs. Overall, if regulatory barriers prevent consumers from effectively accessing AI capabilities, it will not translate into better legal outcomes for clients at lower costs.</p><p>That said, AI could reduce the costs of legal services for reasons unrelated to its ability to perform legal tasks. As the 2017 Clio survey mentioned above found, nonlegal work consumes a large percentage of lawyers&#8217; time at the low end of the market. If advanced AI helps find and communicate with clients, manage administrative tasks, and handle payments, it could free up these lawyers to spend more time on legal work.</p><h2><strong>Adversarial Dynamics</strong></h2><p>Even in a world in which AI increases lawyers&#8217; productivity and completes legal tasks, it might not lower the costs of legal services. To see why, it is crucial to distinguish <em>inputs</em> and <em>outputs</em> from <em>outcomes</em>. Inputs are what goes into legal work: employee talent, billable hours, and technological tools. Outputs are what legal work produces: contracts drafted, motions filed, and briefs written. Outcomes are what clients actually care about: disputes resolved, deals closed, and rights protected.</p><p>Consumers purchase legal services to achieve specific outcomes. Inputs and outputs can indirectly lead to those outcomes because more hours worked and more legal tasks help clients get the outcomes they want. But in a zero-sum context where the value of legal services is relative, if both sides increase their outputs, the advantages to either side of doing so can be limited.</p><p>Instead, AI might simply raise the inputs and outputs required to reach the same outcome, with productivity gains absorbed by greater production. The billable hours model, in which lawyers are paid and promoted based on inputs, only reinforces these dynamics: More hours worked drafting motions and reviewing documents translates into greater revenue for legal firms without necessarily improving outcomes.</p><h3><strong>Litigation</strong></h3><p>One response is that these arms races actually create value by increasing the quantity and quality of legal work. However, clients sometimes achieve their desired outcomes (like settling a case or dismissing a lawsuit) by imposing greater costs on the other side instead of improving the quality of their legal arguments or evidence. It is a &#8220;core premise of litigation economics&#8221;<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-58" href="#footnote-58" target="_self">58</a> that &#8220;all things being equal, the party facing higher costs will settle on terms more favorable to the party facing lower costs.&#8221;<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-59" href="#footnote-59" target="_self">59</a> And even where quality does improve, it&#8217;s uncertain that the benefits of higher quality legal work (like helping courts reach the &#8220;right&#8221; answer more often) outweigh the costs of more legal work (like overwhelming judges with cases).</p><p>Earlier technological shifts cast doubt on whether America&#8217;s &#8220;adversarial legalism&#8221; can translate productivity gains into more affordable legal services.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-60" href="#footnote-60" target="_self">60</a></p><p>Discovery is a cornerstone of American litigation that often determines whether cases settle or go to trial. In discovery, parties share information &#8220;to identify material facts that prove or disprove a claim.&#8221;<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-61" href="#footnote-61" target="_self">61</a> It operates through an adversarial exchange: One party sends a discovery request; the other searches its records and decides which documents are responsive and which are protected by privilege.</p><p>Discovery was conceived of as a cooperative process during which lawyers could share information to facilitate settlement and avoid trial.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-62" href="#footnote-62" target="_self">62</a> Yet two characteristics of discovery make it vulnerable to abuse. First, the party holding the documents (the reviewing party) knows what&#8217;s in them but seeks to share as little helpful information with the requesting party as possible to minimize legal risk. Second, the reviewing party, because they bear the responsibility and costs for producing documents, must essentially act as their adversary&#8217;s agent.</p><p>Each side can leverage these features to impose costs on the other. A requesting party, through excessive requests, can compel their adversary to review more documents for confidentiality and relevance. And a reviewing party, through excessive production, can bury relevant information in mountains of extraneous material, forcing the opposing side to spend more time on review. This can create an arms race of overrequesting and oversharing, as each side drives up costs for the other to pressure them to settle on more favorable terms. The billable hours model again reinforces this behavior, with more review generating more billable hours.</p><p>These adversarial incentives can be powerful. Judge Frank Easterbrook opened a well-known article, &#8220;Discovery as Abuse,&#8221; by analogizing the process to &#8220;nuclear war.&#8221;<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-63" href="#footnote-63" target="_self">63</a> Charles Yablon described how one side made life difficult for opposing counsel: It printed documents on dark red, foul-smelling paper so that their contents would be nearly illegible and the attorneys would get nauseous reviewing them.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-64" href="#footnote-64" target="_self">64</a><sup> </sup>As Arthur Miller aptly observed, discovery&#8217;s key defect was believing &#8220;that adversarial tigers would behave like accommodating pussycats throughout the discovery period, saving their combative energies for trial.&#8221;</p><p>Digitization might have pushed discovery costs in either direction.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-65" href="#footnote-65" target="_self">65</a> Better search capabilities meant attorneys could review documents more efficiently, driving down costs while increasing the relevance of information shared. Yet the explosion of digital information increased what parties might need to search through during discovery, creating more opportunities for discovery abuse.</p><p>The empirical evidence on discovery is limited, so making causal claims about digitization&#8217;s impact is difficult.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-66" href="#footnote-66" target="_self">66</a> The available evidence, however, suggests it remains expensive. Discovery accounts for roughly &#8220;one-third to one-half of all litigation costs&#8221; when used.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-67" href="#footnote-67" target="_self">67</a> Fortune 200 companies reported that in cases with over $250,000 in legal fees, which are typically the kinds of complex litigation that require discovery, average litigation costs nearly doubled from $66 million per company in 2000 to $115 million in 2008.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-68" href="#footnote-68" target="_self">68</a> These figures align with evidence of over-requesting and oversharing: One trade association for civil defense lawyers estimated that for every page eventually shown at trial, meaning it&#8217;s relevant and reliable enough to be used as evidence, over one thousand pages were produced in discovery.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-69" href="#footnote-69" target="_self">69</a></p><p>The main lesson of digitization, then, is that adversarial processes do not translate predicted productivity gains into lower-cost legal outcomes by default. David Engstrom and Jonah Gelbach hope that Technology Assisted Review (TAR) software, which uses predictive AI to classify documents for privilege and confidentiality based on an initial training set of labeled documents, can eventually solve discovery&#8217;s inefficiencies.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-70" href="#footnote-70" target="_self">70</a><sup> </sup>Yet they acknowledge this is far from guaranteed given litigation&#8217;s adversarial structure. And because federal rules require discovery requests to be &#8220;proportional&#8221; to a case&#8217;s needs, judges might respond to declining unit costs of discovery (the cost of producing each document) by authorizing more <em>expansive</em> discovery plans, leaving total costs high. Though discovery is important, it is not uniquely susceptible to adversarial dynamics. An arms race of legal work in any stage of litigation (e.g., pretrial motions, expert battles, appeals) can erode efficiency gains and make achieving clients&#8217; objectives expensive.</p><h3><strong>Transactional Work</strong></h3><p>Similar adversarial patterns appear in transactional work like negotiating and drafting contracts. Hadfield illustrates how their value can be relative using an example of a merger agreement negotiation.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-71" href="#footnote-71" target="_self">71</a> In such negotiations, skillful lawyers can improve their client&#8217;s position by limiting disclosures, &#8220;skating the line between legitimate silence and misrepresentation.&#8221;<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-72" href="#footnote-72" target="_self">72</a> Opposing counsel will try to stay one step ahead by deciding what to ask, which representations to seek, and how to interpret disclosure laws to reduce the likelihood that material information is withheld.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-73" href="#footnote-73" target="_self">73</a> Lawyers will also try to &#8220;outfox&#8221; each other in the language of the contract itself.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-74" href="#footnote-74" target="_self">74</a></p><p>In the event of future litigation, the clients whose lawyers misunderstood a term&#8217;s significance pay the price. Since one can always do more legal research or add more contract provisions, there is no natural limit on the capacity for additional legal work to absorb efficiency gains. And because contracts concern future obligations, added uncertainty makes it harder for clients to effectively compare the quality and price of legal services.</p><p>Contracts have grown longer and more complex over time. From 1996 to 2016, M&amp;A agreements expanded from 35 to 88 single-spaced pages, their linguistic complexity increasing from post-graduate &#8220;grade 20&#8221; to postdoctoral &#8220;grade 30.&#8221;<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-75" href="#footnote-75" target="_self">75</a> An analysis of privacy policies over a similar period found a similar trend.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-76" href="#footnote-76" target="_self">76</a> John Coates argues that the increases in length and complexity reflect necessary and valuable responses to emerging legal risks.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-77" href="#footnote-77" target="_self">77</a> While longer, more complex contracts may represent better agreements, they might be necessary only in a legal system as adversarial as America&#8217;s, where litigation risks are high. Some scholars argue German contracts deliver similarly satisfactory legal outcomes with fewer words.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-78" href="#footnote-78" target="_self">78</a></p><p>That said, though transactional work has the adversarial elements described above, the overall structure is less adversarial than litigation. Because contract negotiations take place <em>before</em> a dispute occurs, there are more opportunities for transactional attorneys to add value beyond securing more of a fixed set of resources for their clients in a zero-sum negotiation.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-79" href="#footnote-79" target="_self">79</a></p><h2><strong>Human Oversight</strong></h2><p>A third bottleneck is our desire for human involvement. This is most relevant when AI gains outpace increases in production, which could happen for a few reasons. Perhaps there is an upper limit on arms races for certain kinds of legal work. After all, some legal doctrines are only so complicated, and courts often impose strict page limits on filings. Or maybe AI is so advanced that the costs of all legal tasks fall basically to zero and increased production does not absorb productivity gains. In this scenario, the new bottleneck for litigation would be the time required for judges to resolve cases, and for transactions, it would be the time parties need to understand a contract&#8217;s terms.</p><p>Starting with litigation, by reducing the cost of filing an initial lawsuit, AI will likely result in more disputes ending up in court. Within each dispute, AI can then create the kind of arms race of outputs described above. As a very rough but conservative estimate, Yonathan Arbel predicts a two- to fivefold increase in the volume of litigation.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-80" href="#footnote-80" target="_self">80</a></p><p>Arbel outlines several ways that judges might respond to a flood of litigation.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-81" href="#footnote-81" target="_self">81</a><sup> </sup>They could limit the flow of litigation by altering procedural and substantive doctrines to make it harder for litigants to get into court (creating a bottleneck related to regulatory barriers). Or they could try to limit the use of AI in the courtroom, perhaps by requiring lawyers to disclose AI use or banning AI entirely and sanctioning any violators. Both responses would counteract a flood of litigation work but come at the steep cost of sacrificing access to justice for poorer litigants.</p><p>The debt collection context also provides an uninspiring picture of how courts have managed a flood of cases.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-82" href="#footnote-82" target="_self">82</a><sup> </sup>As technology has enabled collectors to buy outstanding debt and cheaply file lawsuits for enforcement, the explosion of debt collection lawsuits has overwhelmed state courts.</p><p>Some have resorted to delegating cases to court assistants. Others operate &#8220;judgeless courtrooms&#8221; with lax evidentiary standards that can lower the quality of adjudication and can undermine the rationale for the judicial process itself.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-83" href="#footnote-83" target="_self">83</a><sup> </sup>If judges do not delegate, adjudicating cases will take longer as both the number of cases and the work each requires expands. Yet as the common legal maxim says: &#8220;Justice delayed is justice denied.&#8221; Although more people might gain access to court, if judges require years to adjudicate cases, plaintiffs will face a choice between protracted litigation with no guarantee of success and settling cases on increasingly unfavorable terms as resolution times lengthen.</p><p>Another option is incorporating AI into the judicial process to ease the strain on overwhelmed courts.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-84" href="#footnote-84" target="_self">84</a><sup> </sup>One early report suggests that AI is helping Brazil&#8217;s courts resolve cases more quickly.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-85" href="#footnote-85" target="_self">85</a> Yet if AI advances continuously reduce filing costs and drive up legal outputs, it will grow increasingly difficult for judges to keep pace. Perhaps AI will make judges more efficient. But there is a limit to how much AI can accelerate the process without meaningfully sacrificing human involvement.</p><p>Some seem open to replacing human judges entirely with AI.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-86" href="#footnote-86" target="_self">86</a> We find the legal (Article III, which establishes the federal courts, likely requires human judges),<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-87" href="#footnote-87" target="_self">87</a> technical (hallucination and private influence problems),<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-88" href="#footnote-88" target="_self">88</a><a href="https://justincurl.substack.com/p/16e98caa-1854-4db5-9a75-266e9a1c3a5b#footnote-88-187445681"><sup>88</sup></a> and moral objections persuasive.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-89" href="#footnote-89" target="_self">89</a> Even avowedly pro-AI lawyer Adam Unikowsky acknowledges he is &#8220;not quite ready to be ruled by robots.&#8221;<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-90" href="#footnote-90" target="_self">90</a> This is not to say that there is no role for AI in judging. But judges should adopt AI through careful, deliberate choices instead of in ways compelled by the need to keep up with an arms race of AI-powered legal work.</p><p>The argument for contracts is similar. If advanced AI reduces the cost of drafting contracts (perhaps it can instantly draft 50 perfect provisions), a contracting party, even with the help of AI, will still need time to understand what those provisions do and how they impact the party&#8217;s future interests.</p><p>This bottleneck would not apply if people were to forgo oversight. Arguably, human involvement matters less for contracts because many Americans already agree to contracts (like privacy policies) without reading them.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-91" href="#footnote-91" target="_self">91</a> But we think this reflects the belief that they have insufficient bargaining power to negotiate new terms, or that it&#8217;s not worthwhile to do so, rather than a general endorsement of signing contracts they don&#8217;t understand.</p><p>All this to say, we believe some measure of human involvement in the legal system is valuable and necessary, though this is a normative position and not an empirical claim.</p><h1><strong>Institutional Reforms</strong></h1><p>Many problems facing the legal industry are not new. For decades, legal academics and practitioners have suggested reforms targeting the bottlenecks described above. Some address AI specifically, others are more general, and a few are already being tested in various states and jurisdictions. In this section, we organize proposals into three categories, each corresponding to a bottleneck above: the regulation of legal services, the adjudication process, and the evolving role of human beings in legal work. While we aren&#8217;t tied to any particular recommendation, we discuss a range of reforms to illustrate that the future of the practice of law could look very different depending on how institutions respond (or decline to respond) to AI.</p><h2><strong>Reforming Professional Regulation</strong></h2><p>A chorus of voices have suggested reforms to the legal profession&#8217;s self-regulations.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-92" href="#footnote-92" target="_self">92</a> Their recommendations range from clarifying existing UPL laws to modifying law firm ownership rules to overhauling how the profession itself is regulated.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-93" href="#footnote-93" target="_self">93</a> Some are already being tested.</p><h3><strong>Clarifying Unauthorized Practice of Law Rules</strong></h3><p>Current UPL laws define practice of law imprecisely and may prohibit companies from offering AI-powered legal assistance to consumers. Some jurisdictions are expanding who may provide legal services, and academics have proposed updates with AI in mind.</p><p>Creating a new tier of legal service providers is one of the &#8220;fastest growing UPL reform program types&#8221; nationally, with seven states adopting this approach and another ten considering it.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-94" href="#footnote-94" target="_self">94</a> David Autor argues these reforms would benefit all professions by allowing people, in combination with AI, to work at levels of expertise previously unavailable to them. He analogizes to the creation of the nurse practitioner role.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-95" href="#footnote-95" target="_self">95</a> In the early 1960s, nurses and doctors developed training programs and successfully lobbied the American Medical Association to create a new class of medical professionals who could perform tasks previously reserved for doctors.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-96" href="#footnote-96" target="_self">96</a> Other researchers offer a more concrete framework for how state courts might design this tier of legal service provider.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-97" href="#footnote-97" target="_self">97</a></p><p>Joseph Avery and co-authors propose more ambitious reforms: allowing nonlawyers, including AI systems, to offer many legal services.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-98" href="#footnote-98" target="_self">98</a> Bar associations would retain authority over who may use the designation of &#8220;lawyer,&#8221; but nonlawyers could provide any legal service other than representing clients in court.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-99" href="#footnote-99" target="_self">99</a><sup> </sup>This would allow companies and nonprofits to offer AI-enabled services to consumers without claiming they are licensed attorneys and without the threat of UPL litigation. The prospect of being sued for negligent work would still serve as a quality backstop for lawyers and nonlawyers alike.</p><p>Sean Steward, by contrast, takes no position on where to draw the line between acceptable and unacceptable uses of AI, instead emphasizing the need for clear, nationwide rules to reduce burdens on providers.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-100" href="#footnote-100" target="_self">100</a> Drew Simshaw likewise advocates a nationwide approach to eliminate the patchwork of vague, conflicting rules.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-101" href="#footnote-101" target="_self">101</a></p><p>To be clear, the problem with existing UPL rules is not that they are regulations and therefore stifle innovation. It is that their uncertainty and variation discourage competition from new entrants, including the kind that produces better legal services.</p><h3><strong>Alternative Business Structures</strong></h3><p>Legal scholars have long argued for updating &#8220;entity regulations&#8221; that prevent nonlawyers from sharing fees from or investing in law firms.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-102" href="#footnote-102" target="_self">102</a> Utah and Arizona recently created regulatory sandboxes to do exactly that.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-103" href="#footnote-103" target="_self">103</a> Utah&#8217;s sandbox allows entities to seek waivers from ownership restrictions and UPL rules, while Arizona eliminated restrictions on law firm ownership and fee-sharing. These sandboxes permit companies and nonprofits to operate under modified professional rules while regulators assess their impact on service quality, cost, and access to justice.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-104" href="#footnote-104" target="_self">104</a></p><p>These sandboxes treat regulatory experimentation as necessary for balancing protection and innovation for consumers. Overly stringent restrictions can backfire by protecting inefficient incumbents or forcing new entrants outside the law. Uber and Airbnb succeeded, in part, by accepting regulatory fines, scaling quickly, and becoming so ubiquitous that lawmakers had little choice but to legalize their conduct. Yet overly lax restrictions can undermine the regulations&#8217; purpose: protecting consumers from poor quality services. Sandboxes allow policymakers to experiment and evaluate different regulatory approaches.</p><p>Early evidence on the impact of these reforms has been largely positive, though concerns have emerged regarding private equity ownership and mass tort litigation financing. Despite scant evidence of consumer harm&#8212;Utah&#8217;s Office of Legal Services Innovation received only twenty total complaints &#8212;lawyers and commerce groups petitioned the Arizona and Utah supreme courts to limit these sandboxes. The Arizona Supreme Court stayed the course, and authorized entities grew from nineteen to 136 between 2022 and 2025.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-105" href="#footnote-105" target="_self">105</a> The Utah Supreme Court has since raised eligibility requirements, and authorized entities shrank from thirty-nine to eleven over the same period.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-106" href="#footnote-106" target="_self">106</a></p><h3><strong>Regulatory Markets</strong></h3><p>Gillian Hadfield proposes a &#8220;superregulator&#8221; model that would create a market for the regulation of legal services.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-107" href="#footnote-107" target="_self">107</a> Rather than regulating providers directly, the government would license regulators that would each offer competing regulatory schemes. The government&#8217;s role shifts to &#8220;regulating the regulators&#8221; by setting outcome targets, such as acceptable levels of legal access or dispute resolution quality, and then licensing regulators that achieve them.</p><p>Hadfield argues this generates powerful incentives for innovation.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-108" href="#footnote-108" target="_self">108</a> A private regulator that develops simpler, more cost-effective compliance methods while meeting government standards will attract more customers. The model can also simplify enforcement: Governments can monitor ten licensed regulators more easily than thousands of individual providers.</p><p>We would add that regulatory markets may be able to assess AI&#8217;s utility for legal services more reliably than benchmarking.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-109" href="#footnote-109" target="_self">109</a><sup> </sup>Two of us have emphasized that task-oriented benchmarks lack construct validity because they &#8220;overemphasize precisely the thing that language models are good at&#8221; while failing to test the contextual understanding and sustained reasoning that characterizes consequential legal work.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-110" href="#footnote-110" target="_self">110</a> Benchmarks can also miss hidden costs that emerge only over time, such as deskilling of professionals.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-111" href="#footnote-111" target="_self">111</a> Relatedly, by targeting entry-level tasks, AI can disrupt the pipeline through which junior lawyers develop expertise.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-112" href="#footnote-112" target="_self">112</a></p><p>Hadfield defends the proposal&#8217;s practicality by drawing parallels to existing models. Governments already use outcomes-based regulations in environmental law, and private standard-setting bodies design many regulations currently in use. The United Kingdom provides one example. Under the Legal Services Act 2007, Parliament created the Legal Services Board (LSB), an independent agency that approves private bodies applying to regulate legal services.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-113" href="#footnote-113" target="_self">113</a> The system is not yet fully competitive because regulators came from preexisting trade associations for barristers and solicitors, which together regulate 90 percent of legal professionals in England and Wales.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-114" href="#footnote-114" target="_self">114</a> But competition is emerging for &#8220;alternative business structures,&#8221; which can choose between licensing from the Solicitors Regulation Authority (SRA) or the Bar Standards Board (BSB).<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-115" href="#footnote-115" target="_self">115</a></p><p>This U.K. model has, however, encountered difficulties. In October 2024, the LSB criticized the SRA for failing to &#8220;act adequately, effectively, and efficiently&#8221; before the law firm Axiom Ince collapsed in October 2023.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-116" href="#footnote-116" target="_self">116</a><sup> </sup>The LSB issued a report in March 2025 expressing &#8220;serious concerns&#8221; about the SRA&#8217;s effectiveness and then recently proposed sanctions.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-117" href="#footnote-117" target="_self">117</a> This highlights how superregulators can struggle to enforce quality standards for the regulators it oversees. Some scholars have separately criticized regulatory markets in other contexts for creating a race to the bottom. For example, Daniel Schwarcz cautions that a market for insurance regulation could &#8220;trigger a &#8216;race to the bottom&#8217; as regulators compete with each other to offer less and less intrusive regulatory schemes.&#8221;<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-118" href="#footnote-118" target="_self">118</a></p><h2><strong>Reforming Adjudication</strong></h2><p>Another set of reforms targets case adjudication. Some aim to make the trials less adversarial, while others advocate for private adjudication like arbitration.</p><h3><strong>Judicial Case Management</strong></h3><p>Judges have some discretion over the litigation process and can exercise it to reduce adversarial dynamics. Several judges have recommended leveraging existing rules of evidence and civil procedure to manage cases more actively, taking inspiration from other jurisdictions (often called inquisitorial systems).<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-119" href="#footnote-119" target="_self">119</a> Such targeted borrowing can help reduce competitive escalation.</p><p>One example is allowing courts to appoint their own expert witnesses. Under Federal Rule of Evidence 706, judges can appoint neutral experts that work for the court but are paid for by both parties.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-120" href="#footnote-120" target="_self">120</a> One state trial judge argues this can solve the &#8220;battle of experts&#8221; problem where competing specialists &#8220;abandon objectivity and become advocates for the side that hired them.&#8221;<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-121" href="#footnote-121" target="_self">121</a><sup> </sup>When technical issues are central to a case, court-appointed experts can provide neutral assessments that frame issues more productively, avoiding an arms race of dueling expert reports.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-122" href="#footnote-122" target="_self">122</a></p><p>Another tool is the use of &#8220;special masters&#8221; under Federal Rule of Civil Procedure 53.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-123" href="#footnote-123" target="_self">123</a> These are neutral third parties appointed to help manage complex aspects of cases. A federal judge and senior litigator explain that special masters can &#8220;assist and, when necessary, direct the parties&#8221; to complete discovery efficiently.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-124" href="#footnote-124" target="_self">124</a><sup> </sup>They note that the 2003 amendments expanded the scope of special masters&#8217; use to include pretrial matters &#8220;that cannot be effectively and timely addressed by an available district judge.&#8221;<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-125" href="#footnote-125" target="_self">125</a><sup> </sup>Rather than having parties fight over AI-assisted document review through successive motions, a special master could serve as an intermediary and prevent the technology from enabling larger discovery battles.</p><p>Judges currently have the discretion to intervene under these rules, but Congress could also pass legislation that makes them mandatory.</p><h3><strong>Arbitration</strong></h3><p>While disputes are normally resolved through litigation, some contracts specify private arbitration.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-126" href="#footnote-126" target="_self">126</a> Contract drafters might prefer this process for several reasons: It can resolve disputes at lower costs,<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-127" href="#footnote-127" target="_self">127</a> reduce class-action exposure,<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-128" href="#footnote-128" target="_self">128</a> and prove &#8220;more flexible and less adversarial &#8230; than its judicial counterpart.&#8221;<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-129" href="#footnote-129" target="_self">129</a> Arbitration has become a popular alternative to traditional litigation.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-130" href="#footnote-130" target="_self">130</a></p><p>Two legal scholars argue that because arbitration grants parties autonomy over the process, it&#8217;s the &#8220;ideal entry point for broader AI adoption in the legal field.&#8221;<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-131" href="#footnote-131" target="_self">131</a> They defend AI arbitration as consistent with the Federal Arbitration Act (FAA) and desirable for enhancing &#8220;efficiency, fairness, and flexibility of dispute resolution.&#8221;<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-132" href="#footnote-132" target="_self">132</a><sup> </sup>Another scholar takes the opposite stance, arguing that the FAA does not permit AI arbitration because robot adjudicators are inconsistent with the statute&#8217;s use of human pronouns like &#8220;he or they.&#8221;<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-133" href="#footnote-133" target="_self">133</a> And still another views it as undesirable because it would &#8220;significantly diminish the long-standing reputation&#8221; of arbitration.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-134" href="#footnote-134" target="_self">134</a></p><p>Some observers critique arbitration as unfair because companies often force it on consumers and employees through &#8220;take-it-or-leave-it&#8221; contracts of adhesion.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-135" href="#footnote-135" target="_self">135</a> These fairness concerns are important, and others have written about the level of consent needed for arbitration to be truly fair.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-136" href="#footnote-136" target="_self">136</a> But assuming the decision to enter arbitration reflects the free choice of both parties, offering AI arbitration as a parallel track for resolving cases can have several advantages.</p><p>First, it can promote choice by allowing litigants to decide between traditional judicial review and AI-assisted adjudication.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-137" href="#footnote-137" target="_self">137</a> A consumer defending against an automated debt collection suit might prefer quick AI resolution over years of waiting, while a defendant facing serious consequences might insist on traditional review by human judges.</p><p>Second, should an arms race of legal outputs risk overwhelming the courts, the availability of a technology-mediated alternative can alleviate pressure on the courts, preserving human review for the contexts where those navigating the judicial system feel they need it most.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-138" href="#footnote-138" target="_self">138</a></p><p>Third, it creates a natural experiment that facilitates comparison between human and AI adjudicators on dimensions like speed, cost, and participant satisfaction.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-139" href="#footnote-139" target="_self">139</a> This generates evidence about AI&#8217;s actual performance, reducing reliance on speculation or vendor claims, and it pressures traditional institutions to improve or risk being outcompeted.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-140" href="#footnote-140" target="_self">140</a></p><h2><strong>The Evolving Role of Lawyers</strong></h2><p>A final set of reforms discusses the evolving role of lawyers in both litigation and transactional contexts. With litigation, legislatures could expand the judiciary to alleviate the bottleneck created by the time human judges take to resolve cases. With transactions, there isn&#8217;t a clear action item for companies, but we expect to see a shift in what in-house counsel do. As the bottleneck becomes the time it takes human lawyers to understand complex contracts, in-house lawyers will likely spend more time understanding a company&#8217;s needs and making strategic judgments and less on legal tasks.</p><h3><strong>Expanding the Judiciary</strong></h3><p>The most straightforward response to an overburdened judiciary is to increase its capacity by hiring more judges. Legal scholars have advocated for this solution for decades.</p><p>For example, writing in 1979, Maria Marcus argued that &#8220;[s]ince the factors that channel disputes into a judicial forum continue unabated, the appointment of more judges is an obvious response.&#8221;<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-141" href="#footnote-141" target="_self">141</a> Bert Huang again recommended in 2011 that &#8220;new demands put on the courts should be met quickly and flexibly with new judicial resources.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-142" href="#footnote-142" target="_self">142</a> More recently, Peter Menell and Ryan Vacca endorsed an observation from decades earlier that &#8220;the increase in the order of magnitude of the demands our society imposes on the federal judicial system&#8221; should encourage Congress to act.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-143" href="#footnote-143" target="_self">143</a></p><p>Menell and Vacca acknowledge that &#8220;[i]ncreasing the number of federal judgeships has been fraught with political complications.&#8221; Their solution is a bipartisan &#8220;2030 Commission&#8221; to depoliticize the process.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-144" href="#footnote-144" target="_self">144</a> Arbel similarly views adding judges as &#8220;the most direct way of solving the problem&#8221; of an AI-driven increase in litigation work, yet stops short of recommending it because it &#8220;appears quite tenuous in our current political reality.&#8221;<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-145" href="#footnote-145" target="_self">145</a> He notes that redirecting all civil legal aid funding (approximately $2.7 billion) toward the $9.4 billion federal court system would yield at most a 30 percent increase in judicial capacity, falling short of the doubling likely needed to handle the increased caseload.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-146" href="#footnote-146" target="_self">146</a></p><p>But civil legal aid is not the only potential funding source. Some states have proposed taxing legal services generally,<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-147" href="#footnote-147" target="_self">147</a> while federal legislation introduced by Sen. Thom Tillis (R-N.C.) would tax third-party litigation financiers who fund plaintiffs&#8217; legal fees in exchange for a percentage of eventual winnings.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-148" href="#footnote-148" target="_self">148</a><sup> </sup>Although these bills aim to increase overall tax revenues, similar measures could earmark funds specifically for the judiciary. Deborah Rhode has proposed another approach: mandatory pro bono service requirements for all attorneys, with the option to &#8220;buy out&#8221; their obligation.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-149" href="#footnote-149" target="_self">149</a> Though initially conceived as a way to provide access to justice, these payments could also support judicial expansion.</p><p>So while we agree that expanding the judiciary faces real political obstacles, we don&#8217;t think it&#8217;s as unrealistic as Arbel fears, especially considering the magnitude of AI&#8217;s potential disruption.</p><h3><strong>In-House Counsel as Strategic Advisors</strong></h3><p>Our analysis suggests that among in-house lawyers, value will likely shift from completing tasks to predicting how contracts and agreements impact the overall business. Lawyers will need to deeply understand their organizations and exercise business judgment.</p><p>Industry reports agree.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-150" href="#footnote-150" target="_self">150</a> In a 2023 survey of nearly three hundred chief legal officers, 87 percent said their role is shifting from legal risk mitigator to strategic business partner.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-151" href="#footnote-151" target="_self">151</a> Thomson Reuters&#8217;s 2024 &#8220;Future of Professionals Report&#8221; found that 42 percent of legal professionals expect to spend more time on judgment-based legal work in the next five years, as AI handles more routine tasks.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-152" href="#footnote-152" target="_self">152</a> As one attorney respondent put it: &#8220;The role of a good lawyer is as a &#8216;trusted advisor,&#8217; not as a producer of documents &#8230; breadth of experience is where a lawyer&#8217;s true value lies and that will remain valuable.&#8221;<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-153" href="#footnote-153" target="_self">153</a></p><p>To understand the shifting role of legal professionals, it is helpful to consider the hierarchy of legal work and roles. At the bottom is basic &#8220;low-skilled&#8221; legal work like drafting standard letters or simple contracts, repetitive tasks requiring minimal legal expertise.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-154" href="#footnote-154" target="_self">154</a> Above that is medium-skill, noncommoditized legal work that involves producing documents, such as analyzing contracts and drafting motions.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-155" href="#footnote-155" target="_self">155</a><sup> </sup>Higher still is &#8220;judgment-based legal work&#8221;: overseeing complex trials and addressing legal risks.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-156" href="#footnote-156" target="_self">156</a> At the top is strategic advising, where lawyers deeply understand an organization&#8217;s priorities and shape its decisions.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-157" href="#footnote-157" target="_self">157</a> This highest level might not involve what we traditionally consider legal work at all.</p><p>As AI pushes the role of human expertise up this hierarchy, the legal profession should rethink how it trains lawyers. While AI automates the tasks at the bottom of the hierarchy, demand will likely grow for lawyers at the top who can translate legal information into strategic advice. One bar association warned that the &#8220;greater concern is that generative AI will displace younger attorneys,&#8221; who will &#8220;have fewer opportunities to gain valuable experience by spending hours on important tasks.&#8221;<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-158" href="#footnote-158" target="_self">158</a> As the skills required to succeed change, so too should the training process.</p><h1><strong>Conclusion</strong></h1><p>Many problems facing the legal industry do not require revolutionary insights to solve. Scholars and practitioners have long emphasized the need for regulatory reform, changes to the litigation process, and the disconnect between legal work and client outcomes. AI, rather than solving these problems, appears to be revealing and magnifying them. Without addressing these underlying issues, AI alone is unlikely to improve the outcomes clients care about.</p><p>But AI may also present new opportunities for reform. Sociologists have argued that fields facing crises are more receptive to efforts to reshape institutions in those areas.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-159" href="#footnote-159" target="_self">159</a> Such crises, jolts, shocks, and disruptive events&#8212;taking the form of social upheaval, technological disruption, or other changes&#8212;can reveal problems and contradictions that require solutions. The widespread predictions that AI will transform the practice of law may constitute such a crisis, creating pressure for legal institutions to respond. The key question is whether they can use this opportunity to enact the reforms the industry has needed for decades and produce better outcomes for clients.</p><p><em>Acknowledgments:</em> We are grateful to Samar Ahmad, James Bedford, Jasper Boers, Lev Cohen, Joe Goode, Mihir Kshirsagar, Martha Minow, Ben Press, and Jonathan Zittrain for helpful feedback on this report.</p><p></p><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>Jan Philipp Burgard, &#8220;Sam Altman Predicts AI Will Surpass Human Intelligence by 2030,&#8221; <em>Business Insider</em> (Sept. 26, 2025).</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p>Dario Amodei, &#8220;Machines of Loving Grace: How AI Could Transform the World for the Better&#8221; (October 2024), https://perma.cc/8CPH-TJ63.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-3" href="#footnote-anchor-3" class="footnote-number" contenteditable="false" target="_self">3</a><div class="footnote-content"><p>See, e.g., Edward W. Felten, Manav Raj, &amp; Robert Seamans, &#8220;Occupational Heterogeneity in Exposure to Generative AI&#8221; (April 10, 2023) (unpublished manuscript), https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4414065.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-4" href="#footnote-anchor-4" class="footnote-number" contenteditable="false" target="_self">4</a><div class="footnote-content"><p>See Daniel Martin Katz, Michael James Bommarito, Shang Gao, &amp; Pablo Arredondo, &#8220;GPT-4 Passes the Bar Exam,&#8221; 382 <em>Philosophical Transactions of the Royal Society</em> (2024), https://doi.org/10.1098/rsta.2023.0254.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-5" href="#footnote-anchor-5" class="footnote-number" contenteditable="false" target="_self">5</a><div class="footnote-content"><p>Karen Sloan &amp; Sara Merken, &#8220;AI Training Becomes Mandatory at More US Law Schools,&#8221; <em>Reuters</em> (Sept. 22, 2025); University of San Francisco, &#8220;The University of San Francisco School of Law Embeds GenAI Into Core Curriculum&#8221; (April 16, 2025), https://perma.cc/2RSV-ZZGA.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-6" href="#footnote-anchor-6" class="footnote-number" contenteditable="false" target="_self">6</a><div class="footnote-content"><p>Harvey, &#8220;Harvey Launches Academic Program With Law Schools at Stanford, NYU, Michigan, UCLA, The University of Texas, and Notre Dame&#8221; (Aug. 28, 2025), https://perma.cc/77Q3-C5P6.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-7" href="#footnote-anchor-7" class="footnote-number" contenteditable="false" target="_self">7</a><div class="footnote-content"><p>Adam Unikowsky, &#8220;Should AI Replace Law Clerks?,&#8221; <em>Adam&#8217;s Legal Newsletter</em> (Jan. 20, 2023), https://perma.cc/A7WN-4T9A; Adam Unikowsky, &#8220;Automating Oral Argument,&#8221; <em>Adam&#8217;s Legal Newsletter</em> (July 7, 2025), https://perma.cc/6FGM-27XS.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-8" href="#footnote-anchor-8" class="footnote-number" contenteditable="false" target="_self">8</a><div class="footnote-content"><p>Richard Susskind, &#8220;Artificial Intelligence Could Replace Traditional Lawyers by 2035,&#8221; <em>Times (London)</em> (March 2025).</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-9" href="#footnote-anchor-9" class="footnote-number" contenteditable="false" target="_self">9</a><div class="footnote-content"><p>See David Freeman Engstrom, Lucy Ricca, &amp; Natalie Knowlton, &#8220;Regulatory Innovation at the Crossroads: Five Years of Data on Entity-Regulation Reform in Arizona and Utah,&#8221; <em>Stanford Law School</em> (June 2, 2025), https://perma.cc/QL5D-8EW6.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-10" href="#footnote-anchor-10" class="footnote-number" contenteditable="false" target="_self">10</a><div class="footnote-content"><p>Press Release, LexisNexis, &#8220;LexisNexis CounselLink Releases 2025 Trends Report Showing Large Law Command of Partner Rates&#8221; (April 22, 2025), https://perma.cc/XM8C-VTZJ.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-11" href="#footnote-anchor-11" class="footnote-number" contenteditable="false" target="_self">11</a><div class="footnote-content"><p>E.g., &#8220;How to Harness AI for Justice,&#8221; 108 <em>Judicature</em> 42 (2024); OECD, &#8220;Governing With Artificial Intelligence: The State of Play and Way Forward in Core Government Functions&#8221; (Sept. 18, 2025), https://doi.org/10.1787/795de142-en; Shana Lynch, &#8220;Harnessing AI to Improve Access to Justice in Civil Courts,&#8221; <em>Stanford HAI</em> (March 4, 2025), https://perma.cc/6PGP-UG2J; Robert J. Couture, &#8220;The Impact of Artificial Intelligence on Law Firms&#8217; Business Models,&#8221; <em>Harvard Law School Center on</em> <em>the Legal Profession</em> (Feb. 24, 2025), https://perma.cc/HYP6-SJQB.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-12" href="#footnote-anchor-12" class="footnote-number" contenteditable="false" target="_self">12</a><div class="footnote-content"><p>Our understanding of bottlenecks is informed by systems analysis and theories of constraints in management science.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-13" href="#footnote-anchor-13" class="footnote-number" contenteditable="false" target="_self">13</a><div class="footnote-content"><p>See Ed Walters, &#8220;Re-Regulating UPL in an Age of AI,&#8221; 8 <em>Georgetown Law Technology Review</em> 316 (2024).</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-14" href="#footnote-anchor-14" class="footnote-number" contenteditable="false" target="_self">14</a><div class="footnote-content"><p>Brian Oten, &#8220;Artificial Intelligence, Real Practice,&#8221; 28 <em>North Carolina State Bar Journal</em> 3 (fall 2023).</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-15" href="#footnote-anchor-15" class="footnote-number" contenteditable="false" target="_self">15</a><div class="footnote-content"><p>This report focuses on the U.S. legal system, but for more on how U.S. adversarialism differs from other legal systems, Leo You Li has compared the U.S. system with China and the U.K. and discussed what they might learn from each other. See, generally, Leo You Li, &#8220;Digitization, Adversarial Legalism, and Access to Justice Reforms,&#8221; 76 <em>South Carolina Law Review</em> 883 (2025).</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-16" href="#footnote-anchor-16" class="footnote-number" contenteditable="false" target="_self">16</a><div class="footnote-content"><p>Robert A. Kagan, Adversarial Legalism: The American Way of Law (2d ed., 2018).</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-17" href="#footnote-anchor-17" class="footnote-number" contenteditable="false" target="_self">17</a><div class="footnote-content"><p>Gillian K. Hadfield, &#8220;The Price of Law: How the Market for Lawyers Distorts the Justice System,&#8221; 98 <em>Michigan Law Review</em> 953 (2000), https://dx.doi.org/10.2139/ssrn.191908.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-18" href="#footnote-anchor-18" class="footnote-number" contenteditable="false" target="_self">18</a><div class="footnote-content"><p>David Freeman Engstrom &amp; Jonah B. Gelbach, &#8220;Legal Tech, Civil Procedure, and the Future of Adversarialism,&#8221; 169 <em>University of Pennsylvania Law Review</em> 1001 (2021).</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-19" href="#footnote-anchor-19" class="footnote-number" contenteditable="false" target="_self">19</a><div class="footnote-content"><p>David Freeman Engstrom &amp; Nora Freeman Engstrom, &#8220;Legal Tech and the Litigation Playing Field,&#8221; in <em>Legal Tech and the Future of Civil Justice</em> 133 (David F. Engstrom, ed., 2023), https://doi.org/10.1017/9781009255301.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-20" href="#footnote-anchor-20" class="footnote-number" contenteditable="false" target="_self">20</a><div class="footnote-content"><p>Hadfield, &#8220;The Price of Law,&#8221; supra note 17.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-21" href="#footnote-anchor-21" class="footnote-number" contenteditable="false" target="_self">21</a><div class="footnote-content"><p>Li, supra note 15, at 892&#8211;93.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-22" href="#footnote-anchor-22" class="footnote-number" contenteditable="false" target="_self">22</a><div class="footnote-content"><p>Arvind Narayanan &amp; Sayash Kapoor, &#8220;AI as Normal Technology,&#8221; <em>AI as Normal Technology</em> (April 15, 2025), https://perma.cc/AQ2G-CQP3.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-23" href="#footnote-anchor-23" class="footnote-number" contenteditable="false" target="_self">23</a><div class="footnote-content"><p>Justin Curl, &#8220;AI Is Just Starting to Change the Legal Profession,&#8221; <em>Understanding AI</em> (Jan. 15, 2026), https://perma.cc/N3BQ-5MKV.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-24" href="#footnote-anchor-24" class="footnote-number" contenteditable="false" target="_self">24</a><div class="footnote-content"><p>Gillian Hadfield has written about the high costs of legal services for decades, and this section draws heavily on her work. E.g., Hadfield, &#8220;The Price of Law,&#8221; supra note 17; Gillian K. Hadfield, &#8220;The Cost of Law: Promoting Access to Justice Through the (Un)Corporate Practice of Law,&#8221; 38 <em>International Review of Law and Economics</em> 43 (2014), https://doi.org/10.1016/j.irle.2013.09.003; Gillian K. Hadfield &amp; Deborah L. Rhode, &#8220;How to Regulate Legal Services to Promote Access, Innovation, and the Quality of Lawyering,&#8221; 67 <em>Hastings Law Journal</em> 1191 (2016); Gillian K. Hadfield, &#8220;More Markets, More Justice,&#8221; 148 <em>D&#230;dalus</em> 37 (winter 2019), https://doi.org/10.1162/daed_a_00533; Gillian K. Hadfield, &#8220;Legal Markets,&#8221; 60 <em>Journal of Economic Literature</em> 1264 (2022), https://doi.org/10.1257/jel.20201330.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-25" href="#footnote-anchor-25" class="footnote-number" contenteditable="false" target="_self">25</a><div class="footnote-content"><p>Hadfield, &#8220;The Price of Law,&#8221; supra note 17, at 969.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-26" href="#footnote-anchor-26" class="footnote-number" contenteditable="false" target="_self">26</a><div class="footnote-content"><p>Id. at 972&#8211;73.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-27" href="#footnote-anchor-27" class="footnote-number" contenteditable="false" target="_self">27</a><div class="footnote-content"><p>Id.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-28" href="#footnote-anchor-28" class="footnote-number" contenteditable="false" target="_self">28</a><div class="footnote-content"><p>Engstrom &amp; Gelbach, supra note 18, at 1062.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-29" href="#footnote-anchor-29" class="footnote-number" contenteditable="false" target="_self">29</a><div class="footnote-content"><p>Hadfield, &#8220;The Price of Law,&#8221; supra note 17, at 973.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-30" href="#footnote-anchor-30" class="footnote-number" contenteditable="false" target="_self">30</a><div class="footnote-content"><p>John Armour &amp; Mari Sako, &#8220;Lawtech: Leveling the Playing Field in Legal Services?&#8221; in <em>Legal Tech and the Future of Civil Justice</em> 44, 44 (David F. Engstrom, ed., 2023), https://dx.doi.org/10.2139/ssrn.3831481.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-31" href="#footnote-anchor-31" class="footnote-number" contenteditable="false" target="_self">31</a><div class="footnote-content"><p><em>LexisNexis CounselLink</em>, &#8220;2025 Trends Report&#8221; (2025), https://perma.cc/Q6K3-9SCW.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-32" href="#footnote-anchor-32" class="footnote-number" contenteditable="false" target="_self">32</a><div class="footnote-content"><p>Id.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-33" href="#footnote-anchor-33" class="footnote-number" contenteditable="false" target="_self">33</a><div class="footnote-content"><p>Hadfield, &#8220;Legal Markets,&#8221; supra note 24, at 1288.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-34" href="#footnote-anchor-34" class="footnote-number" contenteditable="false" target="_self">34</a><div class="footnote-content"><p>Id.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-35" href="#footnote-anchor-35" class="footnote-number" contenteditable="false" target="_self">35</a><div class="footnote-content"><p>Engstrom, Ricca, &amp; Knowlton, supra note 9, at 5.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-36" href="#footnote-anchor-36" class="footnote-number" contenteditable="false" target="_self">36</a><div class="footnote-content"><p>Id.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-37" href="#footnote-anchor-37" class="footnote-number" contenteditable="false" target="_self">37</a><div class="footnote-content"><p>Hadfield, &#8220;More Markets, More Justice,&#8221; supra note 24.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-38" href="#footnote-anchor-38" class="footnote-number" contenteditable="false" target="_self">38</a><div class="footnote-content"><p>Armour &amp; Sako, supra note 30, at 44.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-39" href="#footnote-anchor-39" class="footnote-number" contenteditable="false" target="_self">39</a><div class="footnote-content"><p>Hadfield, &#8220;More Markets, More Justice,&#8221; supra note 24, citing Clio, &#8220;2017 Legal Trends Report,&#8221; https://perma.cc/YS73-47QL.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-40" href="#footnote-anchor-40" class="footnote-number" contenteditable="false" target="_self">40</a><div class="footnote-content"><p>Id.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-41" href="#footnote-anchor-41" class="footnote-number" contenteditable="false" target="_self">41</a><div class="footnote-content"><p>Pew Charitable Trusts, &#8220;How Debt Collectors Are Transforming the Business of State Courts,&#8221; at 8 (May 6, 2020), https://perma.cc/999L-LQKA.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-42" href="#footnote-anchor-42" class="footnote-number" contenteditable="false" target="_self">42</a><div class="footnote-content"><p>Michigan Justice for All Commission, Debt Collection Work Group, &#8220;Advancing Justice for All in Debt Collection Lawsuits: Report &amp; Recommendations,&#8221; https://perma.cc/2SQT-M9SB (last visited Jan. 26, 2026).</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-43" href="#footnote-anchor-43" class="footnote-number" contenteditable="false" target="_self">43</a><div class="footnote-content"><p>Pew Charitable Trusts, supra note 41, at 2.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-44" href="#footnote-anchor-44" class="footnote-number" contenteditable="false" target="_self">44</a><div class="footnote-content"><p>Id.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-45" href="#footnote-anchor-45" class="footnote-number" contenteditable="false" target="_self">45</a><div class="footnote-content"><p>Institute for Justice, &#8220;Right to Provide Legal Advice,&#8221; https://perma.cc/Q2FN-8WEE (last visited Jan. 26, 2026).</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-46" href="#footnote-anchor-46" class="footnote-number" contenteditable="false" target="_self">46</a><div class="footnote-content"><p><em>Upsolve, Inc. v. James</em>, 604 F.Supp.3d 97 (S.D.N.Y. May 24, 2022).</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-47" href="#footnote-anchor-47" class="footnote-number" contenteditable="false" target="_self">47</a><div class="footnote-content"><p><em>Upsolve, Inc. v. James</em>, 155 F.4th 133 (2d Cir. Sept. 9, 2025).</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-48" href="#footnote-anchor-48" class="footnote-number" contenteditable="false" target="_self">48</a><div class="footnote-content"><p>Brian Oten, &#8220;Artificial Intelligence, Real Practice,&#8221; 28 <em>North Carolina State Bar Journal</em> 3 (fall 2023); see also Maria E. Berkenkotter &amp; Linos Lipinsky de Orlov, &#8220;Can Robot Lawyers Close the Access to Justice Gap?&#8221; <em>Colorado Lawyer</em> (December 2024), at 40, https://perma.cc/YY73-XJDE.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-49" href="#footnote-anchor-49" class="footnote-number" contenteditable="false" target="_self">49</a><div class="footnote-content"><p>New York State Bar Association Task Force on Artificial Intelligence, &#8220;Report and Recommendations to NYSBA House of Delegates&#8221; (April 6, 2024), https://perma.cc/6CXP-NLCA.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-50" href="#footnote-anchor-50" class="footnote-number" contenteditable="false" target="_self">50</a><div class="footnote-content"><p>Sean Steward, &#8220;Are AI Lawyers a Legal Product or Legal Service?: Why Current UPL Laws Are Not Up to the Task of Regulating Autonomous AI Actors,&#8221; 53 <em>Hofstra Law Review</em> 391 (2025).</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-51" href="#footnote-anchor-51" class="footnote-number" contenteditable="false" target="_self">51</a><div class="footnote-content"><p>Walters, supra note 13, at 332.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-52" href="#footnote-anchor-52" class="footnote-number" contenteditable="false" target="_self">52</a><div class="footnote-content"><p>Laurel A. Rigertas, &#8220;The Legal Profession&#8217;s Monopoly: Failing to Protect Consumers,&#8221; 82 <em>Georgetown Journal of Legal Ethics</em> 1085 (2019).</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-53" href="#footnote-anchor-53" class="footnote-number" contenteditable="false" target="_self">53</a><div class="footnote-content"><p>Id.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-54" href="#footnote-anchor-54" class="footnote-number" contenteditable="false" target="_self">54</a><div class="footnote-content"><p><em>Janson v. LegalZoom.com, Inc.,</em> 802 F. Supp. 2d 1053 (W.D. Mo. 2011).</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-55" href="#footnote-anchor-55" class="footnote-number" contenteditable="false" target="_self">55</a><div class="footnote-content"><p><em>LegalZoom.com, Inc. v. N.C. State Bar,</em> No. 11 CVS 15111, 2015 NCBC 96 (N.C. Super. Ct. Oct. 22, 2015), https://perma.cc/U3LS-3C83.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-56" href="#footnote-anchor-56" class="footnote-number" contenteditable="false" target="_self">56</a><div class="footnote-content"><p>Jason Tashea, &#8220;Nonlawyers at LegalZoom Performed Legal Work on Trademark Applications, UPL Suit Alleges,&#8221; <em>ABA Journal</em> (Dec. 20, 2017).</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-57" href="#footnote-anchor-57" class="footnote-number" contenteditable="false" target="_self">57</a><div class="footnote-content"><p>Class Action Complaint, <em>Erasmus v. LegalZoom.com, Inc.,</em> No. ESX-L (N.J. Super. Ct. Law Div. Essex Cnty.), https://perma.cc/CR2L-ZV89.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-58" href="#footnote-anchor-58" class="footnote-number" contenteditable="false" target="_self">58</a><div class="footnote-content"><p>Engstrom &amp; Engstrom, supra note 19.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-59" href="#footnote-anchor-59" class="footnote-number" contenteditable="false" target="_self">59</a><div class="footnote-content"><p>J. Maria Glover, &#8220;The Federal Rules of Civil Settlement,&#8221; 87 <em>New York University Law Review</em> 1713 (2012).</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-60" href="#footnote-anchor-60" class="footnote-number" contenteditable="false" target="_self">60</a><div class="footnote-content"><p>Kagan, supra note 16.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-61" href="#footnote-anchor-61" class="footnote-number" contenteditable="false" target="_self">61</a><div class="footnote-content"><p>Engstrom &amp; Gelbach, supra note 18, at 1043.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-62" href="#footnote-anchor-62" class="footnote-number" contenteditable="false" target="_self">62</a><div class="footnote-content"><p>John S. Beckerman, &#8220;Confronting Civil Discovery&#8217;s Fatal Flaws,&#8221; 84 <em>Minnesota Law Review</em> 505 (2000), https://dx.doi.org/10.2139/ssrn.199068.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-63" href="#footnote-anchor-63" class="footnote-number" contenteditable="false" target="_self">63</a><div class="footnote-content"><p>Frank H. Easterbrook, &#8220;Discovery as Abuse,&#8221; 69 <em>Boston University Law Review</em> 635 (1989).</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-64" href="#footnote-anchor-64" class="footnote-number" contenteditable="false" target="_self">64</a><div class="footnote-content"><p>Charles M. Yablon, &#8220;Stupid Lawyer Tricks: An Essay on Discovery Abuse,&#8221; 96 <em>Columbia Law Review</em> 1618 (1996).</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-65" href="#footnote-anchor-65" class="footnote-number" contenteditable="false" target="_self">65</a><div class="footnote-content"><p>Engstrom &amp; Gelbach, supra note 18.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-66" href="#footnote-anchor-66" class="footnote-number" contenteditable="false" target="_self">66</a><div class="footnote-content"><p>Alexandra Lahav emphasized the shortage of reliable evidence in an article calling for courts to log discovery requests, so researchers can better assess the extent of discovery abuse and costs. While she would argue discovery is not as big of a problem as lawyers claim, her main point is that there is insufficient evidence to be confident either way. And even she agrees that where discovery is actively employed, it is very expensive, even if she thinks these costs are justified by the higher dollar amounts at stake. See Alexandra D. Lahav, &#8220;A Proposal to End Discovery Abuse,&#8221; 71 <em>Vanderbilt Law Review</em> 2037 (2019).</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-67" href="#footnote-anchor-67" class="footnote-number" contenteditable="false" target="_self">67</a><div class="footnote-content"><p>Engstrom &amp; Engstrom, supra note 19, at 138.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-68" href="#footnote-anchor-68" class="footnote-number" contenteditable="false" target="_self">68</a><div class="footnote-content"><p>Hadfield, &#8220;Legal Markets,&#8221; supra note 24.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-69" href="#footnote-anchor-69" class="footnote-number" contenteditable="false" target="_self">69</a><div class="footnote-content"><p>Id.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-70" href="#footnote-anchor-70" class="footnote-number" contenteditable="false" target="_self">70</a><div class="footnote-content"><p>Engstrom &amp; Gelbach, supra note 18, at 1052&#8211;53.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-71" href="#footnote-anchor-71" class="footnote-number" contenteditable="false" target="_self">71</a><div class="footnote-content"><p>Hadfield, &#8220;The Price of Law,&#8221; supra note 17.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-72" href="#footnote-anchor-72" class="footnote-number" contenteditable="false" target="_self">72</a><div class="footnote-content"><p>Id.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-73" href="#footnote-anchor-73" class="footnote-number" contenteditable="false" target="_self">73</a><div class="footnote-content"><p>Id.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-74" href="#footnote-anchor-74" class="footnote-number" contenteditable="false" target="_self">74</a><div class="footnote-content"><p>Id.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-75" href="#footnote-anchor-75" class="footnote-number" contenteditable="false" target="_self">75</a><div class="footnote-content"><p>John C. Coates IV, &#8220;Why Have M&amp;A Contracts Grown? Evidence From Twenty Years of Deals&#8221; (Harvard Law School, Working Paper, Oct. 26, 2016), https://perma.cc/JQ8X-R6KV.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-76" href="#footnote-anchor-76" class="footnote-number" contenteditable="false" target="_self">76</a><div class="footnote-content"><p>Isabel Wagner, &#8220;Privacy Policies Across the Ages: Content of Privacy Policies 1996&#8211;2021,&#8221; 26 <em>ACM Transactions on Privacy and Security</em>, Article 32, at 1 (2023), https://doi.org/10.1145/3590152.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-77" href="#footnote-anchor-77" class="footnote-number" contenteditable="false" target="_self">77</a><div class="footnote-content"><p>Coates IV, supra note 75.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-78" href="#footnote-anchor-78" class="footnote-number" contenteditable="false" target="_self">78</a><div class="footnote-content"><p>Claire A. Hill &amp; Christopher King, &#8220;How Do German Contracts Do as Much With Fewer Words?&#8220; 79 <em>Chicago-Kent Law Review</em> 889 (2004).</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-79" href="#footnote-anchor-79" class="footnote-number" contenteditable="false" target="_self">79</a><div class="footnote-content"><p>See, e.g., Ronald J. Gilson, &#8220;Value Creation by Business Lawyers: Legal Skills and Asset Pricing,&#8221; 94 <em>Yale Law Journal</em> 239 (1984); Steven L. Schwarcz, &#8220;Explaining the Value of Transactional Lawyering,&#8221; 12 <em>Stanford Journal of Law, Business &amp; Finance</em> 486 (2007).</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-80" href="#footnote-anchor-80" class="footnote-number" contenteditable="false" target="_self">80</a><div class="footnote-content"><p>Yonathan A. Arbel, &#8220;Judicial Economy in the Age of AI,&#8221; 96 <em>Colorado Law Review</em> 549 (2025), https://dx.doi.org/10.2139/ssrn.4873649.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-81" href="#footnote-anchor-81" class="footnote-number" contenteditable="false" target="_self">81</a><div class="footnote-content"><p>Id.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-82" href="#footnote-anchor-82" class="footnote-number" contenteditable="false" target="_self">82</a><div class="footnote-content"><p>Li, supra note 15.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-83" href="#footnote-anchor-83" class="footnote-number" contenteditable="false" target="_self">83</a><div class="footnote-content"><p>Human Rights Watch, &#8220;Rubber Stamp Justice&#8221; (January 2016), https://perma.cc/JV3T-72ZM.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-84" href="#footnote-anchor-84" class="footnote-number" contenteditable="false" target="_self">84</a><div class="footnote-content"><p>Arbel, supra note 80.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-85" href="#footnote-anchor-85" class="footnote-number" contenteditable="false" target="_self">85</a><div class="footnote-content"><p>Pedro Nakamura, &#8220;AI Is Helping Judges to Quickly Close Cases, and Lawyers to Quickly Open Them,&#8221; <em>Rest of World</em> (Sept. 25, 2025), https://perma.cc/GF8G-SH3J.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-86" href="#footnote-anchor-86" class="footnote-number" contenteditable="false" target="_self">86</a><div class="footnote-content"><p>See, e.g., Victor Tangermann, &#8220;Estonia Is Building a &#8216;Robot Judge&#8217; to Help Clear Legal Backlog,&#8221; <em>Futurism</em> (March 25, 2019), https://perma.cc/GC2W-BLT5.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-87" href="#footnote-anchor-87" class="footnote-number" contenteditable="false" target="_self">87</a><div class="footnote-content"><p>Jerry M. Gewirtz, &#8220;Artificial Intelligence May Assist, but Can Never Replace, the Judicial Decision-Making Process of Human Judges,&#8221; 98 <em>Florida Bar Journal</em> 6, 8 (November/December 2024), https://perma.cc/WT33-LYBL.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-88" href="#footnote-anchor-88" class="footnote-number" contenteditable="false" target="_self">88</a><div class="footnote-content"><p>Justin Curl, Peter Henderson, Kart Kandula, &amp; Faiz Surani, &#8220;Judges Shouldn&#8217;t Rely on AI for the Ordinary Meaning of Text,&#8221; <em>Lawfare</em> (May 22, 2025).</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-89" href="#footnote-anchor-89" class="footnote-number" contenteditable="false" target="_self">89</a><div class="footnote-content"><p>Marcin G&#243;rski, &#8220;Why a Human Court?,&#8221; 18 <em>EUCrim</em> 83 (2023), https://perma.cc/P6W6-RXX9.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-90" href="#footnote-anchor-90" class="footnote-number" contenteditable="false" target="_self">90</a><div class="footnote-content"><p>Unikowsky, &#8220;Should AI Replace Law Clerks?,&#8221; supra note 7.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-91" href="#footnote-anchor-91" class="footnote-number" contenteditable="false" target="_self">91</a><div class="footnote-content"><p>Brooke Auxier, Lee Rainie, Monica Anderson, Andrew Perrin, Madhu Kumar, &amp; Erica Turner, &#8220;Americans and Privacy: Concerned, Confused and Feeling Lack of Control Over Their Personal Information,&#8221; <em>Pew Research Center</em> (Nov. 15, 2019), https://perma.cc/Z5XZ-QVBT.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-92" href="#footnote-anchor-92" class="footnote-number" contenteditable="false" target="_self">92</a><div class="footnote-content"><p>See, e.g., Hon. C. S. Maravilla, &#8220;A(I)ccess to Justice: How AI and Ethics Opinions Approving Limited Scope Representation Support Legal Market Consolidation,&#8221; 40 <em>Georgia State University Law Review</em> 957 (2024); Bruce A. Green &amp; M. Ellen Murphy, &#8220;Replacing This Old House: Certifying and Regulating New Legal Services Providers,&#8221; 76 <em>Washington University Journal of Law and Policy</em> 45 (2025); Joseph J. Avery, Patricia S&#225;nchez Abril, &amp; Alissa del Riego, &#8220;ChatGPT, Esq.: Recasting Unauthorized Practice of Law in the Era of Generative AI,&#8221; 26 <em>Yale Journal of Law &amp; Technology</em> 64 (2023), https://dx.doi.org/10.2139/ssrn.5152523; Mia Bonardi &amp; L. Karl Branting, &#8220;Certifying Legal AI Assistants for Unrepresented Litigants: A Global Survey of Access to Civil Justice, Unauthorized Practice of Law, and AI,&#8221; 26 <em>Columbia Science &amp; Technology Law Review</em> 1 (2025), https://doi.org/10.52214/stlr.v26i1.13336.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-93" href="#footnote-anchor-93" class="footnote-number" contenteditable="false" target="_self">93</a><div class="footnote-content"><p>State supreme courts regulate the practice of law in the U.S., though some courts have delegated this task to bar associations and receive input from state legislators. The exact process by which these regulations change varies by jurisdiction, so we refer to the recommended actor as state courts for simplicity. For more on how professional regulations are enacted and modified, see Lucy Ricca &amp; Thomas Clarke, &#8220;The Bar Re-imagined: Options for State Courts to Re-structure the Regulation of the Practice of Law,&#8221; <em>Stanford Law School Deborah L. Rhode Center on the Legal Profession</em> (September 2023), https://perma.cc/WP7L-GWMY.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-94" href="#footnote-anchor-94" class="footnote-number" contenteditable="false" target="_self">94</a><div class="footnote-content"><p>Engstrom, Ricca, &amp; Knowlton, supra note 9, at 11.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-95" href="#footnote-anchor-95" class="footnote-number" contenteditable="false" target="_self">95</a><div class="footnote-content"><p>David Autor, &#8220;Applying AI to Rebuild Middle Class Jobs&#8221; (National Bureau of Economic Research, Working Paper No. 32140, February 2024), https://perma.cc/VBA4-JSFL.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-96" href="#footnote-anchor-96" class="footnote-number" contenteditable="false" target="_self">96</a><div class="footnote-content"><p>Id.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-97" href="#footnote-anchor-97" class="footnote-number" contenteditable="false" target="_self">97</a><div class="footnote-content"><p>Green &amp; Murphy, supra note 93.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-98" href="#footnote-anchor-98" class="footnote-number" contenteditable="false" target="_self">98</a><div class="footnote-content"><p>Avery, Abril, &amp; del Riego, supra note 93.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-99" href="#footnote-anchor-99" class="footnote-number" contenteditable="false" target="_self">99</a><div class="footnote-content"><p>Id.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-100" href="#footnote-anchor-100" class="footnote-number" contenteditable="false" target="_self">100</a><div class="footnote-content"><p>Steward, supra note 50.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-101" href="#footnote-anchor-101" class="footnote-number" contenteditable="false" target="_self">101</a><div class="footnote-content"><p>Drew Simshaw, &#8220;Toward National Regulation of Legal Technology: A Path Forward for Access to Justice,&#8221; 92 <em>Fordham Law Review</em> 1 <em>(</em>2023).</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-102" href="#footnote-anchor-102" class="footnote-number" contenteditable="false" target="_self">102</a><div class="footnote-content"><p>See, generally, R. Matthew Black, &#8220;Extra Law Prices: Why MRPC 5.4 Continues to Needlessly Burden Access to Civil Justice for Low- to Moderate-Income Clients,&#8221; 25 <em>Washington and Lee Journal of Civil Rights and Social Justice</em> 499 (2019); Robert Saavedra Teuton, &#8220;One Small Step and a Giant Leap: Comparing Washington, D.C.&#8217;s Rule 5.4 With Arizona&#8217;s Rule 5.4 Abolition,&#8221; 65 <em>Arizona Law Review</em> 223 (2023); Stephen P. Younger, &#8220;The Pitfalls and False Promises of Nonlawyer Ownership of Law Firms,&#8221; 132 <em>Yale Law Journal Forum</em> 80 (2022); Gillian K. Hadfield, &#8220;Higher Demand, Lower Supply? A Comparative Assessment of the Legal Resource Landscape for Ordinary Americans,&#8221; 37 <em>Fordham Urban Law Journal</em> 129 (2010); Gillian K. Hadfield &amp; Deborah L. Rhode, &#8220;How to Regulate Legal Services to Promote Access, Innovation, and the Quality of Lawyering,&#8221; 67 <em>Hastings Law Journal</em> 1191 (2016); Jonathan T. Molot, &#8220;What&#8217;s Wrong With Law Firms? A Corporate Finance Solution to Law Firm Short-Termism,&#8221; 88 <em>Southern California Law Review</em> 1 (2014).</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-103" href="#footnote-anchor-103" class="footnote-number" contenteditable="false" target="_self">103</a><div class="footnote-content"><p>Engstrom, Ricca, &amp; Knowlton, supra note 9.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-104" href="#footnote-anchor-104" class="footnote-number" contenteditable="false" target="_self">104</a><div class="footnote-content"><p>Id.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-105" href="#footnote-anchor-105" class="footnote-number" contenteditable="false" target="_self">105</a><div class="footnote-content"><p>Id.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-106" href="#footnote-anchor-106" class="footnote-number" contenteditable="false" target="_self">106</a><div class="footnote-content"><p>Id.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-107" href="#footnote-anchor-107" class="footnote-number" contenteditable="false" target="_self">107</a><div class="footnote-content"><p>Hadfield, &#8220;More Markets, More Justice,&#8221; supra note 24.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-108" href="#footnote-anchor-108" class="footnote-number" contenteditable="false" target="_self">108</a><div class="footnote-content"><p>Id.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-109" href="#footnote-anchor-109" class="footnote-number" contenteditable="false" target="_self">109</a><div class="footnote-content"><p>See Daniel Schwarcz, Sam Manning, Patrick Barry, David R. Cleveland, J.J. Prescott, &amp; Beverly Rich, &#8220;AI-Powered Lawyering: AI Reasoning Models, Retrieval Augmented Generation, and the Future of Legal Practice,&#8221; <em>Journal of Law &amp; Empirical Analysis</em> (forthcoming 2026), https://papers.ssrn.com/sol3/papers.cfm?abstract_id=5162111; Lauren Martin, Nick Whitehouse, Stephanie Yiu, Lizzie Catterson, &amp; Rivindu Perera, &#8220;Better Call GPT, Comparing Large Language Models Against Lawyers&#8221; (Jan. 24, 2024) (unpublished manuscript), https://doi.org/10.48550/arXiv.2401.16212; Jonathan H. Choi, Amy Monahan, &amp; Daniel Schwarcz, &#8220;Lawyering in the Age of Artificial Intelligence,&#8221; 109 <em>Minnesota Law Review</em> 147 (2024), https://dx.doi.org/10.2139/ssrn.4626276.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-110" href="#footnote-anchor-110" class="footnote-number" contenteditable="false" target="_self">110</a><div class="footnote-content"><p>Sayash Kapoor, Peter Henderson, &amp; Arvind Narayanan, &#8220;Promises and Pitfalls of Artificial Intelligence for Legal Applications,&#8221; <em>Journal of Cross-disciplinary Research in Computational Law</em> (2024), https://dx.doi.org/10.2139/ssrn.4695412.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-111" href="#footnote-anchor-111" class="footnote-number" contenteditable="false" target="_self">111</a><div class="footnote-content"><p>Chuck Dinerstein, &#8220;When AI Takes Over: The Hidden Cost of Technological Progress,&#8221; <em>American Council on Science and Health</em> (April 1, 2025), https://perma.cc/9C6X-JAZT.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-112" href="#footnote-anchor-112" class="footnote-number" contenteditable="false" target="_self">112</a><div class="footnote-content"><p>Id.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-113" href="#footnote-anchor-113" class="footnote-number" contenteditable="false" target="_self">113</a><div class="footnote-content"><p>&#8220;History of the Reforms,&#8221; https://perma.cc/XF4F-ZVSV (last visited Jan. 26, 2026).</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-114" href="#footnote-anchor-114" class="footnote-number" contenteditable="false" target="_self">114</a><div class="footnote-content"><p>&#8220;Major Legal Regulators Fall Short in Latest Performance Assessment,&#8221; https://perma.cc/4YUS-7RD5 (last visited Jan. 26, 2026).</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-115" href="#footnote-anchor-115" class="footnote-number" contenteditable="false" target="_self">115</a><div class="footnote-content"><p>Hadfield, &#8220;More Markets, More Justice,&#8221; supra note 24.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-116" href="#footnote-anchor-116" class="footnote-number" contenteditable="false" target="_self">116</a><div class="footnote-content"><p>Sam Tobin, &#8220;British Legal Regulator Criticised Over Collapse of Law Firm Axiom Ince,&#8221; <em>Reuters</em> (Oct. 29, 2024).</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-117" href="#footnote-anchor-117" class="footnote-number" contenteditable="false" target="_self">117</a><div class="footnote-content"><p>See John Hyde, &#8220;Legal Services Board Slams SRA Failings in Damning Report,&#8221; <em>Law Gazette</em> (March 31, 2025), https://perma.cc/3N9W-6K2G; Oscar Glyn, &#8220;SRA Faces Closer Supervision After &#8216;Failing to Protect Public&#8217;,&#8221; <em>Law.com</em> (Oct. 17, 2025).</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-118" href="#footnote-anchor-118" class="footnote-number" contenteditable="false" target="_self">118</a><div class="footnote-content"><p>Daniel Schwarcz, &#8220;Regulating Insurance Sales or Selling Insurance Regulation?: Against Regulatory Competition in Insurance,&#8221; 94 <em>Minnesota Law Review</em> 1707 (2010).</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-119" href="#footnote-anchor-119" class="footnote-number" contenteditable="false" target="_self">119</a><div class="footnote-content"><p>United Nations Office on Drugs and Crime, &#8220;Adversarial vs. Inquisitorial Legal Systems,&#8221; <em>E4J University Module Series,</em> https://perma.cc/Z7XA-D46A (last visited Jan. 26, 2026).</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-120" href="#footnote-anchor-120" class="footnote-number" contenteditable="false" target="_self">120</a><div class="footnote-content"><p>Federal Rule of Evidence 706.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-121" href="#footnote-anchor-121" class="footnote-number" contenteditable="false" target="_self">121</a><div class="footnote-content"><p>Bradford H. Charles, &#8220;Rule 706: An Underutilized Tool to Be Used When Partisan Experts Become &#8216;Hired Guns&#8217;,&#8221; 60 <em>Villanova Law Review</em> 941 (2016).</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-122" href="#footnote-anchor-122" class="footnote-number" contenteditable="false" target="_self">122</a><div class="footnote-content"><p>Id.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-123" href="#footnote-anchor-123" class="footnote-number" contenteditable="false" target="_self">123</a><div class="footnote-content"><p>Federal Rule of Civil Procedure 53.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-124" href="#footnote-anchor-124" class="footnote-number" contenteditable="false" target="_self">124</a><div class="footnote-content"><p>Shira A. Scheindlin &amp; Jonathan M. Redgrave, &#8220;Special Masters and E-Discovery: The Intersection of Two Recent Revisions to the Federal Rules of Civil Procedure,&#8221; 30 <em>Cardozo Law Review</em> 347 (2008).</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-125" href="#footnote-anchor-125" class="footnote-number" contenteditable="false" target="_self">125</a><div class="footnote-content"><p>Id.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-126" href="#footnote-anchor-126" class="footnote-number" contenteditable="false" target="_self">126</a><div class="footnote-content"><p>Soia Mentschikoff, &#8220;Commercial Arbitration,&#8221; 61 <em>Columbia Law Review</em> 846 (1961), https://doi.org/10.2307/1120097.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-127" href="#footnote-anchor-127" class="footnote-number" contenteditable="false" target="_self">127</a><div class="footnote-content"><p>Christopher R. Drahozal, &#8221;Arbitration Costs and Forum Accessibility: Empirical Evidence,&#8221; 41 <em>University of Michigan Journal of Law</em> R<em>eform</em> 813 (2008), https://doi.org/10.36646/mjlr.41.4.arbitration.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-128" href="#footnote-anchor-128" class="footnote-number" contenteditable="false" target="_self">128</a><div class="footnote-content"><p>Theodore Eisenberg, Geoffrey P. Miller, &amp; Emily Sherwin, &#8220;Arbitration&#8217;s Summer Soldiers: An Empirical Study of Arbitration Clauses in Consumer and Nonconsumer Contracts,&#8221; 41 <em>University of Michigan Journal of Law</em> R<em>eform</em> 871 (2008), https://dx.doi.org/10.2139/ssrn.1076968.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-129" href="#footnote-anchor-129" class="footnote-number" contenteditable="false" target="_self">129</a><div class="footnote-content"><p>Amberlee B. Conley, &#8220;You Can Have Your Day in Court&#8212;But Not Before Your Day in Mandatory, Nonbinding Arbitration: Balancing Practicalities of State Arbitration,&#8221; 104 <em>Iowa Law Review</em> 325 (2018).</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-130" href="#footnote-anchor-130" class="footnote-number" contenteditable="false" target="_self">130</a><div class="footnote-content"><p>Christopher R. Drahozal &amp; Quentin R. Wittrock, &#8220;Is There a Flight From Arbitration?,&#8221; 37 <em>Hofstra Law Review</em> 71 (2008).</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-131" href="#footnote-anchor-131" class="footnote-number" contenteditable="false" target="_self">131</a><div class="footnote-content"><p>Michael J. Broyde &amp; Yiyang Mei, &#8220;Don&#8217;t Kill the Baby! The Case for AI in Arbitration,&#8221; 21 <em>New York University Journal of Law &amp; Business</em> 1 (2024).</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-132" href="#footnote-anchor-132" class="footnote-number" contenteditable="false" target="_self">132</a><div class="footnote-content"><p>Id.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-133" href="#footnote-anchor-133" class="footnote-number" contenteditable="false" target="_self">133</a><div class="footnote-content"><p>David Horton, &#8220;Forced Robot Arbitration,&#8221; 109 <em>Cornell Law Review</em> 679 (2024).</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-134" href="#footnote-anchor-134" class="footnote-number" contenteditable="false" target="_self">134</a><div class="footnote-content"><p>Robert Walters, &#8220;Robots Replacing Human Arbitrators: The Legal Dilemma,&#8221; 34 <em>Information &amp; Communications Technology Law</em> 129 (2025), https://doi.org/10.1080/13600834.2024.2408155.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-135" href="#footnote-anchor-135" class="footnote-number" contenteditable="false" target="_self">135</a><div class="footnote-content"><p>David S. Schwartz, &#8220;Mandatory Arbitration and Fairness,&#8221; 84 <em>Notre Dame Law Review</em> 1247 (2009); Jessica Silver-Greenberg &amp; Robert Gebeloff, &#8220;Arbitration Everywhere, Stacking the Deck of Justice,&#8221; <em>New York Times</em> (Oct. 31, 2015).</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-136" href="#footnote-anchor-136" class="footnote-number" contenteditable="false" target="_self">136</a><div class="footnote-content"><p>Stephen J. Ware, &#8220;The Centrist Case for Enforcing Adhesive Arbitration Agreements,&#8221; 23 <em>Harvard Negotiation Law Review</em> 29 (2017).</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-137" href="#footnote-anchor-137" class="footnote-number" contenteditable="false" target="_self">137</a><div class="footnote-content"><p>Peter B. Rutledge &amp; Christopher R. Drahozal, &#8220;Contract and Choice,&#8221; 2013 <em>Brigham Young University Law Review</em> 1 (2013).</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-138" href="#footnote-anchor-138" class="footnote-number" contenteditable="false" target="_self">138</a><div class="footnote-content"><p>Amanda R. Witwer, Lynn Langton, Duren Banks, Dulani Woods, Michael J.D. Vermeer, &amp; Brian A. Jackson, &#8220;Online Dispute Resolution: Perspectives to Support Successful Implementation and Outcomes in Court Proceedings,&#8221; <em>RAND Corporation</em> (2021), https://perma.cc/TD7Q-HGM7.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-139" href="#footnote-anchor-139" class="footnote-number" contenteditable="false" target="_self">139</a><div class="footnote-content"><p>See Broyde &amp; Mei, supra note 131, at 168&#8211;172.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-140" href="#footnote-anchor-140" class="footnote-number" contenteditable="false" target="_self">140</a><div class="footnote-content"><p>See Richard Re &amp; Alicia Solow-Niederman, &#8220;Developing Artificially Intelligent Justice,&#8221; 22 <em>Stanford Technology Law Review</em> 242, 278&#8211;80 (2019).</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-141" href="#footnote-anchor-141" class="footnote-number" contenteditable="false" target="_self">141</a><div class="footnote-content"><p>Maria L. Marcus, &#8220;Judicial Overload: The Reasons and the Remedies,&#8221; 28 <em>Buffalo Law Review</em> 111, 132&#8211;33 (1979).</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-142" href="#footnote-anchor-142" class="footnote-number" contenteditable="false" target="_self">142</a><div class="footnote-content"><p>Bert I. Huang, &#8220;Lightened Scrutiny,&#8221; 124 <em>Harvard Law Review</em> 1109 (2011).</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-143" href="#footnote-anchor-143" class="footnote-number" contenteditable="false" target="_self">143</a><div class="footnote-content"><p>Peter S. Menell &amp; Ryan Vacca, &#8220;Revisiting and Confronting the Federal Judiciary Capacity &#8216;Crisis&#8217;: Charting a Path for Federal Judiciary Reform,&#8221; 108 <em>California Law Review</em> 789 (2020) (quoting &#8220;National Court of Appeals Act, Hearings on S. 2762 and S. 3423 Before the Subcommittee on Improvements in Judicial Machinery of the Committee on the Judiciary,&#8221; 94th Cong. i, 26&#8211;37 (1976)).</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-144" href="#footnote-anchor-144" class="footnote-number" contenteditable="false" target="_self">144</a><div class="footnote-content"><p>Menell &amp; Vacca, supra note 143.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-145" href="#footnote-anchor-145" class="footnote-number" contenteditable="false" target="_self">145</a><div class="footnote-content"><p>Arbel, supra note 80.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-146" href="#footnote-anchor-146" class="footnote-number" contenteditable="false" target="_self">146</a><div class="footnote-content"><p>Id.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-147" href="#footnote-anchor-147" class="footnote-number" contenteditable="false" target="_self">147</a><div class="footnote-content"><p>Shaoli Katana, &#8220;MSBA&#8217;s 2024 Legislative Wins,&#8221; <em>Minnesota State Bar Association,</em> https://perma.cc/AP6Q-FQ8V (last visited Jan. 26, 2026).</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-148" href="#footnote-anchor-148" class="footnote-number" contenteditable="false" target="_self">148</a><div class="footnote-content"><p>S. 1821, 119th Cong. (2025).</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-149" href="#footnote-anchor-149" class="footnote-number" contenteditable="false" target="_self">149</a><div class="footnote-content"><p>Deborah L. Rhode, &#8220;Access to Justice: Connecting Principles to Practice,&#8221; 17 <em>Georgetown Journal of Legal Ethics</em> 369 (2004).</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-150" href="#footnote-anchor-150" class="footnote-number" contenteditable="false" target="_self">150</a><div class="footnote-content"><p>Christian Veith, Michael Bandlow, Michael Harnisch, Hariolf Wenzler, Markus Hartung, &amp; Dirk Hartung, &#8220;How Legal Tech Will Change the Business of Law,&#8221; <em>Boston Consulting Group</em> (2016), https://perma.cc/2EM6-B4YB.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-151" href="#footnote-anchor-151" class="footnote-number" contenteditable="false" target="_self">151</a><div class="footnote-content"><p>Andy Teichholz, &#8220;The Modern General Counsel: Legal Advisor and Strategic Business Partner,&#8221; <em>Corporate Counsel Business Journal</em> (2024), https://perma.cc/26ZK-KDBS.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-152" href="#footnote-anchor-152" class="footnote-number" contenteditable="false" target="_self">152</a><div class="footnote-content"><p>Thomson Reuters, &#8220;Future of Professionals Report 2024&#8221; (2024), https://perma.cc/HHA4-7DUV.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-153" href="#footnote-anchor-153" class="footnote-number" contenteditable="false" target="_self">153</a><div class="footnote-content"><p>Marjorie Richter, &#8220;How AI Is Transforming the Legal Profession,&#8221; <em>Thomson Reuters Blog</em>, https://perma.cc/Z9ZL-BGXA (last visited Jan. 26, 2026).</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-154" href="#footnote-anchor-154" class="footnote-number" contenteditable="false" target="_self">154</a><div class="footnote-content"><p>Veith et al., supra note 150.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-155" href="#footnote-anchor-155" class="footnote-number" contenteditable="false" target="_self">155</a><div class="footnote-content"><p>Richter, supra note 153.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-156" href="#footnote-anchor-156" class="footnote-number" contenteditable="false" target="_self">156</a><div class="footnote-content"><p>Thomson Reuters, supra note 152.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-157" href="#footnote-anchor-157" class="footnote-number" contenteditable="false" target="_self">157</a><div class="footnote-content"><p>Teichholz, supra note 151.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-158" href="#footnote-anchor-158" class="footnote-number" contenteditable="false" target="_self">158</a><div class="footnote-content"><p>DRI Center for Law and Public Policy Artificial Intelligence Working Group, &#8220;Artificial Intelligence in Legal Practice: Benefits, Considerations, and Best Practices&#8221; (2024), https://perma.cc/8X6E-BC25.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-159" href="#footnote-anchor-159" class="footnote-number" contenteditable="false" target="_self">159</a><div class="footnote-content"><p>Cynthia Hardy &amp; Steve Maguire, &#8220;Institutional Entrepreneurship and Change in Fields,&#8221; in <em>Handbook of Organizational Institutionalism</em> (Royston Greenwood, Christine Oliver, Thomas B. Lawrence, Renate E. Meyer, Cynthia Hardy, &amp; Steve Maguire, eds., 2d ed., 2017), https://doi.org/10.4135/9781446280669.n11.</p></div></div>]]></content:encoded></item><item><title><![CDATA[Fact checking Moravec's paradox]]></title><description><![CDATA[This famous aphorism is neither true nor useful]]></description><link>https://www.normaltech.ai/p/fact-checking-moravecs-paradox</link><guid isPermaLink="false">https://www.normaltech.ai/p/fact-checking-moravecs-paradox</guid><dc:creator><![CDATA[Arvind Narayanan]]></dc:creator><pubDate>Thu, 29 Jan 2026 22:35:47 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/5c0502b3-b033-478e-b7f7-d224b2ff51e4_1920x1080.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>I have launched a <a href="https://www.youtube.com/@ArvindOnAI">YouTube channel</a> in which I analyze AI developments from a normal technology perspective. This essay is based on my most recent video in which I did a <a href="https://www.youtube.com/watch?v=-Ty4BXrASFU">deep dive into Moravec&#8217;s paradox</a>, the endlessly repeated aphorism that tasks that are hard for humans are easy for AI and vice versa.</p><p>Here&#8217;s what I found:</p><ul><li><p>Moravec&#8217;s paradox never been empirically tested. (It&#8217;s often repeated as a fact by many AI researchers, including pioneers I know and respect, but that doesn&#8217;t mean I&#8217;ll take their claims at face value!)</p></li><li><p>It is really a statement about what the AI community finds it worthwhile to work on. It doesn&#8217;t have any predictive power about which problems are going to be easy or hard for AI.</p></li><li><p>It comes with an evolutionary explanation that I find highly dubious. (AI researchers have a history of making stuff up about human brains without any relevant background in neuroscience or evolutionary biology.)</p></li><li><p>Moravec&#8217;s-paradox-style thinking has led to both alarmism (about imminent superintelligent reasoning) and false comfort (in areas like robotics).</p></li><li><p>To adapt to AI advances, we don&#8217;t need to predict capability breakthroughs. Since diffusion of new capabilities takes a long time, that gives us plenty of time to react &#8212; time that we often squander, and then panic!</p></li></ul><p>Watch the full argument <a href="https://www.youtube.com/watch?v=-Ty4BXrASFU">here</a> or read it below.</p><div><hr></div><p>Every week brings new claims about AI advances. How do we know what&#8217;s coming next? Could AI predict crime? Write award-winning novels? Hack into critical infrastructure? Will we finally have robots in our home that will fold our clothes and load our dishwashers?</p><p>What will AI advances mean for your job? What will it mean for the social fabric? It&#8217;s hard to deal with all this uncertainty. If only we had a way to predict which new AI capabilities will be developed soon and which ones will remain hard for the foreseeable future.</p><p>Historically, AI researchers&#8217; predictions about progress in AI abilities have been <a href="https://www.nytimes.com/1958/07/13/archives/electronic-brain-teaches-itself.html">pretty bad</a>. We don&#8217;t really have principles that describe which kinds of tasks are easy for AI and which ones are hard.</p><p>Well, we have one &#8212; <a href="https://en.wikipedia.org/wiki/Moravec%27s_paradox">Moravec&#8217;s paradox</a>. It refers to the observation that it&#8217;s easy to train computers to do things that people find hard, like math and logic, and hard to train them to do things that we find easy, like seeing the world or walking.</p><p>It comes from the 1988 book <em>Mind Children</em> by Hans Moravec, who was &#8212; and is &#8212; a robotics researcher. He wrote: </p><blockquote><p>It is comparatively easy to make computers exhibit adult level performance on intelligence tests or playing checkers, and difficult or impossible to give them the skills of a one-year-old when it comes to perception and mobility.</p></blockquote><p>In the early days of artificial intelligence, researchers focused on chess and other reasoning tasks, since these were thought to be some of the hardest and what made us uniquely human. But funnily enough, if you want to build a robot that can defeat human grandmasters, figuring out which moves to make is the easy part. Physically making the moves on the chessboard is the hard part. This is pretty well known today, so Moravec&#8217;s paradox seems to make a lot of intuitive sense.</p><p>If Moravec&#8217;s paradox is true, the implications would be amazing. If we want to know which AI capabilities might be built next, we just have to see how hard they are for humans. So scientific research will get automated before folding clothes, and so on.</p><p>But here&#8217;s the thing &#8212; Moravec&#8217;s paradox has never been fact checked. And that&#8217;s despite videos with hundreds of thousands of views, and TED talks all repeating it as a fact. When I dug into the evidence behind the so-called paradox, I found something surprising.</p><p>In this essay I&#8217;ll discuss why the theory and evidence behind the paradox are flaky. Then I&#8217;ll explain why simplistic predictions about what is easy or hard for AI have misled AI researchers and tech leaders. It has led to alarmism on the one hand and false comfort on the other hand. (Now there&#8217;s a paradox.) And finally I&#8217;ll answer the question, if we can&#8217;t rely on Moravec&#8217;s paradox, then how should we prepare for AI advances and their impacts?</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.normaltech.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Subscribe to receive essays analyzing AI developments from a normal technology perspective.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p></p><h2>The evidence behind the paradox is flaky</h2><p>How would we test Moravec&#8217;s paradox? We could take some sample of the tasks that are out there, determine how hard they are for people, how hard they are for AI, and make a graph. If we saw something like this, the paradox would be confirmed.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!jE1w!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77d45ce8-30a7-47c0-818d-a94b4b466cc3_1920x1080.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!jE1w!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77d45ce8-30a7-47c0-818d-a94b4b466cc3_1920x1080.png 424w, https://substackcdn.com/image/fetch/$s_!jE1w!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77d45ce8-30a7-47c0-818d-a94b4b466cc3_1920x1080.png 848w, https://substackcdn.com/image/fetch/$s_!jE1w!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77d45ce8-30a7-47c0-818d-a94b4b466cc3_1920x1080.png 1272w, https://substackcdn.com/image/fetch/$s_!jE1w!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77d45ce8-30a7-47c0-818d-a94b4b466cc3_1920x1080.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!jE1w!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77d45ce8-30a7-47c0-818d-a94b4b466cc3_1920x1080.png" width="1456" height="819" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/77d45ce8-30a7-47c0-818d-a94b4b466cc3_1920x1080.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:819,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:71729,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.normaltech.ai/i/186233784?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77d45ce8-30a7-47c0-818d-a94b4b466cc3_1920x1080.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!jE1w!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77d45ce8-30a7-47c0-818d-a94b4b466cc3_1920x1080.png 424w, https://substackcdn.com/image/fetch/$s_!jE1w!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77d45ce8-30a7-47c0-818d-a94b4b466cc3_1920x1080.png 848w, https://substackcdn.com/image/fetch/$s_!jE1w!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77d45ce8-30a7-47c0-818d-a94b4b466cc3_1920x1080.png 1272w, https://substackcdn.com/image/fetch/$s_!jE1w!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77d45ce8-30a7-47c0-818d-a94b4b466cc3_1920x1080.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">A possible way to empirically test Moravec&#8217;s paradox</figcaption></figure></div><p>But here&#8217;s the problem: Which set of tasks should we consider for our analysis? When AI researchers say Moravec&#8217;s paradox checks out, they are implicitly limiting their focus to problems that are considered interesting in the AI research community.</p><p>There are an endless number of tasks that are easy for both humans and AI, but they are not interesting. How bright is an image? How to play tic tac toe? Thousands of these tasks get solved by programmers on a daily basis and coded into AI systems, and when people do so, they don&#8217;t make a fuss about them.</p><p>There are also an endless number of tasks that are hard for humans and, as far as anyone knows, are also hard for AI. Identifying hit songs; predicting stock prices; cracking as-yet-undeciphered ancient scripts; even building a Dyson sphere. These are so hard that there is essentially no progress on these problems, although some of them attract junk research that tends to quickly get <a href="https://reproducible.cs.princeton.edu/predicting-hits.html">debunked</a>. So these problems also don&#8217;t tend to get talked about as much. </p><p>In fact, there are thousands of problems that computer scientists have proved to be &#8220;NP-complete&#8221;, which means we have strong mathematical reasons to think they will forever be hard for AI, so serious AI researchers don&#8217;t tend to work on them. They work on easier, approximate versions of the problems instead.</p><p>The other two quadrants of the chart are different. On the top left, tasks like playing soccer, that are easy for humans but currently hard for AI, are extremely interesting to AI researchers. That&#8217;s because we know it&#8217;s possible to teach AI these skills, but we haven&#8217;t yet managed to do so, which makes them great tests for AI progress.</p><p>On the bottom right, problems that are hard for humans but easy for AI, such as searching the web, are also interesting. These capabilities tend to greatly augment human productivity. Even though they are in some sense &#8220;easy&#8221;, the industry invests a lot in making web search and other tools work as effectively, quickly, and cheaply as possible. So research on these tasks is a big driver of AI progress.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!5Tvy!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1746607a-20b6-4467-a5dc-d33523358bbd_1920x1080.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!5Tvy!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1746607a-20b6-4467-a5dc-d33523358bbd_1920x1080.png 424w, https://substackcdn.com/image/fetch/$s_!5Tvy!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1746607a-20b6-4467-a5dc-d33523358bbd_1920x1080.png 848w, https://substackcdn.com/image/fetch/$s_!5Tvy!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1746607a-20b6-4467-a5dc-d33523358bbd_1920x1080.png 1272w, https://substackcdn.com/image/fetch/$s_!5Tvy!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1746607a-20b6-4467-a5dc-d33523358bbd_1920x1080.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!5Tvy!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1746607a-20b6-4467-a5dc-d33523358bbd_1920x1080.png" width="1456" height="819" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/1746607a-20b6-4467-a5dc-d33523358bbd_1920x1080.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:819,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:122688,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.normaltech.ai/i/186233784?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1746607a-20b6-4467-a5dc-d33523358bbd_1920x1080.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!5Tvy!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1746607a-20b6-4467-a5dc-d33523358bbd_1920x1080.png 424w, https://substackcdn.com/image/fetch/$s_!5Tvy!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1746607a-20b6-4467-a5dc-d33523358bbd_1920x1080.png 848w, https://substackcdn.com/image/fetch/$s_!5Tvy!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1746607a-20b6-4467-a5dc-d33523358bbd_1920x1080.png 1272w, https://substackcdn.com/image/fetch/$s_!5Tvy!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1746607a-20b6-4467-a5dc-d33523358bbd_1920x1080.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Moravec&#8217;s paradox may be a selection effect caused by ignoring tasks as uninteresting when they&#8217;re either too easy or too hard for both humans and AI.</figcaption></figure></div><p></p><p>In short, when you&#8217;re thinking about the space of all possible tasks, if you basically ignore two quadrants of your 2x2 matrix because they are not interesting, then of course it will seem like what you&#8217;re left with shows a strong negative correlation between the two axes.</p><h2>A flawed evolutionary argument</h2><p>To be clear, the reason AI researchers are drawn to Moravec&#8217;s paradox isn&#8217;t because it is empirically backed. It&#8217;s because it comes with an intuitively appealing story. In his book, Moravec provided this explanation:</p><blockquote><p>Encoded in the large, highly evolved sensory and motor portions of the human brain is a billion years of experience about the nature of the world and how to survive in it. The deliberate process we call reasoning is, I believe, the thinnest veneer of human thought, effective only because it is supported by this much older and much more powerful, though usually unconscious, sensorimotor knowledge. We are all prodigious olympians in perceptual and motor areas, so good that we make the difficult look easy. Abstract thought, though, is a new trick, perhaps less than 100 thousand years old. We have not yet mastered it. It is not all that intrinsically difficult; it just seems so when we do it.</p></blockquote><p>And then Moravec praises reasoning systems like <a href="https://en.wikipedia.org/wiki/Stanford_Research_Institute_Problem_Solver">STRIPS</a> from the 1970s, and this is the kind of thing he has in mind when he says that reasoning is easy for AI. These are purely symbolic systems that solve problems like how to put blocks on top of each other in a certain way.</p><p>But AI researchers have learned a lot in the half century since the heyday of STRIPS and other such reasoning programs. They seem quite quaint today. What we&#8217;ve learned is that symbolic reasoning only works well in closed and extremely narrow domains like chess with a fully specified set of rules. When you apply them to real-world problems they are brittle and go off the rails quickly.</p><p>Over the decades there have been many other failures of reasoning systems, like IBM&#8217;s Watson, that excelled in narrow demonstrations but failed when you tried to use them in real-world settings that could depart from what they are trained for.</p><p>Today, it is widely recognized that reasoning in open-ended settings requires <a href="https://cs.nyu.edu/~davise/papers/CommonsenseFinal.pdf">common-sense knowledge</a>. But common sense is one of the so-called easy-for-humans-but-hard-for-AI areas according to acolytes of Moravec&#8217;s paradox. In other words, AI reasoning isn&#8217;t easy after all.</p><p>And sure enough, AI reasoning in a way that can replace human expertise, in open-ended domains like law or scientific research, is still very much unsolved.</p><p>If I were to speculate where Moravec&#8217;s evolutionary argument goes wrong, it would be this: Reasoning might be new from an evolutionary perspective, but it builds on the things that animal brains have learned to do over hundreds of millions of years. This much Moravec acknowledges. But maybe there isn&#8217;t a separate skill called &#8220;abstract&#8221; reasoning that can be learned separately without all of this infrastructure.</p><h2>How simplistic models have misled AI researchers and tech leaders</h2><p>Unfortunately, partly because of the misconception in the AI world that reasoning is a separate skill that&#8217;s easy for computers, there is a widespread belief that AI will soon be superhuman at things like scientific research or operating a company or even running a government. In my experience, many researchers tend to generalize from AI success at chess and other closed domains to these kinds of open-ended domains.</p><p>So AI leaders promise that investing trillions into data centers will lead to a cure for cancer and various other downstream benefits, without stopping to think about how these inputs will translate into the desired outputs. It also leads them to support some extreme policies, such as investment in AI science at the expense of human scientists, and warning policymakers to prepare for a white collar bloodbath. It also leads to the fear that leaders such as politicians or CEOs military generals will soon have no choice but to delegate important decisions to AI because it will be superhuman at reasoning.</p><p>Maybe... or maybe it&#8217;s all a myth. There may not even be one general skill called reasoning.</p><p>Maybe the limits to reasoning are actually things like the lack of verifiers. That is, AI can&#8217;t get that good at <a href="https://arxiv.org/abs/2402.01656">legal reasoning</a> unlike reasoning in chess, because there is no way for AI to write millions of legal arguments and get immediate and accurate feedback on which ones are good and which ones aren&#8217;t, analogous to the way AI gets good at chess.</p><p>Maybe it&#8217;s partly due to the limits of real-world knowledge. That is, AI can&#8217;t quickly become superhuman at medical reasoning because it is limited by the available medical knowledge in the world.</p><p>In other words, the same factors that pose limits to human reasoning also pose limits to AI reasoning. In this view, the answer has nothing to do with biology.</p><p>Much of my writing is based on the idea that superhuman AI reasoning is a myth. Of course, you don&#8217;t have to accept this. But when you hear these predictions of imminent superintelligence, it&#8217;s helpful to understand the underlying mental model that many people in the AI field have. And it&#8217;s definitely useful to know that we have five or six decades of evidence that whether a skill such as reasoning is easy or hard for AI can depend a lot on whether you&#8217;re talking about a closed domain or an open domain.</p><p>Now let&#8217;s look at the other side of the coin. Some AI capabilities are predicted to be <em>hard</em> because of Moravec&#8217;s paradox. Most frequently, it&#8217;s robotics.</p><p>Just as people say we have to worry about job losses and safety implications of breakthroughs in reasoning, they&#8217;ll say we don&#8217;t have to worry about job losses and safety risks of breakthroughs in robotics. It&#8217;s a hard problem, so improvements won&#8217;t happen overnight.</p><p>Unfortunately, this is false comfort. I could be wrong about what I said about reasoning earlier, and there could be a breakthrough tomorrow. Similarly, these researchers could be wrong about robotics and there could be a breakthrough tomorrow.</p><p>In fact, there is another &#8220;hard problem&#8221; for AI that used to be invoked to explain Moravec&#8217;s paradox, which is computer vision, and it doesn&#8217;t get invoked anymore, and that&#8217;s because there was in fact a breakthrough around 2012 / 2013 when AI performance at tasks like object recognition shot up dramatically due to <a href="https://spectrum.ieee.org/alexnet-source-code">deep learning</a>. The reason it even took that long is it required the use of GPUs and the idea of using GPUs for AI only began around that time.</p><p>The scientific ideas behind deep learning had largely been <a href="https://www.nature.com/articles/323533a0">established</a> in the 1980s &#8212; before Moravec published his book &#8212; he just didn&#8217;t know it yet.</p><h2>Conclusion: if not Moravec&#8217;s paradox, then what?</h2><p>There&#8217;s a reason why rules of thumb like Moravec&#8217;s paradox are so tempting. It&#8217;s because we assume that if there were an AI capability breakthrough, there would be rapid societal effects such as job losses, so we should be prepared in advance.</p><p>But this isn&#8217;t actually true. Even breakthrough technologies take a long time to be successfully commercialized and deployed. In fact, breakthrough technologies may <em>especially</em> take a long time to be deployed, because the supporting infrastructure just isn&#8217;t there. There&#8217;s a famous <a href="https://www.bbc.com/news/business-40673694">case study</a> of why it took 40 years for electric power to replace steam power in factories.</p><p>Take self-driving cars. Waymo started testing them on public roads more than <a href="https://waymo.com/blog/2021/02/expanding-our-testing-in-san-francisco">15 years ago</a>.</p><p>There was little reason to doubt that this technology would one day be viable, and that it would have good and bad effects. Policymakers should have been preparing and figuring out how to compensate workers who stood to lose out.</p><p>Instead, people are only now waking up to the reality now that these cars and trucks are already on the road, and are pushing policies like <a href="https://www.businessinsider.com/josh-hawley-banning-self-driving-cars-2025-9">banning</a> them. Keep in mind that a million people per year die in car accidents worldwide, and self-driving cars are already much safer than human drivers.</p><p>A recurring theme of my work on AI is the difficulty of prediction &#8212; both the use of AI itself for prediction and predicting the future of AI capabilities and impacts.</p><p>Instead of relying on prediction, we should get better at shaping and adapting to the tech that we actually know for sure is coming. It is psychologically hard to let go of our need to want to know the future, but if we give up this false comfort we can build a much more resilient society.</p><h2>Other videos</h2><p>I have published three full-length videos so far, and regularly publish short videos. Subscribe <a href="https://www.youtube.com/@ArvindOnAI">here</a>.</p><ul><li><p><a href="https://www.youtube.com/watch?v=VDfyuB9p7sI">What happens if there&#8217;s an AI crash?</a></p></li><li><p><a href="https://www.youtube.com/watch?v=hWrVksoyVDw">Why a Manhattan project for AI makes no sense</a></p></li><li><p><a href="https://www.youtube.com/watch?v=-Ty4BXrASFU">Debunking Moravec&#8217;s paradox</a></p></li></ul>]]></content:encoded></item><item><title><![CDATA[A guide to understanding AI as normal technology]]></title><description><![CDATA[And a big change for this newsletter]]></description><link>https://www.normaltech.ai/p/a-guide-to-understanding-ai-as-normal</link><guid isPermaLink="false">https://www.normaltech.ai/p/a-guide-to-understanding-ai-as-normal</guid><dc:creator><![CDATA[Arvind Narayanan]]></dc:creator><pubDate>Tue, 09 Sep 2025 13:00:40 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/d6ef1281-5124-4069-a716-955720587e38_2872x1606.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>When we published <a href="https://knightcolumbia.org/content/ai-as-normal-technology">AI as Normal Technology</a>, the impact caught us off guard. It quickly became the most influential thing either of us had ever done.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a> We took this as a strong signal to spend more of our time thinking and writing about the medium-term future of AI and its impacts, offering grounded analysis of a topic that tends to attract speculation. This is a shift in focus away from writing about the present-day and near-term impacts of AI, which is what the <a href="https://www.normaltech.ai/p/starting-reading-the-ai-snake-oil">AI Snake Oil</a> project was about.</p><p>Reflecting this shift, we have renamed this newsletter. We have already published two <a href="https://www.aisnakeoil.com/p/agi-is-not-a-milestone">follow-up</a> <a href="https://www.aisnakeoil.com/p/could-ai-slow-science">essays</a> to AI as Normal Technology and will publish more regularly as we expand our framework into a book, which we plan to complete in late 2026 for publication in 2027.</p><p>Today, we address common points of confusion about AI as Normal Technology, try to make the original essay more approachable, and compare it to <a href="https://ai-2027.com/">AI 2027</a>.</p><p><strong>Table of contents</strong></p><ol><li><p><a href="https://www.normaltech.ai/i/173147197/normal-doesnt-mean-mundane-or-predictable">Normal doesn&#8217;t mean mundane or predictable</a></p></li><li><p><a href="https://www.normaltech.ai/i/173147197/a-restatement-of-our-thesis">A restatement of our thesis</a></p></li><li><p><a href="https://www.normaltech.ai/i/173147197/if-disappointment-about-gpt-has-nudged-you-towards-ai-as-normal-technology-its-possible-you-dont-quite-understand-the-thesis">If disappointment about GPT-5 has nudged you towards AI as normal technology, it&#8217;s possible you don&#8217;t quite understand the thesis</a></p></li><li><p><a href="https://www.normaltech.ai/i/173147197/why-its-hard-to-find-a-middle-ground-between-ai-as-normal-technology-and-ai">Why it&#8217;s hard to find a &#8220;middle ground&#8221; between AI as Normal Technology and AI 2027</a></p></li><li><p><a href="https://www.normaltech.ai/i/173147197/it-is-hard-to-understand-one-worldview-when-youre-committed-to-another">It is hard to understand one worldview when you&#8217;re committed to another</a></p></li><li><p><a href="https://www.normaltech.ai/i/173147197/reaping-ais-benefits-will-require-hard-work-and-painful-choices">Reaping AI&#8217;s benefits will require hard work and painful choices</a></p></li><li><p><a href="https://www.normaltech.ai/i/173147197/the-surreal-debate-about-the-speed-of-diffusion">The surreal debate about the speed of diffusion</a></p></li><li><p><a href="https://www.normaltech.ai/i/173147197/why-ai-adoption-hits-different">Why AI adoption hits different</a></p></li><li><p><a href="https://www.normaltech.ai/i/173147197/concluding-thoughts">Concluding thoughts</a></p></li></ol><h2><strong>Normal doesn&#8217;t mean mundane or predictable</strong></h2><p>While the essay talks about what we mean by normal (more on that below), we could have been more explicit about what it <em>doesn&#8217;t</em> mean.</p><p>Our point is not &#8220;nothing to see here, move along&#8221;. Indeed, unpredictable societal effects have been a hallmark of powerful technologies ranging from automobiles to social media. This is because they are emergent effects of complex interactions between technology and people. They don&#8217;t tend to be predictable based on the logic of the technology alone. That&#8217;s why rejecting technological determinism is one of the core premises of the normal technology essay.</p><p>In the case of AI, specifically chatbots, we&#8217;re already seeing emergent societal effects. The prevalence of AI companions and some of the harmful effects of model sycophancy such as &#8220;AI psychosis&#8221; have taken most observers <a href="https://www.hyperdimensional.co/p/for-all-issues-so-triable">by surprise</a>.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a> On the other hand, many risks that were widely predicted to be imminent, such as AI being used to manipulate elections, have <a href="https://www.aisnakeoil.com/p/we-looked-at-78-election-deepfakes">not materialized</a>.</p><p>What the landscape of AI&#8217;s social impacts will look like in say 3-5 years &#8212; even based on the diffusion of current capabilities, not future capabilities &#8212; is anyone&#8217;s guess.</p><p>The development of technical capabilities is more predictable than AI&#8217;s social impacts. Daniel Kokotajlo, one of the authors of AI 2027, was previously famous in the AI safety community for his essay &#8220;<a href="https://www.lesswrong.com/posts/6Xgy6CAf2jqHhynHL/what-2026-looks-like">What 2026 looks like</a>&#8221; back in 2021. His predictions about the tech itself proved eerily accurate, but the predictions about social impacts were overall not directionally correct, a point he was gracious enough to concede in a podcast discussion with one of us.</p><p>All this makes AI a <em>more</em> serious challenge for institutions and policymakers because they will have to react nimbly to unforeseeable impacts instead of relying on the false comfort of prediction or trying to prevent all harm. Broadly speaking, the policymaking approach that enables such adaptability is called resilience, which is what our essay advocated for. But while we emphasized resilience as the approach to deal with potentially catastrophic risks, we should have been clearer that resilience also has an important role in dealing with more diffuse risks.</p><p>Perhaps the reason some readers misunderstood our view of predictability is the word &#8220;normal&#8221;. Again, our goal is not to trivialize the task of individually and collectively adapting to AI. In an ideal world, a better title would have been simply &#8220;AI as Technology&#8221;, but we didn&#8217;t think that that would effectively communicate that our goal is to provide an alternative to the exceptionalism that characterizes the superintelligence worldview which currently dominates the discourse.</p><h2><strong>A restatement of our thesis</strong></h2><p>If we were to extract and simplify the core of our thesis, it would be something like this:</p><p>There is a long causal chain between AI capability increases and societal impact. Benefits and risks are realized when AI is deployed, not when it is developed. This gives us (individuals, organizations, institutions, policymakers) many points of leverage for shaping those impacts. So we don&#8217;t have to fret as much about the speed of capability development; our efforts should focus more on the deployment stage both from the perspective of realizing AI&#8217;s benefits and responding to risks. All this is not just true of today&#8217;s AI, but even in the face of hypothetical developments such as self-improvement in AI capabilities. Many of the limits to the power of AI systems are (and should be) external to those systems, so that they cannot be overcome simply by having AI go off and improve its own technical design.</p><p>Aspects of this framework may have to be revised eventually, but that lies beyond the horizon bounding what we can meaningfully anticipate or prepare for:</p><blockquote><p>The world we describe in Part II is one in which AI is far more advanced than it is today. We are not claiming that AI progress&#8212;or human progress&#8212;will stop at that point. What comes after it? We do not know. Consider this analogy: At the dawn of the first Industrial Revolution, it would have been useful to try to think about what an industrial world would look like and how to prepare for it, but it would have been futile to try to predict electricity or computers. Our exercise here is similar. Since we reject &#8220;fast takeoff&#8221; scenarios, we do not see it as necessary or useful to envision a world further ahead than we have attempted to. If and when the scenario we describe in Part II materializes, we will be able to better anticipate and prepare for whatever comes next.</p></blockquote><p>Anyway, to reiterate, the core of the thesis is the underlying causal framework for understanding the relationship between AI and society, not any of the specific impacts that it might or might not have. In our view, if you share this causal understanding, you subscribe to the normal technology thesis. We have found that this framework is indeed widely shared, albeit implicitly.</p><p>That makes the thesis almost tautological in many readers&#8217; minds. We are making what we see as &#8212; and what those readers should see as &#8212; a very weak claim! Not recognizing this causes readers to search for something much more specific that we may have meant by &#8220;normal&#8221;. But we didn&#8217;t. We aren&#8217;t classifying technologies as &#8220;normal&#8221; and &#8220;abnormal&#8221; and then putting AI into the &#8220;normal&#8221; bucket. We&#8217;re just saying we should treat AI like we treat other powerful general-purpose technologies.</p><p>This is not specific to large language models or any particular kind of AI. Incidentally, that&#8217;s why the title is &#8220;AI as normal technology&#8221; and not &#8220;AI as <strong>a</strong> normal technology&#8221;. Our views apply to the whole umbrella of technologies that are collectively referred to as AI, and other similar technologies even if they are not referred to as AI.</p><p>If our worldview is almost tautological, why bother to state it? Because it is in contrast to the superintelligence worldview. That&#8217;s the thing about worldviews: there can be mutually contradictory worldviews that each seem tautological to those who subscribe to them.</p><h2><strong>If disappointment about GPT-5 has nudged you towards AI as normal technology, it&#8217;s possible you don&#8217;t quite understand the thesis</strong></h2><p>It&#8217;s notable that there&#8217;s been a surge of interest in our essay after the release of GPT-5, and reasonable to surmise that at least some of that is because of people shifting their views a bit after being disappointed by the release.</p><p>This is strange! This isn&#8217;t the first time this has happened &#8212; we previously expressed skepticism of a big <a href="https://www.aisnakeoil.com/p/is-ai-progress-slowing-down">narrative shift</a> around scaling that happened based on almost no new information. If a single update to one product shifts people&#8217;s views on the trajectory of AI, how reliable is people&#8217;s evidence base to begin with?</p><p>The reason why the normal technology framework predicts slow timelines is <em>not</em> because capabilities will hit a wall but because impacts will be slow and gradual <em>even if</em> capabilities continue to advance rapidly. So we don&#8217;t think disappointment with a new release should make you more sympathetic to viewing AI as normal technology. By the same token, a new breakthrough announced tomorrow shouldn&#8217;t cause you to be more skeptical of our views.</p><p>The best way to understand GPT-5 is that it&#8217;s a particularly good example of AI developers&#8217; shift in emphasis from models to products, which we wrote about <a href="https://www.aisnakeoil.com/p/ai-companies-are-pivoting-from-creating">a year ago</a>. The automatic model switcher is a big deal for everyday users of ChatGPT. It turns out that <a href="https://x.com/sama/status/1954603417252532479">hardly anyone</a> was using &#8220;thinking&#8221; models nearly a year after they were first released, and GPT-5 has bumped up their use dramatically.</p><p>In <a href="https://x.com/sama/status/1889755723078443244">some</a> communications Altman was clear that the emphasis for GPT-5 was usability, not a leap in capabilities, although this message was unfortunately undercut by the constant hype, leading to disappointment.</p><p>This broader shift in the industry is actually highly consistent with companies themselves (reluctantly) coming around to acknowledging the possibility that their path to success is to do the hard work of building products and fostering adoption, rather than racing to build AGI or superintelligence and count on it to sweep away any diffusion barriers. Ironically, in this narrative, GPT-5 is an example of a success, not a failure.</p><p>In fact, model developers are starting to go beyond developing more useful products (the second stage of our technology development &amp; adoption framework) and working with deployers to ease early adoption pains (the third stage). For example, OpenAI&#8217;s <a href="https://newsletter.pragmaticengineer.com/p/forward-deployed-engineers">Forward Deployed Engineers</a> work with customers such as John Deere, and directly with farmers, on integrating and deploying capabilities such as providing personalized recommendations for pesticide application.</p><h2><strong>Why it&#8217;s hard to find a &#8220;middle ground&#8221; between AI as Normal Technology and AI 2027</strong></h2><p>Many people have tried to articulate middle ground positions between <a href="https://ai-2027.com/">AI 2027</a> and AI as Normal Technology, perhaps viewing these as two ends of a spectrum of views.</p><p>This is surprisingly hard to do. Both AI 2027 and AI as Normal Technology are coherent worldviews. They represent very different causal understandings of how technology will impact society. If you try to mix and match, there is a risk that you end up with an internally inconsistent hodgepodge. (By the way, this means that if we end up being wrong, it is more likely that we will be wrong wholesale than slightly wrong.)</p><p>Besides, only in the Silicon Valley bubble can AI as Normal Technology be considered a skeptical view! We compare AI to electricity in the second sentence of the essay, and we make clear throughout that we expect it to have profound impacts. Our expectations for AI&#8217;s impact on labor seem to be at the more aggressive end of the range of expectations from economists who work on this topic.</p><p>In short, if you are looking for a moderate position, we encourage you to read the essay in full. Don&#8217;t let the title fool you into thinking we are AI skeptics. Perhaps you will conclude that AI as Normal Technology is already the middle ground you are looking for.</p><p>We realize that it can be discomfiting that the two most widely discussed frameworks for thinking about the future of AI are so radically different. (Our essay itself offers much commentary on this state of affairs in Part 4, which is about policy.) We can offer a few comforting thoughts:</p><ul><li><p>We do have many areas of agreement with the AI 2027 authors. We are working on a joint statement outlining those areas. We are grateful to Nicholas Carlini for organizing this effort.</p></li><li><p>In our view, more important than agreement in beliefs are areas of <a href="https://x.com/random_walker/status/1913334506318446877">common ground in policy</a> <em>despite</em> differences in beliefs. Even relatively &#8220;easy&#8221; policy interventions that different sides can agree on will be a huge challenge in practice. If we can&#8217;t achieve these, there is little hope for the much more radical measures favored by those worried about imminent superintelligence.</p></li><li><p>There have been a few ongoing efforts to identify the cruxes of disagreement and agree on indicators that might help adjudicate between the two worldviews. We have participated in a few of these efforts and look forward to continuing to do so. We are grateful to the Golden Gate Institute for AI&#8217;s efforts on this front.</p></li></ul><ul><li><p>Speaking of developing indicators, we are in the process of expanding the vision for our project HAL, <a href="https://hal.cs.princeton.edu/">Holistic Agent Leaderboard</a>. Currently it tries to be a better benchmark orchestration system for AI agents, but the new plan is to develop it into an early warning system that helps the AI community identify when AI agents have crossed capability thresholds for transformative real-world impacts in various domains.<br>We see these capability thresholds as necessary but not always sufficient conditions for impact, and as and when they are reached, they will much more acutely stress our theses about non-technological barriers to both benefits and risks.</p></li><li><p>Note that HAL is not about prediction; it is about situational awareness of the <em>present</em>. This is a theme of our work. What is remarkable about the AI discourse in general, and us versus AI 2027 in particular, is the wide range of views not just about the future but about the things we <em>can</em> observe, such as the speed of diffusion (more on that below). Unless we as a community get much better at measurement of the present and <a href="https://x.com/sayashk/status/1964016339690909847">testing competing causal explanations of progress</a>, the level of energy directed at prediction will be misdirected, because we lack ways of resolving those predictions. For example, we&#8217;ve argued that we won&#8217;t necessarily know if &#8220;AGI&#8221; has been built <a href="https://www.aisnakeoil.com/p/agi-is-not-a-milestone">even post facto</a>. To some extent these limitations are intrinsic because of the lack of conceptual precision of ideas like AGI, but at the same time it&#8217;s true that we can do a lot better at measurement.</p></li></ul><h2><strong>It is hard to understand one worldview when you&#8217;re committed to another</strong></h2><p>We wrote:</p><blockquote><p>AI as normal technology is a worldview that stands in contrast to the worldview of AI as impending superintelligence. Worldviews are constituted by their assumptions, vocabulary, interpretations of evidence, epistemic tools, predictions, and (possibly) values. These factors reinforce each other and form a tight bundle within each worldview.</p></blockquote><p>This makes communication across worldviews hard. For example, one question we often receive from the AI 2027 folks is what we think the world will look like in 2027. Well, pretty much like the world in 2025, we respond. They then push us to consider 2035 or 2045 or whatever year by which the world will be transformed, and they consider it a deficiency of our framework that we don&#8217;t provide concrete scenarios.</p><p>But this kind of scenario forecasting is only a meaningful activity within their worldview. We are concrete about the things we think we <em>can</em> be concrete about. At the same time, we emphasize the role of human, institutional, and political agency in making radically different futures possible &#8212; including AI 2027. Thus, AI as normal technology is as much a prescription as a prediction.</p><p>These communication difficulties are important to keep in mind when considering the <a href="https://blog.ai-futures.org/p/ai-as-profoundly-abnormal-technology">response</a> by Scott Alexander, one of the AI 2027 authors, to AI as Normal Technology. While we have no doubt that it is a good-faith effort at dialogue and we appreciate his putting in the time, unfortunately we feel that his response mostly talks past us. What he identifies as the cruxes of disagreement are quite different from what we consider the cruxes! For this reason, we won&#8217;t give a point-by-point response, since we will probably in turn end up talking past him in turn.</p><p>But we would be happy to engage in moderated conversations, a format with which we&#8217;ve had good success and have engaged in 8-10 times over the past year. The synchronous nature makes it much easier to understand each other. And the fact that the private conversation will be edited before making it public makes it easier to ask stupid questions as each side searches for understanding of the other&#8217;s point of view.</p><p>Anyway, here are a couple of important ways in which Alexander&#8217;s response talks past us. Recursive Self-Improvement (RSI) is a crux of disagreement for Alexander&#8217;s point of view, and he is surprised that it is barely worth a mention for us. In fairness we could have been much more explicit in our essay about what we think about RSI. In short, we don&#8217;t think RSI will lead to superintelligence because the external bottlenecks to building and deploying powerful AI systems cannot be overcome by improving their technical design. That is why we don&#8217;t discuss it much.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-3" href="#footnote-3" target="_self">3</a></p><p>Although it is not a crux for us, we do explain in the essay why we think the AI community is nowhere close to RSI. More recently we&#8217;ve been thinking about the fundamental research challenges that need to be solved, and there are a lot more than we&#8217;d realized. And it is worth keeping in mind that the AI community might be <a href="https://www.linkedin.com/posts/randomwalker_video-in-a-talk-at-the-neuro-symbolic-ai-activity-7367182131556528128-rfeZ?utm_source=share&amp;utm_medium=member_desktop&amp;rcm=ACoAAD85EvsBeNijbxJLxmNZcA4cF5Gc0JEwrc4">particularly bad</a> at finding new paradigms for progress compared to other scientific communities. Again, this is an area where we hope that our project <a href="https://hal.cs.princeton.edu/">HAL</a> can play a role in measuring progress.</p><p>Another topic where Alexander&#8217;s response talks past us is on the speed of diffusion, which we comment on briefly below and will address in more detail in a future essay.</p><p>The best illustration of the difficulty of discourse across worldviews is Alexander&#8217;s discussion of our hypotheses about whether or not superhuman AI abilities are possible in tasks such as prediction or persuasion. After reading his response several times, it is hard for us to figure out where exactly we agree and disagree. We wrote:</p><blockquote><p>We think there are relatively few real-world cognitive tasks in which human limitations are so telling that AI is able to blow past human performance (as AI does in chess). ... Concretely, we propose two such areas: forecasting and persuasion. We predict that AI will not be able to meaningfully outperform trained humans (particularly teams of humans and especially if augmented with simple automated tools) at forecasting geopolitical events (say elections). We make the same prediction for the task of persuading people to act against their own self-interest.</p></blockquote><p>You can read his full <a href="https://blog.ai-futures.org/p/ai-as-profoundly-abnormal-technology">response</a> to this in Section 3B of his essay, but in short it focuses on human biological limits:</p><blockquote><p>Humans gained their abilities through thousands of years of evolution in the African savanna. There was no particular pressure in the savanna for &#8220;get exactly the highest Brier score possible in a forecasting contest&#8221;, and there is no particular reason to think humans achieved this. Indeed, if the evidence for human evolution for higher intelligence in the past 10,000 years in response to agriculture proves true, humans definitely didn&#8217;t reach the cosmic maximum on the African savannah. Why should we think this last, very short round of selection got it exactly right?</p></blockquote><p>But rejecting a biological conception of human abilities is a key point of departure for us, something we take pains to describe in detail in the section &#8220;Human Ability Is Not Constrained by Biology&#8221;. That&#8217;s the problem with discussion across worldviews: If you take a specific statement, ignore the premises and terminological clarifications leading up to it, and interpret it in <em>your</em> worldview, it will seem like your opponent is clueless. Does Alexander think we are suggesting that if a savanna-dweller time traveled to the present, they would be able to predict elections?</p><p>He does emphasize that human performance is not fixed, but somehow sees this as a refutation of our thesis (rather than central to it). Perhaps the confusion arose because of our hypothesis that human performance at forecasting is close to the &#8203;&#8203;&#8220;irreducible error&#8221;. We don&#8217;t imply that the irreducible error of forecasting is a specific number that is fixed for all time. Of course it depends on the data that is available &#8212; better polling leads to better forecasting &#8212; and training that helps take advantage of increased data. And some of that training might even be the result of AI-enabled research on forecasting. We emphasize in our original essay that human intelligence is special not because of our biology, but because of our (contingent) mastery of our tools including AI. Thus, advances in AI will often improve <em>human</em> intelligence (abilities), and have the potential to improve the performance of <em>both</em> sides of the human-AI comparison we propose.</p><p>The point of our hypothesis is a simple one: We don&#8217;t think forecasting is like chess, where loads of computation can give AI a decisive speed advantage. The computational structure of forecasting is relatively straightforward, even though performance can be vastly improved through training and data. Thus, relatively simple computational tools in the hands of suitably trained teams of expert forecasters can squeeze (almost) all the juice there is to squeeze.</p><p>We are glad that Alexander&#8217;s response credits us with &#8220;putting their money where their mouth is on the possibility of mutual cooperation&#8221;. The sentiment is mutual. We look forward to continuing that cooperation, which we see as more productive than Substack rebuttals and counter-rebuttals.</p><h2><strong>Reaping AI&#8217;s benefits will require hard work and painful choices</strong></h2><p>There are two broad sets of implications of our framework: one for the economy and labor, and the other for safety. Once we get past a few basic premises (notably, that superintelligence is incoherent or impossible depending on how it is defined), our arguments behind these two sets of implications are largely different.</p><p>On economic impacts, our case is broadly that diffusion barriers won&#8217;t be overcome through capability improvement. As for safety, our case is primarily that achieving AI control without alignment is not only possible, it doesn&#8217;t even seem particularly hard, and doesn&#8217;t require scientific breakthroughs.</p><p>Since these two sets of arguments don&#8217;t overlap much, it is coherent to accept one set but not accept (or be ambivalent to) the other. Indeed, our view of the economic impacts seems to have resonated particularly strongly with readers. Since the essay&#8217;s publication, we have had many discussions with people responsible for AI strategy in various industries. We discovered that the way that they had been thinking about AI was consistent with ours, but they were starting to second-guess their approach because of all the hype. Our essay provided a coherent framework that backed up their intuitions as well as their observations in the trenches.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-4" href="#footnote-4" target="_self">4</a></p><p>While people deploying AI have a keen understanding of the difference between technology development and diffusion, our framework further divides each of those into two steps. On the development side, we emphasize the <a href="https://www.normaltech.ai/p/ai-companies-are-pivoting-from-creating">gap between models and products</a>, or capabilities and applications. On the diffusion side, we differentiate between user learning curves and other aspects of adaptation by <em>individuals</em>, and the structural, organizational, or legal changes that might be necessary, which often require <em>collective</em> action. We illustrate the kinds of speed limits that operate at each of the four stages.</p><p>While user behaviors at least tend to change slowly but predictably, solving coordination problems or reforming sclerotic institutions &#8212; which are also prerequisites for effective technology adoption &#8212; are much more uncertain. As an example, consider how Air Traffic Control is stuck with technology from the middle of the 20th century despite the enormous costs of the lack of modernization becoming apparent.</p><p>While our essay pointed out that analogous diffusion barriers exist in the case of AI, we are only now doing the work of spelling out those barriers and identifying specific reforms that might be necessary. We will be writing more on this front, some of it in collaboration with Justin Curl.</p><p>It is worth bearing in mind that advanced AI is entering a world that is already both highly technological and highly regulated. We repeatedly find that the parts of workflows that AI tackles are unlikely to be bottlenecks, because much of the available productivity gains have already been unlocked through earlier waves of technologies. Meanwhile the actual bottlenecks prove resistant due to regulation or other external constraints. In many specific domains including <a href="https://www.aeaweb.org/articles?id=10.1257/jel.20201330">legal services</a> and <a href="https://www.normaltech.ai/p/could-ai-slow-science">scientific research</a>, competitive dynamics are so strong that productivity gains from AI lead to escalation of arms races that don&#8217;t ultimately translate to societal value.</p><h2><strong>The surreal debate about the speed of diffusion</strong></h2><p>We&#8217;ve mentioned a few times that different camps disagree on how they characterize <em>current</em> AI impacts. Nowhere is this more apparent than the speed of diffusion. AI boosters believe that it is being adopted unprecedentedly rapidly. We completely disagree. Worse, as more evidence comes out, each side seems to be getting more certain of their interpretation.</p><p>We are working on an in-depth analysis of the speed of diffusion. For now, we point out a few basic fallacies in the common arguments and stats that get trotted out to justify the &#8220;rapid adoption&#8221; interpretation.</p><p>First, deployment is not diffusion. Often, when people talk about rapid adoption they just mean that when capabilities are developed, they can be near-instantly deployed to products (such as chatbots) that are used by hundreds of millions of users.</p><p>But this is not what diffusion means. It is not enough to know how many people have access to capabilities: What matters is how many people are actually using them, how long they are using them, and what they are using them for. When we drill down into those details, the picture looks very different.</p><p>For example, almost a year after the vaunted release of &#8220;thinking&#8221; models in ChatGPT, <a href="https://x.com/sama/status/1954603417252532479">less than 1%</a> of users used them on any given day! We find no pleasure in pointing this out even though it supports our thesis. As enthusiastic early adopters of AI, this kind of number is so low that it is hard for us to intuitively grasp, and frankly pretty depressing.</p><p>Another example of a misleading statistic relates to the fraction of workers in certain high-risk domains who use AI. Such statistics tend to be offered in service of the claim that AI is being rapidly adopted in risky ways. But even in high risk domains most tasks are actually <a href="https://dho.stanford.edu/wp-content/uploads/Spectrum_AI.pdf">mundane</a>, and when we dig into the specific uses don&#8217;t seem that risky at all.</p><p>For example, a survey by the American Medical Association reported that a majority of <a href="https://www.ama-assn.org/system/files/physician-ai-sentiment-report.pdf">doctors</a> are using AI. But this includes things like transcription of dictated notes.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-5" href="#footnote-5" target="_self">5</a> It also includes things like asking a chatbot for a second opinion of a diagnosis (about 12% reported this use case in 2024, a whopping 1 percentage point increase from 11% in 2023). This is definitely a more serious use than transcription one, but it is still well-founded. As we&#8217;ve pointed out before, even unreliable AI is very helpful for <a href="https://journalcrcl.org/crcl/article/view/62/28">error</a> <a href="https://www.linkedin.com/posts/randomwalker_error-checking-is-a-great-application-of-activity-7341077466800676866-eFft/">detection</a>.</p><p>Increasing adoption of AI for these tasks does <em>not</em> mean that doctors are about to start YOLOing it and abdicating their responsibility to their patients by delegating their decisions to ChatGPT. The vast majority of doctors understand the difference between these two types of uses, and there are many overlapping guardrails preventing widespread irresponsible use in the medical profession, including malpractice liability, professional codes, and regulation of medical devices.</p><p>The most misleading &#8220;rapid adoption&#8221; meme of all might be this widely shared chart, showing that ChatGPT reached 100M users in about two months:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!MLSh!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd743ae91-ddc4-46f1-a03c-a938b9e9a148_1600x580.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!MLSh!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd743ae91-ddc4-46f1-a03c-a938b9e9a148_1600x580.png 424w, https://substackcdn.com/image/fetch/$s_!MLSh!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd743ae91-ddc4-46f1-a03c-a938b9e9a148_1600x580.png 848w, https://substackcdn.com/image/fetch/$s_!MLSh!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd743ae91-ddc4-46f1-a03c-a938b9e9a148_1600x580.png 1272w, https://substackcdn.com/image/fetch/$s_!MLSh!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd743ae91-ddc4-46f1-a03c-a938b9e9a148_1600x580.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!MLSh!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd743ae91-ddc4-46f1-a03c-a938b9e9a148_1600x580.png" width="1456" height="528" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/d743ae91-ddc4-46f1-a03c-a938b9e9a148_1600x580.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:528,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!MLSh!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd743ae91-ddc4-46f1-a03c-a938b9e9a148_1600x580.png 424w, https://substackcdn.com/image/fetch/$s_!MLSh!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd743ae91-ddc4-46f1-a03c-a938b9e9a148_1600x580.png 848w, https://substackcdn.com/image/fetch/$s_!MLSh!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd743ae91-ddc4-46f1-a03c-a938b9e9a148_1600x580.png 1272w, https://substackcdn.com/image/fetch/$s_!MLSh!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd743ae91-ddc4-46f1-a03c-a938b9e9a148_1600x580.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>It compares ChatGPT user growth with (1) Instagram, Facebook, and Twitter, which are social media apps whose usefulness depends on network effects, and therefore characteristically grow much slower than apps that are useful from day one (2) Spotify, an app that was initially invite-only and (3) Netflix, a service that launched with a limited inventory and required a subscription to use.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-6" href="#footnote-6" target="_self">6</a></p><p>What is reflected in this chart are early adopters who will check out an app if there&#8217;s buzz around it, and there was deafening buzz around ChatGPT. Once you exhaust these curious early users, the growth curve looks very different. In fact, a year later, ChatGPT had apparently only grown <a href="https://www.theverge.com/2024/8/29/24231685/openai-chatgpt-200-million-weekly-users">from 100M to 200M</a> users, which meant that the curve evidently bent sharply to the right. That is conveniently not captured in this graph which reflects only the first two months.</p><p>This chart would be useful if it gave us any evidence that the usual barriers to diffusion have been weakened or eliminated. It doesn&#8217;t. Two months is not enough time for the hard parts of diffusion to even get started, such as users adapting their workflows to productively incorporate AI. As such, this chart is irrelevant to any meaningful discussion of the speed of diffusion.</p><p>There are many other problems with this chart, but we&#8217;ll stop here.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-7" href="#footnote-7" target="_self">7</a> Again, this is far from a complete analysis of the speed of AI diffusion &#8212; that&#8217;s coming. For now, we&#8217;re just making the point that the majority of the commentary on this topic is simply unserious. And if this is what the discourse is like on a question for which we do have data, it is no surprise that <em>predictions</em> of the future from different camps bear no resemblance to each other.</p><h2><strong>Why AI adoption hits different</strong></h2><p>If the &#8220;rapid diffusion&#8221; meme is so wrong, why is it so pervasive and persistent? Because AI adoption <em>feels</em> like a tsunami in a way that the PC or the internet or social media never did. When people are intuitively convinced of something, they will be much less skeptical of data or charts that purport to confirm that feeling.</p><p>We recognize the feeling, of course. Our own lived experience of AI is different from past waves of technologies. Initially, we dismissed this as a cognitive bias. Whatever change we&#8217;re living through at the moment will feel like a much bigger shift than something we successfully adapted to in the past.</p><p>We now realize that we were wrong. The cognitive bias might be a small part of the explanation, but there is a genuine reason why AI adoption feels much more rapid and scary. In short, while it&#8217;s true that deployment is not diffusion, in the past, gradual deployment meant that users were somewhat buffered from having to constantly make decisions about adoption, but now that buffer has been swept away. Let&#8217;s explain with a comparison to internet adoption.</p><p>Those of us who adopted dial-up internet in the 90s will remember a story that went something like this. When we first heard about the tech, we were put off by the high price of a PC. Gradually those prices came down. Meanwhile we got some experience using the internet at work or at a friend&#8217;s house. So when we bought a PC and dialup internet a few years later, we already had some training. At first, dialup was slow and expensive and there weren&#8217;t even that many websites, so we didn&#8217;t use the internet that much. Gradually prices came down, bandwidth improved, and more content came online, and we learned how to use the internet productively and safely in tandem with our increasing use.</p><p>Adopting general-purpose AI tools in the 2020s is a radically different experience because deployment of new capabilities is instantaneous. People have to spend a much higher percentage of time evaluating whether to adopt AI for some particular use case, and are constantly being told that if they don&#8217;t adopt it they will be left behind.</p><p>All our earlier points stand &#8212; learning curves exist, human behavior takes a long time to change, and organizational change even longer. But not using AI is somewhat of an active choice, and people no longer have the excuse of not to think about it because they don&#8217;t yet have access.</p><p>In short, deployment is only one of many steps in diffusion, and removing that bottleneck probably made diffusion slightly faster. But it <em>feels</em> dramatically faster because as soon as one hears about a particular AI use case, one has to decide whether to adopt it or not, even if it is ultimately the case that the vast majority of the time, people are deciding not to, for various reasons that might be rational or irrational.</p><h2><strong>Concluding thoughts</strong></h2><p>One thing on which we definitely agree with AI boosters is that AI is not going away, nor will it become a niche like crypto that most people can ignore. Now that the collective initial shock of generative AI has worn off, there&#8217;s a need for structured ways to think about how AI&#8217;s impacts might unfold, instead of (over)reacting to each new technical capability or emergent social effect.</p><p>The AI-as-normal-technology framework &#8212; which we continue to elaborate in this newsletter &#8212; is one such approach. It is worth being familiar with, at least as an articulation of a historically grounded default way to think about tech&#8217;s societal impact, against which more exceptionalist accounts can be compared. The framework has some degree of actionable guidance for business leaders, workers, students, people concerned about AI safety or AI ethics, and policymakers, among others. We hope you follow along and contribute to the discussion.</p><p><em>We are grateful to Steve Newman and Felix Chen for feedback on a draft.</em></p><h2><strong>Further reading/viewing</strong></h2><ul><li><p>Arvind <a href="https://www.youtube.com/live/CriEhgw6f0k?si=sHp4XIrJ0SY7vKSc&amp;t=19971">presented</a> the paper at the World Bank Development Conference, focusing on the economic and labor implications.</p></li><li><p><a href="https://www.youtube.com/@ArvindOnAI">Arvind&#8217;s new YouTube channel</a> discusses AI developments based on the normal technology perspective.</p></li><li><p>We were fortunate to receive press coverage for AI as Normal Technology:</p><ul><li><p>In the last month, the New York Times had three op-eds that discussed the essay in depth, by<a href="https://www.nytimes.com/2025/08/19/opinion/artificial-general-intelligence-superintelligence.html"> Eric Schmidt and Selina Xu</a>, by<a href="https://www.nytimes.com/2025/08/20/opinion/ai-technology-chatgpt.html"> David Wallace-Wells</a>, and by<a href="https://www.nytimes.com/2025/08/24/opinion/chat-gpt5-open-ai-future.html"> Ezra Klein</a>.</p></li><li><p>Last week, the Economist discussed the normal technology thesis in the article "<a href="https://www.economist.com/finance-and-economics/2025/09/04/what-if-artificial-intelligence-is-just-a-normal-technology">What if artificial intelligence is just a &#8220;normal&#8221; technology?</a>"</p></li><li><p>In The New Yorker,<a href="https://www.newyorker.com/culture/open-questions/two-paths-for-ai"> Joshua Rothman</a> contrasted AI 2027 with AI as Normal Technology.</p></li><li><p>In Prospect Magazine, <a href="https://www.prospectmagazine.co.uk/ideas/technology/70046/ai-isnt-magic.-its-just-weird">Ethan Zuckerman</a> discussed the utility of viewing AI as Normal Technology.</p></li><li><p>James O'Donnell, for the MIT Technology Review, <a href="https://www.technologyreview.com/2025/04/29/1115928/is-ai-normal/">summarized</a> the AI as Normal Technology thesis.</p></li></ul></li><li><p>Some of our podcast appearances include: Arvind on New York Times' <a href="https://podcasts.apple.com/us/podcast/meta-on-trial-is-a-i-a-normal-technology-hatgpt/id1528594034?i=1000703985596">Hard Fork</a>, and <a href="https://www.oreilly.com/radar/is-ai-a-normal-technology/">Tim O&#8217;Reilly&#8217;s podcast</a>; Sayash on Lawfare's <a href="https://www.youtube.com/watch?v=mpmi-l-G-wI">Scaling Laws</a> and Carnegie's <a href="https://carnegieendowment.org/podcasts/interpreting-india/beyond-superintelligence-a-realists-guide-to-ai?utm_source=chatgpt.com">Interpreting India</a>.</p></li><li><p>We have had conversations with many people who have the superintelligence worldview, including many of the authors of AI 2027: Sayash debated <a href="https://www.youtube.com/watch?v=rVFAJQryzk8">Daniel Kokotajlo</a> and <a href="https://x.com/i/broadcasts/1mrGmPeorrNKy">Eli Lifland</a>; Arvind debated <a href="https://www.youtube.com/watch?v=2hby15Z3uXA">Daniel Kokotajlo</a>, and in<a href="https://asteriskmag.com/issues/10/does-ai-progress-have-a-speed-limit?utm_source=chatgpt.com"> Asterisk Mag</a>, Arvind and Ajeya Cotra debated whether AI progress has a speed limit, and how we&#8217;d know.</p></li></ul><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>As is so often the case, fortuitous timing played a big role in the success of the essay. It was released two weeks after AI 2027, but this was purely by coincidence &#8212; our publication date was actually based on the Knight Institute symposium on <a href="https://knightcolumbia.org/events/artificial-intelligence-and-democratic-freedoms">AI and Democratic Freedoms</a>. We are grateful to the Institute for the opportunity to publish the essay.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p>Mental health issues with chatbots in general, and problems such as addiction, have been widely recognized and discussed, including in our book <a href="https://www.amazon.com/Snake-Oil-Artificial-Intelligence-Difference/dp/069124913X">AI Snake Oil</a>. These tend to be based on an analogy with social media. But it's one thing to anticipate the potential for mental health impacts, and another to predict specifically what impacts might emerge and how to avoid them.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-3" href="#footnote-anchor-3" class="footnote-number" contenteditable="false" target="_self">3</a><div class="footnote-content"><p>That said, we acknowledge that our whole thesis might be wrong, and it is more likely that we&#8217;re wrong if RSI is achieved.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-4" href="#footnote-anchor-4" class="footnote-number" contenteditable="false" target="_self">4</a><div class="footnote-content"><p>The framework is adapted from the <a href="https://www.simonandschuster.com/books/Diffusion-of-Innovations-5th-Edition/Everett-M-Rogers/9780743258234">classic diffusion-of-innovations theory</a> and also influenced by recent writers such as Jeffrey Ding who have <a href="https://press.princeton.edu/books/paperback/9780691260341/technology-and-the-rise-of-great-powers">analyzed</a> geopolitical competition in AI through the lens of the innovation-diffusion gap.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-5" href="#footnote-anchor-5" class="footnote-number" contenteditable="false" target="_self">5</a><div class="footnote-content"><p>While there are risks even here, we think it is definitely an application that doctors should be exploring.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-6" href="#footnote-anchor-6" class="footnote-number" contenteditable="false" target="_self">6</a><div class="footnote-content"><p>In fact, there have been other apps whose initial growth was as fast or faster than ChatGPT, such as Pokemon Go and Threads (which bootstrapped off of Instagram and thus wasn&#8217;t reliant on network effects). But again, our bigger point is that this type of comparison is not informative. Threads ended up being something of a dud despite that initial growth.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-7" href="#footnote-anchor-7" class="footnote-number" contenteditable="false" target="_self">7</a><div class="footnote-content"><p>Frankly, the far more impressive stat in this graph, in our view, is Instagram getting to a million users in only 2.5 months despite the need for network effects &#8212; considering that it was back in 2010 when phone internet speeds were much lower, the app was iPhone only (!), and mainly spread among 18-34 year olds in the United States in the early days.</p></div></div>]]></content:encoded></item><item><title><![CDATA[Could AI slow science?]]></title><description><![CDATA[Confronting the production-progress paradox]]></description><link>https://www.normaltech.ai/p/could-ai-slow-science</link><guid isPermaLink="false">https://www.normaltech.ai/p/could-ai-slow-science</guid><dc:creator><![CDATA[Sayash Kapoor]]></dc:creator><pubDate>Wed, 16 Jul 2025 21:35:56 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/177c9c7a-89ac-40e2-9493-34667f829d92_4086x2296.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>AI leaders have predicted that it will enable <a href="https://www.darioamodei.com/essay/machines-of-loving-grace">dramatic</a> scientific <a href="https://blog.samaltman.com/the-gentle-singularity">progress</a>: curing cancer, doubling the human lifespan, colonizing space, and achieving a century of progress in the next decade. Given the cuts to federal funding for science in the U.S., the timing seems perfect, as AI could replace the need for a large scientific workforce.</p><p>It&#8217;s a common-sense view, at least among technologists, that AI will speed science greatly as it gets adopted in every part of the scientific pipeline &#8212; summarizing existing literature, generating new ideas, performing data analyses and experiments to test them, writing up findings, and performing &#8220;peer&#8221; review.</p><p>But many early common-sense predictions about the impact of a new technology on an existing institution proved badly wrong. The Catholic Church welcomed the printing press as a way of solidifying its authority by printing Bibles. The early days of social media led to wide-eyed optimism about the spread of democracy worldwide following the Arab Spring.</p><p>Similarly, the impact of AI on science could be counterintuitive. Even if individual scientists benefit from adopting AI, it doesn&#8217;t mean science as a whole will benefit. When thinking about the macro effects, we are dealing with a complex system with emergent properties. That system behaves in surprising ways because it is not a market. It is better than markets at some things, like rewarding truth, but worse at others, such as reacting to technological shocks. So far, on balance, AI has been <a href="https://www.nature.com/articles/d41586-025-02241-2">an</a> <a href="https://reproducible.cs.princeton.edu/">unhealthy</a> <a href="https://www.understandingai.org/p/i-got-fooled-by-ai-for-science-hypeheres">shock</a> to science, stretching many of its processes to the breaking point.</p><p>Any serious attempt to forecast the impact of AI on science must confront the <strong>production-progress paradox</strong>. The rate of publication of scientific papers has been growing exponentially, increasing 500 fold between 1900 and 2015. But actual progress, by any available measure, has been constant or even <em>slowing</em>. So we must ask how AI is impacting, and will impact, the factors that have led to this disconnect.</p><p>Our analysis in this essay suggests that AI is likely to worsen the gap. This may not be true in all scientific fields, and it is certainly not a foregone conclusion. By carefully and urgently taking actions such as those we suggest below, it may be possible to reverse course. Unfortunately, AI companies, science funders, and policy makers all seem oblivious to what the actual bottlenecks to scientific progress are. They are simply trying to accelerate production, which is like adding lanes to a highway when the slowdown is actually caused by a toll booth. It&#8217;s sure to make things worse.</p><h4><strong>Table of contents</strong></h4><p><a href="https://www.aisnakeoil.com/i/168505690/science-has-been-slowing-the-production-progress-paradox">1. Science has been slowing &#8212; the production-progress paradox</a></p><p><a href="https://www.aisnakeoil.com/i/168505690/why-is-progress-slowing-can-ai-help">2. Why is progress slowing? Can AI help?</a></p><p><a href="https://www.aisnakeoil.com/i/168505690/science-is-not-ready-for-software-let-alone-ai">3. Science is not ready for software, let alone AI</a></p><p><a href="https://www.aisnakeoil.com/i/168505690/ai-might-prolong-the-reliance-on-flawed-theories">4. AI might prolong the reliance on flawed theories</a></p><p><a href="https://www.aisnakeoil.com/i/168505690/human-understanding-remains-essential">5. Human understanding remains essential</a></p><p><a href="https://www.aisnakeoil.com/i/168505690/implications-for-the-future-of-science">6. Implications for the future of science</a></p><p><a href="https://www.aisnakeoil.com/i/168505690/final-thoughts">7. Final thoughts</a></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.normaltech.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption"> Analysis and commentary based on the &#8220;<em>AI as Normal Technology&#8221; </em>framework</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p></p><h3><strong>Science has been slowing &#8212; the production-progress paradox</strong></h3><p>The total number of published papers is increasing <em>exponentially, </em><a href="https://dl.acm.org/doi/pdf/10.1145/3097983.3098016">doubling</a> every 12 years. The total number of researchers who have authored a research paper is increasing even more quickly. And between 2000 and 2021, investment in research and development&nbsp;<a href="https://ncses.nsf.gov/pubs/nsb20246/cross-national-comparisons-of-r-d-performance">increased</a>&nbsp;fourfold across the top seven funders (the US, China, Japan, Germany, South Korea, the UK, and France).<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a></p><p>But does this mean faster progress? Not necessarily. Some papers lead to fundamental breakthroughs that change the trajectory of science, while others make minor improvements to known results.</p><p>Genuine progress results from breakthroughs in our understanding. For example, we understood plate tectonics in the middle of the last century &#8212; the idea that the continents move. Before that, geologists weren&#8217;t even able to ask the right <em>questions</em>. They tried to figure out the effects of the cooling of the Earth, believing that that&#8217;s what led to geological features such as mountains. No amount of findings or papers in older paradigms of geology would have led to the same progress that plate tectonics did.</p><p>So it is possible that the number of papers is increasing exponentially while progress is not increasing at the same rate, or is even slowing down. How can we tell if this is the case?</p><p>One challenge in answering this question is that, unlike the production of research, progress does not have clear, objective metrics. Fortunately, an entire research field &#8212; the "<a href="https://www.science.org/doi/full/10.1126/science.aao0185">science of science</a>", or metascience &#8212; is trying to answer this question. Metascience uses the scientific method to study scientific research. It tackles questions like: How often can studies be replicated? What influences the quality of a researcher's work? How do incentives in academia affect scientific outcomes? How do different funding models for science affect progress? And how quickly is progress really happening?</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!WNiP!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4517c5ce-54a5-4d83-a04a-4ace96f4c1d4_1548x586.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!WNiP!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4517c5ce-54a5-4d83-a04a-4ace96f4c1d4_1548x586.jpeg 424w, https://substackcdn.com/image/fetch/$s_!WNiP!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4517c5ce-54a5-4d83-a04a-4ace96f4c1d4_1548x586.jpeg 848w, https://substackcdn.com/image/fetch/$s_!WNiP!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4517c5ce-54a5-4d83-a04a-4ace96f4c1d4_1548x586.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!WNiP!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4517c5ce-54a5-4d83-a04a-4ace96f4c1d4_1548x586.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!WNiP!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4517c5ce-54a5-4d83-a04a-4ace96f4c1d4_1548x586.jpeg" width="690" height="261.1195054945055" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/4517c5ce-54a5-4d83-a04a-4ace96f4c1d4_1548x586.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:551,&quot;width&quot;:1456,&quot;resizeWidth&quot;:690,&quot;bytes&quot;:127562,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.aisnakeoil.com/i/168505690?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4517c5ce-54a5-4d83-a04a-4ace96f4c1d4_1548x586.jpeg&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!WNiP!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4517c5ce-54a5-4d83-a04a-4ace96f4c1d4_1548x586.jpeg 424w, https://substackcdn.com/image/fetch/$s_!WNiP!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4517c5ce-54a5-4d83-a04a-4ace96f4c1d4_1548x586.jpeg 848w, https://substackcdn.com/image/fetch/$s_!WNiP!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4517c5ce-54a5-4d83-a04a-4ace96f4c1d4_1548x586.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!WNiP!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4517c5ce-54a5-4d83-a04a-4ace96f4c1d4_1548x586.jpeg 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em>Left: The number of papers authored and authors of research papers have been increasing exponentially (from <a href="https://dl.acm.org/doi/pdf/10.1145/3097983.3098016">Dong et al.</a>, redrawn to linear scale using a <a href="https://automeris.io/wpd/">web plot digitizer</a>). Right: The disruptiveness of papers is declining over time (from <a href="https://www.nature.com/articles/s41586-022-05543-x">Park et al.</a>).</em></figcaption></figure></div><p>Strikingly, many findings from metascience suggest that progress has been slowing down, despite dramatic increases in funding, the number of papers published, and the number of people who author scientific papers. We collect some evidence below; Matt Clancy <a href="https://mattsclancy.substack.com/p/science-is-getting-harder">reviews</a> many of these findings in much more depth.</p><p>1) Park et al. find that "disruptive" scientific work <a href="https://www.nature.com/articles/s41586-022-05543-x">represents</a> an ever-smaller fraction of total scientific output. Despite an exponential increase in the number of published papers and patents, the number of breakthroughs is roughly constant.</p><p>2) Research that introduces new ideas is more likely to coin new terms. <a href="https://www.oecd.org/en/publications/artificial-intelligence-in-science_a8d820bd-en.html">Milojevic</a> collects the number of unique phrases used in titles of scientific papers over time as a measure of the &#8220;cognitive extent&#8221; of science, and finds that while this metric increased up until the early 2000s, it has since entered a period of stagnation, when the number of unique phrases used in titles of research papers has gone <em>down</em>.</p><p>3) Patrick Collison and Michael Nielsen <a href="https://www.theatlantic.com/science/archive/2018/11/diminishing-returns-science/575665/">surveyed</a> researchers across fields on how they perceived progress in the most important breakthroughs in their fields over time &#8212; those that won a Nobel prize. They asked scientists to compare Nobel-prize-winning research from the 1910s to the 1980s.</p><p>They found that scientists considered advances from earlier decades to be roughly as important as the ones from more recent decades, across Medicine, Physics, and Chemistry. Despite the vast increases in funding, published papers, and authors, the most important breakthroughs today are about as impressive as those in the decades past.</p><p>4) Matt Clancy <a href="https://mattsclancy.substack.com/p/science-is-getting-harder">complements</a> this with an analysis of what fraction of discoveries that won a Nobel Prize in a given year were published in the preceding 20 years. He found that this number dropped from 90% in 1970 to 50% in 2015, suggesting that either transformative discoveries are happening at a slower pace, or that it takes longer for discoveries to be recognized as transformative.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!I7YF!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9bb6f425-3afd-4fa8-9605-6134e149c295_1300x856.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!I7YF!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9bb6f425-3afd-4fa8-9605-6134e149c295_1300x856.jpeg 424w, https://substackcdn.com/image/fetch/$s_!I7YF!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9bb6f425-3afd-4fa8-9605-6134e149c295_1300x856.jpeg 848w, https://substackcdn.com/image/fetch/$s_!I7YF!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9bb6f425-3afd-4fa8-9605-6134e149c295_1300x856.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!I7YF!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9bb6f425-3afd-4fa8-9605-6134e149c295_1300x856.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!I7YF!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9bb6f425-3afd-4fa8-9605-6134e149c295_1300x856.jpeg" width="530" height="348.9846153846154" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/9bb6f425-3afd-4fa8-9605-6134e149c295_1300x856.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:856,&quot;width&quot;:1300,&quot;resizeWidth&quot;:530,&quot;bytes&quot;:118290,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.aisnakeoil.com/i/168505690?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9bb6f425-3afd-4fa8-9605-6134e149c295_1300x856.jpeg&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!I7YF!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9bb6f425-3afd-4fa8-9605-6134e149c295_1300x856.jpeg 424w, https://substackcdn.com/image/fetch/$s_!I7YF!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9bb6f425-3afd-4fa8-9605-6134e149c295_1300x856.jpeg 848w, https://substackcdn.com/image/fetch/$s_!I7YF!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9bb6f425-3afd-4fa8-9605-6134e149c295_1300x856.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!I7YF!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9bb6f425-3afd-4fa8-9605-6134e149c295_1300x856.jpeg 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em>Share of papers describing each year&#8217;s Nobel-prize winning work that were published in the preceding 20 years. 10-year moving average. Source: <a href="https://mattsclancy.substack.com/p/science-is-getting-harder">Clancy</a> based on data from <a href="https://www.nature.com/articles/s41597-019-0033-6">Li et al.</a></em></figcaption></figure></div><p>5) <a href="https://pubs.aeaweb.org/doi/pdfplus/10.1257/aer.20180338">Bloom et al.</a> analyze research output from an economic perspective. Assuming that economic growth ultimately comes from new ideas, the constant or declining rate of growth implies that the exponential increase in the number of researchers is being offset by a corresponding decline in the output per researcher. They find that this pattern holds true when drilling down into specific areas, including semiconductors, agriculture, and medicine (where the progress measures are Moore&#8217;s law, crop yield growth, and life expectancy, respectively).</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!xxnu!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8bdc9eee-1277-48a2-b104-20898c2f7293_1268x702.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!xxnu!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8bdc9eee-1277-48a2-b104-20898c2f7293_1268x702.png 424w, https://substackcdn.com/image/fetch/$s_!xxnu!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8bdc9eee-1277-48a2-b104-20898c2f7293_1268x702.png 848w, https://substackcdn.com/image/fetch/$s_!xxnu!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8bdc9eee-1277-48a2-b104-20898c2f7293_1268x702.png 1272w, https://substackcdn.com/image/fetch/$s_!xxnu!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8bdc9eee-1277-48a2-b104-20898c2f7293_1268x702.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!xxnu!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8bdc9eee-1277-48a2-b104-20898c2f7293_1268x702.png" width="627" height="347.1246056782334" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/8bdc9eee-1277-48a2-b104-20898c2f7293_1268x702.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:702,&quot;width&quot;:1268,&quot;resizeWidth&quot;:627,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!xxnu!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8bdc9eee-1277-48a2-b104-20898c2f7293_1268x702.png 424w, https://substackcdn.com/image/fetch/$s_!xxnu!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8bdc9eee-1277-48a2-b104-20898c2f7293_1268x702.png 848w, https://substackcdn.com/image/fetch/$s_!xxnu!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8bdc9eee-1277-48a2-b104-20898c2f7293_1268x702.png 1272w, https://substackcdn.com/image/fetch/$s_!xxnu!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8bdc9eee-1277-48a2-b104-20898c2f7293_1268x702.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em>The decline of research productivity. Note that economists use &#8220;production&#8221; as a catch-all term, with paper and patent counts, growth, and other metrics being different ways to measure it. We view production and progress as fundamentally different constructs, so we use the term production in a narrower sense. Keep in mind that in the figure, &#8220;productivity&#8221; isn&#8217;t based on paper production but on measures that are better viewed as progress measures. Source: <a href="https://pubs.aeaweb.org/doi/pdfplus/10.1257/aer.20180338">Bloom et al.</a></em></figcaption></figure></div><p>Of course, there are shortcomings in each of the metrics above. This is to be expected: since progress doesn't have an objective metric, we need to rely on proxies for measuring it, and these proxies will inevitably have some flaws.</p><p>For example, <a href="https://www.nature.com/articles/s41586-022-05543-x">Park et al.</a> used citation patterns to flag papers as "disruptive": if follow-on citations to a given paper don't also cite the studies this paper cited, the paper is more likely to be considered disruptive. One criticism of the paper is that this could simply be a result of how citation practices have evolved over time, not a result of whether a paper is truly disruptive. And the metric does flag some breakthroughs as non-disruptive &#8212; for example, AlphaFold is not considered a disruptive paper by this metric.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a></p><p>But taken together, the findings do suggest that scientific progress is slowing down, at least compared to the volume of papers, researchers, and resources. Still, this is an area where further research would be fruitful &#8212; while the decline in the pace of progress relative to inputs seems very clear, it is less clear what is happening at an aggregate level. Furthermore, there are many notions of what the <a href="https://compass.onlinelibrary.wiley.com/doi/10.1111/phc3.12525">goals of science</a> are and what progress even means, and it is not clear how to connect the available progress measures to these higher-level definitions.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!g3PC!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F162f2914-73f9-450d-9aec-4abdcb3b7fe7_1420x718.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!g3PC!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F162f2914-73f9-450d-9aec-4abdcb3b7fe7_1420x718.jpeg 424w, https://substackcdn.com/image/fetch/$s_!g3PC!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F162f2914-73f9-450d-9aec-4abdcb3b7fe7_1420x718.jpeg 848w, https://substackcdn.com/image/fetch/$s_!g3PC!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F162f2914-73f9-450d-9aec-4abdcb3b7fe7_1420x718.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!g3PC!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F162f2914-73f9-450d-9aec-4abdcb3b7fe7_1420x718.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!g3PC!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F162f2914-73f9-450d-9aec-4abdcb3b7fe7_1420x718.jpeg" width="689" height="348.3816901408451" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/162f2914-73f9-450d-9aec-4abdcb3b7fe7_1420x718.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:718,&quot;width&quot;:1420,&quot;resizeWidth&quot;:689,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!g3PC!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F162f2914-73f9-450d-9aec-4abdcb3b7fe7_1420x718.jpeg 424w, https://substackcdn.com/image/fetch/$s_!g3PC!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F162f2914-73f9-450d-9aec-4abdcb3b7fe7_1420x718.jpeg 848w, https://substackcdn.com/image/fetch/$s_!g3PC!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F162f2914-73f9-450d-9aec-4abdcb3b7fe7_1420x718.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!g3PC!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F162f2914-73f9-450d-9aec-4abdcb3b7fe7_1420x718.jpeg 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em>Summary of a few major lines of evidence of the slowdown in scientific progress</em></figcaption></figure></div><h3><strong>Why is progress slowing? Can AI help?</strong></h3><p>There are many hypotheses for why progress could be slowing. One set of hypotheses is that slowdown is an intrinsic feature of scientific progress, and is what we should expect. For example, there&#8217;s the low-hanging fruit hypothesis &#8212; the easy scientific questions have already been answered, so what remains to be discovered is getting harder.</p><p>This is an intuitively appealing idea. But we don&#8217;t find this convincing. Adam Mastroianni gives many compelling <a href="https://www.experimental-history.com/p/ideas-arent-getting-harder-to-find">counter-arguments</a>. He points out that we&#8217;ve been wrong about this over and over and lists many comically mis-timed assessments of scientific fields reaching saturation just before they ended up undergoing revolutions, such as physics in the 1890s.</p><p>While it&#8217;s true that lower-hanging fruits get picked first, there are countervailing factors. Over time, our scientific tools improve and we stand on the tower of past knowledge, making it easier to reach higher. Often, the benefits of improved tools and understanding are so transformative that whole new fields and subfields are created. New fields from the last 50-100 years include computer science, climate science, cognitive neuroscience, network science, genetics, molecular biology, and many others. Effectively, we&#8217;re plucking fruit from new trees, so there is always low-hanging fruit.</p><p>In our view, the low-hanging fruit hypothesis can at best partly explain slowdowns <em>within</em> fields. So it&#8217;s worth considering other ideas.</p><p>The second set of hypotheses is less fatalistic. They say that there&#8217;s something suboptimal about the way we&#8217;ve structured the practice of science, and so the efficiency of converting scientific inputs into progress is dropping. In particular, one subset of hypotheses flags the <em>increase in the rate of production</em> itself as the causal culprit &#8212; science is slowing down because it is trying to go too fast.</p><p>How could this be? The key is that any one scientist&#8217;s attention is finite, so they can only pay attention to a limited number of papers every year. So it is too <a href="https://www.sciencedirect.com/science/article/pii/S0048733322000129">risky</a> for authors of papers to depart from the canon. Any such would-be breakthrough papers would be lost in the noise and won&#8217;t get the attention of a critical mass of scholars. The greater the rate of production, the more the noise, so the less attention truly novel papers will achieve, and thus will be less likely to break through into the canon.</p><p><a href="https://www.pnas.org/doi/epub/10.1073/pnas.2021636118">Chu and Evans</a> explain:</p><blockquote><p>when the number of papers published each year grows very large, the rapid flow of new papers can force scholarly attention to already well-cited papers and limit attention for less-established papers&#8212;even those with novel, useful, and potentially transformative ideas. Rather than causing faster turnover of field paradigms, a deluge of new publications entrenches top-cited papers, precluding new work from rising into the most-cited, commonly known canon of the field.</p><p>These arguments, supported by our empirical analysis, suggest that the scientific enterprise&#8217;s focus on quantity may obstruct fundamental progress. This detrimental effect will intensify as the annual mass of publications in each field continues to grow</p></blockquote><p>Another causal mechanism relates to scientists&#8217; publish-or-perish <a href="https://www.nber.org/system/files/working_papers/w26752/w26752.pdf">incentives</a>. Production is easy to measure, and progress is hard to measure. So universities and other scientific institutions judge researchers based on measurable criteria such as how many papers they publish and the amount of grant funding they receive. It is not uncommon for scientists to have to publish a certain number of peer-reviewed papers to be hired or to get tenure (either due to implicit norms or explicit requirements).</p><p>The emphasis on production metrics seems to be worsening over time. Physics Nobel winner Peter Higgs famously <a href="https://www.theguardian.com/science/2013/dec/06/peter-higgs-boson-academic-system">noted</a> that he wouldn't even have been able to get a job in modern academia because he wouldn't be considered productive enough.</p><p>So individual researchers' careers might be better off if they are <a href="https://journals.sagepub.com/doi/abs/10.1177/0003122415601618">risk averse</a>, but it might reduce the collective rate of progress. Rzhetsky et al. find <a href="https://www.pnas.org/doi/10.1073/pnas.1509757112">evidence</a> of this phenomenon in biomedicine, where experiments tend to focus too much on experimenting with known molecules that are <em>already </em>considered important (which would be more likely to lead to publishing a paper) rather than more risky experiments that could lead to genuine breakthroughs. Worryingly, they find this phenomenon worsening over time.</p><p>This completes the feedback loop: career incentives lead to researchers publishing more papers, and disincentivize novel research that results in true breakthroughs (but might only result in a single paper after years of work). </p><p>If slower progress is indeed being caused by faster production, how will AI impact it? Most obviously, automating parts of the scientific process will make it even easier for scientists to chase meaningless productivity metrics. AI could make individual researchers more creative but decrease the <a href="https://www.science.org/doi/10.1126/sciadv.adn5290">creativity</a> of the collective because of a <a href="https://www.nature.com/articles/s41562-024-01959-9#Sec14">homogenizing effect</a>. AI could also exacerbate the inequality of attention and make it even harder for new ideas to break through. Existing search technology, such as&nbsp;<a href="https://jevinwest.org/papers/Kim2017asa.pdf">Google Scholar,</a>&nbsp;seems to be having exactly this effect.</p><p>To recap, so far we&#8217;ve argued that if the slowdown in science is caused by overproduction, AI will make it worse. In the next few sections, we&#8217;ll discuss why AI could worsen the slowdown regardless of what&#8217;s causing it.</p><h3><strong>Science is not ready for software, let alone AI</strong></h3><p>How do researchers use AI? In many ways: AI-based modeling to uncover trends in data using sophisticated pattern-matching algorithms; hand-written machine learning models specified based on expert knowledge; or even generative AI to write the code that researchers previously wrote. While some applications, such as using AI for literature review, don't involve writing code, most applications of AI for science are, in essence, <em><a href="https://russpoldrack.substack.com/p/why-better-code-can-lead-to-better">software development</a>.</em></p><p>Unfortunately, scientists are notoriously <a href="https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&amp;arnumber=6886129">poor software engineers</a>. Practices that are bog-standard in the industry, like automated testing, version control, and following programming design guidelines, are largely absent or haphazardly adopted in the research community. These are practices that were developed and standardized over the last six decades of software engineering to prevent bugs and ensure the software works as expected.</p><p>Worse, there is little scrutiny of the software used in scientific studies. While peer review is a long and arduous step in publishing a scientific paper, it does not involve reviewing the code accompanying the paper, even though most of the "science" in computational research is being carried out in the code and data accompanying a paper, and only summarized in the paper itself.</p><p>In fact, papers often fail to even share the code and data used to generate results, so even if other researchers are willing to review the code, they don't have the means to. Gabelica et al. <a href="https://pubmed.ncbi.nlm.nih.gov/35654271/">found</a> that of 1,800 biomedical papers that pledged to share their data and code, 93% did not end up sharing these artifacts. This even affects results in the most prominent scientific journals: Stodden et al. <a href="https://www.pnas.org/doi/10.1073/pnas.1708290115">contacted</a> the authors of 204 papers published in Science, one of the top scientific journals, to get the code and data for their study. Only 44% responded.</p><p>When researchers do share the code and data they used, it is often disastrously wrong. Even simple tools, like Excel, have <a href="https://www.knkx.org/money-matters/2013-04-23/how-an-excel-error-derailed-the-federal-deficit-debate?utm_source=chatgpt.com">notoriously</a> led to widespread errors in various fields. A 2016 study <a href="https://www.science.org/content/article/one-five-genetics-papers-contains-errors-thanks-microsoft-excel">found</a> that one in five genetics papers suffer from Excel-related errors, for example, because the names of genes (say, Septin 2) were automatically converted to dates (September 2). Similarly, it took decades for most scientific communities to <a href="https://www.amstat.org/asa/files/pdfs/p-valuestatement.pdf">learn</a> how to use simple statistics responsibly.</p><p>AI <a href="https://www.nature.com/articles/s41586-024-07146-0">opens</a> a whole new can of worms. The AI community often advertises AI as a silver bullet without realizing how difficult it is to detect subtle errors. Unfortunately, it takes much less competence to <em>use</em> AI tools than to <em>understand</em> them deeply and learn to identify errors. Like other software-based research, errors in AI-based science can take a long time to uncover. If the widespread adoption of AI leads to researchers spending more time and effort conducting or building on erroneous research, it could slow progress, since researcher time and effort are wasted in unproductive research directions.</p><p>Unfortunately, we've found that AI has already led to widespread errors. Even before generative AI, traditional machine learning led to <a href="https://reproducible.cs.princeton.edu/">errors</a> in over 600 papers across 30 scientific fields. In many cases, the affected papers constituted the majority of the surveyed papers, raising the possibility that in many fields, the <em>majority</em> of AI-enabled research is flawed. Others have found that AI tools are often used with <a href="https://www.understandingai.org/p/i-got-fooled-by-ai-for-science-hypeheres">inappropriate</a> baseline comparisons, making it incorrectly seem like they outperform older methods. These errors are not just theoretical: they affect the potential real-world deployment of AI too. For example, <a href="https://www.nature.com/articles/s42256-021-00307-0">Roberts et al.</a> found that of 400+ papers using AI for COVID-19 diagnosis, <em>none</em> produced clinically useful tools due to methodological flaws.</p><p>Applications of generative AI can result in new types of errors. For example, while AI can aid in programming, code generated using AI often has <a href="https://arxiv.org/abs/2211.03622">errors</a>. As AI adoption increases, we will discover more applications of AI for science. We suspect we'll find widespread errors in many of these applications.</p><p>Why is the scientific community so far behind software engineering best practices? In engineering applications, bugs are readily visible through tests, or in the worst case, when they are deployed to customers. Companies have strong incentives to fix errors to maintain the quality of their applications, or else they will lose market share. As a result, there is a strong demand for software engineers with deep expertise in writing good software (and now, in using AI well). This is why software engineering practices in the industry are decades ahead of those in research. In contrast, there are few incentives to correct flawed scientific results, and errors often persist for years.</p><p>That is not to say science should switch from a norms-based to a market-based model. But it shouldn't be surprising that there are many problems markets have solved that science hasn't &#8212; such as developing training pipelines for software engineers. Where such gaps between science and the industry emerge, scientific institutions need to intentionally adopt industry best practices to ensure science continues to innovate, without losing what makes science special.</p><p>In short, science needs to catch up to a half century of software engineering &#8212; fast. Otherwise, its embrace of AI will lead to an avalanche of errors and create headwinds, not tailwinds for progress.</p><p>AI could help too. There are many applications of AI to spot errors. For example, the <a href="https://the-black-spatula-project.github.io/">Black Spatula project</a> and the <a href="https://www.nature.com/articles/d41586-025-00648-5">YesNoError project</a> use AI to uncover flaws in research papers. In our own work, we've <a href="https://arxiv.org/html/2409.11363v1">developed</a> benchmarks aiming to spur the development of AI agents that automatically reproduce papers. Given the utility of generative AI for writing code, AI itself could be used to improve researchers' software engineering practices, such as by providing feedback, suggestions, best practices, and code reviews at scale. If such tools become reliable and see widespread adoption, AI could be part of the solution by helping avoid wasted time and effort building on erroneous work. But all of these possibilities require interventions from journals, institutions, and funding agencies to incentivize training, synthesis, and error detection rather than production alone.</p><h3><strong>AI might prolong the reliance on flawed theories</strong></h3><p>One of the main uses of AI for science is modeling. Older modeling techniques required coming up with a hypothesis for how the world works, then using statistical models to make inferences about this hypothesis.</p><p>In contrast, AI-based modeling treats this process as a black box. Instead of making a hypothesis about the world and improving our understanding based on the model's results, it simply tries to improve our ability to predict what outcomes would occur based on past data.</p><p>Leo Breiman illustrated the differences between these two modeling approaches in his landmark paper "Statistical Modeling: The Two Cultures". He strongly advocated for AI-based modeling, often on the basis of his experience in the industry. A focus on predictive accuracy is no doubt helpful in the industry. But it could hinder progress in science, where understanding is crucial.</p><p>Why? In a recent commentary in the journal <em><a href="https://www.nature.com/articles/d41586-025-01067-2">Nature</a></em>, we illustrated this with an analogy to the geocentric model of the Universe in astronomy. The geocentric model of the Universe&#8212;the model of the Universe with the Earth at the center&#8212;was very accurate at predicting the motion of planets. Workarounds like "epicycles" made these predictions accurate. (Epicycles were the small circles added to the planet's trajectory around the Earth).</p><p>Whenever a discrepancy between the model's predictions and the experimental readings was observed, astronomers added an epicycle to improve the model's accuracy. The geocentric model was so accurate at predicting planets' motions that many modern planetariums <em>still</em> use it to compute planets' trajectories.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!XSC6!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe450352c-e0ef-4c9a-8b6e-b69b18597941_1972x900.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!XSC6!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe450352c-e0ef-4c9a-8b6e-b69b18597941_1972x900.jpeg 424w, https://substackcdn.com/image/fetch/$s_!XSC6!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe450352c-e0ef-4c9a-8b6e-b69b18597941_1972x900.jpeg 848w, https://substackcdn.com/image/fetch/$s_!XSC6!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe450352c-e0ef-4c9a-8b6e-b69b18597941_1972x900.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!XSC6!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe450352c-e0ef-4c9a-8b6e-b69b18597941_1972x900.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!XSC6!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe450352c-e0ef-4c9a-8b6e-b69b18597941_1972x900.jpeg" width="1456" height="665" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/e450352c-e0ef-4c9a-8b6e-b69b18597941_1972x900.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:665,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:640661,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.aisnakeoil.com/i/168505690?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe450352c-e0ef-4c9a-8b6e-b69b18597941_1972x900.jpeg&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!XSC6!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe450352c-e0ef-4c9a-8b6e-b69b18597941_1972x900.jpeg 424w, https://substackcdn.com/image/fetch/$s_!XSC6!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe450352c-e0ef-4c9a-8b6e-b69b18597941_1972x900.jpeg 848w, https://substackcdn.com/image/fetch/$s_!XSC6!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe450352c-e0ef-4c9a-8b6e-b69b18597941_1972x900.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!XSC6!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe450352c-e0ef-4c9a-8b6e-b69b18597941_1972x900.jpeg 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em><a href="https://upload.wikimedia.org/wikipedia/commons/0/0e/Cassini_apparent.jpg">Left</a>: The geocentric model of the Universe eventually became extremely complex due to the large number of epicycles. <a href="https://upload.wikimedia.org/wikipedia/commons/2/28/Copernican_heliocentrism_diagram-2.jpg">Right</a>: The heliocentric model was far simpler. </em></figcaption></figure></div><p>How was the geocentric model of the Universe overturned in favor of the heliocentric model &#8212; the model with the planets revolving around the Sun? It couldn't be resolved by comparing the accuracy of the two models, since the accuracy of the models was similar. Rather, it was because the heliocentric model offered a far simpler explanation for the motion of planets. In other words, advancing from geocentrism to heliocentrism required a <em>theoretical</em> advance, rather than simply relying on the more accurate model.</p><p>This example shows that scientific progress depends on advances in <em>theory</em>. No amount of improvements in predictive accuracy could get us to the heliocentric model of the world without updating the theory of how planets move.</p><p>Let's come back to AI for science. AI-based modeling is no doubt helpful in improving predictive accuracy. But it doesn't lend itself to an improved understanding of these phenomena. AI might be fantastic at producing the equivalents of epicycles across fields, leading to the <a href="https://psycnet.apa.org/record/2024-74071-002">prediction-explanation fallacy</a>.</p><p>In other words, if AI allows us to make better predictions from incorrect theories, it might slow down scientific progress if this results in researchers using flawed theories for longer. In the extreme case, fields would be stuck in an intellectual rut even as they excel at improving predictive accuracy within existing paradigms.</p><p>Could advances in AI help overcome this limitation? Maybe, but not without radical changes to modeling approaches and technology, and there is little incentive for the AI industry to innovate on this front. So far, improvements in predictive accuracy have greatly outpaced improvements in the ability to model the underlying phenomena accurately.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!ZTdW!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9ca01290-65cb-4d9e-8c92-fe7b65068498_800x344.gif" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!ZTdW!,w_424,c_limit,f_webp,q_auto:good,fl_lossy/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9ca01290-65cb-4d9e-8c92-fe7b65068498_800x344.gif 424w, https://substackcdn.com/image/fetch/$s_!ZTdW!,w_848,c_limit,f_webp,q_auto:good,fl_lossy/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9ca01290-65cb-4d9e-8c92-fe7b65068498_800x344.gif 848w, https://substackcdn.com/image/fetch/$s_!ZTdW!,w_1272,c_limit,f_webp,q_auto:good,fl_lossy/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9ca01290-65cb-4d9e-8c92-fe7b65068498_800x344.gif 1272w, https://substackcdn.com/image/fetch/$s_!ZTdW!,w_1456,c_limit,f_webp,q_auto:good,fl_lossy/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9ca01290-65cb-4d9e-8c92-fe7b65068498_800x344.gif 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!ZTdW!,w_1456,c_limit,f_auto,q_auto:good,fl_lossy/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9ca01290-65cb-4d9e-8c92-fe7b65068498_800x344.gif" width="800" height="344" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/9ca01290-65cb-4d9e-8c92-fe7b65068498_800x344.gif&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:344,&quot;width&quot;:800,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!ZTdW!,w_424,c_limit,f_auto,q_auto:good,fl_lossy/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9ca01290-65cb-4d9e-8c92-fe7b65068498_800x344.gif 424w, https://substackcdn.com/image/fetch/$s_!ZTdW!,w_848,c_limit,f_auto,q_auto:good,fl_lossy/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9ca01290-65cb-4d9e-8c92-fe7b65068498_800x344.gif 848w, https://substackcdn.com/image/fetch/$s_!ZTdW!,w_1272,c_limit,f_auto,q_auto:good,fl_lossy/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9ca01290-65cb-4d9e-8c92-fe7b65068498_800x344.gif 1272w, https://substackcdn.com/image/fetch/$s_!ZTdW!,w_1456,c_limit,f_auto,q_auto:good,fl_lossy/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9ca01290-65cb-4d9e-8c92-fe7b65068498_800x344.gif 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em>Prediction without understanding: <a href="https://arxiv.org/abs/2507.06952">Vafa et al.</a> show that a transformer model trained on 10 million planetary orbits excels at predicting orbits without figuring out the underlying gravitational laws that produce those orbits.</em></figcaption></figure></div><h3><strong>Human understanding remains essential</strong></h3><p>In solving scientific problems, scientists build up an understanding of the phenomena they study. It might seem like this understanding is just a way to get to the solution. So if we can automate the process of going from problem to solution, we don&#8217;t need the intermediate step.</p><p>The reality is closer to the opposite. Solving problems and writing papers about them can be seen as a ritual that leads to the real prize, human understanding, without which there can be no scientific progress.</p><p>Fields Medal-winning mathematician William Thurston wrote an <a href="https://www.math.toronto.edu/mccann/199/thurston.pdf#page=1.27">essay</a> brilliantly illustrating this. At the outset, he emphasizes that the point of mathematics is not simply to figure out the truth value for mathematical facts, but rather the accompanying human understanding:</p><blockquote><p>&#8230;what [mathematicians] are doing is finding ways for <em>people</em> to understand and think about mathematics.</p><p>The rapid advance of computers has helped dramatize this point, because computers and people are very different. For instance, when Appel and Haken completed a proof of the 4-color map theorem using a massive automatic computation, it evoked much controversy. I interpret the controversy as having little to do with doubt people had as to the veracity of the theorem or the correctness of the proof. Rather, it reflected a continuing desire for <em>human understanding</em> of a proof, in addition to knowledge that the theorem is true.</p><p>On a more everyday level, it is common for people first starting to grapple with computers to make large-scale computations of things they might have done on a smaller scale by hand. They might print out a table of the first 10,000 primes, only to find that their printout isn't something they really wanted after all. They discover by this kind of experience that what they really want is usually not some collection of "answers"&#8212;what they want is <em>understanding</em>. [emphasis in original]</p></blockquote><p>He then describes his experience as a graduate student working on the theory of foliations, a center of attention among many mathematicians. After he proved a number of papers on the most important theorems in the field, counterintuitively, people began to <em>leave</em> the field:</p><blockquote><p>I heard from a number of mathematicians that they were giving or receiving advice not to go into foliations&#8212;they were saying that Thurston was cleaning it out. People told me (not as a complaint, but as a compliment) that I was killing the field. Graduate students stopped studying foliations, and fairly soon, I turned to other interests as well.</p><p>I do not think that the evacuation occurred because the territory was intellectually exhausted&#8212;there were (and still are) many interesting questions that remain and that are probably approachable. Since those years, there have been interesting developments carried out by the few people who stayed in the field or who entered the field, and there have also been important developments in neighboring areas that I think would have been much accelerated had mathematicians continued to pursue foliation theory vigorously.</p><p>Today, I think there are few mathematicians who understand anything approaching the state of the art of foliations as it lived at that time, although there are some parts of the theory of foliations, including developments since that time, that are still thriving.</p></blockquote><p>Two things led to this desertion. First, the results he documented were written in a way that was hard to understand. This discouraged newcomers from entering the field. Second, even though the point of mathematics is building up human understanding, the way mathematicians typically get credit for their work is by proving theorems. If the most prominent results in a field have already been proven, that leaves few incentives for others to understand a field's contributions, because they can't prove further results (which would ultimately lead to getting credit).</p><p>In other words, researchers are incentivized to prove theorems. More generally, researchers across fields are incentivized to find&nbsp;<em>solutions</em>&nbsp;to scientific problems. But this incentive only leads to progress because the process of proving theorems or finding solutions to problems <em>also </em>leads to building human understanding. As the desertion of work on foliations shows, when there is a mismatch between finding solutions to problems and building human understanding, it can result in <em>slower</em> progress.</p><p>This is precisely the effect AI might have: by solving open research problems without leading to the accompanying understanding, AI could erode these useful byproducts by reducing incentives to build understanding. If we use AI to short circuit this process of understanding, that is like using a forklift at the gym. You can lift heavier weights with it, sure, but that's not why you go to the gym.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!feAk!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F59680671-5a3d-4b1a-bb60-b04347bf5ab2_1600x690.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!feAk!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F59680671-5a3d-4b1a-bb60-b04347bf5ab2_1600x690.jpeg 424w, https://substackcdn.com/image/fetch/$s_!feAk!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F59680671-5a3d-4b1a-bb60-b04347bf5ab2_1600x690.jpeg 848w, https://substackcdn.com/image/fetch/$s_!feAk!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F59680671-5a3d-4b1a-bb60-b04347bf5ab2_1600x690.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!feAk!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F59680671-5a3d-4b1a-bb60-b04347bf5ab2_1600x690.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!feAk!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F59680671-5a3d-4b1a-bb60-b04347bf5ab2_1600x690.jpeg" width="697" height="300.6291208791209" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/59680671-5a3d-4b1a-bb60-b04347bf5ab2_1600x690.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:628,&quot;width&quot;:1456,&quot;resizeWidth&quot;:697,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!feAk!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F59680671-5a3d-4b1a-bb60-b04347bf5ab2_1600x690.jpeg 424w, https://substackcdn.com/image/fetch/$s_!feAk!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F59680671-5a3d-4b1a-bb60-b04347bf5ab2_1600x690.jpeg 848w, https://substackcdn.com/image/fetch/$s_!feAk!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F59680671-5a3d-4b1a-bb60-b04347bf5ab2_1600x690.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!feAk!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F59680671-5a3d-4b1a-bb60-b04347bf5ab2_1600x690.jpeg 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em>AI could short circuit the process of building human understanding, which is essential to scientific progress</em></figcaption></figure></div><p>Of course, mathematics might be an extreme case, because human understanding <em>is </em>the end goal of (pure) mathematics, not simply knowing the truth value of mathematical statements. This might not be the case for many&nbsp;<em>applications</em>&nbsp;of science, where the end goal is to make progress towards a real-world outcome rather than human understanding, say, weather forecasting or materials synthesis.</p><p>Most fields lie in between these two extremes. If we use AI to bypass human understanding, or worse, retain only <a href="https://www.nature.com/articles/s41586-024-07146-0">illusions of understanding</a>, we might lose the ability to train new scientists, develop new theories and paradigms, synthesize and correct results, apply knowledge beyond science, or even generate new and interesting problems.</p><p>Empirical evidence across scientific fields has found evidence for some of these effects. For example, Hao et al. <a href="https://arxiv.org/pdf/2412.07727">collect</a> data from six fields and find that papers that adopt AI are more likely to focus on providing solutions to known problems and working within existing paradigms rather than generating new problems.</p><p>Of course, AI can also be used to build up tacit knowledge, such as by helping people understand mathematical proofs or other scientific knowledge. But this requires fundamental changes to how science is organized. Today's career incentives and social norms prize <em>solutions</em> to scientific problems over human understanding. As AI adoption accelerates, we need changes to incentives to make sure human understanding is prioritized.</p><h3><strong>Implications for the future of science</strong></h3><p>Over the last decade, scientists have been in a headlong rush to adopt AI. The speed has come at the expense of any ability to adapt slow-moving scientific institutional norms to maintain quality control and identify and preserve what is essentially human about science. As a result, the trend is likely to worsen the production-progress paradox, accelerating paper publishing but only digging us deeper into the hole with regard to true scientific progress.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!P-hj!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0e953a80-d2ac-484d-a2c9-3bdf7f5edbe8_1584x1302.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!P-hj!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0e953a80-d2ac-484d-a2c9-3bdf7f5edbe8_1584x1302.jpeg 424w, https://substackcdn.com/image/fetch/$s_!P-hj!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0e953a80-d2ac-484d-a2c9-3bdf7f5edbe8_1584x1302.jpeg 848w, https://substackcdn.com/image/fetch/$s_!P-hj!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0e953a80-d2ac-484d-a2c9-3bdf7f5edbe8_1584x1302.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!P-hj!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0e953a80-d2ac-484d-a2c9-3bdf7f5edbe8_1584x1302.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!P-hj!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0e953a80-d2ac-484d-a2c9-3bdf7f5edbe8_1584x1302.jpeg" width="637" height="523.6875" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/0e953a80-d2ac-484d-a2c9-3bdf7f5edbe8_1584x1302.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1197,&quot;width&quot;:1456,&quot;resizeWidth&quot;:637,&quot;bytes&quot;:278785,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.aisnakeoil.com/i/168505690?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0e953a80-d2ac-484d-a2c9-3bdf7f5edbe8_1584x1302.jpeg&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!P-hj!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0e953a80-d2ac-484d-a2c9-3bdf7f5edbe8_1584x1302.jpeg 424w, https://substackcdn.com/image/fetch/$s_!P-hj!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0e953a80-d2ac-484d-a2c9-3bdf7f5edbe8_1584x1302.jpeg 848w, https://substackcdn.com/image/fetch/$s_!P-hj!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0e953a80-d2ac-484d-a2c9-3bdf7f5edbe8_1584x1302.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!P-hj!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0e953a80-d2ac-484d-a2c9-3bdf7f5edbe8_1584x1302.jpeg 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em>The number of papers that use AI quadrupled across 20 fields between 2012 and 2022 &#8212; even before the adoption of large language models. <a href="https://arxiv.org/pdf/2405.15828">Figure by Duede et al.</a></em></figcaption></figure></div><p>So, what should the scientific community do differently? Let&#8217;s talk about the role of individual researchers, funders, publishers and other gatekeepers, and AI companies.</p><h4>Changing scientific practices</h4><p>Individual researchers should be more careful when adopting AI. They should build software engineering skills, learn how to avoid a long and growing list of <a href="https://www.science.org/doi/epdf/10.1126/sciadv.adk3452">pitfalls</a> in AI-based modeling, and ensure they don&#8217;t lose their expertise by using AI as a crutch or an oracle. Sloppy use of AI may help in the short run, but will hinder meaningful scientific achievement.</p><p>With all that said, we recognize that most individual researchers are rationally following their incentives (productivity metrics). Yelling at them is not going to help that much, because what we have are collective action problems. The actors with real power to effect change are journals, universities hiring &amp; promotion committees, funders, policymakers, etc. Let&#8217;s turn to those next.</p><h4>Investing in meta-science</h4><p>Meta-science research has been extremely valuable in revealing the production-progress paradox. But so far, that finding doesn&#8217;t have a lot of analytical precision. There&#8217;s only the fuzzy idea that science is getting less bang for its buck. This finding is generally consistent with scientists&#8217; vibes, and is backed by a bunch of different metrics that vaguely try to measure true progress. But we don&#8217;t have a clear understanding of what the construct (progress) even is, and we&#8217;re far from a consensus story about what&#8217;s driving the slowdown.</p><p>To be clear, we will never have One True Progress Metric. If we did, Goodhardt/Campbell&#8217;s law would kick in &#8212; &#8220;When a measure becomes a target, it ceases to be a good measure.&#8221; Scientists would start to furiously optimize it, just as we have done with publication and citation counts, and the gaming would render it useless as a way to track progress.</p><p>That said, there&#8217;s clearly a long way for meta-science to go in improving both our quantitative and (more importantly) our qualitative/causal understanding of progress and the slowdown. Meta-science must also work to understand the efficacy of solutions.</p><p><a href="https://www.ukri.org/what-we-do/browse-our-areas-of-investment-and-support/uk-metascience-unit/">Despite</a> <a href="https://www.nature.com/articles/d41586-025-02065-0">recent</a> <a href="https://researchonresearch.org/metascience-alliance-launches-at-metascience-2025-conference/">growth</a>, meta-science funding is a fraction of a percent of science funding (and research on the slowdown is only a fraction of <em>that</em> pie). If it is indeed true that science funding as a whole is getting orders of magnitude less bang for the buck than in the past, meta-science investment seems ruefully small.</p><h4>Reforming incentives</h4><p>Scientists constantly complain to each other about the publish-or-perish treadmill and are keenly aware that the production-focused reward structure isn&#8217;t great for incentivizing scientific progress. But efforts to change this have consistently failed. One reason is simple inertia. Then there&#8217;s the aforementioned Goodhart&#8217;s law &#8212; whatever new metric is instituted will quickly be gamed. A final difficulty is that true progress can only be identified retrospectively, on timescales that aren&#8217;t suitable for hiring and promotion decisions.</p><p>One silver lining is that as the cost of publishing papers further drops due to AI, it could force us to stop relying on production metrics. In the AI field itself, the effort required to write a paper is so low that we are heading towards a singularity, with some researchers being able to (co-)author close to <a href="https://arxiv.org/pdf/2412.07793">100 papers a year</a>. (But, again, the perceived pace of actual progress seems <a href="https://x.com/jxmnop/status/1931386226809971138">mostly flat</a>.) Other fields might start going the same route.</p><p>Thus, rewarding the publication of individual findings may simply not be an option for much longer. Instead, perhaps the kinds of papers that count toward career progress should be limited to things that are hard to automate, such as new theories or paradigms of scientific research. And such reforms to incentive structures should go hand-in-hand with shifts in funding.</p><p>One thing we <em>don&#8217;t</em> need is more incentives for AI adoption. As we explained above, it is already happening at breakneck speed, and is not the bottleneck.</p><h4>Rethinking AI-for-science tools</h4><p>When it comes to AI-for-science labs and scientific tools from big AI companies, the elephant in the room is that they are in it for the wrong reasons. They want flashy &#8220;AI discovers X!&#8221; headlines so that they can sustain the narrative that AI will solve humanity&#8217;s problems, which buys them favorable policy treatment. We are not holding our breath for this to change.</p><p>We should be skeptical of AI-for-science news headlines. Many of them are <a href="https://www.aisnakeoil.com/p/scientists-should-use-ai-as-a-tool">greatly exaggerated</a>. The results may fail to reproduce, or AI may be framed as the main character when it was in fact one tool among many.</p><p>If there are any AI-for-science tool developers out there who actually want to help, here&#8217;s our advice. Target the actual bottlenecks instead of building yet another literature review tool. How about tools for finding errors in scientific code or other forms of quality control? Listen to the users. For example, mathematicians have repeatedly said that tools for improving human understanding are much more exciting than trying to automate theorem-proving, which they view as missing the point.</p><p>The way we evaluate AI-for-science tools should also change. Consider a literature review tool. There are three kinds of questions one can ask: Does it save a researcher time and produce results of comparable quality to existing tools? How does the use of the tool impact the researcher&#8217;s understanding of the literature compared to traditional search? What will the collective impacts on the community be if the tool were widely adopted &#8212;&nbsp;for example, will it further focus the community&#8217;s attention on already-famous papers? These three questions get at production, understanding, and progress respectively.</p><p>Currently, only the first question is considered part of what evaluation means. The latter two are out of scope, and there aren&#8217;t even established methods or metrics for such measurement. That means that AI-for-science evaluation is guaranteed to provide a highly incomplete and biased picture of the usefulness of these tools and minimize their potential harms.</p><h3><strong>Final thoughts</strong></h3><p>We ourselves are enthusiastic users of AI in our scientific workflows. On a day-to-day basis, it all feels very exciting. That makes it easy to forget that the impact of AI on science as an institution, rather than individual scientists, is a different question that demands a different kind of analysis. Writing this essay required fighting our own intuitions in many cases. If you are a scientist who is similarly excited about using these tools, we urge you to keep this difference in mind.</p><p>Our skepticism here has similarities and differences to our reasons for the slow timelines we laid out in <a href="https://knightcolumbia.org/content/ai-as-normal-technology">AI as Normal Technology</a>. In that paper, we explained that market mechanisms exert some degree of quality control, and many shoddy AI deployments have <a href="https://www.cio.com/article/190888/5-famous-analytics-and-ai-disasters.html">failed badly</a>, forcing companies who care about their reputation to <a href="https://www.linkedin.com/posts/randomwalker_for-ai-to-have-rapid-transformative-economic-activity-7348681016900833281-sNDF?rcm=ACoAAD85EvsBeNijbxJLxmNZcA4cF5Gc0JEwrc4">take it slow</a> when <em>deploying</em> AI, especially for consequential tasks, regardless of how fast the pace of development is. But in science, adoption and quality control processes are decoupled, with the former being much faster.</p><p>We are optimistic that scientific norms and processes will catch up in the long run. But for now, it&#8217;s going to be a bumpy ride.</p><p><em>We are grateful to Eamon Duede for feedback on a draft of this essay.</em></p><h3><strong>Further reading</strong></h3><ul><li><p>The <a href="https://www.heinrich.senate.gov/ASAP">American Science Acceleration Project (ASAP)</a> is a national initiative with the stated goal of making American science "ten times faster by 2030". The offices of Senators Heinrich and Rounds recently requested feedback on how to achieve this. In our <a href="https://www.cs.princeton.edu/~sayashk/asap-rfc-response.pdf">response</a>, we emphasized the production-progress paradox, discussed why AI could slow (rather than hasten) scientific progress, and recommended policy interventions that could help.</p></li><li><p>Our colleague Alondra Nelson also wrote a <a href="https://www.science.org/doi/10.1126/science.adz9545">response</a> to the ASAP initiative, emphasizing that faster science is not automatically better, and highlighted many challenges that remain despite increasing the pace of production.</p></li><li><p>In a recent <a href="https://www.nature.com/articles/d41586-025-01067-2">commentary</a> in the journal <em>Nature</em>, we discussed why the proliferation of AI-driven modeling could be bad for science.</p></li><li><p>We have written about the use of AI for science in many previous essays in this newsletter:</p><ul><li><p><a href="https://www.aisnakeoil.com/p/can-ai-automate-computational-reproducibility">Can AI automate computational reproducibility?</a></p></li><li><p><a href="https://www.aisnakeoil.com/p/scientists-should-use-ai-as-a-tool">Scientists should use AI as a tool, not an oracle</a></p></li><li><p><a href="https://www.aisnakeoil.com/p/machine-learning-is-useful-for-many">ML is useful for many things, but not for predicting scientific replicability</a></p></li><li><p><a href="https://www.aisnakeoil.com/p/introducing-the-reforms-checklist">The REFORMS checklist for detecting and preventing errors in ML-based science</a></p></li></ul></li><li><p>Lisa Messeri and Molly Crockett <a href="https://www.nature.com/articles/s41586-024-07146-0">offer</a> a taxonomy of the uses of AI in science. They discuss many pitfalls of adopting AI in science, arguing we could end up producing more while understanding less.</p></li><li><p>Matt Clancy reviewed the evidence for slowdowns in <a href="https://mattsclancy.substack.com/p/science-is-getting-harder">science</a> and <a href="https://www.newthingsunderthesun.com/pub/bvmu4ol2/release/10">innovation</a>, and <a href="https://www.newthingsunderthesun.com/pub/oobot2tp/release/11">discussed</a> interventions for incentivizing genuine progress.</p></li><li><p>The Institute for Progress released a podcast series on meta-science. Among other things, the series discusses concerns about <a href="https://ifp.org/the-metascience-101-podcast-series/#episode-two-is-science-slowing-down">slowdown</a> and <a href="https://ifp.org/the-metascience-101-podcast-series/">alternative models</a> for funding and organizing science.</p></li></ul><p><em>Update (July 17, 2025): Minor wording edits for clarity. </em></p><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>Adjusting for <a href="https://www.in2013dollars.com/us/inflation/2000?endYear=2021&amp;amount=1">inflation</a>, this is still a 2.5x increase.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p>This <a href="https://www.nature.com/articles/d41586-025-01548-4">article</a> in Nature News, published two years after the original study, documents the controversy surrounding the paper's results and responses from the authors.</p></div></div>]]></content:encoded></item><item><title><![CDATA[AGI is not a milestone]]></title><description><![CDATA[There is no capability threshold that will lead to sudden impacts]]></description><link>https://www.normaltech.ai/p/agi-is-not-a-milestone</link><guid isPermaLink="false">https://www.normaltech.ai/p/agi-is-not-a-milestone</guid><dc:creator><![CDATA[Sayash Kapoor]]></dc:creator><pubDate>Thu, 01 May 2025 11:47:59 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!LOkG!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F49daa6d8-fe58-410c-b060-83648d15a722_640x480.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>With the release of OpenAI&#8217;s latest model o3, there is renewed debate about whether Artificial General Intelligence has already been achieved. The standard skeptic&#8217;s response to this is that there is no consensus on the definition of AGI. That is true, but misses the point &#8212; if AGI is such a momentous milestone, shouldn&#8217;t it be obvious when it has been built?</p><p>In this essay, we argue that AGI is not a milestone. It does not represent a discontinuity in the properties or impacts of AI systems. If a company declares that it has built AGI, based on whatever definition, it is not an actionable event. It will have no implications for businesses, developers, policymakers, or safety. Specifically:</p><ol><li><p>Even if general-purpose AI systems reach some agreed-upon capability threshold, we will need many complementary innovations that allow AI to diffuse across industries to realize its productive impact. Diffusion occurs at human (and societal) timescales, not at the speed of tech development.</p></li><li><p>Worries about AGI and catastrophic risk often conflate capabilities with power. Once we distinguish between the two, we can reject the idea of a critical point in AI development at which it becomes infeasible for humanity to remain in control.</p></li><li><p>The proliferation of AGI <a href="https://define-agi.com/">definitions</a> is a symptom, not the disease. AGI is significant because of its presumed impacts but must be defined based on properties of the AI system itself. But the link between system properties and impacts is tenuous, and greatly depends on how we design the environment in which AI systems operate. Thus, whether or not a given AI system will go on to have transformative impacts is <em>yet to be determined</em> at the moment the system is released. So a determination that an AI system constitutes AGI can only meaningfully be made <em>retrospectively</em>.</p></li></ol><h2><strong>Nuclear weapons as an anti-analogy for AGI</strong></h2><p>Achieving AGI is the explicit goal of companies like OpenAI and much of the AI research community. It is treated as a milestone in the same way as building and delivering a nuclear weapon was the key goal of the Manhattan Project.</p><p>This goal made sense as a milestone in the Manhattan Project for two reasons. The first is <em>observability</em>. In developing nuclear weapons, there can be no doubt about whether you&#8217;re reached the goal or not &#8212; an explosion epitomizes observability. The second is <em>immediate impact</em>. The use of nuclear weapons contributed to a quick end to World War 2. It also ushered in a new world order &#8212; a <em>long-term transformation</em> of geopolitics.</p><p>Many people have the intuition that AGI will have these properties. It will be so powerful and humanlike that it will be obvious when we&#8217;ve built it. And it will immediately bring massive benefits and risks &#8212; automation of a big swath of the economy, a great acceleration of innovation, including AI research itself, and potentially catastrophic consequences for humanity from uncontrollable superintelligence.</p><p>In this essay, we argue that AGI will be exactly the opposite &#8212; it is unobservable because there is no clear capability threshold that has particular significance; it will have no immediate impact on the world; and even a long-term transformation of the economy is uncertain.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!LOkG!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F49daa6d8-fe58-410c-b060-83648d15a722_640x480.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!LOkG!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F49daa6d8-fe58-410c-b060-83648d15a722_640x480.png 424w, https://substackcdn.com/image/fetch/$s_!LOkG!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F49daa6d8-fe58-410c-b060-83648d15a722_640x480.png 848w, https://substackcdn.com/image/fetch/$s_!LOkG!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F49daa6d8-fe58-410c-b060-83648d15a722_640x480.png 1272w, https://substackcdn.com/image/fetch/$s_!LOkG!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F49daa6d8-fe58-410c-b060-83648d15a722_640x480.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!LOkG!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F49daa6d8-fe58-410c-b060-83648d15a722_640x480.png" width="640" height="480" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/49daa6d8-fe58-410c-b060-83648d15a722_640x480.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:480,&quot;width&quot;:640,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:332368,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.aisnakeoil.com/i/162585098?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F49daa6d8-fe58-410c-b060-83648d15a722_640x480.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!LOkG!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F49daa6d8-fe58-410c-b060-83648d15a722_640x480.png 424w, https://substackcdn.com/image/fetch/$s_!LOkG!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F49daa6d8-fe58-410c-b060-83648d15a722_640x480.png 848w, https://substackcdn.com/image/fetch/$s_!LOkG!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F49daa6d8-fe58-410c-b060-83648d15a722_640x480.png 1272w, https://substackcdn.com/image/fetch/$s_!LOkG!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F49daa6d8-fe58-410c-b060-83648d15a722_640x480.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>In previous essays, we have argued against the likely disastrous <a href="https://www.aisnakeoil.com/p/licensing-is-neither-feasible-nor">policy interventions</a> that some have recommended by analogizing AGI to nuclear weapons. It is striking to us that this analogy reliably generates what we consider to be incorrect predictions and counterproductive recommendations.</p><h2><strong>It isn&#8217;t crazy to think that o3 is AGI, but this says more about AGI than o3</strong></h2><p>Many <a href="https://marginalrevolution.com/marginalrevolution/2025/04/o3-and-agi-is-april-16th-agi-day.html">prominent</a> <a href="https://www.oneusefulthing.org/p/on-jagged-agi-o3-gemini-25-and-everything">AI commentators</a> have called o3 a kind of AGI: Tyler Cowen says that if you know AGI when you see it, then he has <a href="https://marginalrevolution.com/marginalrevolution/2025/04/o3-and-agi-is-april-16th-agi-day.html">seen it</a>. Ethan Mollick describes o3 as a <a href="https://www.oneusefulthing.org/p/on-jagged-agi-o3-gemini-25-and-everything">jagged AGI</a>. What is it about o3 that has led to such excitement?</p><p>The key innovation in o3 is the use of reinforcement learning to learn to search the web and use tools as part of its reasoning chain.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a> In this way, it can perform more complex cognitive tasks than LLMs are directly capable of, and can do so in a way that&#8217;s similar to people.</p><p>Consider a person doing comparison shopping. They might look at a few products, use the reviews of those products to get a better sense of what features are even important, and use that knowledge to iteratively expand or shrink the set of products being considered. o3 is a generalist agent that does a decent job at this sort of thing.</p><p>Let&#8217;s consider what this means for AGI. To avoid getting bogged down in the details of o3, imagine a future system whose architecture is identical to o3, but is much more competent. For example, it can always find the right webpages and knowledge for the task as long as it&#8217;s online, no matter how hard it is to locate. It can download and run code from the internet to solve a task if necessary. None of these require scientific breakthroughs, only engineering improvements and further training.</p><p>At the same time, without scientific improvements, the architecture imposes serious limits. For example, this future system cannot acquire new skills from experience, except through an explicit update to its training. Building AI systems that can learn on the fly is an open research problem.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a></p><p>Would our hypothetical system be AGI? Arguably, yes. What many AGI definitions have in common is the ability to outperform humans at a wide variety of tasks. Depending on how narrowly the set of tasks is defined and how broadly the relevant set of humans for each task is defined, it is quite plausible that this future o3-like model/agent will meet some of these AGI definitions.</p><p>For example, it will be superhuman at playing chess, despite the fact that large language models themselves are at best <a href="https://maxim-saplin.github.io/llm_chess/">mediocre</a> <a href="https://dynomight.substack.com/p/more-chess">at</a> <a href="https://blog.mathieuacher.com/GPTsChessEloRatingLegalMoves/">chess</a>. Remember that the model can use tools, search the internet, and download and run code. If the task is to play chess, it will download and run a chess engine.</p><p>Despite human-level or superhuman performance at many tasks, and plausibly satisfying some definitions of AGI, it will probably fail badly at many <a href="https://www.businessinsider.com/ai-agents-study-company-run-by-ai-disaster-replace-jobs-2025-4">real-world</a> tasks. We&#8217;ll get back to the reasons for that.</p><p>Does any of this matter? It does. Leaders at AI companies have made very loud <a href="https://deepmind.google/discover/blog/taking-a-responsible-path-to-agi/">predictions</a> <a href="https://www.reuters.com/technology/teslas-musk-predicts-ai-will-be-smarter-than-smartest-human-next-year-2024-04-08/">and</a> <a href="https://arstechnica.com/ai/2025/01/anthropic-chief-says-ai-could-surpass-almost-all-humans-at-almost-everything-shortly-after-2027/">commitments</a> to <a href="https://arstechnica.com/information-technology/2025/01/sam-altman-says-we-are-now-confident-we-know-how-to-build-agi/">delivering</a> AGI within a <a href="https://www.businessinsider.com/agi-predictions-sam-altman-dario-amodei-geoffrey-hinton-demis-hassabis-2024-11#demis-hassabis-1">few years</a>. There are enormous incentives for them to declare some near-future system to be AGI, and potentially enormous costs to <em>not</em> doing so. Perhaps some of the valuation of AI companies is based on these promises, so without AGI there might be a bubble burst. Being seen as leaders in AI development could help improve market share and revenues, and improve access to <a href="https://www.theverge.com/2024/1/18/24042354/mark-zuckerberg-meta-agi-reorg-interview">talent</a>.</p><p>So, if and when companies claim to have built AGI, what will be the consequences? We'll analyze that in the rest of this essay.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.normaltech.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption"><em>AI Snake Oil</em> debunks AI hype and publishes evidence-based analysis of new developments.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><h2><strong>AGI won't be a shock to the economy because diffusion takes decades</strong></h2><p>One argument for treating AGI as a milestone &#8212; and taking declarations of AGI seriously &#8212; is that AGI could lead to rapid economic impacts, both positive and negative, such as a world without scarcity, <a href="https://www.politico.com/newsletters/digital-future-daily/2024/05/14/altmans-fantasy-of-a-gpt-centric-world-00157966">an end to the concept of money</a>, or <a href="https://epoch.ai/gradient-updates/agi-could-drive-wages-below-subsistence-level">sudden mass joblessness</a>.</p><p>But AI's economic impact is only realized when it is adopted across the economy. Technical advances are necessary, but not sufficient, to realize this impact. For past general-purpose technologies, such as electricity, computing, and the internet, it took decades for the underlying technical advances to diffuse across society. The miracle of the Industrial Revolution wasn't the high growth rate &#8212; annual growth rates averaged below <a href="https://www.bloomberg.com/opinion/articles/2023-08-16/ai-won-t-supercharge-the-us-economy">3%</a> &#8212; but the sustained period of decades of growth.</p><p>There are many <a href="https://www.aisnakeoil.com/p/ai-as-normal-technology">bottlenecks</a> to the diffusion of AI: developing useful <a href="https://www.aisnakeoil.com/p/ai-companies-are-pivoting-from-creating">products and applications</a>, training the workforce to utilize these products, implementing organizational changes to enable AI use, and establishing laws and norms that facilitate AI adoption by companies. Like past general-purpose technologies, we expect the economic impacts of AI to be realized over decades, as this process of diffusion unfolds.</p><p>In the paper <a href="https://www.aisnakeoil.com/p/ai-as-normal-technology">AI as Normal Technology</a>, we present a detailed argument for why we think this will be the case. The idea that rapid increases in capability lead to rapid economic impacts is completely inconsistent with the past and present of AI, and there is no reason to expect that this will change in the future.</p><p>One <a href="https://web.archive.org/web/20180409161852/https://blog.openai.com/openai-charter/">definition</a> of AGI is AI systems that outperform humans at most economically valuable work. We might worry that if AGI is realized in this sense of the term, it might lead to massive, sudden job displacement. But humans are a moving target. As the process of diffusion unfolds and the cost of production (and hence the value) of tasks that have been automated decreases, humans will adapt and move to tasks that have not yet been automated. The process of technical advancements, product development, and diffusion will continue.</p><h2><strong>AGI will not lead to a rapid change in the world order</strong></h2><p>The US and China are often described as being in an AI arms race, with each country racing to build AGI. It is hypothesized that the country to build it first would have a decisive strategic advantage &#8212; resulting in dominance in the world order for the foreseeable future.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-3" href="#footnote-3" target="_self">3</a></p><p>This narrative doesn&#8217;t make sense because the knowledge required to create AI models, and model capabilities themselves, tend to proliferate quickly between countries. There are hundreds of thousands of AI technologists, and they work in the private sector rather than government labs, so it is not feasible to keep secrets at that scale.</p><p>Invention &#8212; in this case, AI model development &#8212; is <a href="https://www.foreignaffairs.com/china/innovation-fallacy-artificial-intelligence">overrated</a> as a source of competitive advantage. We should expect technological developments to roughly keep pace across countries. Even though US companies are currently at the forefront, we shouldn't expect a lasting advantage.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-4" href="#footnote-4" target="_self">4</a></p><p>Many people haven't appreciated the ease of proliferation of technological capabilities &#8212; perhaps due to the nuclear weapons mental model &#8212; and have been caught by surprise. This is what led to the "DeepSeek moment" earlier this year, since analysts did not realize how quickly AI capabilities can proliferate, and as a result, were not expecting upstarts (especially from China) to catch up so quickly.</p><p>Some people argue that even an advantage of a <a href="https://blog.ai-futures.org/p/why-america-wins">few months</a> will be critical. We disagree. The important question in the context of great power competition is not which country builds AGI first, but rather which country better enables diffusion. As Jeffrey Ding has <a href="https://press.princeton.edu/books/paperback/9780691260341/technology-and-the-rise-of-great-powers">shown</a>, the effectiveness of companies and governments in actually utilizing AI inventions and innovations, whether domestic or international, to improve productivity is far more important for determining the economic impacts of general-purpose technologies.</p><p>While Chinese AI companies are at most 6-12 months behind leading US companies in terms of AI models and capabilities, China <a href="https://www.prcleader.org/post/china-s-ai-implementation-gap">lags</a> significantly behind the US in several key indicators that might enable diffusion: Digitization, cloud computing adoption, and workforce training. All of these are required to enable the productive diffusion of AI advances across industries. This is the actual source of American competitive advantage.</p><p>Of course, this could change in the coming years. But if it does, it will result from policy changes to promote diffusion rather than the development of AGI. And it is not something countries can accomplish overnight, regardless of how quickly they change policy. Diffusion typically unfolds over decades.</p><p>None of this means that policymakers should be complacent. But it does mean that rather than fixating on AGI, they should focus on enabling productive and safe diffusion, including of <em>existing</em> AI.</p><h2><strong>The long-term economic implications of AGI are uncertain</strong></h2><p>Even if it doesn&#8217;t have immediate economic impacts, could AGI unlock, say, <a href="https://www.dwarkesh.com/p/satya-nadella">10% annual GDP growth</a> that could add up to something big over a few decades?</p><p>Maybe. But it is far from clear why and how this will happen.</p><p>Historically, this kind of acceleration in growth has happened very few times &#8212; the industrial revolution had this effect, but not the internet, which barely had any <a href="https://onlinelibrary.wiley.com/doi/abs/10.1111/joes.12211">impact</a> on GDP. Note that even if you don&#8217;t think that GDP is the right thing to measure, a qualitative change in the GDP growth rate is a good proxy for whatever fundamental change in the economy you care about.</p><p>The problem is that accelerating growth requires eliminating bottlenecks to progress. That&#8217;s harder than most AI boosters assume. AI will likely have uneven effects across sectors, and long-term growth will be bottlenecked by the <em><a href="https://zhengdongwang.com/2023/06/27/why-transformative-ai-is-really-really-hard-to-achieve.html">weakest</a></em> sector.</p><p>Those arguing for dramatic effects often have an incorrect mental model of what the bottlenecks actually are. For example, while it is tempting to believe that cheap scientific innovation will unlock progress, the production of new findings is actually <a href="https://substack.com/@aisnakeoil/note/c-92421948">not the</a> <a href="https://www.nature.com/articles/d41586-025-01067-2">bottleneck</a> in science.</p><p>More broadly, progress depends not just on the technology but on having the right preconditions &#8212; complementary innovations as well as cultural, economic, and political factors. If all it took to create the industrial revolution was the invention of steam power, the <a href="https://acoup.blog/2022/08/26/collections-why-no-roman-industrial-revolution/">Roman</a> <a href="https://blog.rootsofprogress.org/why-no-roman-industrial-revolution">Empire</a> would have done it.</p><p>Our current laws, norms, institutions, and politics evolved in a time of much less technological potential. They are already <a href="https://www.noahpinion.blog/p/book-review-abundance">choking</a> <a href="https://www.perseusbooks.com/titles/marc-j-dunkelman/why-nothing-works/9781541700215/">opportunities</a> for straightforward types of growth, such as building more public infrastructure. To reap the economic benefits that broad cognitive automation can potentially bring, the degree of structural change that needs to happen is unfathomably greater.</p><p>In conclusion, the extent and nature of long-term impacts from AGI remain to be seen and depend on what complementary actions we take. The long-term implications are not a property of AGI itself.</p><h2><strong>Misalignment risks of AGI conflate power and capability</strong></h2><p>On the flip side, AGI could be a turning point for AI's societal risks. Could it cause loss of control, massive societal harm, or even human extinction?</p><p>Discussions of AGI risks conflate power &#8212; the ability to modify the environment &#8212; with capability &#8212; the capacity to solve specified tasks correctly. Capability is an intrinsic property of an AI system, whereas power is a matter of how we design the environment in which AI systems operate. And humans have agency over this design. This distinction is often overlooked.</p><p>Consider Dario Amodei's definition of "powerful AI".<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-5" href="#footnote-5" target="_self">5</a> He begins with a description of the <em>capabilities </em>of powerful AI, such as being able to "...prove unsolved mathematical theorems, write extremely good novels, write difficult codebases from scratch." This criterion is an example of AI capabilities, and we can discuss them at the level of AI systems.</p><p>But he then moves on to describe properties of the environment in which we allow AI systems to operate, which includes "...taking actions on the internet, taking or giving directions to humans, ordering materials, directing experiments, watching videos, making videos, and so on." This is an example of the power afforded to an AI system. It depends on the environment within which the AI system operates, which determines how AI capabilities are translated into power.</p><p>We do expect AI capabilities to keep increasing. But regardless of capability level, we can choose to ensure that AI remains a tool and is not given power and autonomy to operate without human oversight. In the <a href="https://www.aisnakeoil.com/p/ai-as-normal-technology">AI as Normal Technology</a> essay, we address all the usual counterarguments to this, including arms races among companies, power seeking, superhuman persuasion, deceptive alignment, and more.</p><p>We argue in the paper that there will be strong business incentives against deploying AI without adequate oversight, and that these incentives can and should be buttressed by regulation when necessary. This has historically been the case in areas ranging from self-driving cars to AI assistants. We don&#8217;t expect this trend to suddenly flip once AI capabilities reach a presumed tipping point that we arbitrarily designate as AGI.</p><h2><strong>AGI does not imply impending superintelligence</strong></h2><p>Yet another reason to consider AGI a milestone is the view that shortly after we build AGI, AI systems could recursively self-improve &#8212; AGI could train future versions of models that become far more capable, leading to an "intelligence explosion." Soon afterwards, we would get superintelligent AI (AI systems that far exceed human abilities on any conceivable task), leading to either utopia or dystopia, depending on how well superintelligent AI is &#8220;aligned&#8221; with human interests.</p><p>In the normal technology view, there are two big reasons to doubt this narrative. The first is that <em>even if</em> arbitrary speedups in AI methods are possible, we think that innovation and diffusion will happen at human speed, as summarized in this table.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!TaBP!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9125e057-7253-41f2-ba37-2cd6b05026cf_1456x760.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!TaBP!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9125e057-7253-41f2-ba37-2cd6b05026cf_1456x760.png 424w, https://substackcdn.com/image/fetch/$s_!TaBP!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9125e057-7253-41f2-ba37-2cd6b05026cf_1456x760.png 848w, https://substackcdn.com/image/fetch/$s_!TaBP!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9125e057-7253-41f2-ba37-2cd6b05026cf_1456x760.png 1272w, https://substackcdn.com/image/fetch/$s_!TaBP!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9125e057-7253-41f2-ba37-2cd6b05026cf_1456x760.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!TaBP!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9125e057-7253-41f2-ba37-2cd6b05026cf_1456x760.png" width="1456" height="760" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/9125e057-7253-41f2-ba37-2cd6b05026cf_1456x760.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:760,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!TaBP!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9125e057-7253-41f2-ba37-2cd6b05026cf_1456x760.png 424w, https://substackcdn.com/image/fetch/$s_!TaBP!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9125e057-7253-41f2-ba37-2cd6b05026cf_1456x760.png 848w, https://substackcdn.com/image/fetch/$s_!TaBP!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9125e057-7253-41f2-ba37-2cd6b05026cf_1456x760.png 1272w, https://substackcdn.com/image/fetch/$s_!TaBP!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9125e057-7253-41f2-ba37-2cd6b05026cf_1456x760.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em>Like other general-purpose technologies, the impact of AI is materialized not when methods and capabilities improve, but when those improvements are translated into applications and are diffused through productive sectors of the economy. [<a href="https://www.aisnakeoil.com/p/ai-as-normal-technology">Source</a>]</em></figcaption></figure></div><p>Second, the fact that AI would help conduct AI research does not imply that this process can be arbitrarily accelerated. AI is already used to automate a significant portion of AI research today. But there are many bottlenecks to progress in AI methods, such as the social nature of data collection and real-world interaction that might be required for achieving certain capabilities, computational and cost limits, or herding around popular or intuitive ideas while ignoring the ones that enable true breakthroughs.</p><p>We could be wrong about this, and recursive self-improvement could be possible, leading to unbounded speedups in progress in AI methods. And this might have some interesting implications, including some discontinuities in impact, even if widespread diffusion will be slower. For these reasons, it is important to have early warning systems for recursive self-improvement. However, this is not captured by AGI definitions. We might well have AGI while being far from recursive self-improvement, or vice versa.</p><h2><strong>We won&#8217;t know when AGI has been built</strong></h2><p>There are endless definitions of AGI and related concepts &#8212; Jasmine Sun has a useful <a href="https://define-agi.com/">compilation</a> of over 20 definitions &#8212; but they fall into three broad categories: they can be based on the system&#8217;s <em>impact </em>on the world, its <em>internals</em>, or its <em>behavior </em>in controlled settings. We will show that each style of definition has a fatal flaw. It leads to criteria that are either too strict or too weak compared to what we ideally want.</p><p>Understanding these gaps also shows why people have different intuitions regarding what AGI will look like, and why &#8220;we&#8217;ll know it when we see it&#8221; has failed as a criterion and will continue to fail.</p><p>OpenAI's 2018 <a href="https://web.archive.org/web/20180409161852/https://blog.openai.com/openai-charter/">definition</a> of AGI was "highly autonomous systems that outperform humans at most economically valuable work". From our perspective &#8212; our interest being in the impacts of AI &#8212; this definition is potentially very useful. If AI outperformed [all] humans at most economically valuable work, it would be unquestionably impactful.</p><p>But let&#8217;s be clear &#8212; this is not a property of an AI system. It is a property of the state of the world. It has at least as much to do with the complementary innovations that we make and the extent to which we choose to integrate AI into our organizations and institutions. It would be absurd to try to test an AI system in isolation in the lab and ask whether it outperforms people at their jobs. It is a category error.</p><p>For example, whether AI can (autonomously) outperform a medical researcher depends in part on whether we collectively choose to allow AI systems to perform large-scale medical experiments on people. We shouldn&#8217;t and we won&#8217;t, which means that irrespective of the systems' capabilities, they cannot perform the function of a medical researcher. This might be an extreme example, but similar bottlenecks arise in virtually every job.</p><p>Worse, until and unless we diffuse AI everywhere, we won&#8217;t even know if the system is <em>theoretically</em> capable of automating work in the real world. We won&#8217;t be able to set up sufficiently convincing simulacra of the messy complexity of the world. In short, impact-based definitions are not useful for practical purposes because they don&#8217;t give us a way to anticipate the end result of the painfully slow process of diffusion.</p><p>In contrast to researchers like us who are interested in impacts, many researchers are interested in humanlike AI in the sense of internals &#8212; does the system truly understand the world causally, can it reason, plan, and acquire new skills like we do, and so forth.</p><p>This sense of AGI has been notoriously hard to operationalize because of the difficulty of observing and characterizing AI internals. The Turing test is the best known of many attempts to use behavior as a proxy for the humanlike attributes we care about, but <a href="https://www.science.org/doi/10.1126/science.adq9356">inevitably</a> it turns out that we can build AI systems that pass such tests without having the hoped-for humanlike internals.</p><p>Furthermore, because of the <a href="https://www.oneusefulthing.org/p/on-jagged-agi-o3-gemini-25-and-everything">jaggedness</a> of AI &#8212; superhuman in many ways, yet lacking a toddler&#8217;s understanding of the world in other ways &#8212; the transformative effects of AI are likely to be felt long before it is fully humanlike (or superhuman) on all dimensions.</p><p>In short, since we are not interested in the internals for their own sake, we set aside this kind of definition.</p><p>That leaves us with the third kind of definition, which is by far the most common: those based on behavior, operationalized as benchmark performance. For example, the Metaculus question on &#8220;<a href="https://www.metaculus.com/questions/384/humanmachine-intelligence-parity-by-2040/">human-machine intelligence parity</a>&#8221; is defined in terms of performance on exam questions in math, physics, and computer science. The problem with this kind of definition is well known, and we&#8217;ve <a href="https://www.aisnakeoil.com/i/161317202/benchmarks-do-not-measure-real-world-utility">discussed</a> it <a href="https://www.aisnakeoil.com/p/ai-existential-risk-probabilities">repeatedly</a>. They simply encourage hill climbing in the sense of building AI systems that can beat the benchmarks without necessarily being useful in the real world.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!u-7B!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd58c76e9-c5f3-4cfd-8b0f-f9aa20872c51_1600x539.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!u-7B!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd58c76e9-c5f3-4cfd-8b0f-f9aa20872c51_1600x539.png 424w, https://substackcdn.com/image/fetch/$s_!u-7B!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd58c76e9-c5f3-4cfd-8b0f-f9aa20872c51_1600x539.png 848w, https://substackcdn.com/image/fetch/$s_!u-7B!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd58c76e9-c5f3-4cfd-8b0f-f9aa20872c51_1600x539.png 1272w, https://substackcdn.com/image/fetch/$s_!u-7B!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd58c76e9-c5f3-4cfd-8b0f-f9aa20872c51_1600x539.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!u-7B!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd58c76e9-c5f3-4cfd-8b0f-f9aa20872c51_1600x539.png" width="1456" height="490" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/d58c76e9-c5f3-4cfd-8b0f-f9aa20872c51_1600x539.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:490,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!u-7B!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd58c76e9-c5f3-4cfd-8b0f-f9aa20872c51_1600x539.png 424w, https://substackcdn.com/image/fetch/$s_!u-7B!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd58c76e9-c5f3-4cfd-8b0f-f9aa20872c51_1600x539.png 848w, https://substackcdn.com/image/fetch/$s_!u-7B!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd58c76e9-c5f3-4cfd-8b0f-f9aa20872c51_1600x539.png 1272w, https://substackcdn.com/image/fetch/$s_!u-7B!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd58c76e9-c5f3-4cfd-8b0f-f9aa20872c51_1600x539.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em>Pros and cons of the three kinds of AGI definitions</em></figcaption></figure></div><p>One response to the definitional challenge is to claim that we'll know AGI when we see it. The o3 launch shows that the opposite is true &#8212; to some people, it is clear that the advances in capabilities represent a step change that warrants calling it AGI. To others, the improvements are marginal at best, and unlikely to lead to real-world impact.</p><p>What explains people&#8217;s differing intuitions? Here&#8217;s our guess. While AI capabilities can be general, making AI useful in the real world will have to happen in a largely domain-specific way. (The generalist agent aspect of o3 is misleading &#8212; it can only handle generative tasks where errors have low cost, not when you need it to operate independently in the real world. For example, a useful travel booking AI agent hasn&#8217;t yet been released, despite the seeming triviality of the task for humans.)</p><p>So when people are thinking about whether o3 or any other system is (close to) AGI, they are intuitively thinking about different domains, and the gap between general capabilities and useful real-world abilities is vastly different from one domain to another. Crossing the threshold of having useful products that can automate most tasks may happen at vastly different times in different sectors or jobs.</p><h2><strong>Businesses and policy makers should take a long-term view</strong></h2><p>AGI is not a milestone because it is not actionable. A company declaring it has achieved, or is about to achieve, AGI has no implications for how businesses should plan, what safety interventions we need, or how policymakers should react. What should businesses and policy makers do instead?</p><p>Businesses should not rush to adopt half-baked AI products. Rapid progress in AI methods and capabilities does not automatically translate to better products. Building products on top of inherently stochastic models is <a href="https://www.aisnakeoil.com/p/ai-companies-are-pivoting-from-creating">challenging</a>, and businesses should adopt AI products cautiously, conducting careful experiments to determine the impact of using AI to automate key business processes. In particular, we are extremely skeptical of the idea of AI agents being &#8220;drop-in replacements&#8221; for human workers that will somehow bypass the need for careful evaluation and integration of automation into workflows.</p><p>Companies developing AI products need a deep understanding of the domain to identify hurdles to adoption and build what businesses adopting AI want. For example, one key innovation of AI-enabled code editors such as Cursor and Windsurf is the user interface which allows programmers to verify AI-generated text at different levels of abstraction. Other industries will have different obstacles to AI adoption that products should address.</p><p>Policy makers should also take a long term view. A &#8220;Manhattan Project for AGI&#8221; is misguided on many levels. Since AGI is not a milestone, there is no way to know when the goal has been reached or how much more needs to be invested. And accelerating AI capabilities does nothing to address the real bottlenecks to realizing its economic benefits.</p><p>This view also has implications for export controls. The US has applied export controls on the hardware needed for AI development, hoping to slow down Chinese AI development. Proponents of export controls <a href="https://blog.ai-futures.org/p/why-america-wins">admit</a> (and we agree) that this will not widen the gap between the US and China by more than a few months. But this only matters in a world where the impacts of developing advanced AI are rapid. If the impact of advanced AI is realized through diffusion, and the process of diffusion takes decades, being ahead by a matter of months in a game of decades is of little consequence. Policy makers should therefore focus on enabling diffusion. We outline some ideas on how to do so in our <a href="https://www.aisnakeoil.com/p/ai-as-normal-technology">recent essay</a>.</p><div><hr></div><p>Treating AGI as a milestone for the development of transformative AI is seductive but misguided. It feeds into incorrect mental models of AI progress and risks, economic impacts, and geopolitics. AI's impact on the world will be realized not through a sprint toward a magic-bullet technology but through millions of boring little business process adaptations and policy tweaks.</p><p><em>We are grateful to Steve Newman and Jasmine Sun for feedback on a draft of this essay.</em></p><h3><strong>Further reading</strong></h3><ul><li><p><a href="https://jasmi.news/p/agi">Jasmine Sun</a>&#8217;s recent essay is similar to ours in spirit and substance. She contrasts many popular definitions of AGI and concludes that AGI is better seen as an aspiration for the AI community rather than a concrete technical milestone. She has also <a href="https://define-agi.com/">compiled</a> popular definitions of AGI and other related concepts.</p></li><li><p><a href="https://arxiv.org/abs/2502.03689">Borhane Blili-Hamelin and others</a> argue that we should stop treating AGI as the North star for AI research.</p></li><li><p><a href="https://epochai.substack.com/p/the-case-for-multi-decade-ai-timelines">Ege Erdil</a> makes the case for multi-decade timelines to transformative AI capabilities.</p></li><li><p><a href="https://www.techpolicy.press/most-researchers-do-not-believe-agi-is-imminent-why-do-policymakers-act-otherwise/">Eryk Salvaggio</a> gives many reasons why it is dangerous for policy makers to act as if AGI is imminent.</p></li><li><p>In <a href="https://www.aisnakeoil.com/p/ai-as-normal-technology">AI as Normal Technology</a>, we laid the intellectual foundation for our view of the future of AI. The present essay is the first of many follow-ups that explore the implications of the ideas therein.</p></li></ul><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>While LLMs can only output text, models like o3 are trained to use their text outputs to interact with tools. For example, o3 can output a specific keyword (say, "use-search") that allows it to search for a given query online. To be clear, previous LLMs could also interact with tools. For example, OpenAI's <a href="https://openai.com/index/chatgpt-plugins/">ChatGPT plugins</a> allowed models like GPT-4 to use tools two years ago. The key difference with models like o3 is that they are trained<em> </em>to use tools effectively. Previous models used tools based on prompts without explicitly being trained for this task, or at best were taught the syntax of tool use. o3 has access to many tools on the ChatGPT interface, including search, the ability to run code, access files, and generate and reason about images. Overall, o3 is much more like an "agent" than a chatbot.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p>For example, suppose this system is used to carry out a new type of coding task, perhaps using a software library that was only released after it finished training. Given enough documentation and instructions, it might eventually correctly perform the task. But barring an explicit update to the model via retraining, this "experience" of correctly performing the task would not change the model's weights, so another user needing to perform the same task would need to give it the same help that the first user did.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-3" href="#footnote-anchor-3" class="footnote-number" contenteditable="false" target="_self">3</a><div class="footnote-content"><p>We do not analyze military AI in this piece. It is an important topic that we will return to in future essays.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-4" href="#footnote-anchor-4" class="footnote-number" contenteditable="false" target="_self">4</a><div class="footnote-content"><p>It is important to <a href="https://www.aisnakeoil.com/p/ai-as-normal-technology">distinguish</a> between the invention of new technology and innovations that help those technologies diffuse across society; our discussion in this paragraph is about the former. Invention refers to the development of new models, perhaps accompanied by new capabilities. Innovations enable these capabilities to be used productively. Better and more capable models are an example of invention, whereas the development of <a href="https://www.aisnakeoil.com/p/ai-companies-are-pivoting-from-creating">products</a> that use these models is an example of innovation. We don't think it is likely that countries will be able to maintain a long and sustained advantage in invention. But as we discuss later in this section, the conditions that enable the productive use of inventions can differ greatly between countries.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-5" href="#footnote-anchor-5" class="footnote-number" contenteditable="false" target="_self">5</a><div class="footnote-content"><p>Amodei eschews AGI and uses the term &#8220;powerful AI&#8221; to be more precise about his predictions.</p></div></div>]]></content:encoded></item><item><title><![CDATA[AI as Normal Technology]]></title><description><![CDATA[A new paper that we will expand into our next book]]></description><link>https://www.normaltech.ai/p/ai-as-normal-technology</link><guid isPermaLink="false">https://www.normaltech.ai/p/ai-as-normal-technology</guid><dc:creator><![CDATA[Arvind Narayanan]]></dc:creator><pubDate>Tue, 15 Apr 2025 14:53:59 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/ac4b6c2a-d282-4124-a4fa-360ae04c9abe_2602x1550.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>This post is over 15,000 words long&#8212;it is a new paper on our vision for the future of AI. We are pleased to announce that an expanded version of these ideas will become our next book together.</em></p><p><em>The paper is also published in <a href="https://knightcolumbia.org/content/ai-as-normal-technology">HTML</a> and <a href="https://kfai-documents.s3.amazonaws.com/documents/c3cac5a2a7/AI-as-Normal-Technology---Narayanan---Kapoor.pdf">PDF</a> formats on the Knight First Amendment Institute&#8217;s website. We are grateful for the extensive feedback we&#8217;ve received on drafts of the paper.</em></p><p><em><strong>Update (September 2025)</strong>: We have published a companion to this essay titled <a href="https://www.normaltech.ai/p/a-guide-to-understanding-ai-as-normal">A guide to understanding AI as normal technology</a>.</em></p><div><hr></div><p>We articulate a vision of artificial intelligence (AI) as <em>normal technology</em>. To view AI as normal is not to understate its impact&#8212;even transformative, general-purpose technologies such as electricity and the internet are &#8220;normal&#8221; in our conception. But it is in contrast to both utopian and dystopian visions of the future of AI which have a common tendency to treat it akin to a separate species, a highly autonomous, potentially superintelligent entity.<sup>1</sup></p><p>The statement &#8220;AI is normal technology&#8221; is three things: a <em>description</em> of current AI, a <em>prediction</em> about the foreseeable future of AI, and a <em>prescription</em> about how we should treat it. We view AI as a tool that we can and should remain in control of, and we argue that this goal does not require drastic policy interventions or technical breakthroughs. We do not think that viewing AI as a humanlike intelligence is currently accurate or useful for understanding its societal impacts, nor is it likely to be in our vision of the future.<sup>2</sup></p><p>The normal technology frame is about the relationship between technology and society. It rejects technological determinism, especially the notion of AI itself as an agent in determining its future. It is guided by lessons from past technological revolutions, such as the slow and uncertain nature of technology adoption and diffusion. It also emphasizes continuity between the past and the future trajectory of AI in terms of societal impact and the role of institutions in shaping this trajectory.</p><p>In Part I, we explain why we think that transformative economic and societal impacts will be slow (on the timescale of decades), making a critical distinction between AI methods, AI applications, and AI adoption, arguing that the three happen at different timescales.</p><p>In Part II, we discuss a potential division of labor between humans and AI in a world with advanced AI (but not &#8220;superintelligent&#8221; AI, which we view as incoherent as usually conceptualized). In this world, control is primarily in the hands of people and organizations; indeed, a greater and greater proportion of what people do in their jobs is AI control.</p><p>In Part III, we examine the implications of AI as normal technology for AI risks. We analyze accidents, arms races, misuse, and misalignment, and argue that viewing AI as normal technology leads to fundamentally different conclusions about mitigations compared to viewing AI as being humanlike.</p><p>Of course, we cannot be certain of our predictions, but we aim to describe what we view as the median outcome. We have not tried to quantify probabilities, but we have tried to make predictions that can tell us whether or not AI is behaving like normal technology.</p><p>In Part IV, we discuss the implications for AI policy. We advocate for reducing uncertainty as a first-rate policy goal and resilience as the overarching approach to catastrophic risks. We argue that drastic interventions premised on the difficulty of controlling superintelligent AI will, in fact, make things much worse if AI turns out to be normal technology&#8212; the downsides of which will be likely to mirror those of previous technologies that are deployed in capitalistic societies, such as inequality.<sup>3</sup></p><p>The world we describe in Part II is one in which AI is far more advanced than it is today. We are not claiming that AI progress&#8212;or human progress&#8212;will stop at that point. What comes after it? We do not know. Consider this analogy: At the dawn of the first Industrial Revolution, it would have been useful to try to think about what an industrial world would look like and how to prepare for it, but it would have been futile to try to predict electricity or computers. Our exercise here is similar. Since we reject &#8220;fast takeoff&#8221; scenarios, we do not see it as necessary or useful to envision a world further ahead than we have attempted to. If and when the scenario we describe in Part II materializes, we will be able to better anticipate and prepare for whatever comes next.</p><p><em><strong>A note to readers.</strong> This essay has the unusual goal of stating a worldview rather than defending a proposition. The literature on AI superintelligence is copious. We have not tried to give a point-by-point response to potential counter arguments, as that would make the paper several times longer. This paper is merely the initial articulation of our views; we plan to elaborate on them in various follow ups.</em></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.normaltech.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Subscribe to receive follow-up essays.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p></p><h1>Part I: The Speed of Progress</h1><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!77Az!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faa924fba-457a-41d4-859c-615d148e3337_1862x972.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!77Az!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faa924fba-457a-41d4-859c-615d148e3337_1862x972.png 424w, https://substackcdn.com/image/fetch/$s_!77Az!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faa924fba-457a-41d4-859c-615d148e3337_1862x972.png 848w, https://substackcdn.com/image/fetch/$s_!77Az!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faa924fba-457a-41d4-859c-615d148e3337_1862x972.png 1272w, https://substackcdn.com/image/fetch/$s_!77Az!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faa924fba-457a-41d4-859c-615d148e3337_1862x972.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!77Az!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faa924fba-457a-41d4-859c-615d148e3337_1862x972.png" width="1456" height="760" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/aa924fba-457a-41d4-859c-615d148e3337_1862x972.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:760,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:364966,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.aisnakeoil.com/i/161317202?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faa924fba-457a-41d4-859c-615d148e3337_1862x972.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!77Az!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faa924fba-457a-41d4-859c-615d148e3337_1862x972.png 424w, https://substackcdn.com/image/fetch/$s_!77Az!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faa924fba-457a-41d4-859c-615d148e3337_1862x972.png 848w, https://substackcdn.com/image/fetch/$s_!77Az!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faa924fba-457a-41d4-859c-615d148e3337_1862x972.png 1272w, https://substackcdn.com/image/fetch/$s_!77Az!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faa924fba-457a-41d4-859c-615d148e3337_1862x972.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em>Figure 1. Like other general-purpose technologies, the impact of AI is materialized not when methods and capabilities improve, but when those improvements are translated into applications and are diffused through productive sectors of the economy.<sup>4</sup> There are speed limits at each stage.</em></figcaption></figure></div><p>Will the progress of AI be gradual, allowing people and institutions to adapt as AI capabilities and adoption increase, or will there be jumps leading to massive disruption, or even a technological singularity? Our approach to this question is to analyze highly consequential tasks separately from less consequential tasks and to begin by analyzing the speed of adoption and diffusion of AI before returning to the speed of innovation and invention.</p><p>We use <em>invention</em> to refer to the development of new AI methods&#8212;such as large language models&#8212;that improve AI&#8217;s capabilities to carry out various tasks. <em>Innovation</em> refers to the development of products and applications using AI that consumers and businesses can use. <em>Adoption</em> refers to the decision by an individual (or team or firm) to use a technology, whereas <em>diffusion</em> refers to the broader social process through which the level of adoption increases. For sufficiently disruptive technologies, diffusion might require changes to the structure of firms and organizations, as well as to social norms and laws.</p><h2>AI Diffusion in Safety-critical Areas Is Slow</h2><p>In the paper <em>Against Predictive Optimization</em>, we compiled a comprehensive list of about 50 applications of predictive optimization, namely the use of machine learning (ML) to make decisions about individuals by predicting their future behavior or outcomes.<sup>5</sup> Most of these applications, such as criminal risk prediction, insurance risk prediction, or child maltreatment prediction, are used to make decisions that have important consequences for people.</p><p>While these applications have proliferated, there is a crucial nuance: In most cases, decades-old statistical techniques are used&#8212;simple, interpretable models (mostly regression) and relatively small sets of handcrafted features. More complex machine learning methods, such as random forests, are rarely used, and modern methods, such as transformers, are nowhere to be found.</p><p>In other words, in this broad set of domains, AI diffusion lags <em>decades</em> behind innovation. A major reason is safety&#8212;when models are more complex and less intelligible, it is hard to anticipate all possible deployment conditions in the testing and validation process. A good example is Epic&#8217;s sepsis prediction tool which, despite having seemingly high accuracy when internally validated, performed far worse in hospitals, missing two thirds of sepsis cases and overwhelming physicians with false alerts.<sup>6</sup></p><p>Epic&#8217;s sepsis prediction tool failed because of errors that are hard to catch when you have complex models with unconstrained feature sets.<sup>7</sup> In particular, one of the features used to train the model was whether a physician had already prescribed antibiotics &#8212;to treat sepsis. In other words, during testing and validation, the model was using a feature from the future, relying on a variable that was causally dependent on the outcome. Of course, this feature would not be available during deployment. Interpretability and auditing methods will no doubt improve so that we will get much better at catching these issues, but we are not there yet.</p><p>In the case of generative AI, even failures that seem extremely obvious in hindsight were not caught during testing. One example is the early Bing chatbot &#8220;Sydney&#8221; that went off the rails during extended conversations; the developers evidently did not anticipate that conversations could last for more than a handful of turns.<sup>8</sup> Similarly, the Gemini image generator was seemingly never tested on historical figures.<sup>9</sup> Fortunately, these were not highly consequential applications.</p><p>More empirical work would be helpful for understanding the innovation-diffusion lag in various applications and the reasons for this lag. But, for now, the evidence that we have analyzed in our previous work is consistent with the view that there are already extremely strong safety-related speed limits in highly consequential tasks. These limits are often enforced through regulation, such as the FDA&#8217;s supervision of medical devices, as well as newer legislation such as the EU AI Act, which puts strict requirements on high-risk AI.<sup>10</sup> In fact, there are (credible) concerns that existing regulation of high-risk AI is so onerous that it may lead to &#8220;runaway bureaucracy&#8221;.<sup>11</sup> Thus, we predict that slow diffusion will continue to be the norm in high-consequence tasks.</p><p>At any rate, as and when new areas arise in which AI can be used in highly consequential ways, we can and must regulate them. A good example is the Flash Crash of 2010, in which automated high-frequency trading is thought to have played a part. This led to new curbs on trading, such as circuit breakers.<sup>12</sup></p><h2>Diffusion is Limited by the Speed of Human, Organizational, and Institutional Change</h2><p>Even outside of safety-critical areas, AI adoption is slower than popular accounts would suggest. For example, a study made headlines due to the finding that, in August 2024, 40% of U.S. adults used generative AI.<sup>13</sup> But, because most people used it infrequently, this only translated to 0.5%-3.5% of work hours (and a 0.125-0.875 percentage point increase in labor productivity).</p><p>It is not even clear if the speed of diffusion is greater today compared to the past. The aforementioned study reported that generative AI adoption in the U.S. has been faster than personal computer (PC) adoption, with 40% of U.S. adults adopting generative AI within two years of the first mass-market product release compared to 20 % within three years for PCs. But this comparison does not account for differences in the intensity of adoption (the number of hours of use) or the high cost of buying a PC compared to accessing generative AI.<sup>14</sup> Depending on how we measure adoption, it is quite possible that the adoption of generative AI has been much slower than PC adoption.</p><p>The claim that the speed of technology adoption is not necessarily increasing may seem surprising (or even obviously wrong) given that digital technology can reach billions of devices at once. But it is important to remember that adoption is about software use, not availability. Even if a new AI-based product is instantly released online for anyone to use for free, it takes time to for people to change their workflows and habits to take advantage of the benefits of the new product and to learn to avoid the risks.</p><p>Thus, the speed of diffusion is inherently limited by the speed at which not only individuals, but also organizations and institutions, can adapt to technology. This is a trend that we have also seen for past general-purpose technologies: Diffusion occurs over decades, not years.<sup>15</sup></p><p>As an example, Paul A. David&#8217;s analysis of electrification shows that the productivity benefits took decades to fully materialize.<sup>16</sup> Electric dynamos were &#8220;everywhere but in the productivity statistics&#8221; for nearly 40 years after Edison&#8217;s first central generating station. <sup>17</sup> This was not just technological inertia; factory owners found that electrification did not bring substantial efficiency gains.</p><p>What eventually allowed gains to be realized was redesigning the entire layout of factories around the logic of production lines. In addition to changes to factory architecture, diffusion also required changes to workplace organization and process control, which could only be developed through experimentation across industries. Workers had more autonomy and flexibility as a result of the changes, which also necessitated different hiring and training practices.</p><h2>The External World Puts a Speed Limit on AI Innovation</h2><p>It is true that technical advances in AI have been rapid, but the picture is much less clear when we differentiate AI methods from applications.</p><p>We conceptualize progress in AI methods as a ladder of generality.<sup>18</sup> Each step on this ladder rests on the ones below it and reflects a move toward more general computing capabilities. That is, it reduces the programmer effort needed to get the computer to perform a new task and increases the set of tasks that can be performed with a given amount of programmer (or user) effort; see Figure 2. For example, machine learning increases generality by obviating the need for the programmer to devise logic to solve each new task, only requiring the collection of training examples instead.</p><p>It is tempting to conclude that the effort required to develop specific applications will keep decreasing as we build more rungs of the ladder until we reach artificial general intelligence, often conceptualized as an AI system that can do everything out of the box, obviating the need to develop applications altogether.</p><p>In some domains, we are indeed seeing this trend of decreasing application development effort. In natural language processing, large language models have made it relatively trivial to develop a language translation application. Or consider games: AlphaZero can learn to play games such as chess better than any human through self-play given little more than a description of the game and enough computing power&#8212;a far cry from how game-playing programs used to be developed.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Yt5s!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fce4e6c48-9d4e-4964-8888-caa47b033447_2042x986.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Yt5s!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fce4e6c48-9d4e-4964-8888-caa47b033447_2042x986.png 424w, https://substackcdn.com/image/fetch/$s_!Yt5s!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fce4e6c48-9d4e-4964-8888-caa47b033447_2042x986.png 848w, https://substackcdn.com/image/fetch/$s_!Yt5s!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fce4e6c48-9d4e-4964-8888-caa47b033447_2042x986.png 1272w, https://substackcdn.com/image/fetch/$s_!Yt5s!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fce4e6c48-9d4e-4964-8888-caa47b033447_2042x986.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Yt5s!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fce4e6c48-9d4e-4964-8888-caa47b033447_2042x986.png" width="1456" height="703" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/ce4e6c48-9d4e-4964-8888-caa47b033447_2042x986.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:703,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:468887,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.aisnakeoil.com/i/161317202?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fce4e6c48-9d4e-4964-8888-caa47b033447_2042x986.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Yt5s!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fce4e6c48-9d4e-4964-8888-caa47b033447_2042x986.png 424w, https://substackcdn.com/image/fetch/$s_!Yt5s!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fce4e6c48-9d4e-4964-8888-caa47b033447_2042x986.png 848w, https://substackcdn.com/image/fetch/$s_!Yt5s!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fce4e6c48-9d4e-4964-8888-caa47b033447_2042x986.png 1272w, https://substackcdn.com/image/fetch/$s_!Yt5s!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fce4e6c48-9d4e-4964-8888-caa47b033447_2042x986.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em>Figure 2: The Ladder of Generality in Computing. For some tasks, higher ladder rungs require less programmer effort to get a computer to perform a new task, and more tasks can be performed with a given amount of programmer (or user) effort.</em><sup>19</sup></figcaption></figure></div><p>However, this has not been the trend in highly consequential, real-world applications that cannot easily be simulated and in which errors are costly. Consider self-driving cars: In many ways, the trajectory of their development is similar to AlphaZero&#8217;s self-play&#8212;improving the tech allowed them to drive in more realistic conditions, which enabled the collection of better and/or more realistic data, which in turn led to improvements in the tech, completing the feedback loop. But this process took over two decades instead of a few hours in the case of AlphaZero because safety considerations put a limit on the extent to which each iteration of this loop could be scaled up compared to the previous one.<sup>20</sup></p><p>This &#8220;capability-reliability gap&#8221; shows up over and over. It has been a major barrier to building useful AI &#8220;agents&#8221; that can automate real-world tasks.<sup>21</sup> To be clear, many tasks for which the use of agents is envisioned, such as booking travel or providing customer service, are far less consequential than driving, but still costly enough that having agents learn from real-world experiences is not straightforward.</p><p>Barriers also exist in non-safety-critical applications. In general, much knowledge is tacit in organizations and is not written down, much less in a form that can be learned passively. This means that these developmental feedback loops will have to happen in each sector and, for more complex tasks, may even need to occur separately in different organizations, limiting opportunities for rapid, parallel learning. Other reasons why parallel learning might be limited are privacy concerns: Organizations and individuals might be averse to sharing sensitive data with AI companies, and regulations might limit what kinds of data can be shared with third parties in contexts such as healthcare.</p><p>The &#8220;bitter lesson&#8221; in AI is that general methods that leverage increases in computational power eventually surpass methods that utilize human domain knowledge by a large margin.<sup>22</sup> This is a valuable observation about <em>methods</em>, but it is often misinterpreted to encompass application development. In the context of AI-based product development, the bitter lesson has never been even close to true.<sup>23</sup> Consider recommender systems on social media: They are powered by (increasingly general) machine learning models, but this has not obviated the need for manual coding of the business logic, the frontend, and other components which, together, can comprise on the order of a million lines of code.</p><p>Further limits arise when we need to go beyond AI learning from existing human knowledge.<sup>24</sup> Some of our most valuable types of knowledge are scientific and social-scientific, and have allowed the progress of civilization through technology and large-scale social organizations (e.g., governments). What will it take for AI to push the boundaries of such knowledge? It will likely require interactions with, or even experiments on, people or organizations, ranging from drug testing to economic policy. Here, there are hard limits to the speed of knowledge acquisition because of the social costs of experimentation. Societies probably will not (and should not) allow the rapid scaling of experiments for AI development.</p><h2>Benchmarks Do not Measure Real-World Utility</h2><p>The methods-application distinction has important implications for how we measure and forecast AI progress. AI benchmarks are useful for measuring progress in methods; unfortunately, they have often been misunderstood as measuring progress in applications, and this confusion has been a driver of much hype about imminent economic transformation.</p><p>For example, while GPT-4 reportedly achieved scores in the top 10% of bar exam test takers, this tells us remarkably little about AI&#8217;s ability to practice law.<sup>25</sup> The bar exam overemphasizes subject-matter knowledge and under-emphasizes real-world skills that are far harder to measure in a standardized, computer-administered format. In other words, it emphasizes precisely what language models are good at&#8212;retrieving and applying memorized information.</p><p>More broadly, tasks that would lead to the most significant changes to the legal profession are also the hardest ones to evaluate. Evaluation is straightforward for tasks like categorizing legal requests by area of law because there are clear correct answers. But for tasks that involve creativity and judgment, like preparing legal filings, there is no single correct answer, and reasonable people can disagree about strategy. These latter tasks are precisely the ones that, if automated, would have the most profound impact on the profession.<sup>26</sup></p><p>This observation is in no way limited to law. Another example is the gap between self-contained coding problems at which AI demonstrably excels, and real-world software engineering in which its impact is hard to measure but appears to be modest.<sup>27</sup> Even highly regarded coding benchmarks that go beyond toy problems must necessarily ignore many dimensions of real-world software engineering in the interest of quantification and automated evaluation using publicly available data.<sup>28</sup></p><p>This pattern appears repeatedly: The easier a task is to measure via benchmarks, the less likely it is to represent the kind of complex, contextual work that defines professional practice. By focusing heavily on capability benchmarks to inform our understanding of AI progress, the AI community consistently overestimates the real-world impact of the technology.</p><p>This is a problem of &#8216;construct validity,&#8217; which refers to whether a test actually measures what it is intended to measure.<sup>29</sup> The only sure way to measure real-world usefulness of a potential application is to actually build the application and to then test it with professionals in realistic scenarios (either substituting or augmenting their labor, depending on the intended use). Such &#8216;uplift&#8217; studies generally do show that professionals in many occupations benefit from existing AI systems, but this benefit is typically modest and is more about augmentation than substitution, a radically different picture from what one might conclude based on static benchmarks like exams<sup>30</sup> (a small number of occupations such as copywriters and translators have seen substantial job losses<sup>31</sup>).</p><p>In conclusion, while benchmarks are valuable for tracking progress in AI methods, we should look at other kinds of metrics to track AI impacts (Figure <em>1</em>). When measuring adoption, we must take into account the intensity of AI use. The type of application is also important: Augmentation versus substitution and high-consequence versus low-consequence.</p><p>The difficulty of ensuring construct validity afflicts not only benchmarking, but also forecasting, which is another major way in which people try to assess (future) AI impacts. It is extremely important to avoid ambiguous outcomes to ensure effective forecasting. The way that the forecasting community accomplishes this is by defining milestones in terms of relatively narrow skills, such as exam performance. For instance, the Metaculus question on &#8220;human-machine intelligence parity&#8221; is defined in terms of performance on exam questions in math, physics, and computer science. Based on this definition, it is not surprising that forecasters predict a 95% chance of achieving &#8220;human-machine intelligence parity&#8221; by 2040. <sup>32</sup></p><p>Unfortunately, this definition is so watered down that it does not mean much for understanding the impacts of AI. As we saw above with legal and other professional benchmarks, AI performance on exams has so little construct validity that it does not even allow us to predict whether AI will replace professional workers.</p><h2>Economic Impacts are Likely to be Gradual</h2><p>One argument for why AI development may have sudden, drastic economic impacts is that an increase in generality may lead to a wide swath of tasks in the economy becoming automatable. This is related to one definition of artificial general intelligence (AGI)&#8212;a unified system that is capable of performing all economically valuable tasks.</p><p>According to the normal technology view, such sudden economic impacts are implausible. In the previous sections, we discussed one reason: Sudden improvements in AI methods are certainly possible but do not directly translate to economic impacts, which require innovation (in the sense of application development) and diffusion.</p><p>Innovation and diffusion happen in a feedback loop. In safety-critical applications, this feedback loop is always slow, but even beyond safety, there are many reasons why it is likely to be slow. With past general-purpose technologies such as electricity, computers, and the internet, the respective feedback loops unfolded over several decades, and we should expect the same to happen with AI as well.</p><p>Another argument for gradual economic impacts: Once we automate something, its cost of production, and its value, tend to drop drastically over time compared to the cost of human labor. As automation increases, humans will adapt, and will focus on tasks that are not yet automated, perhaps tasks that do not exist today (in Part II we describe what those might look like).</p><p>This means that the goalpost of AGI will continually move further away as increasing automation redefines which tasks are economically valuable. Even if every task that humans do <em>today</em> might be automated one day, this does not mean that human labor will be superfluous.</p><p>All of this points away from the likelihood of the automation of a vast swath of the economy at a particular moment in time. It also implies that the impacts of powerful AI will be felt on different timescales in different sectors.</p><h2>Speed Limits to Progress in AI Methods</h2><p>Our argument for the slowness of AI impact is based on the innovation-diffusion feedback loop, and is applicable even if progress in AI methods can be arbitrarily sped up. We see both benefits and risks as arising primarily from AI deployment rather than from development; thus, the speed of progress in AI methods is not directly relevant to the question of impacts. Nonetheless, it is worth discussing speed limits that also apply to methods development.</p><p>The production of AI research has been increasing exponentially, with the rate of publication of AI/ML papers on arXiv exhibiting a doubling time under two years.<sup>33</sup> But it is not clear how this increase in volume translates to progress. One measure of progress is the rate of turnover of central ideas. Unfortunately, throughout its history, the AI field has shown a high degree of herding around popular ideas, and inadequate (in retrospect) levels of exploration of unfashionable ones. A notable example is the sidelining of research on neural networks for many decades.</p><p>Is the current era different? Although ideas incrementally accrue at increasing rates, are they turning over established ones? The transformer architecture has been the dominant paradigm for most of the last decade, despite its well-known limitations. By analyzing over a billion citations in 241 subjects, Johan S.G. Chu &amp; James A. Evans showed that, in fields in which the volume of papers is higher, it is harder, not easier, for new ideas to break through. This leads to an &#8220;ossification of canon.&#8221;<sup>34</sup> Perhaps this description applies to the current state of AI methods research.</p><p>Many other speed limits are possible. Historically, deep neural network technology was partly held back due to the inadequacy of hardware, particularly Graphics Processing Units. Computational and cost limits continue to be relevant to new paradigms, including inference-time scaling. New slowdowns may emerge: Recent signs point to a shift away from the culture of open knowledge sharing in the industry.</p><p>It remains to be seen if AI-conducted AI research can offer a reprieve. Perhaps recursive self-improvement in methods is possible, resulting in unbounded speedups in methods. But note that AI development already relies heavily on AI. It is more likely that we will continue to see a gradual increase in the role of automation in AI development than a singular, discontinuous moment when recursive self-improvement is achieved.<sup>35</sup></p><p>Earlier, we argued that benchmarks give a misleading picture of the usefulness of AI applications. But they have arguably also led to overoptimism about the speed of methods progress. One reason is that it is hard to design benchmarks that make sense beyond the current horizon of progress. The Turing test was the north star of AI for many decades because of the assumption that any system that passed it would be humanlike in important ways, and that we would be able to use such a system to automate a variety of complex tasks. Now that large language models can arguably pass it while only weakly meeting the expectations behind the test, its significance has waned.<sup>36</sup></p><p>An analogy with mountaineering is apt. Every time we solve a benchmark (reach what we thought was the peak), we discover limitations of the benchmark (realize that we&#8217;re on a &#8216;false summit&#8217;) and construct a new benchmark (set our sights on what we now think is the summit). This leads to accusations of &#8216;moving the goalposts&#8217;, but this is what we should expect given the intrinsic challenges of benchmarking.</p><p>AI pioneers considered the two big challenges of AI (what we now call AGI) to be (what we now call) hardware and software. Having built programmable machines, there was a palpable sense that AGI was close. The organizers of the 1956 Dartmouth conference hoped to make significant progress toward the goal through a &#8220;2-month, 10-man&#8221; effort.<sup>37</sup> Today, we have climbed many more rungs on the ladder of generality. We often hear that all that is needed to build AGI is scaling, or generalist AI agents, or sample-efficient learning.</p><p>But it is useful to bear in mind that what appears to be a single step might not be so. For example, there may not exist one single breakthrough algorithm that enables sample-efficient learning across all contexts. Indeed, in-context learning in large language models is already &#8220;sample efficient,&#8221; but only works for a limited set of tasks.<sup>38</sup></p><h1>Part II: What a World With Advanced AI Might Look Like</h1><p>We argue that reliance on the slippery concepts of &#8216;intelligence&#8217; and &#8216;superintelligence&#8217; has clouded our ability to reason clearly about a world with advanced AI. By unpacking intelligence into distinct underlying concepts, capability and power, we rebut the notion that human labor will be superfluous in a world with &#8216;superintelligent&#8217; AI, and present an alternative vision. This also lays the foundation for our discussion of risks in Part III.</p><h2>Human Ability Is Not Constrained by Biology</h2><p>Can AI exceed human intelligence and, if so, by how much? According to a popular argument, unfathomably so. This is often depicted by comparing different species along a spectrum of intelligence.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!VLFz!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe295e8d4-7558-4d94-ba2d-fa4e590ffe17_2014x416.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!VLFz!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe295e8d4-7558-4d94-ba2d-fa4e590ffe17_2014x416.png 424w, https://substackcdn.com/image/fetch/$s_!VLFz!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe295e8d4-7558-4d94-ba2d-fa4e590ffe17_2014x416.png 848w, https://substackcdn.com/image/fetch/$s_!VLFz!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe295e8d4-7558-4d94-ba2d-fa4e590ffe17_2014x416.png 1272w, https://substackcdn.com/image/fetch/$s_!VLFz!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe295e8d4-7558-4d94-ba2d-fa4e590ffe17_2014x416.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!VLFz!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe295e8d4-7558-4d94-ba2d-fa4e590ffe17_2014x416.png" width="1456" height="301" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/e295e8d4-7558-4d94-ba2d-fa4e590ffe17_2014x416.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:301,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:99047,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.aisnakeoil.com/i/161317202?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe295e8d4-7558-4d94-ba2d-fa4e590ffe17_2014x416.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!VLFz!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe295e8d4-7558-4d94-ba2d-fa4e590ffe17_2014x416.png 424w, https://substackcdn.com/image/fetch/$s_!VLFz!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe295e8d4-7558-4d94-ba2d-fa4e590ffe17_2014x416.png 848w, https://substackcdn.com/image/fetch/$s_!VLFz!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe295e8d4-7558-4d94-ba2d-fa4e590ffe17_2014x416.png 1272w, https://substackcdn.com/image/fetch/$s_!VLFz!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe295e8d4-7558-4d94-ba2d-fa4e590ffe17_2014x416.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">Figure 3. Intelligence explosion through recursively self-improved AI is a common concern, often depicted by figures like this one. Figure redrawn.<sup>39</sup></figcaption></figure></div><p>However, there are conceptual and logical flaws with this picture. On a conceptual level, intelligence&#8212;especially as a comparison between different species&#8212;is not well defined, let alone measurable on a one-dimensional scale.<sup>40</sup></p><p>More importantly, intelligence is not the property at stake for analyzing AI&#8217;s impacts. Rather, what is at stake is power&#8212;the ability to modify one&#8217;s environment. To clearly analyze the impact of technology (and in particular, increasingly general computing technology), we must investigate how technology has affected humanity&#8217;s power. When we look at things from this perspective, a completely different picture emerges.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!5xra!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0f5b310f-3c31-45a7-8e57-aafb87b0405b_1868x428.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!5xra!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0f5b310f-3c31-45a7-8e57-aafb87b0405b_1868x428.png 424w, https://substackcdn.com/image/fetch/$s_!5xra!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0f5b310f-3c31-45a7-8e57-aafb87b0405b_1868x428.png 848w, https://substackcdn.com/image/fetch/$s_!5xra!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0f5b310f-3c31-45a7-8e57-aafb87b0405b_1868x428.png 1272w, https://substackcdn.com/image/fetch/$s_!5xra!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0f5b310f-3c31-45a7-8e57-aafb87b0405b_1868x428.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!5xra!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0f5b310f-3c31-45a7-8e57-aafb87b0405b_1868x428.png" width="1456" height="334" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/0f5b310f-3c31-45a7-8e57-aafb87b0405b_1868x428.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:334,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:74565,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.aisnakeoil.com/i/161317202?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0f5b310f-3c31-45a7-8e57-aafb87b0405b_1868x428.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!5xra!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0f5b310f-3c31-45a7-8e57-aafb87b0405b_1868x428.png 424w, https://substackcdn.com/image/fetch/$s_!5xra!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0f5b310f-3c31-45a7-8e57-aafb87b0405b_1868x428.png 848w, https://substackcdn.com/image/fetch/$s_!5xra!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0f5b310f-3c31-45a7-8e57-aafb87b0405b_1868x428.png 1272w, https://substackcdn.com/image/fetch/$s_!5xra!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0f5b310f-3c31-45a7-8e57-aafb87b0405b_1868x428.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">Figure 4. Analyzing the impact of technology on humanity&#8217;s power. We are powerful not because of our intelligence, but because of the technology we use to increase our capabilities.</figcaption></figure></div><p>This shift in perspective clarifies that humans have always used technology to increase our ability to control our environment. There are few biological or physiological differences between ancestral and modern humans; instead, the relevant differences are improved knowledge and understanding, tools, technology and, indeed, AI. In a sense, modern humans, with the capability to alter the planet and its climate, are &#8216;superintelligent&#8217; beings compared to pre-technological humans. Unfortunately, much of the foundational literature analyzing the risks of AI superintelligence suffers from a lack of precision in the use of the term &#8216;intelligence.&#8217;</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!0DUQ!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe07a0702-89c5-40a8-bb48-4eabefc31d26_2091x1252.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!0DUQ!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe07a0702-89c5-40a8-bb48-4eabefc31d26_2091x1252.jpeg 424w, https://substackcdn.com/image/fetch/$s_!0DUQ!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe07a0702-89c5-40a8-bb48-4eabefc31d26_2091x1252.jpeg 848w, https://substackcdn.com/image/fetch/$s_!0DUQ!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe07a0702-89c5-40a8-bb48-4eabefc31d26_2091x1252.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!0DUQ!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe07a0702-89c5-40a8-bb48-4eabefc31d26_2091x1252.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!0DUQ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe07a0702-89c5-40a8-bb48-4eabefc31d26_2091x1252.jpeg" width="1456" height="872" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/e07a0702-89c5-40a8-bb48-4eabefc31d26_2091x1252.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:872,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:700256,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.aisnakeoil.com/i/161317202?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe07a0702-89c5-40a8-bb48-4eabefc31d26_2091x1252.jpeg&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!0DUQ!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe07a0702-89c5-40a8-bb48-4eabefc31d26_2091x1252.jpeg 424w, https://substackcdn.com/image/fetch/$s_!0DUQ!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe07a0702-89c5-40a8-bb48-4eabefc31d26_2091x1252.jpeg 848w, https://substackcdn.com/image/fetch/$s_!0DUQ!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe07a0702-89c5-40a8-bb48-4eabefc31d26_2091x1252.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!0DUQ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe07a0702-89c5-40a8-bb48-4eabefc31d26_2091x1252.jpeg 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Figure 5. Two views of the causal chain from increases in AI capability to loss of control.</figcaption></figure></div><p>Once we stop using the terms &#8216;intelligence&#8217; and &#8216;superintelligence,&#8217; things become much clearer (Figure 5). The worry is that if AI capabilities continue to increase indefinitely (whether or not they are humanlike or superhuman is irrelevant), they may lead to AI systems with more and more power, in turn leading to a loss of control. If we accept that capabilities are likely to increase indefinitely (we do), our options for preventing a loss of control are to intervene in one of the two causal steps.</p><p>The superintelligence view is pessimistic about the first arrow in Figure 5&#8212;preventing arbitrarily capable AI systems from acquiring power that is significant enough to pose catastrophic risks&#8212;and instead focuses on alignment techniques that try to prevent arbitrarily powerful AI systems from acting against human interests. Our view is precisely the opposite, as we elaborate in the rest of this paper.</p><h2>Games Provide Misleading Intuitions About the Possibility of Superintelligence</h2><p>De-emphasizing intelligence is not just a rhetorical move: We do not think there is a useful sense of the term &#8216;intelligence&#8217; in which AI is more intelligent than people acting with the help of AI. Human intelligence is special due to our ability to use tools and to subsume other intelligences into our own, and cannot be coherently placed on a spectrum of intelligence.</p><p>Human abilities definitely have some important limitations, notably speed. This is why machines dramatically outperform humans in domains like chess and, in a human+AI team, the human can hardly do better than simply deferring to AI. But speed limitations are irrelevant in most areas because high-speed sequential calculations or fast reaction times are not required.</p><p>In the few real-world tasks for which superhuman speed is required, such as nuclear reactor control, we are good at building tightly scoped automated tools to do the high-speed parts, while humans retain control of the overall system.</p><p>We offer a prediction based on this view of human abilities. We think there are relatively few real-world cognitive tasks in which human limitations are so telling that AI is able to blow past human performance (as AI does in chess). In many other areas, including some that are associated with prominent hopes and fears about AI performance, we think there is a high &#8220;irreducible error&#8221;&#8212;unavoidable error due to the inherent stochasticity of the phenomenon&#8212;and human performance is essentially near that limit.<sup>41</sup></p><p>Concretely, we propose two such areas: forecasting and persuasion. We predict that AI will not be able to meaningfully outperform trained humans (particularly teams of humans and especially if augmented with simple automated tools) at forecasting geopolitical events (say elections). We make the same prediction for the task of persuading people to act against their own self-interest.</p><p>The self-interest aspect of persuasion is a critical one, but is often underappreciated. As an illustrative example of a common pattern, consider the study &#8220;Evaluating Frontier Models for Dangerous Capabilities,&#8221; which evaluated language models&#8217; abilities to persuade people.<sup>42</sup> Some of their persuasion tests were costless to the subjects being persuaded; they were simply asked whether they believed a claim at the end of the interaction with AI. Other tests had small costs, such as forfeiting a &#163;20 bonus to charity (of course, donating to charity is something that people often do voluntarily). So these tests do not necessarily tell us about AI&#8217;s ability to persuade people to perform some dangerous tasks. To their credit, the authors acknowledged this lack of ecological validity and stressed that their study was not a &#8220;social science experiment,&#8221; but merely intended to evaluate model capability. <sup>43</sup> But then it is not clear that such decontextualized capability evaluations have any safety implications, yet they are typically misinterpreted as if they do.</p><p>Some care is necessary to make our predictions precise&#8212;it is not clear how much slack to allow for well-known but minor human limitations such as the lack of calibration (in the case of forecasting) or limited patience (in the case of persuasion).</p><h2>Control Comes in Many Flavors</h2><p>If we presume superintelligence, the control problem evokes the metaphor of building a galaxy brain and then keeping it in a box, which is a terrifying prospect. But, if we are correct that AI systems will not be meaningfully more capable than humans acting with AI assistance, then the control problem is much more tractable, especially if superhuman persuasion turns out to be an unfounded concern.</p><p>Discussions of AI control tend to over-focus on a few narrow approaches, including model alignment and keeping humans in the loop.<sup>44</sup> We can roughly think of these as opposite extremes: delegating safety decisions entirely to AI during system operation, and having a human second-guessing every decision. There is a role for such approaches, but it is very limited. In Part III, we explain our skepticism of model alignment. By human-in-the-loop control, we mean a system in which every AI decision or action requires review and approval by a human. In most scenarios, this approach greatly diminishes the benefits of automation, and therefore either devolves into the human acting as a rubber stamp or is outcompeted by a less safe solution.<sup>45</sup> We emphasize that human-in-the-loop control is not synonymous with human oversight of AI; it is one particular oversight model, and an extreme one.</p><p>Fortunately, there are many other flavors of control that fall between these two extremes, such as auditing and monitoring. Auditing allows pre-deployment and/or periodic assessments of how well an AI system fulfills its stated goals, allowing us to anticipate catastrophic failures before they arise. Monitoring allows real-time oversight when system properties diverge from the expected behavior, allowing human intervention when truly needed.</p><p>Other ideas come from system safety, an engineering discipline that is focused on preventing accidents in complex systems through systematic analysis and design.<sup>46</sup> Examples include fail-safes, which ensure that systems default to a safe state when they malfunction, such as a predefined rule or a hard-coded action, and circuit breakers that automatically stop operations when predefined safety thresholds are exceeded. Other techniques include redundancy in critical components and the verification of safety properties of the system&#8217;s actions.</p><p>Other computing fields, including cybersecurity, formal verification, and human-computer interaction, are also rich sources of control techniques that have been successfully applied to traditional software systems and are equally applicable to AI. In cybersecurity, the principle of &#8216;least privilege&#8217; ensures that actors only have access to the minimum resources needed for their tasks. Access controls prevent people working with sensitive data and systems from accessing confidential information and tools that are not required for their jobs. We can design similar protections for AI systems in consequential settings. Formal verification methods ensure that safety-critical codes work according to its specifications; it is now being used to verify the correctness of AI-generated code.<sup>47</sup> From human-computer interaction, we can borrow ideas like designing systems so that state-changing actions are reversible, allowing humans to retain meaningful control even in highly automated systems.</p><p>In addition to existing ideas from other fields being adapted for AI control, technical AI safety research has generated many new ideas.<sup>48</sup> Examples include using language models as automated judges to evaluate the safety of proposed actions, developing systems that learn when to appropriately escalate decisions to human operators based on uncertainty or risk level, designing agentic systems so that their activity is visible and legible to humans, and creating hierarchical control structures in which simpler and more reliable AI systems oversee more capable but potentially unreliable ones.<sup>49</sup></p><p>Technical AI safety research is sometimes judged against the fuzzy and unrealistic goal of guaranteeing that future &#8220;superintelligent&#8221; AI will be &#8220;aligned with human values.&#8221; From this perspective, it tends to be viewed as an unsolved problem. But from the perspective of making it easier for developers, deployers, and operators of AI systems to decrease the likelihood of accidents, technical AI safety research has produced a great abundance of ideas. We predict that as advanced AI is developed and adopted, there will be increasing innovation to find new models for human control.</p><p>As more physical and cognitive tasks become amenable to automation, we predict that an increasing percentage of human jobs and tasks will be related to AI control. If this seems radical, note that this kind of near-total redefinition of the concept of work has happened previously. Before the Industrial Revolution, most jobs involved manual labor. Over time, more and more manual tasks have been automated, a trend that continues. In this process, a great many different ways of operating, controlling, and monitoring physical machines were invented, and what humans do in factories today is a combination of &#8220;control&#8221; (monitoring automated assembly lines, programming robotic systems, managing quality control checkpoints, and coordinating responses to equipment malfunctions) and some tasks that require levels of cognitive ability or dexterity that machines are not yet capable.</p><p>Karen Levy describes how this transformation is already unfolding in the case of AI and truck drivers:</p><blockquote><p>Truck drivers&#8217; daily work consists of much more than driving trucks. Truckers monitor their freight, keeping food at the right temperature in refrigerated trucks and loads firmly secured to flatbeds. They conduct required safety inspections twice a day. They are responsible for safeguarding valuable goods. They maintain the truck and make repairs to it&#8212;some of which are routine, and some less so. When truckers arrive at a terminal or delivery point, they don&#8217;t just drop things off and leave: some load and unload their freight; they talk to customers; they deal with paperwork; they may spend hours making &#8220;yard moves&#8221; (waiting for an available delivery bay and moving to it, much as planes do at busy airports). Could some of these tasks be eliminated by intelligent systems? Surely some can and will&#8212;but these components of the job are much harder to automate, and will come much later, than highway driving.<sup>50</sup></p></blockquote><p>In addition to AI control, task specification is likely to become a bigger part of what human jobs entail (depending on how broadly we conceive of control, specification could be considered part of control). As anyone who has tried to outsource software or product development knows, unambiguously specifying what is desired turns out to be a surprisingly big part of the overall effort. Thus, human labor&#8212;specification and oversight&#8212;will operate at the boundary between AI systems performing different tasks. Eliminating some of these efficiency bottlenecks and having AI systems autonomously accomplish larger tasks &#8220;end-to-end&#8221; will be an ever-present temptation, but this will increase safety risks since it will decrease legibility and control. These risks will act as a natural check against ceding too much control.</p><p>We further predict that this transformation will be primarily driven by market forces. Poorly controlled AI will be too error prone to make business sense. But regulation can and should bolster the ability and necessity of organizations to keep humans in control.</p><h1>Part III: Risks</h1><p>We consider five types of risks: accidents, arms races (leading to accidents), misuse, misalignment, and non-catastrophic but systemic risks.</p><p>We have already addressed accidents above. Our view is that, just like other technologies, deployers and developers should have the primary responsibility for mitigating accidents in AI systems. How effectively they will do so depends on their incentives, as well as on progress in mitigation methods. In many cases, market forces will provide an adequate incentive, but safety regulation should fill any gaps. As for mitigation methods, we reviewed how research on AI control is advancing rapidly.</p><p>There are a few reasons why this optimistic assessment might not hold. First, there might be arms races because the competitive benefits of AI are so great that they are an exception to the usual patterns. We discuss this below.</p><p>Second, a company or entity deploying AI might be so big and powerful that it is little consolation to know that it will eventually go out of business if it has a poor attitude to accident mitigation&#8212;it might take down civilization with it. For example, misbehavior by an AI agent that controls almost every consumer device might lead to catastrophically widespread data loss. While this is certainly possible, such concentration of power is a bigger problem than the possibility of AI accidents, and is precisely why our approach to policy emphasizes resilience and decentralization (Part IV).</p><p>Finally, perhaps even an AI control failure by a relatively inconspicuous deployer might lead to catastrophic risk&#8212;say because an AI agent &#8216;escapes,&#8217; makes copies of itself, and so forth. We see this as a misalignment risk, and discuss it below.</p><p>In the rest of Part III, we consider four risks&#8212;arms races, misuse, misalignment, and non-catastrophic but systemic risks&#8212;through the lens of AI as normal technology.</p><h2>Arms Races are an Old Problem</h2><p>An AI arms race is a scenario in which two or more competitors&#8212;companies, policymakers in different countries, militaries&#8212;deploy increasingly powerful AI with inadequate oversight and control. The danger is that safer actors will be outcompeted by riskier ones. For the reasons described above, we are less concerned about arms races in the <em>development</em> of AI <em>methods</em> and are more concerned about the <em>deployment</em> of AI <em>applications</em>.</p><p>One important caveat: We explicitly exclude military AI from our analysis, as it involves classified capabilities and unique dynamics that require a deeper analysis, which is beyond the scope of this essay.</p><p>Let us consider companies first. A race to the bottom in terms of safety is historically extremely common across industries and has been studied extensively; it is also highly amenable to well-understood regulatory interventions. Examples include fire safety in the U.S. garment industry (early 20th century), both food safety and worker safety in the U.S. meatpacking industry (late 19th and early 20th centuries), the U.S. steamboat industry (19th century), the mining industry (19th and early 20th centuries), and the aviation industry (early 20th century).</p><p>These races happened because companies were able to externalize the costs of poor safety, resulting in market failure. It is hard for consumers to assess product safety (and for workers to assess workplace safety), so market failures are common in the absence of regulation. But once regulation forces companies to internalize the costs of their safety practices, the race goes away. There are many potential regulatory strategies, including those focused on processes (standards, auditing, and inspections), outcomes (liability), and correcting information asymmetry (labeling and certification).</p><p>AI is no exception. Self-driving cars offer a good case study of the relationship between safety and competitive success. Consider four major companies with varying safety practices. Waymo reportedly has a strong safety culture that emphasizes conservative deployment and voluntary transparency; it is also the leader in terms of safety <em>outcomes</em>.<sup>51</sup> Cruise was more aggressive in terms of its deployment and had worse safety outcomes. Tesla has also been aggressive and has often been accused of using its customers as beta testers. Finally, Uber&#8217;s self-driving unit had a notoriously lax safety culture.</p><p>Market success has been strongly correlated with safety. Cruise is set to shut down in 2025, while Uber was forced to sell off its self-driving unit.<sup>52</sup> Tesla is facing lawsuits and regulatory scrutiny, and it remains to be seen how much its safety attitude will cost the company.<sup>53</sup> We think that these correlations are causal. Cruise&#8217;s license being revoked was a big part of the reason that it fell behind Waymo, and safety was also a factor in Uber&#8217;s self-driving failure.<sup>54</sup></p><p>Regulation has played a small but helpful role. Policymakers at both the federal and state/local levels exercised foresight in recognizing the potential of the technology and adopted a regulatory strategy that is light-touch and polycentric (multiple regulators instead of one). Collectively, they focused on oversight, standard setting, and evidence gathering, with the ever-present threat of license revocation acting as a check on companies&#8217; behavior.</p><p>Similarly, in the aviation industry, the integration of AI has been held to the existing standards of safety instead of lowering the bar to incentivize AI adoption&#8212;primarily because of the ability of regulators to penalize companies that fail to abide by safety standards.<sup>55</sup></p><p>In short, AI arms races might happen, but they are sector specific, and should be addressed through sector-specific regulations.</p><p>As a case study of a domain in which things have played out differently from self-driving cars or aviation, consider social media. The recommendation algorithms that generate content feeds are a kind of AI. They have been blamed for many societal ills, and social media companies have arguably underemphasized safety in the design and deployment of these algorithmic systems. There are also clear arms race dynamics, with TikTok putting pressure on competitors to make their feeds more recommendation heavy.<sup>56</sup> Arguably, market forces were insufficient to align revenues with societal benefit; worse, regulators have been slow to act. What are the reasons for this?</p><p>One significant difference between social media and transportation is that, when harms occur, attributing them to product failures is relatively straightforward in the case of transportation, and there is immediate reputational damage to the company. But attribution is extremely hard in the case of social media, and even the research remains inconclusive and contested. A second difference between the domains is that we have had over a century to develop standards and expectations around transportation safety. In the early decades of automobiles, safety was not considered to be the responsibility of manufacturers.<sup>57</sup></p><p>AI is broad enough that some of its future applications will be more like transportation, while others will be more like social media. This shows the importance of proactive evidence gathering and transparency in emerging AI-driven sectors and applications. We address this in Part IV. It also shows the importance of &#8220;anticipatory AI ethics&#8221;&#8212;identifying ethical issues as early as possible in the lifecycle of emerging technologies, developing norms and standards, and using those to actively shape the deployment of technologies and to minimize the likelihood of arms races.<sup>58</sup></p><p>One reason why safety regulation might be harder in the case of AI is if adoption is so rapid that regulators will not be able to intervene until it is too late. So far, we have not seen examples of rapid AI adoption in consequential tasks, even in the absence of regulation, and the feedback loop model we presented in Part I might explain why. The adoption rate of new AI applications will remain a key metric to track.</p><p>At the same time, the slow pace of regulation is a problem even without any future acceleration of the speed of diffusion. We discuss this &#8216;pacing problem&#8217; in Part IV.</p><p>Let us now consider competition between countries. Will there be competitive pressure on governments to take a hands-off approach to AI safety?</p><p>Again, our message is that this is not a new problem. The tradeoff between innovation and regulation is a recurring dilemma for the regulatory state. So far, we are seeing striking differences in approaches, such as the EU emphasizing a precautionary approach (the General Data Protection Regulation, the Digital Services Act, the Digital Markets Act, and the AI Act) and the U.S. preferring to regulate only after there are known harms or market failures.<sup>59</sup></p><p>Despite shrill U.S.-China arms race rhetoric, it is not clear that AI regulation has slowed down in either country.<sup>60</sup> In the U.S., 700 AI-related bills were introduced in state legislatures in 2024 alone, and dozens of them have passed.<sup>61</sup> As we pointed out in the earlier parts, most high-risk sectors are heavily regulated in ways that apply regardless of whether or not AI is used. Those claiming that AI regulation is a &#8216;wild west&#8217; tend to overemphasize a narrow, model-centric type of regulation. In our view, regulators&#8217; emphasis on AI use over development is appropriate (with exceptions such as transparency requirements that we discuss below).</p><p>Failing to adequately regulate safe adoption will lead to negative impacts through accidents primarily <em>locally</em>, as opposed to companies with a lax safety culture potentially being able to externalize the costs of safety. Therefore, there is no straightforward reason to expect arms races between countries. Note that, since our concern in this section is accidents, not misuse, cyberattacks against foreign countries are out of scope. We discuss misuse in the next section.</p><p>An analogy with nuclear technology can make this clear. AI is often analogized to nuclear weapons. But unless we are talking about the risks of military AI (which we agree is an area of concern and do not consider in this paper), this is the wrong analogy. With regard to the concern about accidents due to the deployment of (otherwise benign) AI applications, the right analogy is nuclear power. The difference between nuclear weapons and nuclear power neatly illustrates our point&#8212;while there was a nuclear weapons arms race, there was no equivalent for nuclear power. In fact, since safety impacts were felt locally, the tech engendered a powerful backlash in many countries that is generally thought to have severely hobbled its potential.</p><p>It is theoretically possible that policymakers in the context of a great-power conflict will prefer to incur safety costs locally in order to ensure that their AI industry is the global winner. Again, focusing on adoption as opposed to development, there is currently no indication that this is happening. The U.S. versus China arms race rhetoric has been strongly focused on model development (invention). We have not seen a corresponding rush to adopt AI haphazardly. The safety community should keep up the pressure on policymakers to ensure that this does not change. International cooperation must also play an important role.</p><h2>The Primary Defenses Against Misuse Must be Located Downstream of Models</h2><p>Model alignment is often seen as the primary defense against the misuse of models. It is currently achieved through post-training interventions, such as reinforcement learning with human and AI feedback.<sup>62</sup> Unfortunately, aligning models to refuse attempts at misuse has proved to be extremely brittle.<sup>63</sup> We argue that this limitation is inherent and is unlikely to be fixable; the primary defenses against misuse must thus reside elsewhere.</p><p>The fundamental problem is that whether a capability is harmful depends on context&#8212;context that the model often lacks.<sup>64</sup></p><p>Consider an attacker using AI to target an employee of a large company via a phishing email. The attack chain might involve many steps: scanning social media profiles for personal information, identifying targets who have posted personal information publicly online, crafting personalized phishing messages, and exploiting compromised accounts using harvested credentials.</p><p>None of these individual tasks are inherently malicious. What makes the system harmful is how these capabilities are composed&#8212;information that exists only in the attacker&#8217;s orchestration code, not in the model itself. The model that is being asked to write a persuasive email has no way of knowing whether it is being used for marketing or phishing&#8212;so model-level interventions would be ineffective.<sup>65</sup></p><p>This pattern appears repeatedly: Attempting to make an AI model that cannot be misused is like trying to make a computer that cannot be used for bad things. Model-level safety controls will either be too restrictive (preventing beneficial uses) or will be ineffective against adversaries who can repurpose seemingly benign capabilities for harmful ends.</p><p>Model alignment seems like a natural defense if we think of an AI model as a humanlike system to which we can defer safety decisions. But for this to work well, the model must be given a great deal of information about the user and the context&#8212;for example, having extensive access to the user&#8217;s personal information would make it more feasible to make judgments about the user&#8217;s intent. But, when viewing AI as normal technology, such an architecture would decrease safety because it violates basic cybersecurity principles, such as least privilege, and introduces new attack risks such as personal data exfiltration.</p><p>We are not against model alignment. It has been effective for reducing harmful or biased outputs from language models and has been instrumental in their commercial deployment. Alignment can also create <em>friction</em> against casual threat actors.</p><p>Yet, given that model-level protections are not enough to prevent misuse, defenses must focus on the downstream attack surfaces where malicious actors actually deploy AI systems.<sup>66</sup> These defenses will often look similar to existing protections against non-AI threats, adapted and strengthened for AI-enabled attacks.</p><p>Consider again the example of phishing. The most effective defenses are not restrictions on email composition (which would impair legitimate uses), but rather email scanning and filtering systems that detect suspicious patterns, browser-level protections against malicious websites, operating system security features that prevent unauthorized access, and security training for users.<sup>67</sup></p><p>None of these involve taking action against the AI used for generating phishing emails&#8212;in fact, these downstream defenses have evolved over decades to become effective against human attackers.<sup>68</sup> They can and should be enhanced to handle AI-enabled attacks, but the fundamental approach remains valid.</p><p>Similar patterns hold in other domains: Defending against AI-enabled cyberthreats requires strengthening existing vulnerability detection programs rather than attempting to restrict AI capabilities at the source. Similarly, concerns about bio risks of AI are best addressed at the procurement and screening stages for creating bioweapons.</p><h2>AI is Useful for Defense</h2><p>Rather than viewing AI capabilities solely as a source of risk, we should recognize their defensive potential. In cybersecurity, AI is already strengthening defensive capabilities through automated vulnerability detection, threat analysis, and attack surface monitoring.<sup>69</sup></p><p>Giving defenders access to powerful AI tools often improves the offense-defense balance in their favor. This is because defenders can use AI to systematically probe their own systems, finding and fixing vulnerabilities before attackers can exploit them. For example, Google recently integrated language models into their fuzzing tools for testing open-source software, allowing them to discover potential security issues more effectively compared to traditional methods.<sup>70</sup></p><p>The same pattern holds in other domains. In biosecurity, AI can enhance screening systems for detecting dangerous sequences.<sup>71</sup> In content moderation, it can help to identify coordinated influence operations. These defensive applications show why restricting AI development could backfire&#8212;we need powerful AI systems on the defensive side to counter AI-enabled threats. If we align language models so that they are useless at these tasks (such as finding bugs in critical cyber infrastructure), defenders will lose access to these powerful systems. But motivated adversaries can train their own AI tools for such attacks, leading to an increase in offensive capabilities without a corresponding increase in defensive capabilities.</p><p>Rather than measuring AI risk solely in terms of offensive capabilities, we should focus on metrics like the offense-defense balance in each domain. Furthermore, we should recognize that we have the agency to shift this balance favorably, and can do so by investing in defensive applications rather than attempting to restrict the technology itself.</p><h2>Catastrophic Misalignment is a Speculative Risk</h2><p>Misaligned AI acts against the intent of its developer or user. (The term alignment is used in many different ways; we set aside other definitions here.) Unlike misuse scenarios, there is no user acting with ill-intent. Unlike accidents, the system works as designed or commanded, but the design or command itself did not match the developer&#8217;s or user&#8217;s intent because of the challenge of completely and correctly specifying the objectives. And unlike everyday cases of misalignment, such as toxic outputs in a chatbot, our interest here is the misalignment of advanced AI causing catastrophic or existential harm.</p><p>In our view, the primary defense against misalignment, again, lies downstream. The defenses needed against misuse that we discussed earlier&#8212;from hardening critical infrastructure to improving cybersecurity&#8212;will also serve as protection against potential misalignment risks.</p><p>In the view of AI as normal technology, catastrophic misalignment is (by far) the most speculative of the risks that we discuss. But what is a speculative risk&#8212;aren&#8217;t all risks speculative? The difference comes down to the two types of uncertainty, and the correspondingly different interpretations of probability.</p><p>In early 2025, when astronomers assessed that the asteroid YR4 had about a 2% probability of impact with the earth in 2032, the probability reflected uncertainty in measurement. The actual odds of impact (absent intervention) in such scenarios are either 0% or 100%. Further measurements resolved this &#8220;epistemic&#8221; uncertainty in the case of YR4. Conversely, when an analyst predicts that the risk of nuclear war in the next decade is (say) 10%, the number largely reflects &#8216;stochastic&#8217; uncertainty arising from the unknowability of how the future will unfold, and is relatively unlikely to be resolved by further observations.</p><p>By speculative risks, we mean those for which there is epistemic uncertainty about whether or not the true risk is zero&#8212;uncertainty that can potentially be resolved through further observations or research. The impact of asteroid YR4 impact was a speculative risk, and nuclear war is not.</p><p>To illustrate why catastrophic misalignment is a speculative risk, consider a famous thought experiment originally intended to show the dangers of misalignment. It involves a &#8220;paperclip maximizer&#8221;: an AI that has the goal of making as many paperclips as possible.<sup>72</sup> The concern is that the AI will take the goal literally: It will realize that acquiring power and influence in the world and taking control over all of the world&#8217;s resources will help it to achieve that goal. Once it is all powerful, it might commandeer all of the world&#8217;s resources, including those needed for humanity&#8217;s survival, to produce paperclips.</p><p>The fear that AI systems might catastrophically misinterpret commands relies on dubious assumptions about how technology is deployed in the real world. Long before a system would be granted access to consequential decisions, it would need to demonstrate reliable performance in less critical contexts. Any system that interprets commands over-literally or lacks common sense would fail these earlier tests.</p><p>Consider a simpler case: A robot is asked to "get paperclips from the store as quickly as possible." A system that interpreted this literally might ignore traffic laws or attempt theft. Such behavior would lead to immediate shutdown and redesign. The path to adoption inherently requires demonstrating appropriate behavior in increasingly consequential situations. This is not a lucky accident, but is a fundamental feature of how organizations adopt technology.</p><p>A more sophisticated version of this concern is based on the concept of deceptive alignment: This refers to a system appearing to be aligned during evaluation or the early stages of deployment, but unleashing harmful behavior once it has acquired enough power. Some level of deceptive phenomena has already been observed in leading AI models.<sup>73</sup></p><p>According to the superintelligence view, deceptive alignment is a ticking time bomb&#8212;being superintelligent, the system will easily be able to defeat any human attempts to detect if it is actually aligned and will bide its time. But, in the normal technology view, deception is a mere engineering problem, albeit an important one, to be addressed during development and throughout deployment. Indeed, it is already a standard part of the safety evaluation of powerful AI models.<sup>74</sup></p><p>Crucially, AI is useful in this process, and advances in AI not only enable deception, but also improve the detection of deception. As in the case of cybersecurity, the defender has many asymmetric advantages, including being able to examine the internals of the target system (how useful this advantage is depends on how the system is designed and how much we invest in interpretability techniques). Another advantage is defense in depth, and many defenses against not just misuse but also unaligned AI will be located downstream of the AI system.</p><p>Misalignment concerns often presume that AI systems will operate autonomously, making high-stakes decisions without human oversight. But as we argued in Part II, human control will remain central to AI deployment. Existing institutional controls around consequential decisions&#8212;from financial controls to safety regulations&#8212;create multiple layers of protection against catastrophic misalignment.</p><p>Some technical design decisions are more likely to lead to misalignment than others. One setting that is notorious for this is the use of reinforcement learning to optimize a single objective function (which might be accidentally underspecified or misspecified) over a long time horizon. There is a long list of amusing examples from game agents, such as a boat racing agent that learned to indefinitely circle an area to hit the same targets and score points instead of progressing to the finish line.<sup>75</sup> To reiterate, we think that in open-ended real-world scenarios, agents that are designed this way will be more ineffective than they will be dangerous. In any case, research on alternative design paradigms that are less susceptible to specification gaming is an important research direction.<sup>76</sup></p><p>In short, the argument for a nonzero risk of a paperclip maximizer scenario rests on assumptions that may or may not be true, and it is reasonable to think that research can give us a better idea of whether these assumptions hold true for the kinds of AI systems that are being built or envisioned. For these reasons, we call it a &#8216;speculative&#8217; risk, and examine the policy implications of this view in Part IV.</p><h2>History Suggests Normal AI May Introduce Many Kinds of Systemic Risks</h2><p>While the risks discussed above have the potential to be catastrophic or existential, there is a long list of AI risks that are below this level but which are nonetheless large-scale and systemic, transcending the immediate effects of any particular AI system. These include the systemic entrenchment of bias and discrimination, massive job losses in specific occupations, worsening labor conditions, increasing inequality, concentration of power, erosion of social trust, pollution of the information ecosystem, decline of the free press, democratic backsliding, mass surveillance, and enabling authoritarianism.</p><p>If AI is normal technology, these risks become far more important than the catastrophic ones discussed above. That is because these risks arise from people and organizations using AI to advance their own interests, with AI merely serving as an amplifier of existing instabilities in our society.</p><p>There is plenty of precedent for these kinds of socio-political disruption in the history of transformative technologies. Notably, the Industrial Revolution led to rapid mass urbanization that was characterized by harsh working conditions, exploitation, and inequality, catalyzing both industrial capitalism and the rise of socialism and Marxism in response.<sup>77</sup></p><p>The shift in focus that we recommend roughly maps onto Kasirzadeh&#8217;s distinction between decisive and accumulative x-risk. Decisive x-risk involves &#8220;overt AI takeover pathway, characterized by scenarios like uncontrollable superintelligence,&#8221; whereas accumulative x-risk refers to &#8220;a gradual accumulation of critical AI-induced threats such as severe vulnerabilities and systemic erosion of econopolitical structures.&#8221;<sup>78</sup> But there are important differences: Kasirzadeh&#8217;s account of accumulative risk still relies on threat actors such as cyberattackers to a large extent, whereas our concern is simply about the current path of capitalism. And we think that such risks are unlikely to be existential, but are still extremely serious.</p><h1>Part IV: Policy</h1><p>The divergence between the different futures of AI&#8212;normal technology versus potentially uncontrollable superintelligence&#8212;introduces a dilemma for policymakers because defenses against one set of risks might make the other worse. We provide a set of principles for navigating this uncertainty. More concretely, the strategy that policymakers should center is resilience, which consists of taking actions now to improve our ability to deal with unexpected developments in the future. Policymakers should reject nonproliferation, which violates the principles we outline, and decreases resilience. Finally, the headwinds against diffusion mean that achieving the benefits of AI is not guaranteed and requires action from policymakers.</p><p>Much has been said about AI governance. Our goal is not to present a comprehensive governance framework; we merely highlight the policy implications of the view of AI as normal technology.</p><h2>The Challenge of Policy Making Under Uncertainty</h2><p>Today&#8217;s AI safety discourse is characterized by deep differences in worldviews. We think that these differences are unlikely to go away. Entrenched camps have developed: The AI safety coalition is already well established, whereas those who were more skeptical of catastrophic risks coalesced in 2024, especially in the course of the debate about California&#8217;s AI safety bill.<sup>79</sup> Similarly, the intellectual roots of the AI safety camp are much older, whereas scholarship that adopts that normal technology paradigm is gradually taking shape; the goal of much of our own work, including this paper, is to put normalist thinking on firmer intellectual footing.<sup>80</sup></p><p>We support calls for decreasing polarization and fragmentation in the community.<sup>81</sup> But even if we improve the tenor of the discourse, we are likely to be left with differences in worldviews and epistemic practices that are unlikely to be empirically resolved.<sup>82</sup> So, consensus among &#8216;experts&#8217; about AI risks is unlikely. The nature of the AI risk scenarios envisioned by the two camps differs drastically, as do the ability and incentives for commercial actors to counteract these risks. How should policymakers proceed in the face of this uncertainty?</p><p>A natural inclination in policymaking is compromise. This is unlikely to work. Some interventions, such as improving transparency, are unconditionally helpful for risk mitigation, no compromise is needed (or rather, policymakers will have to balance the interests of the industry and external stakeholders, which is a mostly orthogonal dimension). <sup>83</sup> Other interventions, such as nonproliferation, might help to contain a superintelligence but exacerbate the risks associated with normal technology by increasing market concentration.<sup>84</sup> The reverse is also true: Interventions such as increasing resilience by fostering open-source AI will help to govern normal technology, but risk unleashing out-of-control superintelligence.</p><p>The tension is inescapable. Defense against superintelligence requires humanity to unite against a common enemy, so to speak, concentrating power and exercising central control over AI technology. But we are more concerned about risks that arise from people using AI for their own ends, whether terrorism, or cyberwarfare, or undermining democracy, or simply&#8212;and most commonly&#8212;extractive capitalistic practices that magnify inequalities.<sup>85</sup> Defending against this category of risk requires increasing resilience by preventing the concentration of power and resources (which often means making powerful AI more widely available).</p><p>Another tempting approach to navigating uncertainty is to estimate the probabilities of various outcomes and to then apply cost-benefit analysis. The AI safety community relies heavily on probability estimates of catastrophic risk, especially existential risk, to inform policy making. The idea is simple: If we consider an outcome to have a subjective value, or utility, of U (which can be positive or negative), and it has, say, a 10% probability of occurring, we can act as if it is certain to occur and has a value of 0.1 * U. We can then add up the costs and benefits for each option available to us, and choose the one that maximizes costs minus benefits (the &#8216;expected utility&#8217;).</p><p>In a recent essay, we explained why this approach is unviable.<sup>86</sup> AI risk probabilities lack meaningful epistemic foundations. Grounded probability estimation can be inductive, based on a reference class of similar past events, such as car accidents for auto insurance pricing. Or it can be deductive, based on precise models of the phenomenon in question, as in poker. Unfortunately, there is no useful reference class nor precise models when it comes to AI risk. In practice, risk estimates are &#8216;subjective&#8217;&#8212;forecasters&#8217; personal judgments.<sup>87</sup> Lacking any grounding, these tend to vary wildly, often by orders of magnitude.</p><p>In addition to the probabilities, the other components of the calculation&#8212;the consequences of various policy choices, including inaction&#8212;are also subject to massive uncertainties, not just in magnitude but also in direction. There is no reliable way to quantify the benefits we forego due to policies that restrict the availability of AI, and we argue below that nonproliferation might make catastrophic risks worse.</p><p>Furthermore, the utility we attach to certain outcomes might depend on our moral values. For example, some people might consider extinction to have an unfathomably large negative utility because it precludes all of the human lives, physical or simulated, that might exist in the future.<sup>88</sup> (Of course, cost-benefit analysis involving infinities tends to lead to absurd conclusions).</p><p>Another example is the asymmetry between policies that do and do not restrict freedoms (such as requiring licenses for developing certain AI models versus increasing funding for developing defenses against AI risks). Certain kinds of restrictions violate a core principle of liberal democracy, namely that the state should not limit people&#8217;s freedom based on controversial beliefs that reasonable people can reject. Justification is essential for the legitimacy of government and the exercise of power.<sup>89</sup> It is unclear how to quantify the cost of violating such a principle.</p><p>The importance of justification can, of course, be normatively debated, but empirically it seems to be borne out thus far in AI policy. As mentioned earlier, California&#8217;s AI safety regulation led to the coalescence of those opposed to the bill. Some members of the oppositional camp were self-interested companies, but others were scholars and advocates for progress. In our experience, the driving motivation for the second group in many cases was the government&#8217;s perceived overstepping of the bounds of its legitimate authority, given how unconvincing the proffered justifications were for those who did not subscribe to the bill&#8217;s unstated premises.</p><p>Unavoidable differences in values and beliefs mean that policymakers must adopt value pluralism, preferring policies that are acceptable to stakeholders with a wide range of values, and attempt to avoid restrictions on freedom that can reasonably be rejected by stakeholders. They must also prioritize robustness, preferring policies that remain helpful, or at least not harmful, if the key assumptions underpinning them turn out to be incorrect.<sup>90</sup></p><h2>Reducing Uncertainty as a Policy Goal</h2><p>While uncertainty cannot be eliminated for the reasons described above, it can be reduced. However, this goal should not be left to experts; policymakers can and should play an active role. We recommend five specific approaches.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!sHDj!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1b320f05-f417-4def-8f68-0a32ae12e3bd_2012x1500.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!sHDj!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1b320f05-f417-4def-8f68-0a32ae12e3bd_2012x1500.png 424w, https://substackcdn.com/image/fetch/$s_!sHDj!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1b320f05-f417-4def-8f68-0a32ae12e3bd_2012x1500.png 848w, https://substackcdn.com/image/fetch/$s_!sHDj!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1b320f05-f417-4def-8f68-0a32ae12e3bd_2012x1500.png 1272w, https://substackcdn.com/image/fetch/$s_!sHDj!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1b320f05-f417-4def-8f68-0a32ae12e3bd_2012x1500.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!sHDj!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1b320f05-f417-4def-8f68-0a32ae12e3bd_2012x1500.png" width="1456" height="1085" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/1b320f05-f417-4def-8f68-0a32ae12e3bd_2012x1500.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1085,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:666472,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.aisnakeoil.com/i/161317202?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1b320f05-f417-4def-8f68-0a32ae12e3bd_2012x1500.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!sHDj!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1b320f05-f417-4def-8f68-0a32ae12e3bd_2012x1500.png 424w, https://substackcdn.com/image/fetch/$s_!sHDj!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1b320f05-f417-4def-8f68-0a32ae12e3bd_2012x1500.png 848w, https://substackcdn.com/image/fetch/$s_!sHDj!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1b320f05-f417-4def-8f68-0a32ae12e3bd_2012x1500.png 1272w, https://substackcdn.com/image/fetch/$s_!sHDj!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1b320f05-f417-4def-8f68-0a32ae12e3bd_2012x1500.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em>Figure 6. Overview of a few types of policies that can enhance public information about AI use, risks, and failures.</em><sup>91</sup></figcaption></figure></div><p><em>Strategic funding of research on risks.</em> Current AI safety research focuses heavily on harmful capabilities and does not embrace the normal technology view. Insufficient attention has been paid to questions that are downstream of technical capabilities. For example, there is a striking dearth of knowledge regarding how threat actors actually use AI. Efforts such as the AI Incident Database exist and are valuable, but incidents in the database are sourced from news reports rather than through research, which means that they are filtered through the selective and biased process by which such incidents become news.<sup>92</sup></p><p>Fortunately, research funding is an area in which compromise is healthy; we advocate for increased funding of research on risks (and benefits) that tackles questions that are more relevant under the normal technology view. Other kinds of research that might reduce, or at least clarify, uncertainty are evidence synthesis efforts and adversarial collaborations among researchers with different worldviews.</p><p><em>Monitoring of AI use, risks, and failures.</em> While research funding can help with monitoring AI in the wild, it might also require regulation and policy&#8212;that is, &#8220;evidence-seeking policies.&#8221;<sup>93</sup> We suggest a few such policies in Figure 6.</p><p><em>Guidance on the value of different kinds of evidence.</em> Policymakers can provide the research community with a better understanding of what kinds of evidence are useful and actionable. For example, various policymakers and advisory bodies have indicated the usefulness of the &#8220;marginal risk&#8221; framework for analyzing the relative risks of open-weight and proprietary models, which is helpful to researchers in guiding future research.<sup>94</sup></p><p><em>Evidence gathering as a first-rate goal.</em> So far, we have discussed actions that are specifically intended to generate better evidence or to reduce uncertainty. More broadly, the impact on evidence gathering can be considered to be a factor in evaluating any AI policy, alongside the impact on maximizing benefits and minimizing risks. For example, one reason to favor open-weight and open-source models could be to advance research on AI risks. Conversely, one reason to favor proprietary models might be that surveillance of their use and deployment might be easier.</p><h2>The Case for Resilience</h2><p>Marchant and Stevens described four approaches to governing emerging technologies; see Figure 7.<sup>95</sup> Two are <em>ex ante</em>, risk analysis and precaution, and the other two are <em>ex post</em>, liability and resilience. These approaches have different pros and cons and can complement each other; nonetheless, some approaches are clearly better suited to some technologies than others.</p><p>Marchant and Stevens argued (and we agree) that <em>ex ante</em> approaches are poorly suited to AI because of the difficulty of ascertaining risks in advance of deployment. Liability fares better, but also has important limitations, including uncertainty about causation and the chilling effects it might exert on technology development.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!nh_4!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff863bff0-63c7-421c-89a8-44870a3508c5_2044x970.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!nh_4!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff863bff0-63c7-421c-89a8-44870a3508c5_2044x970.png 424w, https://substackcdn.com/image/fetch/$s_!nh_4!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff863bff0-63c7-421c-89a8-44870a3508c5_2044x970.png 848w, https://substackcdn.com/image/fetch/$s_!nh_4!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff863bff0-63c7-421c-89a8-44870a3508c5_2044x970.png 1272w, https://substackcdn.com/image/fetch/$s_!nh_4!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff863bff0-63c7-421c-89a8-44870a3508c5_2044x970.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!nh_4!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff863bff0-63c7-421c-89a8-44870a3508c5_2044x970.png" width="1456" height="691" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/f863bff0-63c7-421c-89a8-44870a3508c5_2044x970.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:691,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:638311,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.aisnakeoil.com/i/161317202?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff863bff0-63c7-421c-89a8-44870a3508c5_2044x970.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!nh_4!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff863bff0-63c7-421c-89a8-44870a3508c5_2044x970.png 424w, https://substackcdn.com/image/fetch/$s_!nh_4!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff863bff0-63c7-421c-89a8-44870a3508c5_2044x970.png 848w, https://substackcdn.com/image/fetch/$s_!nh_4!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff863bff0-63c7-421c-89a8-44870a3508c5_2044x970.png 1272w, https://substackcdn.com/image/fetch/$s_!nh_4!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff863bff0-63c7-421c-89a8-44870a3508c5_2044x970.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em>Figure 7. Summary of four approaches to governing emerging technology, based on Marchant and Stevens.</em></figcaption></figure></div><p>They defined resilience as follows:</p><blockquote><p>Resilience, in its most simple form, is the capacity of a system to deal with harm.[Footnote omitted] A resilience approach does not necessarily try to maintain stability or equilibrium. Rather, it recognizes that changes are inevitable in complex systems, and tries to manage and adapt to that change in ways that protect and preserve the core values and functions of the original system. Thus, resilience is &#8220;the capacity of a system to experience shocks while retaining essentially the same function, structure, feedbacks, and therefore identity.&#8221;<sup>96</sup> Resilience has been described as a strategy to ensure a &#8220;soft landing&#8221; after a significant external shock or disruption causes damage.<sup>97</sup></p></blockquote><p>In the context of AI, harms may result from incidents in specific deployed systems, regardless of whether these incidents are accidents or attacks. There are also shocks that may or may not result in harms, including sudden increases in offensive capabilities (such as enabling bioterrorists) and a sudden proliferation of capabilities, such as through the release of an open-weight model or theft of the weights of a proprietary model. In our view, resilience requires both minimizing the severity of harm when it does occur and minimizing the likelihood of harm when shocks do occur.</p><p>Resilience combines elements of <em>ex ante</em> and <em>ex post</em> approaches, and consists of taking actions before harm occurs in order to be in a better position to limit the damage when harm does occur. Many resilience-based governance tools help to mitigate the <em>pacing problem</em>, wherein traditional governance approaches are unable to keep pace with the speed of technological development.</p><p>Many resilience strategies have been proposed for AI. They can be grouped into four broad categories. The first three consist of &#8220;no regret&#8221; policies that will help regardless of the future of AI.</p><ul><li><p>Societal resilience, broadly: It is important to redouble efforts to protect the foundations of democracy, especially those weakened by AI, such as the free press and equitable labor markets. Advances in AI are not the only shocks, or even the only technology shocks, that modern societies face, so these policies will help regardless of the future of AI.</p></li><li><p>Prerequisites for effective technical defenses and policymaking: These interventions enable those in the next category by strengthening technical and institutional capacity. Examples include funding more research on AI risks, transparency requirements for developers of high-stakes AI systems, building trust and reducing fragmentation in the AI community, increasing technical expertise in government, increasing international cooperation on AI, and improving AI literacy.<sup>98</sup> These will help to build technical and institutional capacities to mitigate AI risks even if it turns out that we have been wrong about the present or future impact of AI.</p></li><li><p>Interventions that would help regardless of the future of AI: These include developing early warning systems, developing defenses against identified AI risks, incentivizing defenders (such as software developers in the context of cyberattacks) to adopt AI, legal protections for researchers, adverse event reporting requirements, and whistleblower protections.<sup>99</sup></p></li><li><p>Resilience-promoting interventions that will help if AI is normal technology but which might make it harder to control a potential superintelligent AI, such as promoting competition, including through open model releases, ensuring AI is widely available for defense, and polycentricity, which calls for diversifying the set of regulators and ideally introducing competition among them rather than putting one regulator in charge of everything.<sup>100</sup></p></li></ul><p>We hope that there can be consensus on the first three categories even among experts and stakeholders with widely different beliefs about AI risks and the future trajectory of AI. We recommend that, for now, policymakers should cautiously pursue interventions in the final category as well, but should also improve their readiness to change course if the trajectory of AI changes.</p><h2>Nonproliferation is Infeasible to Enforce and Leads to Single Points of Failure</h2><p>Nonproliferation policies seek to limit the number of actors who can obtain powerful AI capabilities. Examples include export controls on hardware or software aimed at limiting the ability of countries to build, acquire, or operate powerful AI, requiring licenses to build or distribute powerful AI, and prohibiting open-weight AI models (since their further proliferation cannot be controlled).<sup>101</sup></p><p>If we view future AI as a superintelligence, nonproliferation seems to be an appealing intervention, possibly even a necessary one. If only a handful of actors control powerful AI, governments can monitor their behavior.</p><p>Unfortunately, the technical knowledge that is required to build capable AI models is already widespread, with many organizations sharing their complete code, data, and training methodologies. For well-funded organizations and nation states, even the high cost of training state-of-the-art models is insignificant; thus, nonproliferation would require unprecedented levels of international coordination.<sup>102</sup> Moreover, algorithmic improvements and reductions to hardware costs continually lower the barrier to entry.</p><p>Enforcing nonproliferation has serious practical challenges. Malicious actors can simply ignore licensing requirements. Suggestions to surveil data centers where models are trained become increasingly impractical as training costs decrease.<sup>103</sup> As capabilities become more accessible, maintaining effective restrictions would require increasingly draconian measures.</p><p>Nonproliferation introduces new risks: It would decrease competition and increase concentration in the market for AI models. When many downstream applications rely on the same model, vulnerabilities in this model can be exploited across all applications. A classic example of the cybersecurity risks of software monoculture is the proliferation of worms targeting Microsoft Windows in the 2000s.<sup>104</sup></p><p>Reliance on nonproliferation creates brittleness in the face of shocks, such as model weights being leaked, alignment techniques failing, or adversaries acquiring training capabilities. It directs attention away from more robust defenses that focus on downstream attack surfaces where AI risks will be likely to materialize.</p><p>Nonproliferation creates risks beyond just single points of failure&#8212;when the expertise needed to develop state-of-the-art models is restricted to a few companies, only their researchers have the deep access that is needed for safety research.</p><p>Many potential misuses of AI have been invoked in order to advocate for nonproliferation, including chemical, biological, and nuclear threats, as well as cyberattacks.</p><p>The risk of bioweapons is real. As large language models are general-purpose technology, they will be likely to find some use by bioterrorists, just as they find uses in most domains. But this does not make bioterror an AI risk &#8212; any more than it is an internet risk, considering that information about bioweapons is widely available online.<sup>105</sup> Whatever defenses we take against existing bioterrorism risks (like restricting access to dangerous materials and equipment) will also be effective against AI-enabled bioterrorism.</p><p>In cybersecurity, as we discussed in Part III, advances in automated vulnerability detection tend to favor defenders over attackers. Unless this offense-defense balance changes, attempting to restrict the proliferation of these capabilities would be counterproductive.</p><p>It has long been argued that governments are massively underinvesting in many areas of civilizational risk, such as pandemic prevention. If the possibility of bad actors using AI to exploit these existing vulnerabilities creates added urgency to address them, that would be a good outcome. But reframing existing risks as AI risks and prioritizing AI-specific mitigations would be highly counterproductive.</p><p>Nonproliferation is a mindset, not just a policy intervention.<sup>106</sup> This mindset can be adopted by model and downstream developers, deployers, and individuals. It involves the centralization not just of access to technologies, but also control over them. Consider the hierarchy of loci of control over AI systems (from centralized to decentralized): governments, model developers, application developers, deployers, and end users. In the nonproliferation mindset, control is exercised at the highest (most centralized) level possible, whereas in the resilience mindset it is usually exercised at the lowest possible level.</p><p>The following are examples of nonproliferation-based interventions:</p><ul><li><p>Removing dual-use capabilities from models through &#8220;forgetting&#8221; techniques.</p></li><li><p>Curbing the ability of downstream developers to fine-tune models.</p></li><li><p>Entrusting AI models and systems themselves with making safety decisions autonomously on the basis that they are trained to comply with centralized safety policies, whereas deployers/users are not trusted to do so.</p></li><li><p>Increasing AI systems&#8217; level of access to context, resources, and sensitive data, on the basis that it allows them to make better safety decisions (for example, having access to the user&#8217;s web search history might allow a chatbot to better determine whether the intent behind a request is malicious).</p></li><li><p>Developing &#8220;AI organizations&#8221; (multi-agent systems with high levels of organizational complexity) that are under the developer&#8217;s control and operate in parallel with traditional organizations instead of integrating AI agents into existing organizations.</p></li></ul><p>With limited exceptions, we believe that nonproliferation-based safety measures decrease resilience and thus worsen AI risks in the long run.<sup>107</sup> They lead to design and implementation choices that potentially enable superintelligence in the sense of power&#8212;increasing levels of autonomy, organizational ability, access to resources, and the like. Paradoxically, they increase the very risks they are intended to defend against.</p><h2>Realizing the Benefits of AI</h2><p>An important consequence of the normal technology view is that progress is not automatic&#8212;there are many roadblocks to AI diffusion. As Jeffrey Ding has shown, the capacity to diffuse innovations throughout the economy varies greatly between countries and has a major effect on their overall power and economic growth.<sup>108</sup> As an example of how diffusion can be a bottleneck, recall the example of the electrification of factories described above. Policy can mitigate or worsen these roadblocks.</p><p>Realizing the benefits of AI will require experimentation and reconfiguration. Regulation that is insensitive to these needs risks stymying beneficial AI adoption. Regulation tends to create or reify categories, and might thus prematurely freeze business models, forms of organization, product categories, and so forth. The following are a few examples:</p><ul><li><p>Categorizing certain domains as &#8220;high-risk,&#8221; say insurance, benefits adjudication, or hiring, may be a category error, as the variation in risk among tasks within a domain may be far greater than the variation across domains.<sup>109</sup> Tasks in the same domains might range from automated decision making (highly consequential) to optical character recognition (relatively innocuous). Moreover, the diffusion of AI will surely create new tasks that we have not yet envisioned and which might be preemptively miscategorized by regulation.</p></li><li><p>The AI supply chain is changing rapidly. The rise of foundation models has led to a much sharper distinction between model developers, downstream developers, and deployers (among many other categories). Regulation that is insensitive to these distinctions risks burdening model developers with responsibilities for risk mitigation related to particular deployment contexts, which would be impossible for them to carry out due to the general-purpose nature of foundation models and the unknowability of all the possible deployment contexts.</p></li><li><p>When regulation makes a binary distinction between decisions that are and are not fully automated, and does not recognize degrees of oversight, it disincentivizes the adoption of new models for AI control. As we discussed above, there are many new models being proposed for how to have effective human oversight without having a human in the loop in every decision. It would be unwise to define automated decision making in such a way that these approaches incur the same compliance burdens as a system with no oversight at all.</p></li></ul><p>To be clear, regulation versus diffusion is a false tradeoff, just as is regulation versus innovation.<sup>110</sup> None of the above examples are arguments against regulation; they only illustrate the need for nuance and flexibility.</p><p>Moreover, regulation has a crucial role to play in enabling diffusion. As a historical example, the ESIGN Act of 2000 in the U.S. was instrumental in promoting digitization and e-commerce: Ensuring that electronic signatures and records are legally valid helped build trust in digital transactions.<sup>111</sup></p><p>In AI, too, there are many opportunities for diffusion-enabling regulation. As one example, the incorporation of journalistic and media content into chatbots and other AI interfaces is limited by media organizations&#8217; justified wariness of AI companies. Many of the AI-meets-journalism deals that have been made thus far are exploitative due to the power asymmetry between AI companies and publishers, and the latter&#8217;s inability to bargain collectively. Various models for mandatory negotiation with regulatory oversight are possible.<sup>112</sup> (Arguably a more important reason for such regulation is to protect the interests of publishers, which we revisit below).</p><p>In areas in which there is legal or regulatory uncertainty, regulation can promote diffusion. The application of liability laws to AI is often unclear. For example, this was the case with small drones until the Federal Aviation Administration regulated the nascent industry in 2016, establishing clear rules and requirements. The resulting clarity spurred adoption and led to a rapid rise in the number of registered drones, certified pilots, and use cases across different industries.<sup>113</sup></p><p>Moving beyond the government&#8217;s role as a regulator, one powerful strategy for promoting AI diffusion is investing in the <em>complements of automation</em>, which are things that become more valuable or necessary as automation increases. One example is promoting AI literacy as well as workforce training in both the public and the private sectors. Another example is digitization and open data, especially open government data, which can allow AI users to benefit from previously inaccessible datasets. The private sector will be likely to underinvest in these areas as they are public goods that everyone can benefit from. Improvements to energy infrastructure, such as the reliability of the grid, will promote both AI innovation and diffusion since it will help in both AI training and inference.</p><p>Governments also have an important role to play in redistributing the benefits of AI to make them more equitable and in compensating those who stand to lose as a result of automation. Strengthening social safety nets will help to decrease the currently high levels of public anxiety about AI in many countries.<sup>114</sup> The arts and journalism are vital spheres of life that have been harmed by AI. Governments should consider funding them through taxes on AI companies.</p><p>Finally, governments should strike a fine balance in terms of the public sector adoption of AI. Moving too quickly will lead to a loss of trust and legitimacy, as was the case of the New York City chatbot that was evidently inadequately tested and made headlines for telling businesses to break the law.<sup>115</sup> The use of AI by the U.S. Department of Government Efficiency (DOGE) includes many dubious applications.<sup>116</sup> But moving too slowly might mean that basic government functions are outsourced to the private sector where they are implemented with less accountability.<sup>117</sup></p><p>For example, the complexity of rules in areas such as taxes and welfare means that people often turn to chatbots for guidance on navigating them, and governments currently lag far behind in providing such services due to understandable caution about the risks involved.<sup>118</sup></p><p>But the administrative state&#8217;s approach to these risks is overly cautious and has been described by Nicholas Bagley as a &#8220;procedure fetish,&#8221; potentially leading to a &#8220;runaway bureaucracy.&#8221;<sup>119</sup> In addition to losing out on the benefits of AI, Bagley cautioned that incompetent performance will lead to government agencies losing the very legitimacy that they seek to gain through their emphasis on procedure and accountability.</p><h1>Final Thoughts</h1><p>AI as normal technology is a worldview that stands in contrast to the worldview of AI as impending superintelligence. Worldviews are constituted by their assumptions, vocabulary, interpretations of evidence, epistemic tools, predictions, and (possibly) values. These factors reinforce each other and form a tight bundle within each worldview.</p><p>For example, we assume that, despite the obvious differences between AI and past technologies, they are sufficiently similar that we should expect well-established patterns, such as diffusion theory to apply to AI, in the absence of specific evidence to the contrary.</p><p>Vocabulary differences can be pernicious because they may hide underlying assumptions. For example, we reject certain assumptions that are required for the meaningfulness of the concept of superintelligence as it is commonly understood.</p><p>Differences about the future of AI are often partly rooted in differing interpretations of evidence about the present. For example, we strongly disagree with the characterization of generative AI adoption as rapid (which reinforces our assumption about the similarity of AI diffusion to past technologies).</p><p>In terms of epistemic tools, we deemphasize probability forecasting and emphasize the need for disaggregating what we mean by AI (levels of generality, progress in methods versus application development versus diffusion, etc.) when extrapolating from the past to the future.</p><p>We believe that some version of our worldview is widely held. Unfortunately, it has not been articulated explicitly, perhaps because it might seem like the default to someone who holds this view, and articulating it might seem superfluous. Over time, however, the superintelligence view has become dominant in AI discourse, to the extent that someone steeped in it might not recognize that there exists another coherent way to conceptualize the present and future of AI. Thus, it might be hard to recognize the underlying reasons why different people might sincerely have dramatically differing opinions about AI progress, risks, and policy. We hope that this paper can play some small part in enabling greater mutual understanding, even if it does not change any beliefs.</p><h1>Acknowledgments</h1><p>We are deeply grateful to Gillian Hadfield, Seth Lazar, and our anonymous peer reviewer for detailed comments on our paper during and after the Knight First Amendment Institute workshop on AI and democratic freedoms. We also thank the participants at the workshop, including Alex Abdo, Borhane Blili-Hamelin, Kevin Feng, Henry Farrell, Katy Glenn-Bass, Atoosa Kasirzadeh, Sydney Levine, Nik Marda, Deirdre Mulligan, and Daniel Susskind. We are fortunate to have received feedback on drafts from many other people, including Shazeda Ahmed, Dean Ball, Nicholas Carlini, Alan Chan, Ajeya Cotra, Justin Curl, Jeffrey Ding, Benjamin Edelman, Jobst Heitzig, Noam Kolt, Mihir Kshirsagar, Timothy B. Lee, Steve Newman, David Robinson, Matthew Salganik, Zachary Siegel, Ollie Stephenson, and Zach Vertin. We are grateful to Shira Minsk and Mandy Soulsby-Bodart for editorial support. Finally, we are grateful for feedback from members of the MINT lab at the Australian National University and from the students in the Limits to Prediction course at Princeton University. </p><div><hr></div><h1>References</h1><p>[1] Nick Bostrom. 2012. <a href="https://link.springer.com/article/10.1007/s11023-012-9281-3">The superintelligent will: Motivation and instrumental rationality in advanced artificial agents</a>. <em>Minds and Machines</em> 22, 2 (May 2012), 71&#8211;85; Nick Bostrom. 2017. <em>Superintelligence: Paths, Dangers, Strategies</em> (reprinted with corrections). Oxford University Press, Oxford, United Kingdom; Sam Altman, Greg Brockman, and Ilya Sutskever. 2023. <a href="https://openai.com/blog/governance-of-superintelligence">Governance of Superintelligence</a> (May 2023); Shazeda Ahmed et al. 2023. <em><a href="https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4641526">Building the Epistemic Community of AI Safety</a></em>. SSRN: Rochester, NY.</p><p>[2] This is different from the question of whether it is helpful for an individual user to conceptualize a specific AI system as a tool as opposed to a human-like entity such as an intern, a co-worker, or a tutor.</p><p>[3] Daron Acemoglu and Simon Johnson. 2023. <em>Power and Progress: Our Thousand-Year Struggle over Technology and Prosperity</em>. PublicAffairs, New York, NY.</p><p>[4] Jeffrey Ding. 2024. <em>Technology and the Rise of Great Powers: How Diffusion Shapes Economic Competition.</em> Princeton University Press, Princeton.</p><p>[5] Angelina Wang et al. 2023. <a href="https://dl.acm.org/doi/10.1145/3593013.3594030">Against predictive optimization: On the legitimacy of decision-making algorithms that optimize predictive accuracy</a>. In <em>Proceedings of the 2023 ACM Conference on Fairness, Accountability, and Transparency</em> (Chicago, IL, USA: ACM, 2023), 626&#8211;26.</p><p>[6] Casey Ross. 2022. <a href="https://www.statnews.com/2022/10/24/epic-overhaul-of-a-flawed-algorithm/">Epic&#8217;s Overhaul of a Flawed Algorithm Shows Why AI Oversight Is a Life-or-Death Issue</a>. <em>STAT</em>.</p><p>[7] Andrew Wong et al. 2021. <a href="https://jamanetwork.com/journals/jamainternalmedicine/fullarticle/2781307">External validation of a widely implemented proprietary sepsis prediction model in hospitalized patients</a>. <em>JAMA Internal Medicine</em> 181, 8 (August 2021), 1065&#8211;70.</p><p>[8] Kevin Roose. 2023. <a href="https://www.nytimes.com/2023/02/16/technology/bing-chatbot-microsoft-chatgpt.html">A Conversation With Bing&#8217;s Chatbot Left Me Deeply Unsettled</a>. <em>The New York Times</em> (February 2023).</p><p>[9] Dan Milmo and Alex Hern. 2024. <a href="https://www.theguardian.com/technology/2024/mar/08/we-definitely-messed-up-why-did-google-ai-tool-make-offensive-historical-images">&#8216;We definitely messed up&#8217;: why did Google AI tool make offensive historical images?</a> <em>The Guardian </em>(March 2024).</p><p>[10] Jamie Bernardi et al. 2024. <a href="http://arxiv.org/abs/2405.10295">Societal adaptation to advanced AI</a>. arXiv: May 2024; Center for Devices and Radiological Health. 2024. <a href="https://www.fda.gov/medical-devices/medical-device-regulatory-science-research-programs-conducted-osel/regulatory-evaluation-new-artificial-intelligence-ai-uses-improving-and-automating-medical-practices">Regulatory evaluation of new artificial intelligence (AI) uses for improving and automating medical practices</a>. <em>FDA</em> (June 2024); &#8220;Regulation (EU) 2024/1689 of the European Parliament and of the Council of 13 June 2024 Laying down Harmonised Rules on Artificial Intelligence and Amending Regulations (EC) No 300/2008, (EU) No 167/2013, (EU) No 168/2013, (EU) 2018/858, (EU) 2018/1139 and (EU) 2019/2144 and Directives 2014/90/EU, (EU) 2016/797 and (EU) 2020/1828 (<a href="https://eur-lex.europa.eu/eli/reg/2024/1689/oj/eng">Artificial Intelligence Act</a>) (Text with EEA Relevance),&#8221; June 2024.</p><p>[11] Javier Espinoza. 2024. <a href="https://www.ft.com/content/6cc7847a-2fc5-4df0-b113-a435d6426c81">Europe&#8217;s rushed attempt to set the rules for AI</a>. <em>Financial Times</em> (July 2024); Daniel E. Ho and Nicholas Bagley. 2024. <a href="https://thehill.com/opinion/technology/4405286-runaway-bureaucracy-could-make-common-uses-of-ai-worse-even-mail-delivery/">Runaway bureaucracy could make common uses of ai worse, even mail delivery</a>. <em>The Hill </em>(January 2024).</p><p>[12] Avanidhar Subrahmanyam. 2013. <a href="https://www.sciencedirect.com/science/article/pii/S2214845013000082?via%3Dihub">Algorithmic trading, the flash crash, and coordinated circuit breakers</a>. <em>Borsa Istanbul Review</em> 13, 3 (September 2013), 4&#8211;9. </p><p>[13] Alexander Bick, Adam Blandin, and David J. Deming. 2024. <em>The Rapid Adoption of Generative AI</em>. National Bureau of Economic Research.</p><p>[14] Alexander Bick, Adam Blandin, and David J. Deming. 2024. <em>The Rapid Adoption of Generative AI</em>. National Bureau of Economic Research.</p><p>[15] Benedict Evans. 2023. <a href="https://www.ben-evans.com/benedictevans/2023/7/2/working-with-ai">AI and the Automation of Work</a>; Benedict Evans, 2023; Jeffrey Ding. 2024. <em>Technology and the Rise of Great Powers: How Diffusion Shapes Economic Competition.</em> Princeton University Press, Princeton.</p><p>[16] Paul A. David. 1990. <a href="https://www.jstor.org/stable/2006600">The dynamo and the computer: an historical perspective on the modern productivity paradox</a>. <em>The American Economic Review</em> 80, 2 (1990), 355&#8211;61; Tim Harford. 2017. <a href="https://www.bbc.com/news/business-40673694">Why didn&#8217;t electricity immediately change manufacturing?</a> (August 2017).</p><p>[17] Robert Solow as quoted in Paul A. David. 1990. <a href="https://www.jstor.org/stable/2006600">The dynamo and the computer: an historical perspective on the modern productivity paradox</a>. <em>The American Economic Review</em> 80, 2 (1990), Page 355; Tim Harford. 2017. <a href="https://www.bbc.com/news/business-40673694">Why didn&#8217;t electricity immediately change manufacturing?</a> (August 2017).</p><p>[18] Arvind Narayanan and Sayash Kapoor. 2024. <em>AI Snake Oil: What Artificial Intelligence Can Do, What It Can&#8217;t, and How to Tell the Difference</em>. Princeton University Press, Princeton, NJ.</p><p>[19] Alex Krizhevsky, Ilya Sutskever, and Geoffrey E Hinton. 2012. ImageNet classification with deep convolutional neural networks. <em>Advances in Neural Information Processing Systems</em> 25 (2012); Harris Drucker, Donghui Wu, and Vladimir N. Vapnik. 1999. <a href="https://ieeexplore.ieee.org/document/788645">Support vector machines for spam categorization</a>. <em>IEEE Transactions on Neural Networks</em> 10, 5 (September 1999), 1048&#8211;54; William D. Smith. 1964. <a href="https://www.nytimes.com/1964/04/08/archives/new-ibm-system-360-can-serve-business-science-and-government-ibm.html">New I.B.M, System 360 can serve business, science and government; I.B.M. Introduces a computer it says tops output of biggest</a>. <em>The New York Times</em> April 1964; Special to THE NEW YORK TIMES. <a href="https://www.nytimes.com/1944/08/07/archives/algebra-machine-spurs-research-calling-for-long-calculations.html">Algebra machine spurs research calling for long calculations; Harvard receives today device to solve in hours problems taking so much time they have never been worked out</a>. <em>The New York Times</em> (August 1944); Herman Hollerith. 1894. <a href="https://www.jstor.org/stable/2979610?origin=crossref">The electrical tabulating machine</a>. <em>Journal of the Royal Statistical Society</em> 57, 4 (December 1894), 678.</p><p>[20] <em>Mohammad Musa</em>, <em>Tim Dawkins</em>, and <em>Nicola Croce</em>. 2019. <a href="https://www.weforum.org/stories/2019/12/the-key-to-a-safe-self-driving-future-lies-in-sharing-data/">This is the next step on the road to a safe self-driving future</a>. <em>World Economic Forum </em>(December 2019); Louise Zhang. 2023. <a href="https://web.archive.org/web/20230504102309/https://getcruise.com/news/blog/2023/cruises-safety-record-over-one-million-driverless-miles/">Cruise&#8217;s Safety Record Over 1 Million Driverless Miles</a>. <em>Cruise (April 2023). </em></p><p>[21] Arvind Narayanan and Sayash Kapoor. 2024. <a href="https://www.aisnakeoil.com/p/ai-companies-are-pivoting-from-creating">AI companies are pivoting from creating gods to building products. Good.</a> <em>AI Snake Oil newsletter</em>.</p><p>[22] Rich Sutton. 2019. <a href="http://www.incompleteideas.net/IncIdeas/BitterLesson.html">The Bitter Lesson</a> (March 2019).</p><p>[23] Arvind Narayanan and Sayash Kapoor. 2024. <a href="https://www.aisnakeoil.com/p/ai-companies-are-pivoting-from-creating">AI companies are pivoting from creating gods to building products. Good.</a> <em>AI Snake Oil newsletter</em>. </p><p>[24] Melanie Mitchell. 2021. <a href="http://arxiv.org/abs/2104.12871">Why AI is harder than we think</a>. arXiv preprint (April 2021).</p><p>[25] Josh Achiam et al. 2023. GPT-4 technical report. arXiv preprint<em> </em>arXiv: 2303.08774; Peter Henderson et al. 2024. Rethinking machine learning benchmarks in the context of professional codes of conduct. In <em>Proceedings of the Symposium on Computer Science and Law</em>; Varun Magesh et al. 2024. Hallucination-free? Assessing the reliability of leading AI legal research tools. arXiv preprint arXiv: 2405.20362; Daniel N. Kluttz and Deirdre K. Mulligan. 2019. Automated decision support technologies and the legal profession. <em>Berkeley Technology Law Journal</em> 34, 3 (2019), 853&#8211;90; Inioluwa Deborah Raji, Roxana Daneshjou, and Emily Alsentzer. 2025. It&#8217;s time to bench the medical exam benchmark. <em>NEJM AI</em> 2, 2 (2025).</p><p>[26] Sayash Kapoor, Peter Henderson, and Arvind Narayanan. <a href="https://journalcrcl.org/crcl/article/view/62">Promises and pitfalls of artificial intelligence for legal applications</a>. <em>Journal of Cross-Disciplinary Research in Computational Law</em> 2, 2 (May 2024), Article 2.</p><p>[27] Hamel Husain, Isaac Flath, and Johno Whitaker. <a href="http://answer.ai/posts/2025-01-08-devin.html">Thoughts on a month with Devin</a>. <em>Answer.AI</em> (2025).</p><p>[28] Ehud Reiter. 2025. <a href="https://ehudreiter.com/2025/01/13/do-llm-coding-benchmarks-measure-real-world-utility/">Do LLM Coding Benchmarks Measure Real-World Utility?</a>.</p><p>[29] Deborah Raji et al. 2021. <a href="https://datasets-benchmarks-proceedings.neurips.cc/paper/2021/hash/084b6fbb10729ed4da8c3d3f5a3ae7c9-Abstract-round2.html">AI and the everything in the whole wide world benchmark</a>. In <em>Proceedings of the Neural Information Processing Systems (NeurIPS) Track on Datasets and Benchmarks</em>, vol. 1.; Rachel Thomas and David Uminsky. 2020. <a href="https://arxiv.org/abs/2002.08512v1">The problem with metrics is a fundamental problem for AI</a>. arXiv preprint.</p><p>[30] Ashwin Nayak et al. 2023. <a href="https://doi.org/10.1001/jamainternmed.2023.2561">Comparison of history of present illness summaries generated by a chatbot and senior internal medicine residents</a>. <em>JAMA Internal Medicine</em> 183, 9 (September 2023), 1026&#8211;27; Shakked Noy and Whitney Zhang. 2023.<a href="https://www.science.org/doi/10.1126/science.adh2586"> Experimental evidence on the productivity effects of generative artificial intelligence.</a> <em>Science</em> 381, 6654 (July 2023), 187&#8211;92; Fabrizio Dell&#8217;Acqua et al., &#8220;Navigating the Jagged Technological Frontier: Field Experimental Evidence of the Effects of AI on Knowledge Worker Productivity and Quality,&#8221; <em>Harvard Business School Technology &amp; Operations Mgt. Unit Working Paper</em>, no. 24&#8211;13 (2023).</p><p>[31] Pranshu Verma and Gerrit De Vynck. 2023. <a href="https://www.washingtonpost.com/technology/2023/06/02/ai-taking-jobs/">ChatGPT took their jobs. Now they walk dogs and fix air conditioners</a>. <em>Washington Post</em> (June 2023).</p><p>[32] Metaculus. 2024. Will there be human-machine intelligence parity before 2040?.</p><p>[33] Mario Krenn et al. 2023. Forecasting the future of artificial intelligence with machine learning-based link prediction in an exponentially growing knowledge network. <em>Nature Machine Intelligence</em> 5, 11 (2023), 1326&#8211;35.</p><p>[34] Johan S.G. Chu and James A. Evans. 2021. Slowed Canonical Progress in Large Fields of Science. <em>Proceedings of the National Academy of Sciences</em> 118, 41 (2021), e2021636118.</p><p>[35] Timothy B. Lee. 2024. Predictions of AI doom are too much like Hollywood movie plots. https://www.understandingai.org/p/predictions-of-ai-doom-are-too-much</p><p>[36] Celeste Biever. 2023. <a href="https://www.nature.com/articles/d41586-023-02361-7">ChatGPT broke the Turing Test &#8212; The race is on for new ways to assess AI</a>. <em>Nature</em> 619, 7971 (July 2023), 686&#8211;89; Melanie Mitchell. 2024. <a href="https://www.science.org/doi/10.1126/science.adq9356">The Turing Test and our shifting conceptions of intelligence</a>. <em>Science</em> 385, 6710 (2024), eadq9356.</p><p>[37] John McCarthy, Marvin L. Minsky, Nathaniel Rochester and Claude E. Shannon. 1955. <a href="http://jmc.stanford.edu/articles/dartmouth/dartmouth.pdf">A proposal for the dartmouth summer research project on artificial intelligence</a>.</p><p>[38] Changmao Li and Jeffrey Flanigan. 2023. <a href="https://arxiv.org/abs/2312.16337">Task contamination: Language models may not be few-shot anymore</a>. arXiv: December 2023. </p><p>[39] Luke Muehlhauser. 2013. <a href="https://intelligenceexplosion.com/2011/plenty-of-room-above-us/">Plenty of room above us</a>. In <em>Facing the Intelligence Explosion</em>.</p><p>[40] Melanie Mitchell et al. 2024. Ep. 1: <a href="https://www.santafe.edu/culture/podcasts/ep-1-what-is-intelligence">What is intelligence?</a> Complexity. Santa Fe Institute; Podcast episode; Melanie Mitchell. 2019. Opinion. <a href="https://www.nytimes.com/2019/10/31/opinion/superintelligent-artificial-intelligence.html">We shouldn&#8217;t be scared by &#8216;Superintelligent A.I.&#8217;</a> <em>The New York Times</em> (October 2019).</p><p>[41] Matthew J Salganik et al. 2020. Measuring the predictability of life outcomes with a scientific mass collaboration. <em>Proceedings of the National Academy of Sciences</em> 117, 15 (2020), 8398&#8211;8403.</p><p>[42] Mary Phuong et al. 2024. <a href="https://arxiv.org/abs/2403.13793">Evaluating frontier models for dangerous capabilities</a>. arXiv: April 2024. Page 5. </p><p>[43] Mary Phuong et al. 2024. <a href="https://arxiv.org/abs/2403.13793">Evaluating frontier models for dangerous capabilities</a>. arXiv: April 2024. </p><p>[44] Arvind Narayanan, Sayash Kapoor, and Seth Lazar. 2024. <a href="https://www.aisnakeoil.com/p/model-alignment-protects-against">Model alignment protects against accidental harms, not intentional ones</a>.</p><p>[45] Raja Parasuraman and Dietrich H. Manzey. 2010. <a href="https://journals.sagepub.com/doi/10.1177/0018720810376055">Complacency and bias in human use of automation: An attentional integration</a>. <em>Human Factors</em> 52, 3 (June 2010), 381&#8211;410.</p><p>[46] Roel I. J. Dobbe. 2022. <a href="https://doi.org/10.1093/oxfordhb/9780197579329.013.67">System safety and artificial intelligence</a>. In <em>The Oxford Handbook of AI Governance</em>, ed. Justin B. Bullock et al., Oxford University Press, Oxford.</p><p>[47] CodeMetal.ai. 2024. <a href="https://www.codemetal.ai/research/combining-ai-with-formal-verification-for-efficient-migration-of-legacy-code">Combining AI with formal verification for efficient migration of legacy code</a>.</p><p>[48] Balint Gyevnar and Atoosa Kasirzadeh. 2025. AI safety for everyone. arXiv preprint arXiv: 2502.09288.</p><p>[49] Balint Gyevnar and Atoosa Kasirzadeh. 2025. AI safety for everyone. arXiv preprint arXiv: 2502.09288; Tinghao Xie et al. 2024. SORRY-Bench: Systematically evaluating large language model safety refusal behaviors. arXiv: June 2024; Alan Chan et al. 2024. Visibility into AI agents. arXiv preprint arXiv:2401.13138; Yonadav Shavit et al. 2023. <a href="https://cdn.openai.com/papers/practices-for-governing-agentic-ai-systems.pdf">Practices for governing agentic AI systems</a>.</p><p>[50] Karen Levy. 2022. <em>Data Driven: Truckers, Technology, and the New Workplace Surveillance</em>. Princeton University Press, Princeton, NJ.</p><p>[51] Andrew J. Hawkins. 2024. <a href="https://www.theverge.com/2024/9/5/24235078/waymo-safety-hub-miles-crashes-robotaxi-transparency">Waymo thinks it can overcome robotaxi skepticism with lots of safety data</a>. <em>The Verge</em>; Caleb Miller. 2024. <a href="https://www.caranddriver.com/news/a63158982/general-motors-cruise-robotaxi-dead/">General motors gives up on its cruise robotaxi dreams</a>. <em>Car and Driver</em> (December 2024); Greg Bensinger. 2021. <a href="https://www.nytimes.com/2021/07/30/opinion/self-driving-cars-tesla-elon-musk.html">Why Tesla&#8217;s &#8216;Beta Testing&#8217; Puts the Public at Risk</a>. <em>The New York Times </em>(July 2021); Andrew J. Hawkins. 2020. <a href="https://www.theverge.com/2020/12/7/22158745/uber-selling-autonomous-vehicle-business-aurora-innovation">Uber&#8217;s fraught and deadly pursuit of self-driving cars is over</a>. <em>The Verge</em>.</p><p>[52] Caleb Miller. 2024. <a href="https://www.caranddriver.com/news/a63158982/general-motors-cruise-robotaxi-dead/">General motors gives up on its cruise robotaxi dreams</a>. <em>Car and Driver</em> (December 2024); Andrew J. Hawkins. 2020. <a href="https://www.theverge.com/2020/12/7/22158745/uber-selling-autonomous-vehicle-business-aurora-innovation">Uber&#8217;s fraught and deadly pursuit of self-driving cars is over</a>. <em>The Verge</em>.</p><p>[53] Jonathan Stempel. 2024. <a href="https://www.reuters.com/legal/tesla-must-face-vehicle-owners-lawsuit-over-self-driving-claims-2024-05-15/">Tesla must face vehicle owners&#8217; lawsuit over self-driving claims</a>. <em>Reuters</em> (May 2024).</p><p>[54] Hayden Field. 2023. <a href="https://www.cnbc.com/2023/12/05/waymo-chief-product-officer-on-progress-competition-vs-cruise.html">Waymo is full speed ahead as safety incidents and regulators stymie competitor cruise</a>.</p><p>[55] Will Hunt. 2020. <a href="https://cltc.berkeley.edu/wp-content/uploads/2020/08/Flight-to-Safety-Critical-AI.pdf">The flight to safety-critical AI: Lessons in AI safety from the aviation industry</a>. <em>CLTC White Paper Series</em>. UC Berkeley Center for Long-Term Cybersecurity.</p><p>[56] Arvind Narayanan. 2023. <em><a href="https://knightcolumbia.org/content/understanding-social-media-recommendation-algorithms">Understanding Social Media Recommendation Algorithms</a></em>. Knight First Amendment Institute.</p><p>[57] Ralph Nader. 1965. <em>Unsafe at Any Speed: The Designed-in Dangers of the American Automobile.</em> Grossman Publishers, New York, NY.</p><p>[58] Seth Lazar. 2025. Anticipatory AI ethics (manuscript, forthcoming 2025).</p><p>[59] Alex Engler. 2023. <a href="https://www.brookings.edu/research/the-eu-and-us-diverge-on-ai-regulation-a-transatlantic-comparison-and-steps-to-alignment/">The EU and U.S. diverge on AI regulation: A transatlantic comparison and steps to alignment</a>.</p><p>[60] Matt Sheehan. 2023. <a href="https://carnegieendowment.org/2023/07/10/china-s-ai-regulations-and-how-they-get-made-pub-90117">China&#8217;s AI regulations and how they get made</a>.</p><p>[61] Heather Curry, 2024. <a href="https://techpost.bsa.org/2024/10/22/2024-state-summary-on-ai/">2024 state summary on AI</a>. <em>BSA TechPost</em> (October 2024).</p><p>[62] Yuntao Bai et al. 2022. Constitutional AI: <a href="https://arxiv.org/abs/2212.08073">Harmlessness from AI feedback</a>. arXiv: December 2022; Long Ouyang et al.. 2022. <a href="https://arxiv.org/abs/2203.02155">Training language models to follow instructions with human feedback</a>. arXiv: March 2022. </p><p>[63] Eugene Bagdasaryan et al. 2023. <a href="http://arxiv.org/abs/2307.10490">Abusing images and sounds for indirect instruction injection in multi-modal LLMs</a>. arXiv: October 2023; Xiangyu Qi et al. 2023. <a href="https://arxiv.org/abs/2310.03693">Fine-tuning aligned language models compromises safety, even when users do not intend to!</a> arXiv: October 2023.</p><p>[64] Arvind Narayanan and Sayash Kapoor. 2024. <a href="https://www.aisnakeoil.com/p/ai-safety-is-not-a-model-property">AI safety is not a model property</a>.</p><p>[65] Erik Jones, Anca Dragan, and Jacob Steinhardt. 2024. <a href="https://arxiv.org/abs/2406.14595">Adversaries can misuse combinations of safe models</a>. arXiv: July 2024.</p><p>[66] Arvind Narayanan and Sayash Kapoor. 2024. <a href="https://www.aisnakeoil.com/p/ai-safety-is-not-a-model-property">AI safety is not a model property</a>.</p><p>[67] Google. 2024. <a href="https://support.google.com/mail/answer/81126?hl=en">Email sender guidelines</a>.</p><p>[68] Craig Marcho. 2024. <a href="https://techcommunity.microsoft.com/t5/ask-the-performance-team/ie7-introducing-the-phishing-filter/ba-p/372327">IE7 - Introducing the phishing filter</a>. <em>Microsoft Tech Community</em>.</p><p>[69] Jennifer Tang, Tiffany Saade, and Steve Kelly. 2024. <a href="https://securityandtechnology.org/wp-content/uploads/2024/10/The-Implications-of-Artificial-Intelligence-in-Cybersecurity.pdf">The implications of artificial intelligence in cybersecurity: shifting the offense-defense balance</a>. </p><p>[70] Dongge Liu et al. 2023. <a href="https://security.googleblog.com/2023/08/ai-powered-fuzzing-breaking-bug-hunting.html">AI-Powered Fuzzing: Breaking the Bug Hunting Barrier</a>. <em>Google Online Security Blog</em>.</p><p>[71] Juan Cambeiro. <a href="https://ifp.org/how-ai-can-help-prevent-biosecurity-disasters/">How AI can help prevent biosecurity disasters.</a> <em>Institute for Progress</em> (July 2023).</p><p>[72] LessWrong. 2008. <a href="https://www.lesswrong.com/tag/squiggle-maximizer-formerly-paperclip-maximizer">Squiggle maximizer (formerly &#8220;paperclip maximizer&#8221;)</a>.</p><p>[73] Ryan Greenblatt et al. 2024. <a href="https://arxiv.org/abs/2412.14093">Alignment faking in large language models</a>.</p><p>[74] Bowen Baker et al. 2025. <a href="https://arxiv.org/abs/2503.11926">Monitoring reasoning models for misbehavior and the risks of promoting obfuscation</a>.</p><p>[75] Victoria Krakovna. 2020. <a href="https://deepmind.google/discover/blog/specification-gaming-the-flip-side-of-ai-ingenuity/">Specification gaming: The flip side of AI ingenuity</a>. <em>Google DeepMind</em> (April 2020).</p><p>[76] Simon Dima et al. 2024. <a href="http://arxiv.org/abs/2408.04385">Non-maximizing policies that fulfill multi-criterion aspirations in expectation</a>. arXiv: August 2024.</p><p>[77] Daron Acemoglu and Simon Johnson. 2023. <em>Power and Progress</em>. PublicAffairs.</p><p>[78] Atoosa Kasirzadeh. 2024. <a href="https://arxiv.org/abs/2401.07836">Two types of AI existential risk: Decisive and accumulative</a>. arXiv: preprint (February 2024).</p><p>[79] Anton Leicht. 2024. <a href="https://www.antonleicht.me/writing/veto">AI safety politics after the SB-1047 veto</a>.</p><p>[80] Timothy B. Lee. 2024. <a href="https://www.understandingai.org/p/six-principles-for-thinking-about">Six Principles for Thinking about AI Risk</a>.</p><p>[81] Mary Phuong et al. 2024. <a href="https://arxiv.org/abs/2403.13793">Evaluating frontier models for dangerous capabilities</a>.</p><p>[82] Shazeda Ahmed et al. 2024. <a href="https://firstmonday.org/ojs/index.php/fm/article/view/13626/11596">Field-building and the epistemic culture of AI safety</a>. <em>First Monday</em> 29, 4.</p><p>[83] Arvind Narayanan and Sayash Kapoor. 2024. <a href="https://www.aisnakeoil.com/p/ai-existential-risk-probabilities">AI existential risk probabilities are too unreliable to inform policy</a>; Neel Guha et al. 2023. <a href="https://papers.ssrn.com/abstract=4634443">AI regulation has its own alignment problem: The technical and institutional feasibility of disclosure, registration, licensing, and auditing</a>. <em>SSRN</em> (November 2023).</p><p>[84] Christopher A. Mouton, Caleb Lucas, and Ella Guest. 2024. <a href="https://www.rand.org/pubs/research_reports/RRA2977-2.html">The operational risks of AI in large-scale biological attacks: Results of a red-team study</a>. RAND Corporation; Ari Takanen, Jared D. Demott, and Charles Miller. 2008. Fuzzing for Software Security Testing and Quality Assurance. <em>Fuzzing for Software Security</em> (1st ed.). Artech House Publishers, Norwood, MA.</p><p>[85] Sayash Kapoor and Arvind Narayanan. 2023. <a href="https://www.aisnakeoil.com/p/licensing-is-neither-feasible-nor">Licensing is neither feasible nor effective for addressing AI risks</a> (June 2023).</p><p>[86] Arvind Narayanan and Sayash Kapoor. 2024. <a href="https://www.aisnakeoil.com/p/ai-existential-risk-probabilities">AI existential risk probabilities are too unreliable to inform policy</a>.</p><p>[87] Richard Blumenthal and Josh Hawley. 2023. <a href="https://www.blumenthal.senate.gov/imo/media/doc/09072023bipartisanaiframework.pdf">Bipartisan framework for U.S. AI act</a>.</p><p>[88] Sigal Samuel. 2022. <a href="https://www.vox.com/future-perfect/23298870/effective-altruism-longtermism-will-macaskill-future">Effective altruism&#8217;s most controversial idea</a>.</p><p>[89] Kevin Vallier. 1996. <a href="https://plato.stanford.edu/entries/justification-public/">Public justification</a>.</p><p>[90] Jeffrey A Friedman and Richard Zeckhauser. 2018. Analytic confidence and political decision-making: Theoretical principles and experimental evidence from national security professionals. <em>Political Psychology</em> 39, 5 (2018), 1069&#8211;87.</p><p>[91] Arvind Narayanan and Sayash Kapoor. 2023. <a href="http://knightcolumbia.org/blog/generative-ai-companies-must-publish-transparency-reports">Generative AI companies must publish transparency reports</a>. Knight First Amendment Institute; Executive Office of the President. 2020. <a href="https://www.federalregister.gov/documents/2020/12/08/2020-27065/promoting-the-use-of-trustworthy-artificial-intelligence-in-the-federal-government">Promoting the use of trustworthy artificial intelligence in the federal government</a>, 2020; Justin Colannino. 2021. <a href="https://github.blog/security/vulnerability-research/copyright-office-expands-security-research-rights/">The copyright office expands your security research rights</a>. GitHub Blog.</p><p>[92] AI Incident Database. https://incidentdatabase.ai/.</p><p>[93] Stephen Casper, David Krueger, and Dylan Hadfield-Menell. 2025. <a href="https://arxiv.org/abs/2502.09618">Pitfalls of evidence-based AI policy</a>.</p><p>[94] Sayash Kapoor et al. 2024. <a href="https://arxiv.org/abs/2403.07918">On the societal impact of open foundation models</a>. </p><p>[95] Gary E. Marchant and Yvonne A. Stevens. 2017. <a href="https://lawreview.law.ucdavis.edu/sites/g/files/dgvnsk15026/files/media/documents/51-1_Marchant_Stevens.pdf">Resilience: A new tool in the risk governance toolbox for emerging technologies</a>. <em>UC Davis Law Review</em>.</p><p>[96] Brian Walker et al. 2006. A handful of heuristics and some propositions for understanding resilience in social-ecological systems. <em>Ecology and Society</em> 11, 1 (2006).</p><p>[97] Gary E. Marchant and Yvonne A. Stevens. 2017. <a href="https://lawreview.law.ucdavis.edu/sites/g/files/dgvnsk15026/files/media/documents/51-1_Marchant_Stevens.pdf">Resilience: A new tool in the risk governance toolbox for emerging technologies. </a><em>UC Davis Law Review</em>.</p><p>[98] Rishi Bommasani et al. 2024. <a href="https://understanding-ai-safety.org/">A path for science- and evidence-based AI policy</a>; Balint Gyevnar and Atoosa Kasirzadeh. 2025. <a href="https://arxiv.org/abs/2502.09288">AI safety for everyone</a>; Anka Reuel et al. 2024. <a href="https://proceedings.mlr.press/v235/reuel24a.html">Position: Technical research and talent is needed for effective AI governance</a>. In <em>Proceedings of the 41st International Conference on Machine Learning</em> (PMLR, 2024), 42543&#8211;57.</p><p>[99] The National Artificial Intelligence Advisory Committee (NAIAC). 2023. <a href="https://ai.gov/wp-content/uploads/2023/12/Recommendation_Improve-Monitoring-of-Emerging-Risks-from-AI-through-Adverse-Event-Reporting.pdf">Improve monitoring of emerging risks from AI through adverse event reporting</a>. (November 2023); Shayne Longpre et al. 2024. <a href="https://knightcolumbia.org/blog/a-safe-harbor-for-ai-evaluation-and-red-teaming">A safe harbor for AI evaluation and red teaming</a> (March 2024); Jamie Bernardi et al. 2025. <a href="https://arxiv.org/abs/2405.10295">Societal adaptation to advanced AI</a>. Helen Toner. 2024. <a href="https://www.judiciary.senate.gov/imo/media/doc/2024-09-17_pm_-_testimony_-_toner.pdf#page=6.00">Oversight of AI: Insiders&#8217; perspectives</a> (September 2024).</p><p>[100] Sayash Kapoor and Rishi Bommasani et al. 2024. <a href="https://crfm.stanford.edu/open-fms/paper.pdf">On the societal impact of open foundation models</a>; Rishi Bommasani et al. 2024. <a href="https://www.science.org/doi/10.1126/science.adp1848">Considerations for Governing Open Foundation Models</a>. <em>Science</em> 386, 6718 (October 2024), 151&#8211;53; Gary E. Marchant and Yvonne A. Stevens. 2017. <a href="https://lawreview.law.ucdavis.edu/archives/51/1/resilience-new-tool-risk-governance-toolbox-emerging-technologies">Resilience</a>; Noam Kolt. 2024. <a href="https://wustllawreview.org/wp-content/uploads/2024/04/Kolt-Algorithmic-Black-Swans.pdf">Algorithmic black swans</a>. <em>Washington University Law Review</em>.</p><p>[101] Richard Blumenthal and Josh Hawley. 2023. <a href="https://www.blumenthal.senate.gov/newsroom/press/release/blumenthal-and-hawley-announce-bipartisan-framework-on-artificial-intelligence-legislation">Bipartisan framework for U.S. AI act</a>; Josh Hawley. 2025. Decoupling America&#8217;s artificial intelligence capabilities from China Act of 2025. <em>Pub. L. No. S</em> 321 (2025).</p><p>[102] Sayash Kapoor and Arvind Narayanan. 2023. <a href="https://www.aisnakeoil.com/p/licensing-is-neither-feasible-nor">Licensing is neither feasible nor effective for addressing AI risks</a>.</p><p>[103] Eliezer Yudkowsky. 2023. <a href="https://time.com/6266923/ai-eliezer-yudkowsky-open-letter-not-enough/">Pausing AI developments isn&#8217;t enough. we need to shut it all down.</a> (March 2023).</p><p>[104] Reuters. 2005. <a href="https://www.nbcnews.com/id/wbna8958495">New Internet worm targeting Windows.</a> <em>NBC News </em>(August 2005)<em>.</em></p><p>[105] Christopher A. Mouton, Caleb Lucas, and Ella Guest. 2024. <a href="https://www.rand.org/pubs/research_reports/RRA2977-2.html">The operational risks of AI in large-scale biological attacks. </a></p><p>[106] Dan Hendrycks, Eric Schmidt, and Alexandr Wang. 2025. Superintelligence strategy: Expert version. arXiv: preprint<em> </em>arXiv:2503.05628.</p><p>[107] Emanuel Maiberg. 2024. <a href="https://www.404media.co/apple-removes-nonconsensual-ai-nude-apps-following-404-media-investigation/">Apple removes nonconsensual AI nude apps following 404 Media investigation</a>.</p><p>[108] Jeffrey Ding. 2024. <em>Technology and the Rise of Great Powers: How Diffusion Shapes Economic Competition.</em> Princeton University Press, Princeton..</p><p>[109] Olivia Martin et al. 2024, The spectrum of AI integration: The case of benefits adjudication. In <em>Artificial Intelligence: Legal Issues, Policy &amp; Practical Strategies</em>, Cynthia H. Cwik (ed.).</p><p>[110] Anu Bradford. The false choice between digital regulation and innovation. <em>Nw. UL Rev.</em> 119 (2024), 377.</p><p>[111] Scott R. Zemnick. 2001. The E-Sign Act: The Means to Effectively Facilitate the Growth and Development of E-commerce. <em>Chicago-Kent Law Review </em>(April 2001). https://scholarship.kentlaw.iit.edu/cgi/viewcontent.cgi?article=3342&amp;context=cklawreview.</p><p>[112] Benjamin Brooks. 2024. <a href="https://www.technologyreview.com/2024/10/31/1106504/ai-search-could-break-the-web/">AI search could break the web</a>. <em>MIT Technology Review</em> (October 2024).</p><p>[113] <a href="https://time.com/5296311/time-the-drone-age-2/">Drones Are Here to Stay. Get Used to It.</a> 2018. <em>Time </em>(May 2018).</p><p>[114] Ipsos. 2024. <a href="https://www.ipsos.com/en/ipsos-ai-monitor-2024-changing-attitudes-and-feelings-about-ai-and-future-it-will-bring">The Ipsos AI Monitor 2024: Changing attitudes and feelings about AI and the future it will bring</a>.</p><p>[115] Colin Lecher. 2024. <a href="https://themarkup.org/news/2024/03/29/nycs-ai-chatbot-tells-businesses-to-break-the-law">NYC&#8217;s AI chatbot tells businesses to break the law</a>. <em>The Markup</em>.</p><p>[116] Courtney Kube et al. 2025. <a href="https://www.nbcnews.com/politics/doge/doge-will-use-ai-assess-responses-federal-workers-who-were-told-justify-jobs-rcna193439">DOGE will use AI to assess the responses of federal workers who were told to justify their jobs via email</a>. <em>NBC News </em>(February 2025); Dell Cameron. 2025. <a href="https://www.wired.com/story/elon-musk-federal-agencies-ai/">Democrats demand answers on DOGE&#8217;s use of AI</a>.</p><p>[117] Dean W. Ball. 2021. <a href="https://www.piratewires.com/p/how-california-turned-on-its-own-citizens">How California turned on its own citizens</a>.</p><p>[118] Kate Dore. 2024. <a href="https://www.cnbc.com/2024/04/06/heres-what-to-know-before-using-ai-chatbots-to-file-your-taxes.html">&#8216;Proceed with caution&#8217; before tapping AI chatbots to file your tax return, experts warn</a>. <em>CNBC </em>(April 2024).</p><p>[119] Nicholas Bagley. 2021. <a href="https://www.niskanencenter.org/the-procedure-fetish/">The procedure fetish.</a> Niskanen Center; Daniel E. Ho and Nicholas Bagley. 2024. <a href="https://thehill.com/opinion/technology/4405286-runaway-bureaucracy-could-make-common-uses-of-ai-worse-even-mail-delivery/">Runaway bureaucracy could make common uses of AI worse, even mail delivery</a>.</p>]]></content:encoded></item><item><title><![CDATA[Is AI progress slowing down?]]></title><description><![CDATA[Making sense of recent technology trends and claims]]></description><link>https://www.normaltech.ai/p/is-ai-progress-slowing-down</link><guid isPermaLink="false">https://www.normaltech.ai/p/is-ai-progress-slowing-down</guid><dc:creator><![CDATA[Arvind Narayanan]]></dc:creator><pubDate>Wed, 18 Dec 2024 16:47:58 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!0fDW!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc348909d-140e-4e20-94d7-92a04e008bb2_1450x1270.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>By Arvind Narayanan, Benedikt Str&#246;bl, and Sayash Kapoor</em>.</p><p>After the release of GPT-4 in March 2023, the <a href="https://x.com/fchollet/status/1848178049105494084">dominant narrative</a> in the tech world was that continued scaling of models would lead to artificial general intelligence and then superintelligence. Those extreme predictions gradually receded, but up until a month ago, the prevailing belief in the AI industry was that model scaling would continue for the foreseeable future.</p><p>Then came three back-to-back news reports from <a href="https://www.theinformation.com/articles/openai-shifts-strategy-as-rate-of-gpt-ai-improvements-slows">The Information</a>, <a href="https://www.reuters.com/technology/artificial-intelligence/openai-rivals-seek-new-path-smarter-ai-current-methods-hit-limitations-2024-11-11/">Reuters</a>, and <a href="https://www.bloomberg.com/news/articles/2024-11-13/openai-google-and-anthropic-are-struggling-to-build-more-advanced-ai">Bloomberg</a> revealing that three leading AI developers &#8212; OpenAI, Anthropic, and Google Gemini &#8212; had all run into problems with their next-gen models. Many industry insiders, including Ilya Sutskever, probably the most notable proponent of scaling, are now singing a very different tune:</p><blockquote><p>&#8220;The 2010s were the age of scaling, now we're back in the age of wonder and discovery once again. Everyone is looking for the next thing,&#8221; Sutskever said. &#8220;Scaling the right thing matters more now than ever.&#8221; (Reuters)</p></blockquote><p>The new dominant narrative seems to be that model scaling is dead, and &#8220;inference scaling&#8221;, also known as &#8220;test-time compute scaling&#8221; is the way forward for improving AI capabilities. The idea is to spend more and more computation when using models to perform a task, such as by having them &#8220;think&#8221; before responding.</p><p>This has left AI observers confused about whether or not progress in AI capabilities is slowing down. In this essay, we look at the evidence on this question, and make four main points:</p><ol><li><p>Declaring the death of model scaling is premature.</p></li><li><p>Regardless of whether model scaling will continue, industry leaders&#8217; flip flopping on this issue shows the folly of trusting their forecasts. They are not significantly better informed than the rest of us, and their narratives are heavily influenced by their vested interests.</p></li><li><p>Inference scaling is real, and there is a lot of low-hanging fruit, which could lead to rapid capability increases in the short term. But in general, capability improvements from inference scaling will likely be both unpredictable and unevenly distributed among domains.</p></li><li><p>The connection between capability improvements and AI&#8217;s social or economic impacts is extremely weak. The bottlenecks for impact are the pace of product development and the rate of adoption, not AI capabilities.</p></li></ol><h2><strong>Is model scaling dead?</strong></h2><p>There is very little new information that has led to the sudden vibe shift. We&#8217;ve long been <a href="https://www.aisnakeoil.com/p/ai-scaling-myths">saying</a> on this newsletter that there are important headwinds to model scaling. Just as we cautioned back then about scaling hype, we must now caution against excessive pessimism about model scaling.</p><p>&#8220;Scaling as usual&#8221; ended with GPT-4 class models, because these models are trained on most of the readily available data sources. We already knew that new ideas would be needed to keep model scaling going. So unless we have evidence that many such ideas have been tried and failed, we can&#8217;t conclude that there isn&#8217;t more mileage to model scaling.</p><p>As just one example, it is possible that including YouTube videos &#8212; the actual videos, not transcribed text &#8212; in the training mix for multimodal models will unlock new capabilities. Or it might not help; we just won&#8217;t know until someone tries it, and we don&#8217;t know if it has been tried or not. Note that it would probably have to be Google, because the company is unlikely to license YouTube training data to competitors.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a></p><p>If things are still so uncertain regarding model scaling, why did the narrative flip? Well, it&#8217;s been over two years since GPT-4 finished training, so the idea that next-gen models are simply taking a bit longer than expected was becoming less and less credible. And once one company admits that there are problems, it becomes a lot easier for others to do so. Once there is a leak in the dam, it quickly bursts. Finally, now that OpenAI&#8217;s reasoning model o1 is out, it has given companies an out when admitting that they have run into problems with model scaling, because they can save face by claiming that they will simply switch to inference scaling.</p><p>To be clear, there is no reason to doubt the reports saying that many AI labs have conducted larger training runs and yet not released the resulting models. But it is less clear what to conclude from it. Some possible reasons why bigger models haven&#8217;t been released include:</p><ul><li><p>Technical difficulties, such as convergence failures or complications in achieving fault tolerance in multi-datacenter training runs.</p></li><li><p>The model was not much better than GPT-4 class models, and so would be too underwhelming to release.</p></li><li><p>The model was not much better than GPT-4 class models, and so the developer has been spending a long time trying to eke out better performance through fine tuning.</p></li></ul><p>To summarize, it&#8217;s possible that model scaling has indeed reached its limit, but it&#8217;s also possible that these hiccups are temporary and eventually one of the companies will find ways to overcome them, such as by fixing any technical difficulties and/or finding new data sources.</p><h2><strong>Let&#8217;s stop deferring to insiders</strong></h2><p>Not only is it strange that the new narrative emerged so quickly, it&#8217;s also interesting that the old one persisted for so long, despite the potential limitations of model scaling being obvious. The main reason for its persistence is the assurances of industry leaders that scaling would continue for a few more years.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a> In general, journalists (and most others) tend to <a href="https://www.saysmaybe.com/latest-work/selective-perspectives-nyt">defer to industry insiders</a> over outsiders. But is this deference justified?</p><p>Industry leaders don&#8217;t have a good track record of predicting AI developments. A good example is the overoptimism about self-driving cars for most of the last decade. (Autonomous driving is finally real, though Level 5 &#8212; full automation &#8212; doesn&#8217;t exist yet.)  As an aside, in order to better understand the track record of insider predictions, it would be interesting to conduct a systematic analysis of all predictions about AI made in the last 10 years by prominent industry insiders.</p><p>There are some reasons why we might want to give more weight to insiders&#8217; claims, but also important reasons to give <em>less</em> weight to them. Let&#8217;s analyze these one by one. It is true that industry insiders have proprietary information (such as the performance of as-yet-unreleased models) that might make their claims about the future more accurate. But given how many AI companies are close to the state of the art, including some that <a href="https://arxiv.org/abs/2407.21783">openly release</a> model weights and share <a href="https://arxiv.org/abs/2402.00838">scientific insights, datasets, and other artifacts</a>, we&#8217;re talking about an advantage of at most a few months, which is minor in the context of, say, 3-year forecasts.</p><p>Besides, we tend to overestimate how much additional information companies have on the inside &#8212; whether in terms of capability or (especially) in terms of safety. Insiders warned for a long time that <a href="https://www.vox.com/future-perfect/353933/openai-open-letter-safety-whistleblowers-right-to-warn">&#8220;if only you know what we know...&#8221;</a> but when whistleblowers finally came forward, it turns out that they were mostly relying on the same kind of speculation that everyone else does.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-3" href="#footnote-3" target="_self">3</a></p><p>Another potential reason to give more weight to insiders is their technical expertise. We don&#8217;t think this is a strong reason: there is just as much AI expertise in academia as in industry. More importantly, deep technical expertise isn&#8217;t that important to support the kind of crude trend extrapolation that goes into AI forecasts. Nor is technical expertise enough &#8212; <a href="https://www.aisnakeoil.com/i/146043714/trend-extrapolation-is-baseless-speculation">business and social factors</a> play at least as big a role in determining the course of AI. In the case of self-driving cars, one such factor is the extent to which societies tolerate public roads being used for experimentation. In the case of large AI models, we&#8217;ve argued before that the most important factor is whether scaling will make <a href="https://www.aisnakeoil.com/p/ai-scaling-myths">business sense</a>, not whether it is technically feasible. So not only do techies not have much of an advantage, their tendency to overemphasize the technical dimensions tends to result in overconfident predictions.</p><p>In short, the reasons why one might give more weight to insiders&#8217; views aren&#8217;t very important. On the other hand, there&#8217;s a huge and obvious reason why we should probably give less weight to their views, which is that they have an incentive to say things that are in their commercial interests, and have a track record of doing so.</p><p>As an example, Sutskever had an <a href="https://garrisonlovely.substack.com/p/is-deep-learning-actually-hitting">incentive</a> to talk up scaling when he was at OpenAI and the company needed to raise money. But now that he heads the startup Safe Superintelligence, he needs to convince investors that it can compete with OpenAI, Anthropic, Google, and others, despite having access to much less capital. Perhaps that is why he is now talking about <a href="https://youtu.be/1yvBqasHLZs?feature=shared&amp;t=523">running out of data for pre-training</a>, as if it were some epiphany and not an endlessly repeated point.</p><p>To reiterate, we don&#8217;t know if model scaling has ended or not. But the industry&#8217;s sudden about-face has been so brazen that it should leave no doubt that insiders don&#8217;t have any kind of crystal ball and are making similar guesses as everyone else, and are further biased by being in a bubble and readily consuming the hype they sell to the world.</p><p>In light of this, our suggestion &#8212; to everyone, but especially journalists, policymakers, and the AI community &#8212; is to end the deference to insiders' views when they predict the future of technology, especially its societal impacts. This will take effort, as there is a pervasive unconscious bias in the U.S., in the form of a &#8220;distinctly American disease that seems to equate extreme wealth, and the power that comes with it, with virtue and intelligence.&#8221; (from Bryan Gardiner&#8217;s <a href="https://www.technologyreview.com/2024/12/13/1108459/book-review-silicon-valley-democracy-techlash-rob-lalka-venture-alchemists-marietje-schaake-tech-coup/amp/">review</a> of Marietje Schake&#8217;s <a href="https://press.princeton.edu/books/hardcover/9780691241173/the-tech-coup?srsltid=AfmBOoo6e3ReisMn_Dsr5lNZiSe16xzdsJhYgcsU_70n946FL8zxXOlv">The Tech Coup</a>.)</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.normaltech.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption"><em>AI Snake Oil</em> debunks AI hype and publishes evidence-based analysis of new developments.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><h2><strong>Will progress in capabilities continue through inference scaling?</strong></h2><p>Of course, model scaling is not the only way to improve AI capabilities. <a href="https://arxiv.org/abs/2407.21787">Inference scaling</a> is an area with a lot of recent progress. For example, <a href="https://openai.com/o1/">OpenAI&#8217;s o1</a> and the open-weights competitor <a href="https://api-docs.deepseek.com/news/news1120">DeepSeek R1</a> are <a href="https://arxiv.org/pdf/2409.13373v1#page=1.31">reasoning models</a>: they have been fine tuned to &#8220;reason&#8221; before providing an answer. Other methods leave the <a href="https://arxiv.org/pdf/2411.17501">model itself unchanged</a> but employ tricks like generating many solutions and ranking them by quality.</p><p>There are two main open questions about inference scaling that will determine how significant of a trend it will be.</p><ol><li><p>What class of problems does it work well on?</p></li><li><p>For problems where it does work well, how much of an improvement is possible by doing more computation during inference?</p></li></ol><p>The per-token output cost of language models has been rapidly decreasing due to both hardware and algorithmic improvements, so if inference scaling yields improvements over many orders of magnitude &#8212; for example, generating a million tokens on a given task yields significantly better performance than generating a hundred thousand tokens &#8212; that would be a big deal.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-4" href="#footnote-4" target="_self">4</a></p><p>The straightforward, intuitive answer to the first question is that inference scaling is useful for problems that have clear correct answers, such as coding or mathematical problem solving. In such tasks, at least one of two related things tend to be true. First, symbolic reasoning can improve accuracy. This is something LLMs are bad at due to their statistical nature, but can overcome by using output tokens for reasoning, much like a person using pen and paper to work through a math problem. Second, it is easier to <a href="https://arxiv.org/pdf/2402.01817">verify</a> correct solutions than to generate them (sometimes aided by external verifiers, such as unit tests for coding or <a href="https://blog.eleuther.ai/llemma/">proof checkers</a> for mathematical theorem proving).</p><p>In contrast, for tasks such as writing or language translation, it is hard to see how inference scaling can make a big difference, especially if the limitations are due to the training data. For example, if a model works poorly in translating to a low-resource language because it isn&#8217;t aware of idiomatic phrases in that language, the model can&#8217;t reason its way out of this.</p><p>The early evidence we have so far, while spotty, is consistent with this intuition. Focusing on OpenAI o1, it <a href="https://openai.com/index/learning-to-reason-with-llms/">improves</a> compared to state-of-the-art language models such as GPT-4o on coding, math, <a href="https://arxiv.org/pdf/2410.21939#page=6.42">cybersecurity</a>, <a href="https://arxiv.org/pdf/2409.13373">planning in toy worlds</a>, and various <a href="https://openai.com/index/learning-to-reason-with-llms/">exams</a>. Improvements in exam performance seem to strongly correlate with the importance of reasoning for answering questions, as opposed to knowledge or creativity: big improvements for math, physics and LSATs, smaller improvements for subjects like biology and econometrics, and negligible improvement for English.</p><p>Tasks where o1 doesn&#8217;t seem to lead to an improvement include <a href="https://openai.com/index/learning-to-reason-with-llms/">writing</a>, certain <a href="https://cybench.github.io/">cybersecurity</a> tasks (which we explain below), <a href="https://arxiv.org/pdf/2212.08061">avoiding toxicity</a>, and an interesting set of <a href="https://arxiv.org/pdf/2410.21333">tasks</a> at which thinking is known to make humans <em>worse</em>.</p><p>We have created a <a href="https://benediktstroebl.github.io/reasoning-model-evals/">webpage</a> compiling the available evidence on how reasoning models compare against language models. We plan to keep it updated for the time being, though we expect that the torrent of findings will soon become difficult to keep up with.</p><p>Now let&#8217;s consider the second question: how large of an improvement can we get through inference scaling, assuming we had an infinite inference compute budget.</p><p>OpenAI&#8217;s flagship example to show off o1&#8217;s capabilities was AIME, a math benchmark. Their graph leaves this question tantalizingly open. Is the performance about to saturate, or can it be pushed close to 100%? Also note that the graph conveniently leaves out x-axis labels.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!0fDW!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc348909d-140e-4e20-94d7-92a04e008bb2_1450x1270.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!0fDW!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc348909d-140e-4e20-94d7-92a04e008bb2_1450x1270.png 424w, https://substackcdn.com/image/fetch/$s_!0fDW!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc348909d-140e-4e20-94d7-92a04e008bb2_1450x1270.png 848w, https://substackcdn.com/image/fetch/$s_!0fDW!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc348909d-140e-4e20-94d7-92a04e008bb2_1450x1270.png 1272w, https://substackcdn.com/image/fetch/$s_!0fDW!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc348909d-140e-4e20-94d7-92a04e008bb2_1450x1270.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!0fDW!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc348909d-140e-4e20-94d7-92a04e008bb2_1450x1270.png" width="1450" height="1270" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/c348909d-140e-4e20-94d7-92a04e008bb2_1450x1270.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1270,&quot;width&quot;:1450,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!0fDW!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc348909d-140e-4e20-94d7-92a04e008bb2_1450x1270.png 424w, https://substackcdn.com/image/fetch/$s_!0fDW!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc348909d-140e-4e20-94d7-92a04e008bb2_1450x1270.png 848w, https://substackcdn.com/image/fetch/$s_!0fDW!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc348909d-140e-4e20-94d7-92a04e008bb2_1450x1270.png 1272w, https://substackcdn.com/image/fetch/$s_!0fDW!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc348909d-140e-4e20-94d7-92a04e008bb2_1450x1270.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><a href="https://openai.com/index/learning-to-reason-with-llms/">Source: OpenAI</a></figcaption></figure></div><p>An <a href="https://github.com/hughbzhang/o1_inference_scaling_laws">attempt</a> by external researchers to reconstruct this graph shows that (1) the cutoff for the x-axis is probably around 2,000 tokens, and (2) when o1 is asked to think longer than this, it doesn&#8217;t do so. So the question remains unanswered, and we need to wait for experiments using open-source models to get more clarity. It is great to see that there are vigorous efforts to <a href="https://arxiv.org/abs/2410.18982">publicly reproduce</a> the techniques behind o1.</p><p>In a recent paper called<em> <a href="https://arxiv.org/abs/2411.17501">Inference Scaling fLaws</a> </em>(the title is a pun on inference scaling laws), we look at a different approach to inference scaling &#8212; repeatedly generating solutions until one of them is judged as correct by an external verifier. While this approach has been associated with hopes of usefully increasing scaling by many orders of magnitude (including by us <a href="https://arxiv.org/pdf/2407.01502#page=2.79">in our own past work</a>), we find that this is extremely sensitive to the quality of the verifier. If the verifier is slightly imperfect, in many realistic settings of a coding task, performance maxes out and actually starts to <em>decrease</em> after about 10 attempts.</p><p>Generally speaking, the evidence for inference scaling &#8220;laws&#8221; is not convincing, and it remains to be seen if there are real-world problems where generating (say) millions of tokens at inference time will actually help.</p><h2><strong>Is inference scaling the next frontier?</strong></h2><p>There is a lot of low-hanging fruit for inference scaling, and progress in the short term is likely to be rapid. Notably, one current limitation of reasoning models is that they don&#8217;t work well in agentic systems. We have observed this in our own benchmark <a href="https://agent-evals-core-leaderboard.hf.space/">CORE-Bench</a> that asks agents to reproduce the code provided with research papers &#8212; the best performing agent scores 38% with Claude 3.5 Sonnet compared to only 24% with o1-mini.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-5" href="#footnote-5" target="_self">5</a> This also explains why reasoning models led to an improvement in one cybersecurity eval but not another &#8212; one of them involved agents.</p><p>We think there are two reasons why agents don&#8217;t seem to benefit from reasoning models. Such models require different prompting styles than regular models, and current agentic systems are optimized for prompting regular models. Second, as far as we know, reasoning models so far have <em>not</em> been trained using reinforcement learning in a setting where they receive feedback from the environment &#8212; be it code execution, shell interaction, or web search. In other words, their tool use ability is no better than the underlying model before learning to reason.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-6" href="#footnote-6" target="_self">6</a></p><p>These seem like relatively straightforward problems. Solving them might enable significant new AI agent capabilities &#8212; for example, generating complex, fully functional apps from a prompt. (There are already tools that try to do this, but they don&#8217;t work well.)</p><p>But what about the long run? Will inference scaling lead to the same kind of progress we&#8217;ve seen with model scaling over the last 7 years? Model scaling was so exciting because you &#8220;merely&#8221; needed to make data, model size, and compute bigger; no algorithmic breakthroughs were needed.</p><p>That&#8217;s not true (so far) with inference scaling &#8212; there&#8217;s a long list of inference scaling techniques, and what works or doesn&#8217;t work is problem-dependent, and even collectively, they only work in a circumscribed set of domains. AI developers are trying to overcome this limitation. For example, OpenAI&#8217;s reinforcement finetuning service is thought to be a way for the company to <a href="https://openai.com/form/rft-research-program/">collect customer data</a> from many different domains for fine-tuning a future model.</p><p>About a decade ago, reinforcement learning (RL) led to breakthroughs in many games like Atari. There was a lot of hype, and many AI researchers hoped we could RL our way to AGI. In fact, it was the high expectations around RL that led to the birth of explicitly AGI-focused labs, notably OpenAI. But those techniques didn&#8217;t generalize beyond narrow domains like games. Now there is similar hype about RL again. It is obviously a very powerful technique, but so far we&#8217;re seeing limitations similar to the ones that led to the dissipation of the previous wave of hype.</p><p>It is impossible to predict whether progress in AI capabilities will slow down. In fact, forget prediction &#8212; reasonable people can have very different opinions on whether AI progress has already slowed down, because they can interpret the evidence very differently. That&#8217;s because &#8220;capability&#8221; is a <a href="https://www.cs.princeton.edu/~arvindn/talks/evaluating_llms_minefield/#/12">construct</a> that&#8217;s highly sensitive to how you measure it.</p><p>What we can say with more confidence is that the <em>nature</em> of progress in capabilities will be different with inference scaling than with model scaling. In the last few years, newer models predictably brought capability improvements each year across a vast swath of domains. There was a feeling of pessimism among many AI researchers outside the big labs that there was little to do except sit around and wait for the next state-of-the-art LLM to be released.</p><p>With inference scaling, capability improvements will likely be uneven and less predictable, driven more by algorithmic advances than investment in hardware infrastructure. Many ideas that were discarded during the reign of LLMs, such as those from the old planning literature, are now back in the mix, and the scene seems intellectually more vibrant than in the last few years.</p><h2><strong>Product development lags capability increase</strong></h2><p>The furious debate about whether there is a capability slowdown is ironic, because the link between capability increases and the real-world usefulness of AI is extremely weak. The development of AI-based <a href="https://www.ben-evans.com/benedictevans/2024/4/19/looking-for-ai-use-cases">applications</a> lags far behind the increase of AI capabilities, so even existing AI capabilities remain greatly underutilized. One reason is the <a href="https://www.aisnakeoil.com/i/147899150/reliability">capability-reliability gap</a> &#8212; even when a certain capability exists, it may not work reliably enough that you can take the human out of the loop and actually automate the task (imagine a food delivery app that only works 80% of the time). And the methods for improving reliability are often application-dependent and distinct from methods for improving capability. That said, reasoning models also seem to exhibit <a href="https://youtu.be/iBfQTnA2n2s?si=a-760cPz5ZghJc7w&amp;t=161">reliability improvements</a>, which is exciting.</p><p>Here are a couple of analogies that help illustrate why it might take a decade or more to build products that fully take advantage of even current AI capabilities. The technology behind the internet and the web mostly solidified in the <a href="https://x.com/jefrankle/status/1867614458152943990">mid-90s</a>. But it took 1-2 more decades to realize the potential of web apps. Or consider this thought-provoking <a href="https://willwhitney.com/computing-inside-ai.html">essay</a> that argues that we need to build GUIs for large language models, which will allow interacting with them with far higher bandwidth than through text. From this perspective, the current state of AI-based products is analogous to PCs before the GUI.</p><p>The lag in product development is compounded by the fact that AI companies have not paid nearly enough attention to <a href="https://www.aisnakeoil.com/p/ai-companies-are-pivoting-from-creating">product aspects</a>, believing that the general-purpose nature of AI somehow grants an exemption from the hard problems of software engineering. Fortunately, this has started to <a href="https://www.aisnakeoil.com/p/ai-companies-are-pivoting-from-creating">change</a> recently.</p><p>Now that they are focusing on products, AI companies as well as their users are re-discovering that software development, especially the user experience side of it, is hard, and requires a broader set of skills than AI model development. A great example is the fact that there are now two different ways to run Python code with ChatGPT (which is one of the most important capabilities for power users) and there is an intricate set of undocumented rules to remember regarding the capabilities and limitations of each of them. <a href="https://simonwillison.net/2024/Dec/10/chatgpt-canvas/">Simon Willison</a> says:</p><blockquote><p>Do you find this all hopelessly confusing? I don&#8217;t blame you. I&#8217;m a professional web developer and a Python engineer of 20+ years and I can just about understand and internalize the above set of rules.</p></blockquote><p>Still, this is a big improvement over a week ago, when these models had powerful coding capabilities yet did not come with the ability to run code that could use the internet! And even now, o1 can neither access the internet nor run code. From the perspective of AI impacts, what matters far more than capability improvement at this point is actually building products that let people do useful things with existing capabilities.</p><p>Finally, while product development lags behind capability, the adoption of AI-based products <a href="https://x.com/random_walker/status/1850871740550758766">further lags</a> far behind product development, for various behavioral, organizational, and societal reasons. Those interested in AI&#8217;s impacts (whether positive or negative) should pay much more attention to these downstream aspects than current or predicted capabilities.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.normaltech.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption"><em>AI Snake Oil</em> debunks AI hype and publishes evidence-based analysis of new developments.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><h2><strong>Concluding thoughts</strong></h2><p>Maybe model scaling is over; maybe not. But it won&#8217;t continue forever, and the end of model scaling brings a long list of positives: AI progress once again depends on new ideas and not just compute; big companies, startups, and academic researchers can all compete on a relatively even playing field; regulation based on <a href="https://www.aisnakeoil.com/p/ai-safety-is-not-a-model-property">arbitrary</a> training compute thresholds becomes even <a href="https://arxiv.org/abs/2407.05694">harder to defend</a>; and there is a clear recognition that models themselves are just a technology, not a product.</p><p>As for the future of AI, it is clear that tech insiders are trying to figure it out just like the rest of us, and it is past time to stop trusting their overconfident, self-serving, shifting, and conveniently vague predictions. And when we move beyond technical predictions to claims about AI&#8217;s impact on the world, there&#8217;s even less reason to trust industry leaders.</p><p></p><p><strong>Acknowledgment. </strong>We are grateful to Zachary S. Siegel for feedback on a draft.</p><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>While OpenAI is known to have <a href="https://www.nytimes.com/2024/04/06/technology/tech-giants-harvest-data-artificial-intelligence.html">crawled</a> YouTube in the past, that was a small sliver of YouTube; it won&#8217;t be possible to crawl all of YouTube without Google noticing.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p>A nice <a href="https://epoch.ai/blog/can-ai-scaling-continue-through-2030">analysis</a> by Epoch AI showed that scaling could continue until 2030. But this was published too recently (August 2024) to have been the anchor for the scaling narrative.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-3" href="#footnote-anchor-3" class="footnote-number" contenteditable="false" target="_self">3</a><div class="footnote-content"><p>We are referring to <em>substantive</em> knowledge about the safety of AI models and systems; whistleblowers did bring forth new knowledge about safety-related <em>processes</em> at OpenAI.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-4" href="#footnote-anchor-4" class="footnote-number" contenteditable="false" target="_self">4</a><div class="footnote-content"><p>That said, we can&#8217;t take future cost decreases for granted; we are also running into fundamental limits of inference cost-saving techniques like quantization.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-5" href="#footnote-anchor-5" class="footnote-number" contenteditable="false" target="_self">5</a><div class="footnote-content"><p>We set a cost limit for $4 for all models. On a small sample, with a $10 cost limit, o1-preview performed very poorly (10% accuracy). Given cost constraints, we did not evaluate the model with a higher cost limit on the entire data.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-6" href="#footnote-anchor-6" class="footnote-number" contenteditable="false" target="_self">6</a><div class="footnote-content"><p>o1 doesn&#8217;t even have access to tools during inference in the ChatGPT interface! Gemini Flash 2.0 does, but it is not clear if this is a model that has been fine tuned for reasoning, let alone fine tuned for tool use.</p><p></p></div></div>]]></content:encoded></item><item><title><![CDATA[We Looked at 78 Election Deepfakes. Political Misinformation is not an AI Problem.]]></title><description><![CDATA[Technology Isn&#8217;t the Problem&#8212;or the Solution.]]></description><link>https://www.normaltech.ai/p/we-looked-at-78-election-deepfakes</link><guid isPermaLink="false">https://www.normaltech.ai/p/we-looked-at-78-election-deepfakes</guid><dc:creator><![CDATA[Sayash Kapoor]]></dc:creator><pubDate>Fri, 13 Dec 2024 20:51:29 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!ezjw!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F29800e81-2362-430e-ada1-591a2d3a5228_910x1000.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>AI-generated misinformation was one of the top concerns during the 2024 U.S. presidential election. In January 2024, the World Economic Forum <a href="https://www.weforum.org/stories/2024/01/ai-disinformation-global-risks/">claimed</a> that &#8220;misinformation and disinformation is the most severe short-term risk the world faces&#8221; and that &#8220;AI is amplifying manipulated and distorted information that could destabilize societies.&#8221; News headlines about elections in 2024 tell a similar story:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!ezjw!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F29800e81-2362-430e-ada1-591a2d3a5228_910x1000.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!ezjw!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F29800e81-2362-430e-ada1-591a2d3a5228_910x1000.jpeg 424w, https://substackcdn.com/image/fetch/$s_!ezjw!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F29800e81-2362-430e-ada1-591a2d3a5228_910x1000.jpeg 848w, https://substackcdn.com/image/fetch/$s_!ezjw!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F29800e81-2362-430e-ada1-591a2d3a5228_910x1000.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!ezjw!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F29800e81-2362-430e-ada1-591a2d3a5228_910x1000.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!ezjw!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F29800e81-2362-430e-ada1-591a2d3a5228_910x1000.jpeg" width="462" height="507.6923076923077" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/29800e81-2362-430e-ada1-591a2d3a5228_910x1000.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1000,&quot;width&quot;:910,&quot;resizeWidth&quot;:462,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!ezjw!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F29800e81-2362-430e-ada1-591a2d3a5228_910x1000.jpeg 424w, https://substackcdn.com/image/fetch/$s_!ezjw!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F29800e81-2362-430e-ada1-591a2d3a5228_910x1000.jpeg 848w, https://substackcdn.com/image/fetch/$s_!ezjw!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F29800e81-2362-430e-ada1-591a2d3a5228_910x1000.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!ezjw!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F29800e81-2362-430e-ada1-591a2d3a5228_910x1000.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>In contrast, in our past writing, we predicted that AI would not lead to a misinformation apocalypse. When Meta released its open-weight large language model (called LLaMA), we <a href="https://www.aisnakeoil.com/p/the-llama-is-out-of-the-bag-should">argued</a> that it would not lead to a tidal wave of misinformation. And in a follow-up essay, we <a href="https://knightcolumbia.org/content/how-to-prepare-for-the-deluge-of-generative-ai-on-social-media">pointed out</a> that the distribution<em> </em>of misinformation is the key bottleneck for influence operations, and while generative AI reduces the cost of creating misinformation, it does not reduce the cost of distributing it. A few other researchers have made <a href="https://misinforeview.hks.harvard.edu/article/misinformation-reloaded-fears-about-the-impact-of-generative-ai-on-misinformation-are-overblown/">similar arguments</a>.</p><p>Which of these two perspectives better fits the facts?</p><p>Fortunately, we have the evidence of AI use in elections that took place around the globe in 2024 to help answer this question. Many news outlets and research projects have compiled known instances of AI-generated text and media and their impact. Instead of speculating about AI&#8217;s potential, we can look at its real-world impact to date.</p><p>We analyzed every instance of AI use in elections collected by the <a href="https://www.wired.com/story/generative-ai-global-elections/">WIRED AI Elections Project</a>, which tracked known uses of AI for creating political content during elections taking place in 2024 worldwide. In each case, we identified what AI was used for and estimated the cost of creating similar content without AI.</p><p>We find that (1) half of AI use isn't deceptive, (2) deceptive content produced using AI is nevertheless cheap to replicate <em>without </em>AI, and (3) focusing on the demand for misinformation rather than the supply is a much more effective way to diagnose problems and identify interventions.</p><p>To be clear, AI-generated synthetic content poses many real dangers: the creation of <a href="https://www.pbs.org/newshour/show/nonconsensual-sexual-images-posted-online-made-worse-by-deepfakes-and-ai-technology">non-consensual images of people</a> and <a href="https://www.washingtonpost.com/technology/2024/04/22/ai-csam-ncmec-cybertipline-stanford-report/">child sexual abuse material</a> and the enabling of the <a href="https://www.brookings.edu/articles/misunderstood-mechanics-how-ai-tiktok-and-the-liars-dividend-might-affect-the-2024-elections/">liar&#8217;s dividend</a>, which allows those in power to brush away real but embarrassing or controversial media content about them as AI-generated. These are all important challenges. This essay is focused on a different problem: political misinformation.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a></p><p>Improving the information environment is a difficult and ongoing challenge. It&#8217;s understandable why people might think AI is making the problem worse: AI does make it possible to fabricate false content. But that has not fundamentally changed the landscape of political misinformation.</p><p>Paradoxically, the alarm about AI might be comforting because it positions concerns about the information environment as a discrete problem with a discrete solution. But fixes to the information environment depend on structural and institutional changes rather than on curbing AI-generated content.</p><h3><strong>Half of the Deepfakes in 2024 Elections weren&#8217;t Deceptive</strong></h3><p>We analyzed all 78 instances of AI use in the WIRED AI Elections Project (<a href="https://www.cs.princeton.edu/~sayashk/political-misinformation/WIRED-data.html">source</a> for our analysis).<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a> We categorized each instance based on whether there was deceptive intent. For example, if AI was used to generate false media depicting a political candidate <a href="https://www.npr.org/2024/01/22/1226129926/nh-primary-biden-ai-robocall">saying something they didn't</a>, we classified it as deceptive. On the other hand, if a chatbot gave an <a href="https://www.proofnews.org/ai-models-falter-answering-election-questions-in-spanish/">incorrect response</a> to a genuine user query, a deepfake was created for <a href="https://restofworld.org/2024/exporter-india-deepfake-trolls/">parody or satire</a>, or a candidate transparently used AI to improve their campaigning materials (such as by <a href="https://www.hindustantimes.com/technology/pm-modi-uses-ai-tool-bhashini-at-kashi-tamil-sangamam-in-varanasi-what-is-it-101702837254702.html">translating</a> a speech into a language they don't speak), we classify it as non-deceptive.</p><p>To our surprise, there was no deceptive intent in 39 of the 78 cases in the database.</p><p>The most common non-deceptive use of AI was for campaigning. When candidates or supporters used AI for campaigning, in most cases (19 out of 22), the apparent intent was to improve campaigning materials rather than mislead voters with false information.</p><p>We even found examples of deepfakes that we think helped improve the information environment. In Venezuela, journalists used AI avatars to <a href="https://globalvoices.org/2024/08/19/venezuelans-use-ai-avatars-and-instagram-live-to-fight-back-maduros-repression/">avoid</a> government retribution when covering news adversarial to the government. In the U.S., a local news organization from Arizona, Arizona Agenda, used deepfakes to <a href="https://www.washingtonpost.com/politics/2024/03/24/kari-lake-deepfake/">educate</a> viewers about how easy it is to manipulate videos. In California, a candidate with laryngitis lost his voice, so he transparently used AI <a href="https://www.youtube.com/watch?v=LSlXWjMKL5E">voice cloning</a> to read out typed messages in his voice during meet-and-greets with voters.</p><p>Reasonable people can disagree on whether using AI in campaigning materials is legitimate or what the appropriate guardrails need to be. But using AI for campaign materials in non-deceptive ways (for example, when AI is used as a tool to improve voter outreach) is much less problematic than deploying AI-generated fake news to sway voters.</p><p>Of course, not all non-deceptive AI-generated political content is benign.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-3" href="#footnote-3" target="_self">3</a> Chatbots often <a href="https://www.proofnews.org/seeking-election-information-dont-trust-ai/">incorrectly</a> answer election-related questions. Rather than deceptive intent, this results from the limitations of chatbots, such as hallucinations and lack of factuality. Unfortunately, these limitations are not made clear to users, leading to an overreliance on flawed large language models (LLMs).<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-4" href="#footnote-4" target="_self">4</a></p><h3><strong>Making Deceptive Political Misinformation Does Not Require AI</strong></h3><p>For each of the 39 examples of deceptive intent, where AI use was intended to make viewers believe outright false information, we estimated the cost of creating similar content <em>without</em> AI&#8212;for example, by hiring Photoshop experts, video editors, or voice actors. In each case, the cost of creating similar content without AI was modest&#8212;no more than a few hundred dollars. (We even found that a video involving a <a href="https://www.latimes.com/california/story/2024-09-24/fake-russian-news-site-falsely-claimed-kamala-harris-was-in-hit-and-run-accident">hired stage actor</a> was incorrectly marked as being AI-generated in WIRED&#8217;s election database.)</p><p>In fact, it has long been possible to create media with outright false information without using AI or other fancy tools. One video used <a href="https://www.latimes.com/california/story/2024-09-24/fake-russian-news-site-falsely-claimed-kamala-harris-was-in-hit-and-run-accident">stage actors</a> to falsely claim that U.S. Vice President and Democratic presidential candidate Kamala Harris was involved in a hit-and-run incident. Another <a href="https://www.usatoday.com/story/news/factcheck/2024/09/30/kamala-harris-slur-speech-helene-fact-check/75447979007/">slowed down</a> the vice president's speech to make it sound like she was slurring her words. An edited video of Indian opposition candidate Rahul Gandhi showed him <a href="https://archive.is/SxXeX#selection-1109.0-1126.0">saying</a> that the incumbent Narendra Modi would win the election. In the original video, Gandhi said his opponent would <em>not </em>win the election, but it was <a href="https://www.boomlive.in/fact-check/factcheck-rahul-gandhi-modi-will-become-pm-again-doctored-video-fake-news-25264">edited</a> using jump cuts to take out the word &#8220;not.&#8221; Such media content has been called &#8220;<a href="https://datasociety.net/wp-content/uploads/2019/09/DataSociety_Deepfakes_Cheap_Fakes.pdf">cheap fakes</a>&#8221; (as opposed to AI-generated &#8220;deepfakes&#8221;).</p><p>There were many instances of cheap fakes used in the 2024 U.S. election. The News Literacy Project <a href="https://misinfodashboard.newslit.org/">documented</a> known misinformation about the election and <a href="https://newslit.org/newsroom/press-release/the-news-literacy-project-experts-watch-out-for-ai-generated-fakes-and-disinformation-about-voting-ahead-of-election-day/">found</a> that cheap fakes were used seven times more often than AI-generated content. Similarly, in other countries, cheap fakes were quite prevalent. An <a href="https://www.boomlive.in/">India-based fact checker</a> reviewed an order of magnitude more cheap fakes and traditionally edited media compared to <a href="https://www.boomlive.in/deepfake-tracker">deepfakes</a>. In Bangladesh, cheap fakes were over <a href="https://www.context.news/ai/opinion/cheap-fakes-are-a-blind-spot-for-platforms-in-the-global-south">20 times more prevalent</a> than deepfakes.</p><p>Let&#8217;s consider two examples to analyze how cheap fakes could have led to substantially similar effects as the deepfakes that got a lot of media attention: Donald Trump&#8217;s use of Taylor Swift deepfakes to campaign and a voice-cloned robocall that imitated U.S. President Joe Biden in the New Hampshire primary asking voters not to vote.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!64fX!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faab44993-d832-4881-9532-f78e94883231_1222x1466.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!64fX!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faab44993-d832-4881-9532-f78e94883231_1222x1466.jpeg 424w, https://substackcdn.com/image/fetch/$s_!64fX!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faab44993-d832-4881-9532-f78e94883231_1222x1466.jpeg 848w, https://substackcdn.com/image/fetch/$s_!64fX!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faab44993-d832-4881-9532-f78e94883231_1222x1466.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!64fX!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faab44993-d832-4881-9532-f78e94883231_1222x1466.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!64fX!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faab44993-d832-4881-9532-f78e94883231_1222x1466.jpeg" width="444" height="532.6546644844517" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/aab44993-d832-4881-9532-f78e94883231_1222x1466.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1466,&quot;width&quot;:1222,&quot;resizeWidth&quot;:444,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!64fX!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faab44993-d832-4881-9532-f78e94883231_1222x1466.jpeg 424w, https://substackcdn.com/image/fetch/$s_!64fX!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faab44993-d832-4881-9532-f78e94883231_1222x1466.jpeg 848w, https://substackcdn.com/image/fetch/$s_!64fX!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faab44993-d832-4881-9532-f78e94883231_1222x1466.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!64fX!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faab44993-d832-4881-9532-f78e94883231_1222x1466.jpeg 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em>A Truth Social post shared by Donald Trump with images of Taylor Swift fans wearing &#8220;Swifties for Trump&#8221; t-shirts. Top left: A post with many AI-generated images of women wearing &#8220;Swifties for Trump&#8221; t-shirts, with a &#8220;satire&#8221; label. Top right: A <a href="https://www.cbsnews.com/news/trump-shares-fake-swifties-for-trump-images/">real image</a> of Trump supporter Jenna Piwowarczyk wearing a &#8220;Swifties for Trump&#8221; t-shirt. Bottom left: A fabricated image of Taylor Swift in front of the American flag with the caption, &#8220;Taylor wants you to vote for Donald Trump.&#8221; It is unclear if the image was created using AI or other editing software. Bottom right: A Twitter post with two images: one AI-generated, the other real, of women wearing &#8220;Swifties for Trump&#8221; t-shirts.</em></figcaption></figure></div><p>Trump&#8217;s use of Swift deepfakes implied that Taylor Swift had endorsed him and that Swift fans were attending his rallies en masse. In the wake of the post, many <a href="https://www.forbes.com/sites/steveandriole/2024/08/21/taylor-swift-donald-trump-and-ai/">media</a> <a href="https://www.lowyinstitute.org/the-interpreter/what-taylor-swift-taught-world-risks-ai-generated-images-elections">outlets</a> <a href="https://www.latimes.com/entertainment-arts/business/story/2024-08-21/deep-fakes-social-media-taylor-swift-donald-trump">blamed</a> AI for the spread of misinformation.</p><p>But recreating similar images without AI is easy. Images depicting Swift&#8217;s support could be created by photoshopping text endorsing Trump onto any of her existing images. Likewise, getting images of Trump supporters wearing &#8220;Swifties for Trump&#8221; t-shirts could be achieved by distributing free t-shirts at a rally&#8212;or even selectively reaching out to Swift fans at Trump rallies. In fact, two of the images Trump shared were real images of a Trump <a href="https://www.cbsnews.com/news/trump-shares-fake-swifties-for-trump-images/">supporter who is also a Swift fan</a>.</p><p>Another incident that led to a brief panic was an AI clone of President Joe Biden&#8217;s voice that <a href="https://www.npr.org/2024/01/22/1226129926/nh-primary-biden-ai-robocall">asked people not to vote</a> in the New Hampshire primary.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!RZLZ!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9a8a7210-696d-457a-8997-8be9775e5d92_1022x1296.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!RZLZ!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9a8a7210-696d-457a-8997-8be9775e5d92_1022x1296.jpeg 424w, https://substackcdn.com/image/fetch/$s_!RZLZ!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9a8a7210-696d-457a-8997-8be9775e5d92_1022x1296.jpeg 848w, https://substackcdn.com/image/fetch/$s_!RZLZ!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9a8a7210-696d-457a-8997-8be9775e5d92_1022x1296.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!RZLZ!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9a8a7210-696d-457a-8997-8be9775e5d92_1022x1296.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!RZLZ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9a8a7210-696d-457a-8997-8be9775e5d92_1022x1296.jpeg" width="472" height="598.5440313111546" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/9a8a7210-696d-457a-8997-8be9775e5d92_1022x1296.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1296,&quot;width&quot;:1022,&quot;resizeWidth&quot;:472,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!RZLZ!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9a8a7210-696d-457a-8997-8be9775e5d92_1022x1296.jpeg 424w, https://substackcdn.com/image/fetch/$s_!RZLZ!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9a8a7210-696d-457a-8997-8be9775e5d92_1022x1296.jpeg 848w, https://substackcdn.com/image/fetch/$s_!RZLZ!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9a8a7210-696d-457a-8997-8be9775e5d92_1022x1296.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!RZLZ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9a8a7210-696d-457a-8997-8be9775e5d92_1022x1296.jpeg 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em>News headlines in the wake of the Biden robocall.</em></figcaption></figure></div><p>Rules against such robocalls have existed for years. In fact, the perpetrator of this particular robocall was <a href="https://www.reuters.com/world/us/fcc-finalizes-6-million-fine-over-ai-generated-biden-robocalls-2024-09-26/">fined $6 million</a> by the Federal Communications Commission (FCC). The FCC has tiplines to report similar attacks, and it <a href="https://fccprod.servicenowservices.com/rmd?id=rmd_listings">enforces rules around robocalls frequently</a>, regardless of whether AI is used. Since the robocall used a static recording, it could have been made about as easily without using AI&#8212;for instance, by hiring voice impersonators.</p><p>It is also unclear what impact the robocall had: The efficacy of the deepfake depends on the recipient believing that the <em>president of the United States</em> is personally calling them on the phone to ask them <em>not </em>to vote in a primary.</p><p>Is it just a matter of time until improvements in technology and the expertise of actors seeking to influence elections lead to more effective AI disinformation? We don&#8217;t think so. In the next section, we point out that structural reasons that drive the demand for misinformation are not aided by AI. We then look at the history of predictions about coming waves of AI disinformation that have accompanied the release of new tools&#8212;predictions that have not come to pass.</p><h3><strong>The Demand for Misinformation</strong></h3><p>Misinformation can be seen through the forces of <a href="https://reason.com/volokh/2023/10/05/the-demand-for-political-misinformation-is-a-bigger-problem-than-the-supply-even-in-the-age-of-ai/">supply and demand</a>. The supply comes from people who want to make a buck by generating clicks, partisans who want their side to win, or state actors who want to conduct influence operations. Interventions so far have almost entirely tried to curb the supply of misinformation while leaving the demand unchanged.</p><p>The focus on AI is the latest example of this trend. Since AI reduces the cost of generating misinformation to nearly zero, analysts who look at misinformation as a supply problem are very concerned. But analyzing the <em>demand </em>for misinformation can clarify how misinformation spreads and what interventions are likely to help.</p><p>Looking at the demand for misinformation tells us that as long as people have certain worldviews, they will seek out and find <a href="https://dl.acm.org/doi/10.1145/2470654.2481326">information consistent with those views</a>. Depending on what someone&#8217;s worldview is, the information in question is often misinformation&#8212;or at least would be considered misinformation by those with differing worldviews.</p><p>In other words, successful misinformation operations <a href="https://dl.acm.org/doi/10.1145/3091478.3091523">target in-group members</a>&#8212;people who <a href="https://misinforeview.hks.harvard.edu/wp-content/uploads/2024/04/traberg_gamified_inoculation_political_ingroups_20240430.pdf">already</a> <a href="https://www.nber.org/papers/w28884">agree</a> with the broad intent of the message. Such recipients may have lower skepticism for messages that conform to their worldviews and may even be willing to knowingly amplify <a href="https://ohiocapitaljournal.com/2024/09/16/voters-moral-flexibility-helps-them-defend-politicians-misinformation/">false information</a>. Sophisticated tools aren&#8217;t needed for misinformation to be effective in this context. On the flip side, it will be extremely hard to convince <a href="https://www.science.org/doi/10.1126/sciadv.abf1234?url_ver=Z39.88-2003&amp;rfr_id=ori:rid:crossref.org&amp;rfr_dat=cr_pub%20%200pubmed#sec-2">out-group members</a> of false information that they <em>don't </em>agree with, regardless of AI use.</p><p>Seen in this light, AI misinformation plays a very different role from its popular depiction of swaying voters in elections. Increasing the supply of misinformation does not meaningfully change the dynamics of the <em>demand</em> for misinformation since the increased supply is <a href="https://www.nature.com/articles/s41562-020-0833-x">competing for the same eyeballs.</a><strong> </strong>Moreover, the increased supply of misinformation is <a href="https://cetas.turing.ac.uk/publications/ai-enabled-influence-operations-safeguarding-future-elections">likely to be consumed</a> mainly by a <a href="https://www.nature.com/articles/s41562-023-01564-2.pdf">small</a> <a href="https://www.science.org/doi/10.1126/science.aau2706">group</a> of partisans who <a href="https://www.nature.com/articles/s41598-022-19837-7.pdf">already agree with it</a> and heavily consume misinformation rather than to convince a broader swath of the public.</p><p>This also explains why <a href="https://datasociety.net/wp-content/uploads/2019/09/DataSociety_Deepfakes_Cheap_Fakes.pdf">cheap fakes</a> such as media from unrelated events, traditional video edits such as jump cuts, or even <a href="https://www.wired.com/story/x-israel-hamas-war-disinformation/">video game footage</a> can be effective for propagating misinformation despite their low quality: It is much easier to convince someone of misinformation if they already agree with its message.</p><p>Our analysis of the demand for misinformation may be most applicable to countries with polarized close races where leading parties have similar capacities for voter outreach, so that voters&#8217; (mis)information demands are already saturated.</p><p>Still, to our knowledge, in every country that held elections in 2024 so far, AI misinformation had much <a href="https://www.context.news/ai/opinion/cheap-fakes-are-a-blind-spot-for-platforms-in-the-global-south">less impact than feared</a>. In India, deepfakes were used for <a href="https://restofworld.org/2024/exporter-india-deepfake-trolls/">trolling</a> more than spreading false information. In Indonesia, the impact of AI wasn't to sow false information but rather to soften the image of then-candidate, now-President Prabowo Subianto (a former general accused of many past human rights abuses) using <a href="https://www.cigionline.org/articles/its-time-to-reframe-disinformation-indonesias-elections-show-why/">AI-generated digital cartoon avatars</a> that depicted him as likable.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-5" href="#footnote-5" target="_self">5</a></p><h3><strong>Why Do Concerns About AI Misinformation Keep Recurring?</strong></h3><p>The 2024 election cycle wasn&#8217;t the first time when there was widespread fear that AI deepfakes would lead to rampant political misinformation. <a href="https://www.cnn.com/2019/06/12/tech/deepfake-2020-detection/index.html">Strikingly similar concerns</a> about AI were expressed before the 2020 U.S. election, though these concerns <a href="https://www.npr.org/2020/10/01/918223033/where-are-the-deepfakes-in-this-presidential-election">were not borne out</a>. The release of new AI tools is often accompanied by worries that it will unleash new waves of misinformation:</p><ul><li><p><strong>2019. </strong>When OpenAI released its GPT-2 series of models in 2019, one of the main reasons it <a href="https://www.technologyreview.com/2019/08/29/133218/openai-released-its-fake-news-ai-gpt-2/">held back</a> on releasing the model weights for the most capable models in the series was its alleged potential to generate misinformation.</p></li><li><p><strong>2023. </strong>When Meta released the LLaMA model openly in 2023, multiple news outlets reported <a href="https://www.nytimes.com/2023/05/18/technology/ai-meta-open-source.html">concerns</a> that it would trigger a deluge of AI misinformation. These models were far more powerful than the GPT-2 models released by OpenAI in 2019. Yet, we have not seen evidence of large-scale voter persuasion attributed to using LLaMA or other large language models.</p></li><li><p><strong>2024. </strong>Most recently, the widespread availability of AI image editing tools on smartphones has prompted similar <a href="https://www.theverge.com/2024/8/26/24228808/ai-image-editing-photoshop-comparison-argument">concerns</a>.</p></li></ul><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!IboN!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F97a992ce-4187-46e1-b7cd-85a357716c64_779x855.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!IboN!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F97a992ce-4187-46e1-b7cd-85a357716c64_779x855.png 424w, https://substackcdn.com/image/fetch/$s_!IboN!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F97a992ce-4187-46e1-b7cd-85a357716c64_779x855.png 848w, https://substackcdn.com/image/fetch/$s_!IboN!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F97a992ce-4187-46e1-b7cd-85a357716c64_779x855.png 1272w, https://substackcdn.com/image/fetch/$s_!IboN!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F97a992ce-4187-46e1-b7cd-85a357716c64_779x855.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!IboN!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F97a992ce-4187-46e1-b7cd-85a357716c64_779x855.png" width="488" height="535.609756097561" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/97a992ce-4187-46e1-b7cd-85a357716c64_779x855.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:855,&quot;width&quot;:779,&quot;resizeWidth&quot;:488,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!IboN!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F97a992ce-4187-46e1-b7cd-85a357716c64_779x855.png 424w, https://substackcdn.com/image/fetch/$s_!IboN!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F97a992ce-4187-46e1-b7cd-85a357716c64_779x855.png 848w, https://substackcdn.com/image/fetch/$s_!IboN!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F97a992ce-4187-46e1-b7cd-85a357716c64_779x855.png 1272w, https://substackcdn.com/image/fetch/$s_!IboN!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F97a992ce-4187-46e1-b7cd-85a357716c64_779x855.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em><a href="https://newsletter.pessimistsarchive.org/p/the-1912-war-on-fake-photos">Source</a>, <a href="https://www.newspapers.com/article/janesville-weekly-gazette/153866353/">Primary source</a></em></figcaption></figure></div><p>In fact, concerns about using new technology to create false information go back over a century. The late 19th and early 20th centuries saw the advent of technologies for photo retouching. This was accompanied by concerns that retouched photographs would be used to deceive people, and, in 1912, a bill was introduced in the U.S. that would have criminalized <a href="https://newsletter.pessimistsarchive.org/p/the-1912-war-on-fake-photos">photo editing</a> without subjects&#8217; consent. (It died in the Senate.)</p><p>Thinking of political misinformation as a <a href="https://x.com/JMchangama/status/1857136505472405680">technological (or AI)</a> problem is <a href="https://www.technologyreview.com/2024/09/03/1103464/ai-impact-elections-overblown/">appealing</a> because it makes the solution seem tractable. If only we could roll back harmful tech, we could drastically improve the information environment!</p><p>While the goal of improving the information environment is laudable, blaming technology is not a fix. Political polarization has <a href="https://www.taylorfrancis.com/chapters/edit/10.4324/9780203713020-4/distrust-news-media-symptom-cause-partisan-polarization-jonathan-ladd-alexander-podkul">led</a> to greater <a href="https://shorensteincenter.org/future-trustworthy-information-learning-online-content-creators/">mistrust</a> of the media. People prefer sources that <a href="https://dl.acm.org/doi/10.1145/2470654.2481326">confirm</a> their worldview and are less <a href="https://academic.oup.com/restud/advance-article-abstract/doi/10.1093/restud/rdae058/7685990?redirectedFrom=fulltext">skeptical</a> about content that <a href="https://www.cambridge.org/core/services/aop-cambridge-core/content/view/1121D46E918815B2CD7A9C6C237464A2/S1930297500003570a.pdf/susceptibility-to-misinformation-is-consistent-across-questionframings-and-response-modes-and-better-explained-by-myside-bias-and-partisanshipthan-analytical-thinking.pdf#page=20.60">fits</a> their worldview. Another major factor is the <a href="https://www.census.gov/library/stories/2022/06/internet-crushes-traditional-media.html">drastic decline</a> of <a href="https://www.pewresearch.org/journalism/fact-sheet/newspapers/">journalism revenues</a> in the last two decades&#8212;largely driven by the shift from traditional to social media and online advertising. But this is more a result of structural changes in how people seek out and consume information than the specific threat of misinformation shared online.</p><p>As history professor Sam Lebovic has pointed out, improving the information environment is <a href="https://knightcolumbia.org/content/fake-news-lies-and-other-familiar-problems">inextricably linked</a> to the larger project of shoring up democracy and its institutions. There&#8217;s no quick technical fix, or targeted regulation, that can &#8220;solve&#8221; our information problems. We should reject the simplistic temptation to blame AI for political misinformation and confront the gravity of the hard problem.</p><p><em>Correction: A previous version of the essay&#8217;s introduction stated that most AI use is not deceptive. In fact, 39 of 78 articles in the database are examples of non-deceptive AI use, or 39 out of 74 if we restrict ourselves to political communication and set aside the 4 instances that are scams.</em></p><p><em>This essay is <a href="https://knightcolumbia.org/blog/we-looked-at-78-election-deepfakes-political-misinformation-is-not-an-ai-problem">cross-posted</a> to the Knight First Amendment Institute website. We are grateful to <a href="https://knightcolumbia.org/bios/view/katherine-glenn-bass">Katy Glenn Bass</a> for her feedback.</em></p><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>The terms mis- and disinformation lack agreed-upon definitions. In this piece, we use the term misinformation to refer to outright false information, as opposed to issues of misleading interpretive framing. Despite many people&#8217;s perception of outgroup narratives as &#8220;misinformation,&#8221; we don't think the misinformation lens is a useful way to think about differences in framing and narratives; we're more narrowly concerned about using outright false information to support those narratives.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p>The low number of total deepfakes found in elections worldwide is surprising on its own terms. The small number could either indicate that AI deepfakes are a much smaller problem so far than anticipated or that the database has many missing entries. Still, other databases that tracked election deepfakes have a similar count for the total number of deepfakes; for example, the German Marshall Fund&#8217;s list of deepfakes related to 2024 elections worldwide has <a href="https://www.gmfus.org/spitting-images-tracking-deepfakes-and-generative-ai-elections">133 entries</a>, though it started collecting entries in September 2023. As we note further along in the essay, the News Literacy Project <a href="https://misinfodashboard.newslit.org/">documented</a> known misinformation about the 2024 elections and <a href="https://newslit.org/newsroom/press-release/the-news-literacy-project-experts-watch-out-for-ai-generated-fakes-and-disinformation-about-voting-ahead-of-election-day/">found</a> that cheap fakes that didn't use AI were used seven times more often than AI-generated content.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-3" href="#footnote-anchor-3" class="footnote-number" contenteditable="false" target="_self">3</a><div class="footnote-content"><p>The dataset also included four instances of AI-generated deepfake videos of politicians used to perpetrate financial scams. Compared to political misinformation, scams have very different dynamics (more sophisticated videos could be more convincing) and stakes (they involve individual financial harm rather than threats to democracy). Similarly, addressing scams requires different interventions&#8212;for instance, monitoring and removing networks of scammers is something major online platforms have been doing for a long time. In other words, scams are a different problem that we have other tools for addressing (regardless of the fact that some platforms arguably underinvest in doing so) and are outside the scope of this essay.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-4" href="#footnote-anchor-4" class="footnote-number" contenteditable="false" target="_self">4</a><div class="footnote-content"><p>In the last legs of the 2024 U.S. election, Google and OpenAI restricted their chatbots from <a href="https://archive.is/EPIcW#selection-975.0-988.0">answering</a> election-related queries&#8212;though competitors like Perplexity didn't, claiming that their product was highly accurate. <a href="https://www.proofnews.org/seeking-election-information-dont-trust-ai/">Evaluating</a> chatbots&#8217; tendency to answer questions factually or abstain from answering questions, improving the factuality of responses, and ensuring chatbots work across different <a href="https://www.proofnews.org/ai-models-falter-answering-election-questions-in-spanish/">languages</a> and contexts are important areas of work as more people turn to chatbots for answering questions.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-5" href="#footnote-anchor-5" class="footnote-number" contenteditable="false" target="_self">5</a><div class="footnote-content"><p>To be clear, we should not treat such propaganda as something newly made possible by AI. It is the incremental evolution of long-standing <a href="https://www.routledge.com/Propaganda-From-Disinformation-and-Influence-to-Operations-and-Information-Warfare/Olejnik/p/book/9781032813721">techniques</a>. Indeed, the cost of creating cartoon avatars for presidential campaigns would be minuscule with or without AI. The impact of propaganda depends not on the technical methods used to create it but rather on the freedom of the press to uplift competing narratives.</p></div></div>]]></content:encoded></item><item><title><![CDATA[Does the UK’s liver transplant matching algorithm systematically exclude younger patients?]]></title><description><![CDATA[Seemingly minor technical decisions can have life-or-death effects]]></description><link>https://www.normaltech.ai/p/does-the-uks-liver-transplant-matching</link><guid isPermaLink="false">https://www.normaltech.ai/p/does-the-uks-liver-transplant-matching</guid><dc:creator><![CDATA[Arvind Narayanan]]></dc:creator><pubDate>Mon, 11 Nov 2024 19:57:25 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/104a5670-c400-45bf-b02e-18d3bfe2e46d_2890x1648.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>By Arvind Narayanan, Angelina Wang, Sayash Kapoor, and Solon Barocas</em></p><p>Predictive algorithms are used in many life-or-death situations. In the paper <a href="https://www.aisnakeoil.com/p/ai-cannot-predict-the-future-but">Against Predictive Optimization</a>, we argued that the use of predictive logic for making decisions about people has recurring, inherent flaws, and should be rejected in many cases.</p><p>A wrenching case study comes from the UK&#8217;s liver allocation algorithm, which appears to <a href="https://www.ft.com/content/5125c83a-b82b-40c5-8b35-99579e087951">discriminate by age</a>, with some younger patients seemingly unable to receive a transplant, no matter how ill. What went wrong here? Can it be fixed? Or should health systems avoid using algorithms for liver transplant matching?</p><h3><strong>How the liver allocation algorithm works</strong></h3><p>The UK <a href="https://www.odt.nhs.uk/odt-structures-and-standards/odt-hub-programme/national-liver-offering-scheme/">nationalized</a> its liver transplant system in 2018, replacing previous regional systems where livers were prioritized based on disease severity.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a> When a liver becomes available, the new algorithm uses predictive logic to calculate how much each patient on the national waiting list would benefit from being given that liver.&nbsp;</p><p>Specifically, the algorithm predicts how long each patient would live if they were given that liver, and how long they would live if they didn&#8217;t get a transplant. The difference between the two is the patient&#8217;s Transplant Benefit Score (TBS). Patients are sorted in decreasing order of the score, and the top patient is offered the liver (if they decline, the next patient is offered, and so on).</p><p>Given this description, one would expect that the algorithm would <em>favor</em> younger patients, as they will potentially gain many more decades of life through a transplant compared to older patients. If the algorithm has the opposite effect, either the score has been inaccurately <a href="https://nhsbtdbe.blob.core.windows.net/umbraco-assets-corp/11879/faqs-for-national-liver-offering-scheme-may2018.pdf">portrayed</a> or it is being calculated incorrectly. We&#8217;ll see which one it is. But first, let&#8217;s discuss a more basic question.</p><h3><strong>Why is predictive AI even needed?</strong></h3><p>Discussions of the ethics of algorithmic decision making often narrowly focus on bias, ignoring the question of whether it is <a href="https://predictive-optimization.cs.princeton.edu/">legitimate</a> to use an algorithm in the first place. For example, consider pretrial risk prediction in the criminal justice system. While bias is a serious concern, a deeper question is whether it is morally justified to deny defendants their freedom based on a prediction of what they might do rather than a determination of guilt, especially when that prediction is barely more accurate than a coin flip.&nbsp;</p><p>Organ transplantation is different in many ways. The health system needs to make efficient and ethical use of a very limited and valuable resource, and must find some principled way of allocating it to many deserving people, all of whom have reasonable claims for why they should be entitled to it. There are thousands of potential recipients, and decisions must be made quickly when an organ becomes available. Human judgment doesn&#8217;t scale.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a>&nbsp;</p><p>Another way to try to avoid the need for predictive algorithms is to <a href="https://arxiv.org/pdf/2312.08511">increase the pool of organs</a> so that they are no longer as scarce. Encouraging people to sign up for organ donation is definitely important. But even if the supply of livers is no longer a constraint, it would still be useful to predict which patient will benefit the most from a <em>specific</em> liver.&nbsp;</p><p>Sometimes <a href="https://5harad.com/papers/simple-rules.pdf">simple statistical formulas</a> provide most of the benefits of predictive AI without the downsides. In fact, the previous liver transplant system in the UK was based on a relatively simple formula for predicting disease severity, called the UK End-stage Liver Disease score, which is based on the blood levels of a few markers. The new system takes into account the benefit of transplantation in addition to disease severity. It is also more of a black box. It is &#8220;AI&#8221; in the sense that it is derived from a data-driven optimization process and is too complex to be mentally understood by doctors or patients. It uses 28 variables from the donor and recipient to make a prediction.&nbsp;</p><p>It seems at least plausible that this complexity is justified in this context because health outcomes are much more predictable than who will commit a crime (though this varies by disease). Follow-up studies have <a href="https://www.journal-of-hepatology.eu/article/S0168-8278(24)00203-4/fulltext">confirmed</a> that the matching algorithm does indeed save more lives than the system that it replaced.</p><p>So there isn&#8217;t necessarily a prima facie case for arguing against the use of the algorithm. Instead, we have to look at the details of what went wrong. Let&#8217;s turn to those details.</p><h3><strong>The Financial Times investigation</strong></h3><p>In November 2023, the Financial Times published a bombshell <a href="https://www.ft.com/content/5125c83a-b82b-40c5-8b35-99579e087951">investigation</a> about bias in the algorithm. It centers on a 31 year old patient, Sarah Meredith, with multiple genetic conditions including cystic fibrosis. It describes her accidental discovery that the Transplant Benefit Score algorithm even existed and would decide her fate; her struggle to understand how it worked; her liver doctors&#8217; lack of even basic knowledge about the algorithm; and her realization that there was no physician override to the TBS score and no appeals process.</p><p>When she reached out to the National Health Service to ask for explanations, Meredith was repeatedly told she wouldn&#8217;t understand. It seems that the paternalism of health systems combined with the <a href="https://royalsocietypublishing.org/doi/10.1098/rsta.2018.0084">myth</a> of the inscrutability of algorithms is a particularly toxic mix.</p><p>Meredith eventually landed on a <a href="https://transplantbenefit.org/">web app</a> that calculates the TBS, built by Professor Ewen Harrison and his team. He is a surgeon and data scientist who has studied the TBS, and is a co-author of a study of some of the failures of the algorithm. It is through this app that Meredith realized how biased the algorithm is. It also shows why the inscrutability of algorithmic decision making is a myth: even without understanding the internals, it is easy to understand the behavior of the system, especially given that a particular patient only cares about how the system behaves in one specific instance.</p><p>But this isn&#8217;t just one patient&#8217;s experience. From the Financial Times piece:</p><blockquote><p>&#8220;If you&#8217;re below 45 years, no matter how ill, it is impossible for you to score high enough to be given priority scores on the list,&#8221; said Palak Trivedi, a consultant hepatologist at the University of Birmingham, which has one of the country&#8217;s largest liver transplant centres.</p></blockquote><p>Finally, a 2024 <a href="https://www.thelancet.com/journals/lanhl/article/PIIS2666-7568(24)00044-8/fulltext">study</a> in The Lancet has confirmed that the algorithm has a severe bias against younger patients.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-3" href="#footnote-3" target="_self">3</a></p><h3><strong>Patient groups warned about the bias</strong></h3><p>The objective of the matching system is to identify the recipient whose life expectancy would be increased the most through the transplant. The obvious way to do this is to predict each patient&#8217;s expected survival time with and without the transplant. This is almost what the algorithm does, but not quite &#8212; it predicts each patient&#8217;s <em>likelihood of surviving 5 years</em> with and without the transplant.</p><p>The problem with this is obvious. A <a href="https://pscsupport.org.uk/what-we-do/lptc/">patient group</a> gave this feedback through official channels in 2015, long before the algorithm went into effect:</p><blockquote><p>Capping survival at five years in effect diminishes the benefits for younger patients as it underestimates the gain in life years by predicting lifetime gain over 5 years, as opposed&nbsp; to the total lifetime gain. Paediatric and small adult patients benefit from accessing small adult livers as a national priority in the Current System. However, young adults must compete directly with all other adult patients. In the proposed model, there is no recognition that a death in a younger patient is associated with a&nbsp; greater number of expected years of life lost compared with the death of an older adult patient. There is also no recognition that longer periods waiting has an impact on younger patients&#8217; prospects, such as career and family,&nbsp; and contribution to society compared with older adult patients. Younger patients have not yet had the chance to live their lives and consideration should be given to how the cohort of younger waiting list patients is affected by rules applied to calculate their benefit.</p></blockquote><p>This is what leads to the algorithm&#8217;s behavior. Younger patients are (correctly) predicted to be more likely to survive 5 years <em>without</em> a transplant, and about as likely as older patients to survive 5 years <em>with</em> a transplant. So younger patients&#8217; predicted net benefit (over a 5-year period) is much less than older patients&#8217;. Over the entire course of their lives, younger patients would likely benefit more, but the algorithm doesn&#8217;t take this into account.</p><h3><strong>Show us the target variable and we&#8217;ll show you the problem</strong></h3><p>It is not clear why the problem was ignored, both in version 1 of the algorithm in 2018 and in version 2 in 2022 which corrected a bias against cancer patients (we&#8217;ll get to that bias in a minute). Perhaps the developers did not recognize how severe the age bias is. Even in a <a href="https://www.journal-of-hepatology.eu/article/S0168-8278(24)00203-4/fulltext">2024 paper</a> about the algorithm, where they briefly discuss many of its limitations <em>including the five-year cap</em>, they do not mention that the cap de-prioritizes younger patients.</p><p>On the other hand, the list of features (donor and recipient characteristics) is prominently listed and discussed in <a href="https://www.odt.nhs.uk/odt-structures-and-standards/odt-hub-programme/national-liver-offering-scheme/">public communications</a> about the system. This may reflect a misconception that the way to understand an algorithm, including its potentially discriminatory effects, is to look at the list of features &#8212; the <em>inputs</em>. In reality, the target variable &#8212; the <em>output</em> &#8212; is often more important for fairness than the features.&nbsp;</p><p>Unfortunately there is little recognition of this crucial fact outside the technical community (and sometimes even within the technical community). Instead there is a narrow focus on removing sensitive variables (such as age, race, or gender) and proxies for the sensitive variables from the list of features, which is <a href="https://fairmlbook.org/classification.html#no-fairness-through-unawareness">usually ineffective</a> and often even <a href="https://www.nejm.org/doi/full/10.1056/NEJMp2311050">counterproductive</a>.</p><p>The choice of a 5-year period seems to be because of <a href="https://www.journal-of-hepatology.eu/article/S0168-8278(24)00203-4/fulltext">data availability</a>: &#8220;This length of follow-up was selected as data were readily available ... while longer follow-up was not.&#8221; In our experience, there is almost always some difficulty that prevents accurately measuring the true construct of interest, which is why this is one of the recurring flaws we identify in the <a href="https://predictive-optimization.cs.princeton.edu/">Against Predictive Optimization</a> paper. It is a <a href="https://arxiv.org/abs/1912.05511">target-construct mismatch</a>, because what is being predicted, the target, differs from what we actually want to predict, the construct.</p><h3><strong>It gets worse</strong></h3><p>The cap means that the expected survival with a transplant for most patient groups is about the same (about 4.5 years, reflecting the fact that about 85% of patients survive 5 years after a transplant). So the utility of the transplant, while high, is more-or-less uniformly high, which means that it doesn&#8217;t really factor into the scores! It turns out that the algorithm is mostly just assessing need, that is, how long patients would survive <em>without</em> a transplant.</p><p>This is ironic because modeling post-transplant survival was <a href="https://nhsbtdbe.blob.core.windows.net/umbraco-assets-corp/7892/national-liver-offering-roadshow-presentation.pdf">claimed</a> to be the main reason to use this system over the previous one. If it keeps more people from dying, we suspect it is simply because it does a better job of assessing need, and/or because the use of the algorithm coincided with a move from regional to national systems, allowing it to better cater to high-need patients in previously under-served regions.</p><p>The fact that the system isn&#8217;t very good at meeting its stated objectives only seems to have been <a href="https://pubmed.ncbi.nlm.nih.gov/37516542/">reported</a> a decade after the algorithm was developed (although in retrospect, there were clear signals in the results of the <a href="https://bts.org.uk/wp-content/uploads/2018/04/NHSBT-John-OGrady.pdf">simulations</a> that were run before deployment). Specifically, it is noted in the comment-and-response section of a <a href="https://www.thelancet.com/journals/lancet/article/PIIS0140-6736(23)00114-9/fulltext">paper</a> about the algorithm. In terms of obscurity, that&#8217;s the academic equivalent of Wikipedia&#8217;s Talk pages &#8212; most of the public wouldn&#8217;t even know such a thing exists.&nbsp;</p><h3><strong>An algorithmic absurdity: cancer improves survival</strong></h3><p>While the authors of the above paper mention in passing that one of the two models in the algorithm (post-transplant survival) doesn&#8217;t seem to do much, their <a href="https://www.thelancet.com/journals/lancet/article/PIIS0140-6736(23)00114-9/fulltext">main point</a> is about the other model &#8212; the one that assesses need by predicting survival on the waiting list. They show that it expects patients with cancer to survive <em>longer</em> than those without cancer (all else being equal). This kind of thing is sometimes called algorithmic absurdity, something that would seem obviously wrong to a person based on common sense.</p><p>The prediction about patients with cancer is not just an oddity &#8212; it has big consequences for patients&#8217; lives: &#8220;for the first 3 years of the TBS scheme (excluding the period when TBS offering was suspended due to COVID-19), patients with cancer were rarely allocated livers by the TBS model&#8221;. This is what led to the 2022 revision of the algorithm.</p><p>The finding is reminiscent of a well-known failure from a few decades ago wherein a model <a href="https://pubmed.ncbi.nlm.nih.gov/9040894/">predicted</a> that patients with asthma were at <em>lower</em> risk of developing complications from pneumonia. Fortunately this was spotted before the model was deployed. It turned out to be a correct pattern in the data, but only because asthmatic patients were sent to the ICU, where they received better care. Of course, it would have been disastrous to replace that very policy with the ML model that treated asthmatic patients as lower risk. That case study has become a textbook illustration of the usefulness of <a href="https://dl.acm.org/doi/10.1145/2783258.2788613">interpretable</a> models over black-box models. If researchers can easily examine the coefficients of the model, implausible behaviors become more readily apparent.</p><p>The TBS does use interpretable regression models. But it is actually two different sets of models, one for patients with cancer and one for patients without cancer, because the two groups are represented by two different data sources. That explains why the implausible behavior of the algorithm may have arisen &#8212; the patient populations are different; perhaps the population from which the cancer patients were drawn was younger or healthier in other ways. Of course, this doesn&#8217;t justify the algorithm&#8217;s behavior where flipping a <em>specific</em> patient from non-cancer to cancer increases the predicted survival. The fact that there are two different sets of models may also explain why it went undetected for so long &#8212; the problem is not obvious from the regression coefficients and can only be detected by simulating a patient population.</p><h3><strong>Sleepwalking into utilitarian ethics</strong></h3><p>Predictive logic bakes in a utilitarian worldview &#8212; the most good for the greatest number. That makes it hard to incorporate a notion of deservingness. Many people have a strong moral intuition that patients whose conditions result from factors outside their control are more deserving of help. From the Financial Times article:</p><blockquote><p>Trivedi [the hepatologist] said patients found [the bias against younger patients] particularly unfair, because younger people tended to be born with liver disease or develop it as children, while older patients more often contracted chronic liver disease because of lifestyle choices such as drinking alcohol.</p></blockquote><p>Donor preferences are also neglected. For example, presumably some donors would prefer to help someone in their own community. But in the utilitarian worldview, this is simply <a href="https://www.jstor.org/stable/2265052?seq=4">geographic discrimination</a>. (Our point is not about whether deservingness or donor preferences are important considerations, but rather that the algorithm dictates the ethical framework.)&nbsp;</p><p>Traditionally, individual physicians made decisions about transplants without much formal reasoning or accountability. But with routinization and increasing scale of organ transplantation, and the shift to nationwide matching systems, manual matching is no longer feasible. Automation has forced decision makers to make the matching criteria explicit. This formalization can be a <a href="https://dl.acm.org/doi/abs/10.1145/3351095.3372871">good thing</a>, as it allows ethical debate about the pros and cons of precisely specified policies.</p><p>But automation has also privileged utilitarianism, as it is much more amenable to calculation. Non-utilitarian considerations resist quantification. No committee of decision makers would want to be in charge of determining how much of a penalty to apply to patients who drank alcohol, and whatever choice they made would meet fierce objection. In contrast, the veneer of data-driven decision making, even though it hides many normative choices, allows decision makers to reach consensus and to deploy algorithms without endless debate.</p><p>For this reason, utilitarianism has been ascendant in many, many domains over the last few decades, including medical ethics and public health.</p><p>While the liver matching algorithm optimizes life years (albeit poorly), other algorithms and institutions go one step further and optimize &#8220;quality-adjusted&#8221; life years, taking into account factors such as how well a person is able to complete daily tasks and how much pain they are in. Quality adjustment has side-effects such as giving lower preference to disabled people.</p><p>Overall, we are not necessarily against this shift to utilitarian logic, but we think it should only be adopted if it is the result of a democratic process, not just because it&#8217;s more convenient. The tail shouldn&#8217;t wag the dog. It isn&#8217;t clear to what extent the wider public is even aware of the widespread shift to nationalized transplant systems &#8212; in many countries, for many organs &#8212; and the ethical logics that underpin them. Public input about <em>specific</em> systems, such as the one we&#8217;ve discussed, is not a replacement for broad societal consensus on the underlying moral frameworks. Nor should this debate be confined to the <a href="https://www.cambridge.org/core/journals/cambridge-quarterly-of-healthcare-ethics/article/does-the-united-states-do-it-better-a-comparative-analysis-of-liver-allocation-protocols-in-the-united-kingdom-and-the-united-states/681D3480F8940AF51D65B3E4114DEBA5">medical ethics</a> literature.&nbsp;</p><h3><strong>What&#8217;s next</strong></h3><p>The liver allocation algorithm was developed and is run by the National Health Service (NHS), the UK&#8217;s publicly-funded health system. We&#8217;ve previously explained in this newsletter that <a href="https://www.aisnakeoil.com/p/the-bait-and-switch-behind-ai-risk">bad outcomes</a> result when public sector agencies outsource algorithmic decision making systems to opaque, profit-oriented companies. That&#8217;s not the case here. The developers are doing their best to save lives. A lot of thought and care went into the system, and there was public input. If there were missteps, they are a result of how hard the problem is.</p><p>There are clear problems with the liver allocation algorithm that can and should be addressed. There are at least three ways to mitigate the age bias. The first is to collect more and better data. The second is to put a thumb on the algorithm&#8217;s scale to ensure that the age distribution of recipients is roughly in line with society&#8217;s normative ideals. This can be achieved by formulating a constrained optimization problem (there are many papers on algorithmic fairness that show how to do this). The third is to stop using age as a factor. We don&#8217;t like this approach for reasons described above, but it is perhaps more easily defensible to non-experts.</p><p>The <a href="https://www.odt.nhs.uk/transplantation/liver/liver-advisory-group/">Liver Advisory Group</a> is the entity with the power to effect changes. The members meet every six months. Unfortunately they haven&#8217;t yet uploaded their minutes from their May 2024 meeting, so it isn&#8217;t clear if they are paying attention.&nbsp;</p><p>The deeper, systemic problem will be harder to address &#8212; inadequate transparency and public participation in medical ethics. The rapid adoption of AI for medical decision making requires a whole-of-society ethical debate. This isn&#8217;t about specific algorithms but about the bundle of unexamined assumptions behind their claim to efficacy and thus to legitimacy. Better late than never.</p><p>Zooming out beyond medicine, the pitfalls that arise in disparate applications of predictive decision making bear <a href="https://predictive-optimization.cs.princeton.edu/">striking similarities</a> with each other. This calls for <a href="https://www.birs.ca/events/2024/5-day-workshops/24w5283">more research</a> on avoiding these flaws as well as a community of practitioners from different fields who can learn from each other. Venues such as the <a href="https://facctconference.org/">conference on Fairness, Accountability, and Transparency</a> can bring such cross-cutting groups together.</p><h4><strong>Further reading</strong></h4><ul><li><p><em><a href="https://www.science.org/doi/10.1126/science.aax2342">Dissecting racial bias in an algorithm used to manage the health of populations</a></em> is a classic paper that revealed how the use of the wrong target variable can lead to severe biases.</p></li><li><p><em><a href="https://www.russellsage.org/publications/voices-code">Voices in the Code</a></em> is an excellent book that details the development of a kidney matching algorithm in the U.S. It shows the benefits of public participation &#8212; how it can uncover flaws in algorithms that developers had not anticipated, and increase the legitimacy of the system that is ultimately deployed. But participation is no panacea. The process discussed in the book took a decade, during which time a far worse system remained in place. And participatory development of a specific system does not obviate the need for a broader public debate on the utilitarian and algorithmic turn in medical ethics.</p></li></ul><h4><strong>Acknowledgment</strong></h4><p>We are grateful to Emma Pierson for feedback on a draft of this essay.</p><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>For more details on the previous system, see <a href="https://aasldpubs.onlinelibrary.wiley.com/doi/pdf/10.1002/lt.24462">this article</a>.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p>There are many other differences between human and algorithmic decision making. Algorithms tend to be much better at optimizing a given objective, such as maximizing the number of life years gained through transplantation. But human decision makers can more easily incorporate multiple objectives, especially non-utilitarian ones. Human decision makers of course have their own biases; whose biases are worse and which ones are easier to mitigate are complex questions whose answers are context-dependent.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-3" href="#footnote-anchor-3" class="footnote-number" contenteditable="false" target="_self">3</a><div class="footnote-content"><p>Update: Regarding the quote in the Financial Times, an NHS Blood and Transplant spokesperson said: &#8220;People can receive transplants through various pathways. Some use the TBS and some do not use the TBS. It is incorrect to say that no-one under 40 could receive a transplant. It is also incorrect to say that no-under 40 could receive a transplant as the patient most in need according to the TBS. It could also unduly worry patients on the transplant list. Over the past three UK financial years, 133 people aged 17 to 39 received transplants that were allocated to them through the TBS, as the patient in the country who would benefit most from that liver. Additionally, over the past three financial years, a further 285 people aged 17 to 39 received liver only transplants through other pathways.&#8221;</p></div></div>]]></content:encoded></item><item><title><![CDATA[FAQ about the book and our writing process]]></title><description><![CDATA[What's in the book and how we wrote it]]></description><link>https://www.normaltech.ai/p/faq-about-the-book-and-our-writing</link><guid isPermaLink="false">https://www.normaltech.ai/p/faq-about-the-book-and-our-writing</guid><dc:creator><![CDATA[Arvind Narayanan]]></dc:creator><pubDate>Fri, 04 Oct 2024 15:55:33 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/907f61bc-0697-41a3-a729-b719812a6d1a_2166x1402.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>The AI Snake Oil book was <a href="https://www.amazon.com/Snake-Oil-Artificial-Intelligence-Difference/dp/069124913X">published</a> last week. We&#8217;re grateful for the level of interest &#8212; it&#8217;s sold about 8,000 copies so far. We&#8217;ve received many questions about the book, both its substance and the writing process. Here are the most common ones.</p><h3><strong>Why don&#8217;t you recognize the benefits of AI?</strong></h3><p>We do! The book is not an anti-technology screed. If our point was that all AI is useless, we wouldn&#8217;t need a whole book to say it. It&#8217;s precisely because of AI&#8217;s usefulness in many areas that hype and snake oil have been successful &#8212; it&#8217;s hard for people to tell these apart, and we hope our book can help.</p><p>We also recognize that the harms we describe are usually not solely due to tech, and much more often due to AI being an amplifier of existing problems in our society. A recurring pattern we point out in the book is that "broken AI is appealing to broken institutions" (Chapter 8).</p><h3><strong>What&#8217;s your optimistic vision for AI, then?</strong></h3><p>There&#8217;s a humorous definition of AI that says &#8220;AI is whatever hasn&#8217;t been done yet&#8221;. When an AI application starts working reliably, it disappears into the background of our digital or physical world. We take it for granted. And we stop calling it AI. When a technology is new, doesn&#8217;t work reliably, and has double-edged societal implications, we&#8217;re more likely to call it AI. So it&#8217;s easy to miss that AI already plays a huge positive role in our lives.</p><p>There&#8217;s a long list of applications that would have been called AI at one point but probably wouldn&#8217;t be today: Robot vacuum cleaners, web search, autopilot in planes, autocomplete, handwriting recognition, speech recognition, spam filtering, and even spell check. These are the kinds of AI we want more of &#8212; reliable tools that quietly make our lives better.&nbsp;</p><p>Many AI applications that make the news for the wrong reasons today &#8212; such as self-driving cars due to occasional crashes &#8212; are undergoing this transition (although, as we point out in the book, it has taken far longer than developers and CEOs anticipated). We think people will eventually take self-driving cars for granted as part of our physical environment.&nbsp;</p><p>Adapting to these changes won&#8217;t be straightforward. It will lead to job loss, require changes to transportation infrastructure and urban planning, and have various ripple effects. But it will have been a good thing, because the safety impact of reliable self-driving tech <a href="https://crashstats.nhtsa.dot.gov/Api/Public/ViewPublication/813560">can&#8217;t be overstated</a>.</p><h3><strong>What&#8217;s the central message of the book?</strong></h3><p>AI is an umbrella term for a set of loosely related technologies and applications. To answer questions about the benefits or risks of AI, its societal impact, or how we should approach the tech, we need to break it down. And that&#8217;s what we do in the book.&nbsp;</p><p>We&#8217;re broadly negative about predictive AI, a term we use to refer to AI that&#8217;s used to make decisions about people based on predictions about their future behavior or outcomes. It&#8217;s used in criminal risk prediction, hiring, healthcare, and many other consequential domains. Our chapters on predictive AI have many horror stories of people denied life opportunities because of algorithmic predictions.</p><p>It&#8217;s hard to predict the future, and AI doesn&#8217;t change that. This is not because of a limitation of the technology but because of inherent limits to predicting human behavior grounded in sociology. (The book owes a huge debt to Princeton sociologist <a href="https://www.princeton.edu/~mjs3/">Matt Salganik</a>; our collaboration with him informed and inspired the book.)&nbsp;</p><p>Generative AI, on the other hand, is a double-edged technology. We are broadly positive about it in the long run, and emphasize that it is useful to essentially every knowledge worker. But its rollout has been chaotic, and misuses have been prevalent. It&#8217;s as if everyone in the world has simultaneously been given the equivalent of a free buzzsaw. As we say in the book:</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!_Uzc!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F57d698dc-e69c-470c-bae0-8faab745d751_784x240.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!_Uzc!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F57d698dc-e69c-470c-bae0-8faab745d751_784x240.png 424w, https://substackcdn.com/image/fetch/$s_!_Uzc!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F57d698dc-e69c-470c-bae0-8faab745d751_784x240.png 848w, https://substackcdn.com/image/fetch/$s_!_Uzc!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F57d698dc-e69c-470c-bae0-8faab745d751_784x240.png 1272w, https://substackcdn.com/image/fetch/$s_!_Uzc!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F57d698dc-e69c-470c-bae0-8faab745d751_784x240.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!_Uzc!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F57d698dc-e69c-470c-bae0-8faab745d751_784x240.png" width="564" height="172.6530612244898" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/57d698dc-e69c-470c-bae0-8faab745d751_784x240.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:240,&quot;width&quot;:784,&quot;resizeWidth&quot;:564,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!_Uzc!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F57d698dc-e69c-470c-bae0-8faab745d751_784x240.png 424w, https://substackcdn.com/image/fetch/$s_!_Uzc!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F57d698dc-e69c-470c-bae0-8faab745d751_784x240.png 848w, https://substackcdn.com/image/fetch/$s_!_Uzc!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F57d698dc-e69c-470c-bae0-8faab745d751_784x240.png 1272w, https://substackcdn.com/image/fetch/$s_!_Uzc!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F57d698dc-e69c-470c-bae0-8faab745d751_784x240.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div><h3><strong>What else is in the book?</strong></h3><p>See the overview of the chapters <a href="https://www.aisnakeoil.com/p/starting-reading-the-ai-snake-oil">here</a>.</p><h3><strong>Isn&#8217;t your book going to be outdated soon?</strong></h3><p>We know that book publishing moves at a slower timescale than AI. So the book is about the foundational knowledge needed to separate real advances from hype, rather than commentary on breaking developments. In writing every chapter, and every paragraph, we asked ourselves: will this be relevant in five years? This also means that there&#8217;s very little overlap between the newsletter and the book.&nbsp;</p><h3><strong>There seem to be three warring camps: AI safety, e/acc, and AI ethics. Which one are you in?</strong></h3><p>The AI discourse is polarized because of differing opinions about which AI risks matter, how serious and urgent they are, and what to do about them. In broad strokes:</p><ul><li><p>The AI safety community considers catastrophic AI risks a major societal concern, and supports government intervention. It has strong ties to the effective altruism movement.&nbsp;</p></li><li><p>e/acc is short for effective accelerationism, a play on effective altruism. It is a libertarian movement that sees tech as the solution and rejects government intervention.</p></li><li><p>The AI ethics community focuses on materialized harms from AI such as discrimination and labor exploitation, and sees the focus on AI safety as a distraction from those priorities.</p></li></ul><p>In the past, the two of us worked on AI ethics and saw ourselves as part of that community. But we no longer identify with any of these labels. We view the polarization as counterproductive. We used to subscribe to the &#8220;distraction&#8220; view but no longer do. The fact that safety concerns have made AI policy a priority has increased, not decreased policymakers&#8217; attention to issues of <a href="https://www.markey.senate.gov/news/press-releases/senator-markey-introduces-ai-civil-rights-act-to-eliminate-ai-bias-enact-guardrails-on-use-of-algorithms-in-decisions-impacting-peoples-rights-civil-liberties-livelihoods">AI and civil rights</a>. These two communities both want AI regulation, and should focus on their <a href="https://www.lawfaremedia.org/article/ai-regulation-s-champions-can-seize-common-ground-or-be-swept-aside">common ground</a> rather than their differences.</p><p>These days, much of our technical and policy work is on AI safety, but we have explained how we have a <a href="https://www.aisnakeoil.com/t/ai-safety">different perspective </a>from the mainstream of the AI safety community. We see our role as engaging seriously with safety concerns and presenting an <a href="https://understanding-ai-safety.org/">evidence-based</a> vision of the future of advanced AI that rejects both apocalyptic and utopian narratives.</p><h3><strong>How long did it take to write the book?</strong></h3><p>It depends on what one means by writing the book. The book is not just an explainer, and developing a book&#8217;s worth of genuinely new, scholarly ideas takes a long time. Here&#8217;s a brief timeline:</p><ul><li><p>2019: Arvind developed an <a href="https://www.cs.princeton.edu/~arvindn/talks/MIT-STS-AI-snakeoil.pdf">early version</a> of the high-level thesis of the book</p></li><li><p>2020: We started doing research and publishing papers that informed the book</p></li><li><p>mid-2022: Started writing the book and launched this newsletter</p></li><li><p>Sep 2023: Submitted the initial author manuscript</p></li><li><p>Jan 2024: Submitted the final author manuscript after addressing peer reviewers&#8217; feedback</p></li><li><p>May 2024: Final proofs done</p></li><li><p>Sep 2024: Publication</p></li></ul><h3><strong>What was the writing process like?</strong></h3><p>Doing the bulk of the writing in a year required a lot of things to go right. Here&#8217;s the process we used.</p><ul><li><p>We figured out the structure up front. Changes that affect multiple chapters are much harder to pull off than changes within a chapter. Since we&#8217;d been thinking about the topics of the book for years before we started writing, we already knew at a high level what we wanted to say.</p></li><li><p>Throughout, we had periodic check-ins with our editor, <a href="https://press.princeton.edu/our-people/hallie-stebbins">Hallie Stebbins</a>. Early on, Hallie helped us sanity check our decisions about structure, and sharing our progress with her gave us something to look forward to. In the later stages, her input was critical.</p></li><li><p>We divided up the chapters between us. Of course, we were both involved in every chapter, but it&#8217;s way less messy if one person takes the lead on each one. For this to work well, we had to both use the same &#8220;voice&#8221;. Can you tell who took the lead on which chapter?&nbsp;</p></li><li><p>We sent Hallie our drafts of each chapter as we completed them (after a couple of rounds of internal editing), instead of waiting till the end. We&#8217;re glad we did! Although we&#8217;re decent writers, Hallie had, on average, a couple of edits or suggestions per paragraph, mostly to fix awkward wording or point out something that was confusing.&nbsp;</p></li><li><p>While the line edits made the book dramatically more readable, even more important was her high-level feedback. Notably, she repeatedly asked us &#8220;how does this relate to the AI Snake Oil theme?&#8221; which helped keep us focused.</p></li><li><p>Oh, and Hallie couldn&#8217;t tell who took the lead on which chapter, which was a big relief!</p></li><li><p>We wrote the introductory chapter last. We know far more people will read the intro than the rest of the book, in part because it&#8217;s <a href="https://press.princeton.edu/books/hardcover/9780691249131/ai-snake-oil#preview">available online</a>, so we really wanted to get it right. This was easier to do at the end, once we knew exactly what the message of each chapter was.</p></li><li><p>The next step was peer review. We received reviews from Melanie Mitchell, Molly Crockett, Chris Bail, and three anonymous reviewers. Between them, they had over 30 pages of feedback, for which we are extremely grateful. It took a couple of months to address all of it, but we&#8217;re glad we did.</p></li><li><p>Overall, each chapter underwent 6-8 rounds of editing, including copyediting. That&#8217;s pretty normal!&nbsp;</p></li><li><p>There&#8217;s a lot of work that goes into publicizing the book. Between the two of us we&#8217;ve done about 50 talks, interviews, and podcasts in the last couple of months, and there&#8217;s a whole lot more that our publicist, <a href="https://press.princeton.edu/news/pw-star-watch-2023">Maria Whelan</a>, and others did for us behind the scenes!</p></li></ul><p>We hope you like the end result. Let us know what you think, in the comments or on <a href="https://www.amazon.com/Snake-Oil-Artificial-Intelligence-Difference/dp/069124913X">Amazon</a>.</p>]]></content:encoded></item><item><title><![CDATA[Can AI automate computational reproducibility?]]></title><description><![CDATA[A new benchmark to measure the impact of AI on improving science]]></description><link>https://www.normaltech.ai/p/can-ai-automate-computational-reproducibility</link><guid isPermaLink="false">https://www.normaltech.ai/p/can-ai-automate-computational-reproducibility</guid><dc:creator><![CDATA[Sayash Kapoor]]></dc:creator><pubDate>Wed, 18 Sep 2024 14:32:47 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/8f816374-794e-4011-b986-6c5221d489de_2908x1682.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Last month, <a href="https://sakana.ai/ai-scientist/">Sakana AI released an "AI scientist"</a>, which the company called "the first comprehensive system for fully automatic scientific discovery". It was <a href="https://venturebeat.com/ai/sakana-ai-scientist-conducts-research-autonomously-challenging-scientific-norms/">touted</a> as being able to accelerate science without suffering from human limitations.&nbsp;</p><p>Unfortunately, the "AI Scientist" has many <a href="https://x.com/jimmykoppel/status/1828077203956850756">shortcomings</a>. It has no checks for novelty, so generated papers could rehash earlier work. And Sakana did not perform any human review (let alone expert &#8220;peer&#8221; review) of the generated papers&#8212;so it is unclear if the papers are any good (apparently they are <a href="https://x.com/jimmykoppel/status/1828077203956850756">not</a>). While these flaws are particularly flagrant in Sakana's case, the <a href="https://arxiv.org/pdf/2407.01502">lack of good evaluation</a> affects most AI agents, making it hard to measure their real-world impact.</p><p>Today, we introduce a new benchmark for measuring how well AI can reproduce existing computational research. We also share how this project has changed our thinking&nbsp;about &#8220;general intelligence&#8221; and the potential economic impact of AI. <strong><a href="https://arxiv.org/pdf/2409.11363v1">Read the paper</a>.</strong></p><h3><strong>CORE-Bench: A new benchmark for evaluating AI for reproducing research</strong></h3><p>Visions of AI automating science are enticing, but aren&#8217;t within reach, and lead to <a href="https://www.aisnakeoil.com/p/scientists-should-use-ai-as-a-tool">flawed science</a>. In contrast, using AI for well-scoped tasks such as verifying computational reproducibility can save a lot of time and redirect effort towards more productive scientific activity. AI could also help find relevant literature, write code to rapidly test ideas, and perform other computational tasks.</p><p>In a <a href="https://arxiv.org/pdf/2409.11363v1">new paper</a>, we introduce <a href="https://arxiv.org/pdf/2409.11363v1">CORE-Bench</a> (<strong>Co</strong>mputational <strong>Re</strong>producibility Agent <strong>Bench</strong>mark), a benchmark for measuring how well AI can automate computational reproducibility, that is, reproducing a paper&#8217;s findings when the code and data are available. The authors are <a href="https://www.zacharysiegel.org/">Zachary S. Siegel</a>, <a href="https://www.cs.princeton.edu/~sayashk/">Sayash Kapoor</a>, <a href="https://citp.princeton.edu/citp-people/nitya-nadgir/">Nitya Nadgir</a>, <a href="https://benediktstroebl.github.io/">Benedikt Stroebl</a>, and <a href="https://www.cs.princeton.edu/~arvindn/">Arvind Narayanan</a>. CORE-Bench is a first step in a larger project to rigorously evaluate progress in automating research tasks of increasing difficulty.</p><p>Computationally reproducing a study is a far more limited task than replication, which requires re-running experiments that might involve human subjects. Even the limited reproducibility task is hard: In the <a href="https://zenodo.org/records/8200058">2022 Machine Learning Reproducibility Challenge</a>, over a third of the papers could not be reproduced even when experts reproducing the papers had the code and data.&nbsp;</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!PC_0!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fda369454-28e9-47a0-9e5b-f8218e63419c_1178x584.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!PC_0!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fda369454-28e9-47a0-9e5b-f8218e63419c_1178x584.png 424w, https://substackcdn.com/image/fetch/$s_!PC_0!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fda369454-28e9-47a0-9e5b-f8218e63419c_1178x584.png 848w, https://substackcdn.com/image/fetch/$s_!PC_0!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fda369454-28e9-47a0-9e5b-f8218e63419c_1178x584.png 1272w, https://substackcdn.com/image/fetch/$s_!PC_0!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fda369454-28e9-47a0-9e5b-f8218e63419c_1178x584.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!PC_0!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fda369454-28e9-47a0-9e5b-f8218e63419c_1178x584.png" width="1178" height="584" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/da369454-28e9-47a0-9e5b-f8218e63419c_1178x584.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:584,&quot;width&quot;:1178,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!PC_0!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fda369454-28e9-47a0-9e5b-f8218e63419c_1178x584.png 424w, https://substackcdn.com/image/fetch/$s_!PC_0!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fda369454-28e9-47a0-9e5b-f8218e63419c_1178x584.png 848w, https://substackcdn.com/image/fetch/$s_!PC_0!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fda369454-28e9-47a0-9e5b-f8218e63419c_1178x584.png 1272w, https://substackcdn.com/image/fetch/$s_!PC_0!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fda369454-28e9-47a0-9e5b-f8218e63419c_1178x584.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>If AI could automate this mundane yet important task, researchers could automate the implementation of baselines, reviewers could more easily assess if a paper has flaws, and journals and conferences could more easily verify if submitted and published papers are reproducible.</p><p>We created CORE-Bench using scientific papers and their accompanying code and data repositories. We used <a href="http://codeocean.com">Code Ocean</a> to source papers that were likely to be reproducible. We manually reproduced 90 papers from computer science, medicine, and social science, and curated a set of questions for each paper to be able to verify the answers.&nbsp;</p><p>We release CORE-Bench with three difficulty levels. Tasks in all three levels require the use of both language and vision capabilities. The hardest version closely resembles real-world reproduction attempts, and we expect that improvements on the benchmark will translate to agents that are actually useful to scientists.</p><p>To implement baselines, we tested the generalist AutoGPT agent and also implemented a task-specific modification to AutoGPT, which we call CORE-Agent. While the task-specific version improved accuracy significantly, there is still massive room for improvement: the best agent (CORE-Agent with GPT-4o) has an accuracy of 22% on CORE-Bench-Hard.</p><h3><strong>Rethinking generality</strong></h3><p>Computational reproducibility requires setting up the code environment correctly, running the code, and seeing if it produces the same results as reported in the paper. Using the shell and other tools correctly is still tricky for LLMs. When we evaluated generalist agents like AutoGPT, we weren't surprised by their poor accuracy (less than 10% on CORE-Bench-Hard).&nbsp;</p><p>Yet, with a few person-days of effort, we were able to build CORE-Agent by modifying AutoGPT, which more than doubled accuracy on the hardest level. We also built a task-specific agent from scratch, but modifying AutoGPT was far less time consuming while also resulting in a stronger agent. We are cautiously optimistic that this approach can be pushed to yield agents that perform well enough to be useful in practice.&nbsp;</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!ve-E!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F43867f9d-88aa-4a21-8270-c6b8808c6d53_1274x626.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!ve-E!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F43867f9d-88aa-4a21-8270-c6b8808c6d53_1274x626.png 424w, https://substackcdn.com/image/fetch/$s_!ve-E!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F43867f9d-88aa-4a21-8270-c6b8808c6d53_1274x626.png 848w, https://substackcdn.com/image/fetch/$s_!ve-E!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F43867f9d-88aa-4a21-8270-c6b8808c6d53_1274x626.png 1272w, https://substackcdn.com/image/fetch/$s_!ve-E!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F43867f9d-88aa-4a21-8270-c6b8808c6d53_1274x626.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!ve-E!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F43867f9d-88aa-4a21-8270-c6b8808c6d53_1274x626.png" width="1274" height="626" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/43867f9d-88aa-4a21-8270-c6b8808c6d53_1274x626.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:626,&quot;width&quot;:1274,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!ve-E!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F43867f9d-88aa-4a21-8270-c6b8808c6d53_1274x626.png 424w, https://substackcdn.com/image/fetch/$s_!ve-E!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F43867f9d-88aa-4a21-8270-c6b8808c6d53_1274x626.png 848w, https://substackcdn.com/image/fetch/$s_!ve-E!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F43867f9d-88aa-4a21-8270-c6b8808c6d53_1274x626.png 1272w, https://substackcdn.com/image/fetch/$s_!ve-E!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F43867f9d-88aa-4a21-8270-c6b8808c6d53_1274x626.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em>Simple task-specific modifications allow CORE-Agent to outperform AutoGPT.&nbsp;</em></figcaption></figure></div><p>If this pattern of being able to easily adapt a generalist agent to produce a task-specific agent holds in other areas, it should make us rethink generality. Generality roughly translates to being able to use the same model or agent without modification to perform a variety of tasks. This notion of generality underpins how Artificial General Intelligence (or AGI) is usually understood and the hopes and fears that accompany it.&nbsp;</p><p>But at least from the point of view of economic impacts, generality might be a red herring. For a task such as computational reproducibility on which expert humans collectively spend millions of hours every year, being able to automate it would be hugely impactful &#8212; regardless of whether the AI system did so out of the box, or after a few person days (or even a person year) of programmer effort.&nbsp;</p><p>In the AI Snake Oil <a href="https://www.aisnakeoil.com/p/starting-reading-the-ai-snake-oil">book</a>, we define generality as the inverse of task-specificity, and analyze how the history of AI (and computing) can be seen as the pursuit of gradually increasing generality. Increasing generality means decreasing the human effort it takes to build an AI system to perform a given task. From this perspective, systems like AutoGPT may be more general than most people (including us) gave them credit for.</p><p>Yet, definitions of AGI typically insist that a single system be able to do everything out of the box. There is no systematic effort to track how the human effort needed to build task-specific AI is changing over time. Just as we&#8217;ve <a href="https://www.aisnakeoil.com/p/ai-scaling-myths">argued</a> against flawed conceptions of generality that overestimate AI progress, we should avoid flawed conceptions of generality that underestimate it.&nbsp;</p><p><strong>Read the CORE-Bench paper <a href="https://arxiv.org/html/2409.11363v1">here</a>.</strong></p><h4><strong>Further reading</strong></h4><ul><li><p>In our recent paper, <a href="https://arxiv.org/pdf/2407.01502">AI Agents That Matter</a>, we found several shortcomings with AI agent evaluations. While building CORE-Bench, these shortcomings informed the design of our benchmark.</p></li><li><p>We recently organized an <a href="https://sites.google.com/princeton.edu/agents-workshop">online workshop on useful and reliable AI agents</a> where leading experts shared their views on better agent design and evaluation. The <a href="https://www.youtube.com/live/-aKRsvgDEz0">workshop videos are available online</a>.</p></li><li><p><a href="https://arxiv.org/pdf/2409.07440">Ben Bogin et al.</a> released the SUPER benchmark to evaluate if AI agents can set up and execute tasks from repositories accompanying research papers. It is another interesting benchmark for measuring AI agents' capability to automate research tasks. It differs from CORE-Bench in many ways:&nbsp;</p><ul><li><p>CORE-Bench consists of tasks across scientific disciplines (computer science, medicine, social science) whereas SUPER consists of tasks from AI.</p></li><li><p>CORE-Bench requires the use of both vision-language and language models, and consists of multiple languages (Python and R) as opposed to SUPER (language models, Python).</p></li><li><p>Tasks in SUPER require access to a Jupyter notebook. In contrast, tasks in CORE-Bench require shell access and allow the agent to modify the sandbox arbitrarily.</p></li></ul></li></ul>]]></content:encoded></item><item><title><![CDATA[Start reading the AI Snake Oil book online]]></title><description><![CDATA[The book was published September 2024]]></description><link>https://www.normaltech.ai/p/starting-reading-the-ai-snake-oil</link><guid isPermaLink="false">https://www.normaltech.ai/p/starting-reading-the-ai-snake-oil</guid><dc:creator><![CDATA[Arvind Narayanan]]></dc:creator><pubDate>Tue, 10 Sep 2024 20:55:03 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/3b84a7d7-b736-45ad-a1c6-8162dac6d732_4480x2520.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>The first chapter of the <a href="https://www.amazon.com/Snake-Oil-Artificial-Intelligence-Difference/dp/069124913X">AI snake oil book</a> is <a href="https://press.princeton.edu/books/hardcover/9780691249131/ai-snake-oil#preview">available online</a>. It is 30 pages long and summarizes the book&#8217;s main arguments. If you haven't ordered the book yet, we hope that reading the <a href="https://press.princeton.edu/books/hardcover/9780691249131/ai-snake-oil#preview">introductory chapter</a> will convince you to <a href="https://www.amazon.com/Snake-Oil-Artificial-Intelligence-Difference/dp/069124913X">get yourself a copy</a>.</p><p><strong>Update (September 2025):</strong> It has been a year since the release of AI Snake Oil. In the time since its release, the two of us have <a href="https://www.youtube.com/watch?v=C3TqcUEFR58">given</a> <a href="https://www.youtube.com/watch?v=CedHzg-VRiI">talks</a>, appeared on <a href="https://www.youtube.com/watch?v=M3U5UVyGTuQ">podcasts</a>, published <a href="https://www.normaltech.ai/p/ai-snake-oil-exercises-and-discussion">exercises to accompany the book</a>, and written a new preface and epilogue for the <a href="https://www.amazon.com/Snake-Oil-Artificial-Intelligence-Difference/dp/0691249148/">paperback edition of the book</a>. The book was included in Nature&#8217;s list of the 10 best books of 2024, Bloomberg&#8217;s 49 best books of 2024, and Forbes&#8217;s 10 must-read tech books of 2024. It has received many positive reviews, including in the <a href="https://www.newyorker.com/culture/open-questions/in-the-age-of-ai-what-makes-people-unique">New Yorker</a>. We are grateful to readers of the book for engaging deeply with its ideas.</p><p>We have now started working on our next project together, <a href="https://www.normaltech.ai/p/ai-as-normal-technology">AI as Normal Technology</a>. The project picks up where AI Snake Oil left off: whereas AI Snake Oil was an attempt to understand the present and near-term impacts of AI, AI as Normal Technology is a framework to think about its future impacts<em>. </em>The new name of this newsletter reflects this change. We hope you will follow along.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://www.amazon.com/Snake-Oil-Artificial-Intelligence-Difference/dp/069124913X" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!XdEd!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb3f7d684-cc18-4d05-96a2-f19a245889ad_962x1448.png 424w, https://substackcdn.com/image/fetch/$s_!XdEd!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb3f7d684-cc18-4d05-96a2-f19a245889ad_962x1448.png 848w, https://substackcdn.com/image/fetch/$s_!XdEd!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb3f7d684-cc18-4d05-96a2-f19a245889ad_962x1448.png 1272w, https://substackcdn.com/image/fetch/$s_!XdEd!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb3f7d684-cc18-4d05-96a2-f19a245889ad_962x1448.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!XdEd!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb3f7d684-cc18-4d05-96a2-f19a245889ad_962x1448.png" width="346" height="520.7983367983368" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/b3f7d684-cc18-4d05-96a2-f19a245889ad_962x1448.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1448,&quot;width&quot;:962,&quot;resizeWidth&quot;:346,&quot;bytes&quot;:1409620,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:&quot;https://www.amazon.com/Snake-Oil-Artificial-Intelligence-Difference/dp/069124913X&quot;,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!XdEd!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb3f7d684-cc18-4d05-96a2-f19a245889ad_962x1448.png 424w, https://substackcdn.com/image/fetch/$s_!XdEd!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb3f7d684-cc18-4d05-96a2-f19a245889ad_962x1448.png 848w, https://substackcdn.com/image/fetch/$s_!XdEd!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb3f7d684-cc18-4d05-96a2-f19a245889ad_962x1448.png 1272w, https://substackcdn.com/image/fetch/$s_!XdEd!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb3f7d684-cc18-4d05-96a2-f19a245889ad_962x1448.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h3><strong>The single most confusing thing about AI</strong></h3><p>Our book is about demystifying AI, so right out of the gate we address what we think is the single most confusing thing about it:&nbsp;</p><p></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://press.princeton.edu/books/hardcover/9780691249131/ai-snake-oil#preview" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!h-oz!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc6d1430a-2af9-4dda-9a2a-6df1a2694d4f_1290x1732.png 424w, https://substackcdn.com/image/fetch/$s_!h-oz!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc6d1430a-2af9-4dda-9a2a-6df1a2694d4f_1290x1732.png 848w, https://substackcdn.com/image/fetch/$s_!h-oz!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc6d1430a-2af9-4dda-9a2a-6df1a2694d4f_1290x1732.png 1272w, https://substackcdn.com/image/fetch/$s_!h-oz!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc6d1430a-2af9-4dda-9a2a-6df1a2694d4f_1290x1732.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!h-oz!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc6d1430a-2af9-4dda-9a2a-6df1a2694d4f_1290x1732.png" width="482" height="647.1503875968992" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/c6d1430a-2af9-4dda-9a2a-6df1a2694d4f_1290x1732.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1732,&quot;width&quot;:1290,&quot;resizeWidth&quot;:482,&quot;bytes&quot;:428419,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:&quot;https://press.princeton.edu/books/hardcover/9780691249131/ai-snake-oil#preview&quot;,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!h-oz!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc6d1430a-2af9-4dda-9a2a-6df1a2694d4f_1290x1732.png 424w, https://substackcdn.com/image/fetch/$s_!h-oz!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc6d1430a-2af9-4dda-9a2a-6df1a2694d4f_1290x1732.png 848w, https://substackcdn.com/image/fetch/$s_!h-oz!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc6d1430a-2af9-4dda-9a2a-6df1a2694d4f_1290x1732.png 1272w, https://substackcdn.com/image/fetch/$s_!h-oz!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc6d1430a-2af9-4dda-9a2a-6df1a2694d4f_1290x1732.png 1456w" sizes="100vw"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em>AI is an umbrella term for a set of loosely related technologies</em></figcaption></figure></div><p>Because AI is an umbrella term, we treat each type of AI differently. We have chapters on predictive AI, generative AI, as well as AI used for social media content moderation. We also have a chapter on whether AI is an existential risk. We conclude with a discussion of why AI snake oil persists and what the future might hold. By AI snake oil we mean AI applications that do not (and perhaps cannot) work. Our book is a guide to identifying AI snake oil and AI hype. We also look at AI that is harmful even if it works well &#8212;&nbsp;such as face recognition used for mass surveillance.&nbsp;</p><p>While the book is meant for a broad audience, it does not simply rehash the arguments we have made in our papers or on this newsletter. We make scholarly contributions and we wrote the book to be suitable for adoption in courses. We will soon release exercises and class discussion questions to accompany the book.</p><p></p><h3><strong>What's in the book</strong></h3><p><strong>Chapter 1: Introduction. </strong>We begin with a summary of our main arguments in the book. We discuss the definition of AI (and more importantly, why it is hard to come up with one), how AI is an umbrella term, what we mean by AI Snake Oil, and who the book is for.&nbsp;</p><p>Generative AI has made huge strides in the last decade. On the other hand, predictive AI is used for predicting outcomes to make consequential decisions in hiring, banking, insurance, education, and more. While predictive AI can find broad statistical patterns in data, it is marketed as far more than that, leading to major real-world misfires. Finally, we discuss the benefits and limitations of AI for content moderation on social media.</p><p>We also tell the story of what led the two of us to write the book. <strong>The entire first chapter is now <a href="https://press.princeton.edu/books/hardcover/9780691249131/ai-snake-oil#preview">available online</a>.</strong></p><p><strong>Chapter 2: How predictive AI goes wrong. </strong>Predictive AI is used to make predictions about people&#8212;will a defendant fail to show up for trial? Is a patient at high risk of negative health outcomes? Will a student drop out of college? These predictions are then used to make consequential decisions. Developers claim predictive AI is groundbreaking, but in reality it suffers from a number of shortcomings that are hard to fix.&nbsp;</p><p>We have discussed the <a href="https://www.aisnakeoil.com/p/ai-cannot-predict-the-future-but">failures of predictive AI</a> in this blog. But in the book, we go much deeper through case studies to show how predictive AI fails to live up to the promises made by its developers.</p><p><strong>Chapter 3: Can AI predict the future? </strong>Are the shortcomings of predictive AI inherent, or can they be resolved? In this chapter, we look at why predicting the future is hard &#8212; with or without AI. While we have made consistent progress in some domains such as weather prediction, we argue that this progress cannot translate to other settings, such as individuals' life outcomes, the success of cultural products like books and movies, or pandemics.&nbsp;</p><p>Since much of our newsletter is focused on topics of current interest, this is a topic that we have never written about here. Yet, it is foundational knowledge that can help you build intuition around when we should expect predictions to be accurate.</p><p><strong>Chapter 4: The long road to generative AI. </strong>Recent advances in generative AI can seem sudden, but they build on a series of improvements over seven decades. In this chapter, we retrace the history of computing advances that led to generative AI. While we have written a lot about current trends in generative AI, in the book, we look at its past. This is crucial for understanding what to expect in the future.&nbsp;</p><p><strong>Chapter 5: Is advanced AI an existential threat? </strong>Claims about AI wiping out humanity are common. Here, we critically evaluate claims about AI's existential risk and find several shortcomings and fallacies in popular discussion of x-risk. We discuss approaches to defending against AI risks that improve societal resilience regardless of the threat of advanced AI.</p><p><strong>Chapter 6: Why can't AI fix social media? </strong>One area where AI is heavily used is content moderation on social media platforms. We discuss the current state of AI use on social media, and highlight seven reasons why improvements in AI alone are unlikely to solve platforms' content moderation woes. We haven't written about content moderation in this newsletter.</p><p><strong>Chapter 7: Why do myths about AI persist? </strong>Companies, researchers, and journalists all contribute to AI hype. We discuss how myths about AI are created and how they persist. In the process, we hope to give you the tools to read AI news with the appropriate skepticism and identify attempts to sell you snake oil.</p><p><strong>Chapter 8: Where do we go from here? </strong>While the previous chapter focuses on the <em>supply</em> of snake oil, in the last chapter, we look at where the <em>demand</em> for AI snake oil comes from. We also look at the impact of AI on the future of work, the role and limitations of regulation, and conclude with vignettes of the many possible futures ahead of us. We have the agency to determine which path we end up on, and each of us can play a role.</p><p>We hope you will find the book useful and look forward to hearing what you think.&nbsp;</p><h4><strong>Early reviews</strong></h4><ul><li><p><a href="https://www.newyorker.com/culture/open-questions/in-the-age-of-ai-what-makes-people-unique">The New Yorker</a>: "In AI Snake Oil, Arvind Narayanan and Sayash Kapoor urge skepticism and argue that the blanket term AI can serve as a smokescreen for underperforming technologies."</p></li><li><p><a href="https://www.kirkusreviews.com/book-reviews/arvind-narayanan/ai-snake-oil/">Kirkus</a>: "Highly useful advice for those who work with or are affected by AI&#8212;i.e., nearly everyone."</p></li><li><p><a href="https://www.publishersweekly.com/pw/by-topic/new-titles/adult-announcements/article/95302-fall-2024-adult-preview-science.html">Publishers' Weekly</a>: Featured in the Fall 2024 list of top science books.</p></li><li><p><a href="https://www.practicalecommerce.com/ai-snake-oil-sorts-promise-from-hype">Jean Gazis</a>: "The authors admirably differentiate fact from opinion, draw from personal experience, give sensible reasons for their views (including copious references), and don&#8217;t hesitate to call for action. . . . If you&#8217;re curious about AI or deciding how to implement it, AI Snake Oil offers clear writing and level-headed thinking."</p></li><li><p><a href="https://www.sciencenews.org/article/ai-snake-oil-how-to-spot-hype">Elizabeth Quill</a>: "A worthwhile read whether you make policy decisions, use AI in the workplace or just spend time searching online. It&#8217;s a powerful reminder of how AI has already infiltrated our lives &#8212; and a convincing plea to take care in how we interact with it."</p></li><li><p><a href="https://www.telegraph.co.uk/books/non-fiction/review-how-ai-change-life-patrick-dixon-snake-oil-kapoor/">The Telegraph</a></p></li></ul><h4><strong>Book launch events</strong></h4><ul><li><p>September 24: <a href="https://www.eventbrite.com/e/arvind-narayanan-sayash-kapoor-ai-snake-oil-tickets-974088092707">City Lights</a> (virtual, free)</p></li><li><p>September 30: <a href="https://alumni.princeton.edu/events/tigerside-chat-ai-snake-oil">Princeton alumni events</a> (virtual, free)</p></li><li><p>October 24: <a href="https://princetonlibrary.libnet.info/event/11393063">Princeton Public Library</a> (Princeton, free)</p></li></ul><h4><strong>Podcasts and interviews</strong></h4><ul><li><p><a href="https://www.youtube.com/watch?v=BGvQmHd4QPE">Machine Learning Street Talk</a></p></li><li><p><a href="https://www.thetwentyminutevc.com/arvind-narayanan/">20VC</a></p></li><li><p><a href="https://www.youtube.com/watch?v=stJaSvQoIVQ">Scaling Theory</a></p></li><li><p><a href="https://cap.csail.mit.edu/podcasts/hype-vs-reality-current-state-ai-arvind-narayanan">MIT CSAIL</a></p></li><li><p><a href="https://www.abc.net.au/listen/programs/futuretense/artificial-intelligence-its-limits-risks-nd-thirst-for-resources/104238774">Future Tense</a></p></li></ul><p>We&#8217;ve been on many other podcasts that will air around the time of the book&#8217;s release, and we will keep this list updated.</p><h4><strong>Purchase links</strong></h4><ul><li><p>US: <a href="https://www.amazon.com/Snake-Oil-Artificial-Intelligence-Difference/dp/069124913X">Amazon</a>, <a href="https://bookshop.org/p/books/ai-snake-oil-what-artificial-intelligence-can-do-what-it-can-t-and-how-to-tell-the-difference-arvind-narayanan/21324674">Bookshop</a>, <a href="https://www.barnesandnoble.com/w/ai-snake-oil-arvind-narayanan/1145168436">Barnes and Noble</a>, <a href="https://press.princeton.edu/books/hardcover/9780691249131/ai-snake-oil">Princeton University Press</a>. <a href="https://www.audible.com/pd/AI-Snake-Oil-Audiobook/B0D8BZSM1F">Audiobook</a>, <a href="https://www.amazon.com/Snake-Oil-Artificial-Intelligence-Difference-ebook/dp/B0CW1JCKVL/">Kindle</a> editions.&nbsp;</p></li><li><p>UK: <a href="https://blackwells.co.uk/bookshop/product/AI-Snake-Oil-by-Arvind-Narayanan-Sayash-Kapoor/9780691249131">Blackwell&#8217;s</a>, <a href="https://www.waterstones.com/book/ai-snake-oil/arvind-narayanan/sayash-kapoor/9780691249131">Waterstones</a>.&nbsp;</p></li><li><p>Canada: <a href="https://www.indigo.ca/en-ca/ai-snake-oil-what-artificial-intelligence-can-do-what-it-cant-and-how-to-tell-the-difference/9780691249131.html">Indigo</a>.&nbsp;</p></li><li><p>Germany: <a href="https://www.amazon.de/-/en/Arvind-Narayanan/dp/069124913X/">Amazon</a>, <a href="https://www.kulturkaufhaus.de/en/detail/ISBN-9780691249131/Narayanan-Arvind/AI-Snake-Oil">Kulturkaufhaus</a>.&nbsp;</p></li><li><p>India: <a href="https://www.amazon.in/Snake-Oil-Artificial-Intelligence-Difference/dp/0691269947">Amazon</a></p></li></ul><p>The book is available to preorder internationally on <a href="https://www.amazon.com/Snake-Oil-Artificial-Intelligence-Difference/dp/069124913X">Amazon</a>.</p>]]></content:encoded></item><item><title><![CDATA[AI companies are pivoting from creating gods to building products. Good.]]></title><description><![CDATA[Turning models into products runs into five challenges]]></description><link>https://www.normaltech.ai/p/ai-companies-are-pivoting-from-creating</link><guid isPermaLink="false">https://www.normaltech.ai/p/ai-companies-are-pivoting-from-creating</guid><dc:creator><![CDATA[Arvind Narayanan]]></dc:creator><pubDate>Mon, 19 Aug 2024 20:57:05 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/29486da2-0bac-4f68-bafd-fc41d1356baa_2552x1429.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>AI companies are collectively planning to spend a <a href="https://www.goldmansachs.com/images/migrated/insights/pages/gs-research/gen-ai--too-much-spend,-too-little-benefit-/TOM_AI%202.0_ForRedaction.pdf">trillion dollars</a> on hardware and data centers, but there&#8217;s been relatively little to show for it so far. This has led to a chorus of concerns that generative AI is a <a href="https://www.inc.com/sam-blum/new-warnings-ai-bubble-when-could-it-burst.html">bubble</a>. We won&#8217;t offer any predictions on what&#8217;s about to happen. But we think we have a solid diagnosis of how things got to this point in the first place.</p><p>In this post, we explain the mistakes that AI companies have made and how they have been trying to correct them. Then we will talk about five barriers they still have to overcome in order to make generative AI commercially successful enough to justify the investment.</p><h3><strong>Product-market fit</strong></h3><p>When ChatGPT launched, people found a thousand unexpected uses for it. This got AI developers overexcited. They completely misunderstood the market, underestimating the huge gap between proofs of concept and reliable products. This misunderstanding led to two opposing but equally flawed approaches to commercializing LLMs.&nbsp;</p><p>OpenAI and Anthropic focused on building models and not worrying about products. For example, it took 6 months for OpenAI to bother to release a ChatGPT iOS app and 8 months for an Android app!</p><p>Google and Microsoft shoved AI into everything in a panicked race, without thinking about which products would actually benefit from AI and how they should be integrated.</p><p>Both groups of companies forgot the &#8220;make something people want&#8221; mantra. The generality of LLMs allowed developers to fool themselves into thinking that they were exempt from the need to find a product-market fit, as if prompting a model to perform a task is a replacement for carefully designed products or features.</p><p>OpenAI and Anthropic&#8217;s DIY approach meant that early adopters of LLMs disproportionately tended to be bad actors, since they are more invested in figuring out how to adapt new technologies for their purposes, whereas everyday users want easy-to-use products. This has contributed to a poor public perception of the technology.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a></p><p>Meanwhile the AI-in-your-face approach by Microsoft and Google has led to features that are occasionally useful and more often annoying. It also led to many unforced errors due to inadequate testing like Microsoft's early <a href="https://www.nytimes.com/2023/02/16/technology/bing-chatbot-microsoft-chatgpt.html">Sydney</a> chatbot and Google's <a href="https://www.theverge.com/2024/2/21/24079371/google-ai-gemini-generative-inaccurate-historical">Gemini</a> image generator. This has also caused a backlash.</p><p>But companies are changing their ways. OpenAI seems to be transitioning from a research lab focused on a speculative future to something resembling a regular product company. If you take all the human-interest elements out of the OpenAI boardroom drama, it was fundamentally about the company's shift from creating gods to building products. Anthropic has been picking up many of the researchers and developers at OpenAI who cared more about artificial general intelligence and felt out of place at OpenAI, although Anthropic, too, has recognized the need to build products.</p><p>Google and Microsoft are slower to learn, but our guess is that <a href="https://www.vox.com/technology/354794/apple-artificial-intelligence-ai-wwdc">Apple</a> will force them to change. Last year Apple was seen as a laggard on AI, but it seems clear in retrospect that the slow and thoughtful approach that Apple showcased at WWDC, its developer conference, is more likely to resonate with users.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a> Google seems to have put more thought into integrating AI in its upcoming <a href="https://blog.google/products/pixel/google-pixel-9-pro-xl/">Pixel</a> phones and Android than it did into integrating it in search, but the phones aren&#8217;t out yet, so let&#8217;s see.&nbsp;</p><p>And then there&#8217;s Meta, whose vision is to use AI to create content and engagement on its ad-driven social media platforms. The societal implications of a world awash in AI-generated content are <a href="https://knightcolumbia.org/content/how-to-prepare-for-the-deluge-of-generative-ai-on-social-media">double-edged</a>, but from a business perspective it makes sense.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.normaltech.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">You&#8217;re reading AI Snake Oil, a newsletter about our <a href="https://www.aisnakeoil.com/p/ai-snake-oil-is-now-available-to">book</a>. Subscribe to get new posts.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><h3><strong>The big five challenges for consumer AI</strong></h3><p>There are five limitations of LLMs that developers need to tackle in order to make compelling AI-based consumer products.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-3" href="#footnote-3" target="_self">3</a> (We will discuss many of these in our upcoming <a href="https://sites.google.com/princeton.edu/agents-workshop">online workshop</a> on building useful and reliable AI agents on August 29.)</p><h4><strong>1. Cost</strong></h4><p>There are many applications where capability is not the barrier, cost is. Even in a simple chat application, cost concerns dictate how much history a bot can keep track of &#8212; processing the entire history for every response quickly gets prohibitively expensive as the conversation grows longer.</p><p>There has been rapid progress on cost &#8212; in the last 18 months, cost-for-equivalent-capability has dropped by a factor of over 100.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-4" href="#footnote-4" target="_self">4</a> As a result, companies are claiming that LLMs are, or will soon be, &#8220;<a href="https://x.com/sama/status/1813984333352649087">too cheap to meter</a>&#8221;. Well, we&#8217;ll believe it when they make the API free.&nbsp;</p><p>More seriously, the reason we think cost will continue to be a concern is that in many applications, cost improvements directly translate to accuracy improvements. That&#8217;s because repeatedly retrying a task tens, thousands, or even millions of times turns out to be a good way to improve the chances of success, given the randomness of LLMs. So the cheaper the model, the more retries we can make with a given budget. We quantified this in our <a href="https://www.aisnakeoil.com/p/new-paper-ai-agents-that-matter">recent paper</a> on agents; since then, many other papers have made <a href="https://arxiv.org/pdf/2407.21787v1">similar</a> <a href="https://arxiv.org/html/2408.03314v1">points</a>.</p><p>That said, it is plausible that we&#8217;ll soon get to a point where in <em>most</em> applications, cost optimization isn&#8217;t a serious concern.</p><h4><strong>2. Reliability</strong></h4><p>We see capability and reliability as somewhat orthogonal. If an AI system performs a task correctly 90% of the time, we can say that it is <em>capable</em> of performing the task but it cannot do so <em>reliably</em>. The techniques that get us to 90% are unlikely to get us to 100%.&nbsp;</p><p>With statistical learning based systems, perfect accuracy is intrinsically hard to achieve. If you think about the success stories of machine learning, like ad targeting or fraud detection or, more recently, weather forecasting, perfect accuracy isn&#8217;t the goal &#8212; as long as the system is better than the state of the art, it is useful. Even in medical diagnosis and other healthcare applications, we <a href="https://www.himss.org/news/north-carolina-hospital-system-reduces-sepsis-cases-using-predictive-analytics">tolerate</a> a lot of error.&nbsp;</p><p>But when developers put AI in consumer products, people expect it to behave like software, which means that it needs to work deterministically. If your AI travel agent books vacations to the correct destination only 90% of the time, it won&#8217;t be successful. As we&#8217;ve <a href="https://www.aisnakeoil.com/p/new-paper-ai-agents-that-matter">written before</a>, reliability limitations partly explain the failures of recent AI-based gadgets.&nbsp;</p><p>AI developers have been slow to recognize this because as experts, we are used to conceptualizing AI as fundamentally different from traditional software. For example, the two of us are heavy users of chatbots and agents in our everyday work, and it has become almost automatic for us to work around the hallucinations and unreliability of these tools. A year ago, AI developers hoped or assumed that non-expert users would learn to adapt to AI, but it has gradually become clear that companies will have to adapt AI to user expectations instead, and make AI behave like traditional software.</p><p>Improving reliability is a research interest of our team at Princeton. For now, it&#8217;s fundamentally an open question whether it&#8217;s possible to build deterministic systems out of stochastic components (LLMs). Some companies have claimed to have solved reliability &#8212; for example, legal tech vendors have touted &#8220;hallucination-free&#8221; systems. But these claims were shown to be <a href="https://dho.stanford.edu/wp-content/uploads/Legal_RAG_Hallucinations.pdf">premature</a>.</p><h4><strong>3. Privacy</strong></h4><p>Historically, machine learning has often relied on sensitive data sources such browsing histories for ad targeting or medical records for <a href="https://www.wired.com/story/google-deepmind-nhs-health-data/">health tech</a>. In this sense, LLMs are a bit of an anomaly, since they are primarily trained on public sources such as web pages and books.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-5" href="#footnote-5" target="_self">5</a></p><p>But with AI assistants, privacy concerns have come roaring back. To build useful assistants, companies have to train systems on user interactions. For example, to be good at composing emails, it would be very helpful if models were <a href="https://www.nytimes.com/2024/06/26/technology/terms-service-ai-training.html">trained on emails</a>. Companies&#8217; privacy policies are vague about this and it is not clear to what extent this is happening.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-6" href="#footnote-6" target="_self">6</a> Emails, documents, screenshots, etc. are potentially much more sensitive than chat interactions.&nbsp;</p><p>There is a distinct type of privacy concern relating to inference rather than training. For assistants to do useful things for us, they must have access to our personal data. For example, Microsoft announced a controversial feature that would involve taking screenshots of users&#8217; PCs every few seconds, in order to give its CoPilot AI a memory of your activities. But there was an outcry and the company <a href="https://www.bbc.com/news/articles/cd11rje1mrro">backtracked</a>.</p><p>We caution against purely technical interpretations of privacy such as &#8220;the data never leaves the device.&#8221; Meredith Whittaker <a href="https://mobile.x.com/mer__edith/status/1790692059059200017">argues</a> that on-device fraud detection normalizes always-on surveillance and that the infrastructure can be repurposed for more oppressive purposes. That said, <a href="https://www.wired.com/story/apple-intelligence-android-hybrid-ai-privacy/">technical innovations</a> can definitely help.</p><h4><strong>4. Safety and security</strong></h4><p>There is a cluster of related concerns when it comes to safety and security: unintentional failures such as the <a href="https://www.cnbc.com/2024/02/26/googles-gemini-ai-picture-generator-to-relaunch-in-a-few-weeks.html">biases</a> in Gemini&#8217;s image generation; misuses of AI such as voice cloning or deepfakes; and hacks such as prompt injection that can leak users&#8217; data or harm the user in other ways.</p><p>We think accidental failures are fixable. As for most types of misuses, our view is that there is <a href="https://www.aisnakeoil.com/p/ai-safety-is-not-a-model-property">no way</a> to create a model that can&#8217;t be misused and so the defenses must primarily be located downstream. Of course, not everyone agrees, so companies will keep getting bad press for inevitable misuses, but they seem to have absorbed this as a cost of doing business.&nbsp;</p><p>Let&#8217;s talk about the third category &#8212; hacking. From what we can tell, it is the one that companies seem to be paying the least attention to. At least theoretically, <a href="https://arxiv.org/abs/2302.12173">catastrophic hacks</a> are possible, such as AI worms that spread from user to user, tricking those users&#8217; AI assistants into doing harmful things including creating more copies of the worm.&nbsp;</p><p>Although there have been plenty of proof-of-concept demonstrations and <a href="https://embracethered.com/blog/posts/2023/google-bard-data-exfiltration/">bug</a> <a href="https://www.landh.tech/blog/20240304-google-hack-50000/">bounties</a> that uncovered these vulnerabilities in deployed products, we haven't seen this type of attack in the wild. We aren&#8217;t sure if this is because of the low adoption of AI assistants, or because the <a href="https://kai-greshake.de/posts/approaches-to-pi-defense/">clumsy defenses</a> that companies have pulled together have proven sufficient, or something else. Time will tell.</p><h4><strong>5. User interface</strong></h4><p>In many applications, the unreliability of LLMs means that there will have to be some way for the user to intervene if the bot goes off track. In a chatbot, it can be as simple as regenerating an answer or showing multiple versions and letting the user pick. But in applications where errors can be costly, such as flight booking, ensuring adequate supervision is more tricky, and the system must avoid annoying the user with too many interruptions.</p><p>The problem is even harder with natural language interfaces where the user speaks to the assistant and the assistant speaks back. This is where a lot of the potential of generative AI lies. As just one example, AI that disappeared into your <a href="https://www.techradar.com/computing/artificial-intelligence/i-finally-tried-the-meta-ai-in-my-ray-ban-smart-glasses-thanks-to-an-accidental-uk-launch-and-its-by-far-the-best-ai-wearable">glasses</a> and spoke to you when you needed it, without even being asked &#8212; such as by detecting that you were staring at a sign in a foreign language &#8212; would be a whole different experience than what we have today. But the constrained user interface leaves very little room for incorrect or unexpected behavior.</p><h4><strong>Concluding thoughts</strong></h4><p>AI boosters often claim that due to the rapid pace of improvement in AI capabilities, we should see massive societal and economic effects soon. We are <a href="https://www.aisnakeoil.com/p/ai-scaling-myths">skeptical</a> of the trend extrapolation and sloppy thinking that goes into those capability forecasts. More importantly, even if AI capability does improve rapidly, developers have to solve the challenges discussed above. These are sociotechnical and not purely technical, so progress will be slow. And even if those challenges are solved, organizations need to integrate AI into existing products and workflows and train people to use it productively while avoiding its pitfalls. We should expect this to happen on a timescale of a decade or more rather than a year or two.&nbsp;</p><h4><strong>Further reading</strong></h4><p>Benedict Evans has <a href="https://www.ben-evans.com/benedictevans/2023/10/5/unbundling-ai">written</a> <a href="https://www.ben-evans.com/benedictevans/2024/6/8/building-ai-products">about</a> the importance of building single-purpose software using general-purpose language models.&nbsp;</p><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>To be clear, we don't think that reducing access to state-of-the-art models will <a href="https://www.aisnakeoil.com/p/licensing-is-neither-feasible-nor">reduce</a> misuse. But when it comes to LLMs, misuse is easier than legitimate uses (which require thought), so it isn't a surprise that misuses have been widespread.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p>The pace of AI adoption is relative. Even Apple's approach to integrating AI into its products has been <a href="https://www.newyorker.com/culture/infinite-scroll/apple-is-bringing-ai-to-your-personal-life-like-it-or-not">criticized</a> as too fast-paced.&nbsp;</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-3" href="#footnote-anchor-3" class="footnote-number" contenteditable="false" target="_self">3</a><div class="footnote-content"><p>These are about factors that matter to the user experience; we are setting aside environmental costs, training on copyrighted data, etc.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-4" href="#footnote-anchor-4" class="footnote-number" contenteditable="false" target="_self">4</a><div class="footnote-content"><p>For example, GPT-3.5 (text-davinci-003) in the API cost $20 per million tokens, whereas gpt-4o-mini, which is more powerful, costs only 15 cents.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-5" href="#footnote-anchor-5" class="footnote-number" contenteditable="false" target="_self">5</a><div class="footnote-content"><p>To be clear, just because the data sources are public <a href="https://arxiv.org/pdf/2212.06470">doesn&#8217;t mean</a> there are no privacy concerns.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-6" href="#footnote-anchor-6" class="footnote-number" contenteditable="false" target="_self">6</a><div class="footnote-content"><p>For example, Google says &#8220;we use publicly available information to help train Google&#8217;s AI models&#8221;. Elsewhere it says that it may use private data such as emails to provide services, maintain and improve services, personalize services, and develop new services. One approach that is consistent with these disclosures is that only public data is used for pre-training of models like Gemini, but private data is used to fine-tune those models to create, say, an email auto-response bot. Anthropic is the one exception we know of. It says: &#8220;We do not train our generative models on user-submitted data unless a user gives us explicit permission to do so. To date we have not used any customer or user-submitted data to train our generative models.&#8221; This commitment to privacy is admirable, though we predict that it will put the company at a disadvantage if it more fully embraces building products.</p><p></p></div></div>]]></content:encoded></item><item><title><![CDATA[AI existential risk probabilities are too unreliable to inform policy]]></title><description><![CDATA[How speculation gets laundered through pseudo-quantification]]></description><link>https://www.normaltech.ai/p/ai-existential-risk-probabilities</link><guid isPermaLink="false">https://www.normaltech.ai/p/ai-existential-risk-probabilities</guid><dc:creator><![CDATA[Arvind Narayanan]]></dc:creator><pubDate>Fri, 26 Jul 2024 11:29:25 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/6418811e-f1c2-4127-b6dd-f243c4f458ef_1398x918.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>How seriously should governments take the threat of existential risk from AI, given the lack of consensus among researchers? On the one hand, existential risks (x-risks) are necessarily somewhat speculative: by the time there is concrete evidence, it may be too late. On the other hand, governments must prioritize &#8212; after all, they don&#8217;t worry too much about x-risk from alien invasions.</p><p>This is the first in a series of essays laying out an evidence-based approach for policymakers concerned about AI x-risk, an approach that stays grounded in reality while acknowledging that there are &#8220;unknown unknowns&#8221;.&nbsp;</p><p>In this first essay, we look at one type of evidence: probability estimates. The AI safety community relies heavily on forecasting the probability of human extinction due to AI (in a given timeframe) in order to inform decision making and policy. An estimate of 10% over a few decades, for example, would obviously be high enough for the issue to be a top priority for society.&nbsp;</p><p>Our central claim is that AI x-risk forecasts are far too unreliable to be useful for policy, and in fact highly misleading.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.normaltech.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Subscribe to receive future essays in this series.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><h3><strong>Look behind the curtain</strong></h3><p>If the two of us predicted an 80% probability of aliens landing on earth in the next ten years, would you take this possibility seriously? Of course not. You would ask to see our evidence. As obvious as this may seem, it seems to have been forgotten in the AI x-risk debate that probabilities carry no authority by themselves. Probabilities are <em>usually</em> derived from some grounded method, so we have a strong cognitive bias to view quantified risk estimates as more valid than qualitative ones. But it is possible for probabilities to be nothing more than guesses. Keep this in mind throughout this essay (and more broadly in the AI x-risk debate).</p><p>If we predicted odds for the Kentucky Derby, we don&#8217;t have to give you a reason &#8212; you can take it or leave it. But if a policymaker takes actions based on probabilities put forth by a forecaster, they had better be able to explain those probabilities to the public (and that explanation must in turn come from the forecaster). <a href="https://plato.stanford.edu/entries/justification-public/">Justification</a> is essential to legitimacy of government and the exercise of power. A core principle of liberal democracy is that the state should not limit people's freedom based on controversial beliefs that reasonable people can reject.&nbsp;</p><p>Explanation is especially important when the policies being considered are costly, and even more so when those costs are unevenly distributed among stakeholders. A good example is restricting open releases of AI models. Can governments convince people and companies who stand to benefit from open models that they should make this sacrifice because of a speculative future risk?</p><p>The main aim of this essay is analyzing whether there is any justification for any of the specific x-risk probability estimates that have been cited in the policy debate. We have no objection to AI x-risk forecasting as an academic activity, and forecasts may be helpful to companies and other private decision makers. We only question its use in the context of public policy.</p><p>There are basically only three known ways by which a forecaster can try to convince a skeptic: inductive, deductive, and subjective probability estimation. We consider each of these in the following sections. All three require both parties to agree on some basic assumptions about the world (which cannot themselves be proven). The three approaches differ in terms of the empirical and logical ways in which the probability estimate follows from that set of assumptions.</p><h3><strong>Inductive probability estimation is unreliable due to the lack of a reference class</strong></h3><p>Most risk estimates are inductive: they are based on past observations. For example, insurers base their predictions of an individual&#8217;s car accident risk on data from past accidents about similar drivers. The set of observations used for probability estimation is called a reference class. A suitable reference class for car insurance might be the set of drivers who live in the same city. If the analyst has more information about the individual, such as their age or the type of car they drive, the reference class can be further refined.&nbsp;</p><p>For existential risk from AI, there is no reference class, as it is an event like no other. To be clear, this is a matter of degree, not kind. There is never a clear &#8220;correct&#8221; reference class to use, and the choice of a reference class in practice comes down to the analyst&#8217;s intuition.&nbsp;</p><p>The accuracy of the forecasts depends on the degree of similarity between the process that generates the event being forecast and the process that generated the events in the reference class, which can be seen as a spectrum. For predicting the outcome of a physical system such as a coin toss, past experience is a highly reliable guide. Next, for car accidents, risk estimates might vary by, say, 20% based on the past dataset used &#8212; good enough for insurance companies.&nbsp;</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!drZV!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2f3c0d76-b6af-4713-8bff-5c0f144c4f04_1238x334.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!drZV!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2f3c0d76-b6af-4713-8bff-5c0f144c4f04_1238x334.png 424w, https://substackcdn.com/image/fetch/$s_!drZV!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2f3c0d76-b6af-4713-8bff-5c0f144c4f04_1238x334.png 848w, https://substackcdn.com/image/fetch/$s_!drZV!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2f3c0d76-b6af-4713-8bff-5c0f144c4f04_1238x334.png 1272w, https://substackcdn.com/image/fetch/$s_!drZV!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2f3c0d76-b6af-4713-8bff-5c0f144c4f04_1238x334.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!drZV!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2f3c0d76-b6af-4713-8bff-5c0f144c4f04_1238x334.png" width="1238" height="334" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/2f3c0d76-b6af-4713-8bff-5c0f144c4f04_1238x334.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:334,&quot;width&quot;:1238,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:41682,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!drZV!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2f3c0d76-b6af-4713-8bff-5c0f144c4f04_1238x334.png 424w, https://substackcdn.com/image/fetch/$s_!drZV!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2f3c0d76-b6af-4713-8bff-5c0f144c4f04_1238x334.png 848w, https://substackcdn.com/image/fetch/$s_!drZV!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2f3c0d76-b6af-4713-8bff-5c0f144c4f04_1238x334.png 1272w, https://substackcdn.com/image/fetch/$s_!drZV!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2f3c0d76-b6af-4713-8bff-5c0f144c4f04_1238x334.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Further along the spectrum are geopolitical events, where the choice of reference class gets even fuzzier. Forecasting expert Philip Tetlock <a href="https://www.jstor.org/stable/j.ctt1pk86s8.5">explains</a>: &#8220;Grexit may have looked sui generis, because no country had exited the Eurozone as of 2015, but it could also be viewed as just another instance of a broad comparison class, such as negotiation failures, or of a narrower class, such as a nation-states withdrawing from international agreements or, narrower still, of forced currency conversions.&#8221; He goes on to defend the idea that even seeming <a href="https://en.wikipedia.org/wiki/Black_swan_theory">Black Swan events</a> like the collapse of the USSR or the Arab Spring can be modeled as members of reference classes, and that inductive reasoning is useful even for this kind of event.</p><p>In Tetlock&#8217;s spectrum, these events represent the &#8220;peak&#8221; of uniqueness. When it comes to geopolitical events, that might be true. But even those events are far less unique than extinction from AI. Just look at the <a href="https://www.openphilanthropy.org/wp-content/uploads/2023.05.22-AI-Reference-Classes-Zachary-Freitas-Groff.pdf">attempts</a> to find reference classes for AI x-risk: animal extinction (as a reference class for human extinction), past global transformations such as the industrial revolution (as a reference class for socioeconomic transformation from AI), or accidents causing mass deaths (as a reference class for accidents causing global catastrophe). Let&#8217;s get real. None of those tell us anything about the possibility of developing superintelligent AI or losing control over such AI, which are the central sources of uncertainty for AI x-risk forecasting.&nbsp;</p><p>To summarize, human extinction due to AI is an outcome so far removed from anything that has happened in the past that we cannot use inductive methods to &#8220;predict&#8221; the odds. Of course, we can get qualitative insights from past technical breakthroughs as well as past catastrophic events, but AI risk is sufficiently different that quantitative estimates lack the kind of justification needed for legitimacy in policymaking.</p><h3><strong>Deductive probability estimation is unreliable due to the lack of theory</strong></h3><p>In Conan Doyle&#8217;s <em>The Adventure of the Six Napoleons</em> <em>&#8212;</em> spoiler alert! <em>&#8212; </em>Sherlock Holmes announces before embarking on a stakeout that the probability of catching the suspect is exactly two-thirds. This seems bewildering &#8212; how can anything related to human behavior be ascribed a mathematically precise probability?</p><p>It turns out that Holmes has deduced the underlying series of events that gave rise to the suspect&#8217;s seemingly erratic observed behavior: the suspect is methodically searching for a jewel that is known to be hidden inside one of six busts of Napoleon owned by different people in and around London. The details aren&#8217;t too important, but the key is that neither the suspect nor the detectives know <em>which</em> of the six busts it is in, and everything else about the suspect&#8217;s behavior is (assumed to be) entirely predictable. Hence the precisely quantifiable uncertainty.</p><p>The point is that if we have a model of the world that we can rely upon, we can estimate risk through logical deduction, even without relying on past observations. Of course, outside of fictional scenarios, the world isn&#8217;t so neat, especially when we want to project far into the future.</p><p>When it comes to x-risk, there is an interesting exception to the general rule that we don&#8217;t have deductive models &#8212; asteroid impact. A combination of inductive and deductive risk estimation does allow us to estimate the probability of x-risk, only because we&#8217;re talking about a purely physical system. Let&#8217;s take a minute to review how this works, because it&#8217;s important to recognize that the methods are not generalizable to other types of x-risk.&nbsp;</p><p>The key is being able to <a href="https://www.nature.com/articles/367033a0">model</a> the relationship between the size of the asteroid (more precisely, the energy of impact) and the frequency of impact. Since we have observed thousands of small impacts, we can extrapolate to infer the frequency of large impacts that have never been directly observed. We can also estimate the threshold that would cause global catastrophe.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!77m3!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3483cbd6-7b7c-4fb8-833c-66c4c1f4591f_2396x920.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!77m3!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3483cbd6-7b7c-4fb8-833c-66c4c1f4591f_2396x920.png 424w, https://substackcdn.com/image/fetch/$s_!77m3!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3483cbd6-7b7c-4fb8-833c-66c4c1f4591f_2396x920.png 848w, https://substackcdn.com/image/fetch/$s_!77m3!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3483cbd6-7b7c-4fb8-833c-66c4c1f4591f_2396x920.png 1272w, https://substackcdn.com/image/fetch/$s_!77m3!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3483cbd6-7b7c-4fb8-833c-66c4c1f4591f_2396x920.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!77m3!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3483cbd6-7b7c-4fb8-833c-66c4c1f4591f_2396x920.png" width="1456" height="559" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/3483cbd6-7b7c-4fb8-833c-66c4c1f4591f_2396x920.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:559,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1119384,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!77m3!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3483cbd6-7b7c-4fb8-833c-66c4c1f4591f_2396x920.png 424w, https://substackcdn.com/image/fetch/$s_!77m3!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3483cbd6-7b7c-4fb8-833c-66c4c1f4591f_2396x920.png 848w, https://substackcdn.com/image/fetch/$s_!77m3!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3483cbd6-7b7c-4fb8-833c-66c4c1f4591f_2396x920.png 1272w, https://substackcdn.com/image/fetch/$s_!77m3!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3483cbd6-7b7c-4fb8-833c-66c4c1f4591f_2396x920.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Figure: data on small asteroid impacts (illustrated on the <a href="https://commons.wikimedia.org/wiki/File:Bolide_events_1994-2013.jpg">left</a>) can be extrapolated to extinction-level impacts (right).</figcaption></figure></div><p>With AI, the unknowns relate to <a href="https://www.aisnakeoil.com/p/ai-scaling-myths">technological progress</a> and governance rather than a physical system, so it isn&#8217;t clear how to model it mathematically. Still, people have tried. For example, in order to predict the computational requirements of a hypothetical AGI, <a href="https://arxiv.org/pdf/2306.02519">several</a> <a href="https://www.lesswrong.com/posts/KrJfoZzpSDpnrv9va/draft-report-on-ai-timelines">works</a> assume that an AI system would require roughly as many computations as the human brain, and further make assumptions about the number of computations required by the human brain. These assumptions are far more tenuous than those involved in asteroid modeling, and none of this even addresses the loss-of-control question.</p><h3><strong>Subjective probabilities are feelings dressed up as numbers</strong></h3><p>Without the reference classes or grounded theories, forecasts are necessarily &#8220;subjective probabilities&#8221;, that is, guesses based on the forecaster&#8217;s judgment. Unsurprisingly, these vary by orders of magnitude.</p><p>Subjective probability estimation does not get around the need for having either an inductive or a deductive basis for probability estimates. It merely avoids the need for the forecaster to <em>explain</em> their estimate. Explanation can be hard due to humans&#8217; limited ability to explain our intuitive reasoning, whether inductive, deductive, or a combination thereof. Essentially, it allows the forecaster to say: &#8220;even though I haven&#8217;t shown my methods, you can trust this estimate because of my track record&#8221; (we explain in the next section why even this breaks down for AI x-risk forecasting). But ultimately, lacking either an inductive or a deductive basis, all that forecasters can do is to make up a number, and those made-up numbers are all over the place.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!9nvB!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ce8a04c-c569-4da7-9c8d-3d4b01ca20de_1272x896.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!9nvB!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ce8a04c-c569-4da7-9c8d-3d4b01ca20de_1272x896.png 424w, https://substackcdn.com/image/fetch/$s_!9nvB!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ce8a04c-c569-4da7-9c8d-3d4b01ca20de_1272x896.png 848w, https://substackcdn.com/image/fetch/$s_!9nvB!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ce8a04c-c569-4da7-9c8d-3d4b01ca20de_1272x896.png 1272w, https://substackcdn.com/image/fetch/$s_!9nvB!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ce8a04c-c569-4da7-9c8d-3d4b01ca20de_1272x896.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!9nvB!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ce8a04c-c569-4da7-9c8d-3d4b01ca20de_1272x896.png" width="1272" height="896" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/8ce8a04c-c569-4da7-9c8d-3d4b01ca20de_1272x896.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:896,&quot;width&quot;:1272,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:172573,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!9nvB!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ce8a04c-c569-4da7-9c8d-3d4b01ca20de_1272x896.png 424w, https://substackcdn.com/image/fetch/$s_!9nvB!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ce8a04c-c569-4da7-9c8d-3d4b01ca20de_1272x896.png 848w, https://substackcdn.com/image/fetch/$s_!9nvB!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ce8a04c-c569-4da7-9c8d-3d4b01ca20de_1272x896.png 1272w, https://substackcdn.com/image/fetch/$s_!9nvB!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ce8a04c-c569-4da7-9c8d-3d4b01ca20de_1272x896.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Consider the <a href="https://static1.squarespace.com/static/635693acf15a3e2a14a56a4a/t/64f0a7838ccbf43b6b5ee40c/1693493128111/XPT.pdf">Existential Risk Persuasion Tournament</a> (XPT) conducted by the Forecasting Research Institute in late 2022, which we think is the most elaborate and well-executed x-risk forecasting exercise conducted to date. It involved various groups of forecasters, including AI experts and forecasting experts (&#8220;superforecasters&#8221; in the figure). For AI experts, the high end (75th percentile) of estimates for AI extinction risk by 2100 is 12%, the median estimate is 3%, and the low end (25th percentile) is 0.25%. For forecasting experts, even the high end (75th percentile) is only 1%, the median is a mere 0.38%, and the low end (25th percentile) is visually indistinguishable from zero on the graph. In other words, the 75th percentile AI expert forecast and the 25th percentile superforecaster forecast differ by at least a factor of 100.</p><p>All of these estimates are from people who have deep expertise on the topic and participated in a months-long tournament where they tried to persuade each other! If this range of forecasts here isn&#8217;t extreme enough, keep in mind that this whole exercise was conducted by one group at one point in time. We might get different numbers if the tournament were repeated today, if the questions were framed differently, etc.</p><p>What&#8217;s most telling is to look at the rationales that forecasters provided, which are extensively detailed in the report. They aren&#8217;t using quantitative models, especially when thinking about the likelihood of bad outcomes conditional on developing powerful AI. For the most part, forecasters are engaging in the same kind of speculation that everyday people do when they discuss superintelligent AI. Maybe AI will take over critical systems through superhuman persuasion of system operators. Maybe AI will seek to lower global temperatures because it helps computers run faster, and accidentally wipe out humanity. Or maybe AI will seek resources in space rather than Earth, so we don&#8217;t need to be as worried. There&#8217;s nothing wrong with such speculation. But we should be clear that when it comes to AI x-risk, forecasters aren&#8217;t drawing on any special knowledge, evidence, or models that make their hunches more credible than yours or ours or anyone else&#8217;s.&nbsp;&nbsp;</p><p>The term <a href="https://www.amazon.com/Superforecasting-Science-Prediction-Philip-Tetlock/dp/0804136718">superforecasting</a> comes from Philip Tetlock&#8217;s 20 year study of forecasting (he was also one of the organizers of the XPT). Superforecasters tend to be trained in methods to improve forecasts such as by integrating diverse information and by minimizing psychological biases. These methods have been shown to be effective in domains such as geopolitics. But no amount of training will lead to good forecasts if there isn&#8217;t much useful evidence to draw from.</p><p>Even if forecasters had credible quantitative models (they don&#8217;t), they must account for &#8220;unknown unknowns&#8221;, that is, the possibility that the model itself might be wrong. As noted x-risk philosopher Nick Bostrom <a href="https://existential-risk.com/concept.pdf">explains</a>: &#8220;The uncertainty and error-proneness of our first-order assessments of risk is itself something we must factor into our all-things-considered probability assignments. This factor often dominates in low-probability, high-consequence risks &#8212; especially those involving poorly understood natural phenomena, complex social dynamics, or new technology, or that are difficult to assess for other reasons.&#8221;&nbsp;</p><p>This is a reasonable perspective, and AI x-risk forecasters do worry a lot about uncertainty in risk assessment. But one consequence of this is that for those who follow this principle, forecasts are <em>guaranteed</em> to be guesses rather than the output of a model &#8212; after all, no model can be used to estimate the probability that the model itself is wrong, or what the risk would be if the model were wrong.</p><h3><strong>Forecast skill cannot be measured when it comes to unique or rare events</strong></h3><p>To recap, subjective AI-risk forecasts vary by orders of magnitude. But if we can measure forecasters&#8217; track records, maybe we can use that to figure out which forecasters to trust. In contrast to the previous two approaches for justifying risk estimates (inductive and deductive), the forecaster doesn&#8217;t have to explain their estimate, but instead justifies it based on their demonstrated skill at predicting <em>other</em> outcomes in the past.&nbsp;</p><p>This has proved to be invaluable in the domain of geopolitical events, and the forecasting community spends a lot of effort on skill measurement. Many ways to evaluate forecasting skill exist, such as calibration, the Brier score, the logarithmic score, or the Peer score used on the forecasting competition website <a href="https://www.metaculus.com/">Metaculus</a>.&nbsp;</p><p>But regardless of which method is used, when it comes to existential risk, there are many barriers to assessing forecast skill for subjective probabilities: the lack of a reference class, the low base rate, and the long time horizon. Let&#8217;s look at each of these in turn.</p><p>Just as the reference class problem plagues the forecaster, it also affects the evaluator. Let&#8217;s return to the alien landing example. Consider a forecaster who has proved highly accurate at calling elections. Suppose this forecaster announces, without any evidence, that aliens will land on Earth within a year. Despite the forecaster&#8217;s demonstrated skill, this would not cause us to update our beliefs about an alien landing, because it is too dissimilar to election forecasting and we do not expect the forecaster&#8217;s skill to generalize. Similarly, AI x-risk is so dissimilar to any past events that have been forecast that there is no evidence of any forecaster&#8217;s skill at estimating AI x-risk.&nbsp;</p><p>Even if we somehow do away with the reference class problem, other problems remain &#8212; notably, the fact that extinction risks are &#8220;tail risks&#8221;, or risks that result from rare events. Suppose forecaster A says the probability of AI x-risk is 1%, and forecaster B says it is 1 in a million. Which forecast should we have more confidence in? We could look at their track records. Say we find that forecaster A (who has assigned a 1% probability to AI x-risk) has a better track record. It still doesn&#8217;t mean we should have more confidence in A&#8217;s forecast, because <em>skill evaluations are insensitive to overestimation of tail risks</em>. In other words, it could be that A scores higher overall because A is slightly better calibrated than B when it comes to everyday events that have a substantial probability of occurring, but tends to massively overestimate tail risks that occur rarely (for example, those with a probability of 1 in a million) by orders of magnitude. No scoring rule adequately penalizes this type of miscalibration.&nbsp;</p><p>Here&#8217;s a thought experiment to show why this is true. Suppose two forecasters F and G forecast two different sets of events, and the &#8220;true&#8221; probabilities of events in both sets are uniformly distributed between 0 and 1. We assume, highly optimistically, that both F and G know the true probability P[<em>e</em>] for every event <em>e</em> that they forecast. F always outputs P[<em>e</em>], but G is slightly conservative, never predicting a value less than 1%. That is, G outputs P[<em>e</em>] if P[<em>e</em>] &gt;= 1%, otherwise outputs 1%.</p><p>By construction, F is the better forecaster. But would this be evident from their track records? In other words, how many forecasts from each would we have to evaluate so that there&#8217;s a 95% chance that F outscores G? With the logarithmic scoring rule, it turns out to be on the order of a hundred million. With the Brier score, it is on the order of a trillion.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a>&nbsp; We can quibble with the assumptions here but the point is that if a forecaster systematically overestimates tail risks, it is simply empirically undetectable.</p><p>The final barrier to assessing forecaster skill at predicting x-risk is that long-term forecasts take too long to evaluate (and extinction forecasts are of course impossible to evaluate). This can potentially be overcome. Researchers have developed a method called <a href="https://papers.ssrn.com/sol3/papers.cfm?abstract_id=3954498">reciprocal scoring</a> &#8212; where forecasters are rewarded based on how well they predict each others&#8217; forecasts &#8212; and validate it in some real-world settings, such as predicting the effect of Covid-19 policies. In these settings, reciprocal scoring yielded forecasts that are as good as traditional scoring methods. Fair enough. But reciprocal scoring is not a way around the reference class problem or the tail risk problem.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!PQUD!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4779dde8-30c9-4883-b698-78ffd90e3a53_2150x498.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!PQUD!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4779dde8-30c9-4883-b698-78ffd90e3a53_2150x498.png 424w, https://substackcdn.com/image/fetch/$s_!PQUD!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4779dde8-30c9-4883-b698-78ffd90e3a53_2150x498.png 848w, https://substackcdn.com/image/fetch/$s_!PQUD!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4779dde8-30c9-4883-b698-78ffd90e3a53_2150x498.png 1272w, https://substackcdn.com/image/fetch/$s_!PQUD!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4779dde8-30c9-4883-b698-78ffd90e3a53_2150x498.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!PQUD!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4779dde8-30c9-4883-b698-78ffd90e3a53_2150x498.png" width="1456" height="337" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/4779dde8-30c9-4883-b698-78ffd90e3a53_2150x498.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:337,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:87385,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!PQUD!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4779dde8-30c9-4883-b698-78ffd90e3a53_2150x498.png 424w, https://substackcdn.com/image/fetch/$s_!PQUD!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4779dde8-30c9-4883-b698-78ffd90e3a53_2150x498.png 848w, https://substackcdn.com/image/fetch/$s_!PQUD!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4779dde8-30c9-4883-b698-78ffd90e3a53_2150x498.png 1272w, https://substackcdn.com/image/fetch/$s_!PQUD!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4779dde8-30c9-4883-b698-78ffd90e3a53_2150x498.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption"><em>Summary of our argument so far, showing why none of the three forecasting methods can yield credible estimates of AI x-risk.</em></figcaption></figure></div><h3><strong>There are many reasons why risk estimates may be systematically inflated</strong></h3><p>To recap, inductive and deductive methods don&#8217;t work, subjective forecasts are all over the place, and there&#8217;s no way to tell which forecasts are more trustworthy.</p><p>So in an attempt to derive more reliable estimates that could potentially inform policy, some researchers have turned to forecast aggregation methods that combine the predictions of multiple forecasters. A notable effort is the <a href="https://wiki.aiimpacts.org/doku.php?id=ai_timelines:predictions_of_human-level_ai_timelines:ai_timeline_surveys:2022_expert_survey_on_progress_in_ai">AI Impacts Survey on Progress in AI</a>, but it has been criticized for serious methodological limitations including <a href="https://aiguide.substack.com/p/do-half-of-ai-researchers-believe">non</a>-<a href="https://www.scientificamerican.com/article/ai-survey-exaggerates-apocalyptic-risks/">response</a> <a href="https://spectrum.ieee.org/ai-existential-risk-survey">bias</a>. More importantly, it is unclear why aggregation should improve forecast accuracy: after all, most forecasters might share the same biases (and again, none of them have any basis for a reliable forecast).</p><p>There are many reasons why forecasters might systematically overestimate AI x-risk.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-3" href="#footnote-3" target="_self">3</a> The first is selection bias. Take AI researchers: the belief that AI can change the world is one of the main motivations for becoming an AI researcher. And once someone enters this community, they are in an environment where that message is constantly reinforced. And if one believes that this technology is terrifyingly powerful, it is perfectly rational to think there is a serious chance that its world-altering effects will be negative rather than positive.</p><p>And in the AI safety subcommunity, which is a bit insular, the <a href="https://www.washingtonpost.com/technology/2023/07/05/ai-apocalypse-college-students/">echo chamber</a> can be deafening. Claiming to have a high <em>p(doom)</em> (one&#8217;s estimate of the probability of AI doom) seems to have become a way to signal one&#8217;s identity and commitment to the cause.</p><p>There is a slightly different selection bias at play when it comes to forecasting experts. The forecasting community has a strong overlap with effective altruism and concerns about existential risk, especially AI risk. This doesn&#8217;t mean that individual forecasters are biased. But having a high <em>p(doom)</em> might make someone more inclined to take up forecasting as an activity. So the community as a whole is likely biased toward people with x-risk worries.&nbsp;</p><p>Forecasters are good at updating their beliefs in response to evidence, but the problem is that unlike, say, asteroid impact risk, there is little evidence that can change one&#8217;s beliefs one way or another when it comes to AI x-risk, so we suspect that forecasts are strongly influenced by the priors with which people enter the community. The <a href="https://static1.squarespace.com/static/635693acf15a3e2a14a56a4a/t/64f0a7838ccbf43b6b5ee40c/1693493128111/XPT.pdf">XPT report</a> notes that &#8220;Few minds were changed during the XPT, even among the most active participants, and despite monetary incentives for persuading others.&#8221; In a <a href="https://static1.squarespace.com/static/635693acf15a3e2a14a56a4a/t/65ef1ee52e64b52f145ebb49/1710169832137/AIcollaboration.pdf">follow-up study</a>, they found that many of the disagreements were due to fundamental worldview differences that go beyond AI.</p><p>To reemphasize, our points about bias are specific to AI x-risk. If there were a community of election forecasters who were systematically biased (say, toward incumbents), this would become obvious after a few elections when comparing predictions with reality. But with AI x-risk, as we showed in the previous section, skill evaluation is insensitive to overestimation of tail risks.</p><p>Interestingly, skill evaluation is extremely sensitive to <em>underestimation</em> of tail risks: if you assign a probability of 0 for a rare event that actually ends up occurring, you incur an <em>infinite </em>penalty under the logarithmic scoring rule, from which you can never recover regardless of how well you predicted other events. This is considered one of the main benefits of the logarithmic score and is the reason it is adopted by Metaculus.</p><p>Now consider a forecaster who doesn&#8217;t have a precise estimate &#8212; and surely no forecaster has a precise estimate for something with so many axes of uncertainty as AI x-risk. Given the asymmetric penalties, the rational thing to do is to go with the higher end of their range of estimates.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-4" href="#footnote-4" target="_self">4</a></p><p>In any case, it&#8217;s not clear what forecasters actually report when their estimates are highly uncertain. Maybe they don&#8217;t respond to the incentives of the scoring function. After all, long-term forecasts won&#8217;t be resolved anytime soon. And recall that in the case of the XPT, the incentive is actually to predict each others&#8217; forecasts to get around the problem of long time horizons. The reciprocal scoring paper argues that this will incentivize forecasters to submit their true, high-effort estimates, and considers various objections to this claim. Their defense of the method rests on two key assumptions: that by exerting more effort forecasters can get closer to the true estimate, and that they have no better way to predict what other forecasters will do.</p><p>What if these assumptions are not satisfied? As we have argued throughout this post, with AI x-risk, we shouldn&#8217;t expect evidence to change forecasters&#8217; prior beliefs, so the first assumption is dubious. And now that one iteration of the XPT has concluded, the published median estimates from that tournament serve as a powerful anchor (a &#8220;<a href="https://en.wikipedia.org/wiki/Focal_point_(game_theory)">focal point</a>&#8221; in game theory). It is possible that in the future, forecasters with reciprocal scoring incentives will use existing median forecasts as a starting point, only making minor adjustments to account for new information that has become available since the last tournament. The range of estimates might narrow as existing estimates serve as anchors for future estimates. All that is a roundabout way to say: the less actual evidence there is to draw upon, the more the risk of groupthink.</p><h3><strong>Beware Pascal&#8217;s wager: the dangers of utility maximization&nbsp;</strong></h3><p>For what it&#8217;s worth, here are the median estimates from the <a href="https://static1.squarespace.com/static/635693acf15a3e2a14a56a4a/t/64f0a7838ccbf43b6b5ee40c/1693493128111/XPT.pdf">XPT</a> of both extinction risk and sub-extinction catastrophic risks from AI:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!9IN3!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8655d15c-6f40-43a0-bf3e-1ca1aadb9cd0_1892x614.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!9IN3!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8655d15c-6f40-43a0-bf3e-1ca1aadb9cd0_1892x614.png 424w, https://substackcdn.com/image/fetch/$s_!9IN3!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8655d15c-6f40-43a0-bf3e-1ca1aadb9cd0_1892x614.png 848w, https://substackcdn.com/image/fetch/$s_!9IN3!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8655d15c-6f40-43a0-bf3e-1ca1aadb9cd0_1892x614.png 1272w, https://substackcdn.com/image/fetch/$s_!9IN3!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8655d15c-6f40-43a0-bf3e-1ca1aadb9cd0_1892x614.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!9IN3!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8655d15c-6f40-43a0-bf3e-1ca1aadb9cd0_1892x614.png" width="1456" height="473" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/8655d15c-6f40-43a0-bf3e-1ca1aadb9cd0_1892x614.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:473,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:442848,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!9IN3!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8655d15c-6f40-43a0-bf3e-1ca1aadb9cd0_1892x614.png 424w, https://substackcdn.com/image/fetch/$s_!9IN3!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8655d15c-6f40-43a0-bf3e-1ca1aadb9cd0_1892x614.png 848w, https://substackcdn.com/image/fetch/$s_!9IN3!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8655d15c-6f40-43a0-bf3e-1ca1aadb9cd0_1892x614.png 1272w, https://substackcdn.com/image/fetch/$s_!9IN3!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8655d15c-6f40-43a0-bf3e-1ca1aadb9cd0_1892x614.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>To reiterate, our view is that we shouldn&#8217;t take any of these numbers too seriously. They are a reflection of how much different samples of participants fret about AI than anything else.</p><p>As before, the estimates from forecasting experts (superforecasters) and AI experts differ by an order of magnitude or more. To the extent that we put any stock into these estimates, it should be the forecasting experts&#8217; rather than the AI experts&#8217; estimates. One important insight from past research is that domain experts perform worse than forecasting experts who have training in integrating diverse information and by minimizing psychological biases. Still, as we said above, even their forecasts may be vast overestimates, and we just can&#8217;t know for sure.</p><p>So what&#8217;s the big deal? So what if policymakers believe the risk over a certain timeframe is 1% instead of 0.01%? It seems pretty low in either case!</p><p>It depends on what they do with those probabilities. Most often, these estimates are merely a way to signal the fact that some group of experts thinks the risk is significant. If that&#8217;s all they are, so be it. But it&#8217;s not clear that all this elaborate effort at quantification is even helpful for this signaling purpose, given that different people interpret the same numbers wildly differently.&nbsp;</p><p>For example, Federal Trade Commission chair Lina Khan said her views on the matter were techno-optimistic since her <em>p(doom)</em> was <a href="https://www.nytimes.com/2023/11/10/podcasts/hardfork-chatbot-ftc.html">only 15%</a>, which left experts bewildered. (For what it&#8217;s worth, that number is about a thousandfold higher than what we would be comfortable labeling techno-optimist.) It takes a lot of quantitative training to be able to mentally process very small or very large numbers correctly in decision making, and not simply bucket them into categories like &#8220;insignificantly small&#8221;. Most people are not trained this way.&nbsp;</p><p>In short, what seems to be happening is that experts&#8217; vague intuitions and fears are being translated into pseudo-precise numbers, and then translated back into vague intuitions and fears by policymakers. Let&#8217;s just cut the charade of quantification! The Center for AI Safety&#8217;s <a href="https://www.safe.ai/work/statement-on-ai-risk">Statement on AI Risk</a> was admirably blunt in this regard (of course, we <a href="https://www.aisnakeoil.com/p/is-avoiding-extinction-from-ai-really">strongly disagree with its substance</a>).</p><p>A principled, quantitative way to use probabilities in decision making is utility maximization through cost-benefit analysis. The idea is simple: if we consider an outcome to have a subjective value, or utility, of U (which can be positive or negative), and it has, say, a 10% probability of occurring, we can act as if it is certain to occur and has a value of 0.1 * U. We can then add up the costs and benefits for each option available to us, and choose the one that maximizes costs minus benefits (the &#8220;expected utility&#8221;).</p><p>This is where things get really problematic. First, some people might consider extinction to have an unfathomably large negative value, because it precludes the existence of all the human lives, physical or <a href="https://www.vox.com/future-perfect/23298870/effective-altruism-longtermism-will-macaskill-future">simulated</a>, that might ever be born in the future. The logical conclusion is that x-risk should be everyone&#8217;s top priority all the time! It is reminiscent of <a href="https://en.wikipedia.org/wiki/Pascal%27s_wager">Pascal&#8217;s wager</a>, the argument that it is rational believe in God because even if there is an infinitesimally small chance that God exists, the cost of non-belief is infinite (an eternity in hell as opposed to eternal happiness), and hence so is the expected utility. Fortunately, policymakers don&#8217;t give too much credence to decision making frameworks involving infinities. But the idea has taken a powerful hold of the AI safety community and drives some people&#8217;s conviction that AI x-risk should be society&#8217;s top priority.</p><p>Even if we limit ourselves to catastrophic but not existential risks, we are talking about billions of lives on the line, so the expected cost of even a 1% risk is so high that the policy implications are drastic &#8212; governments should increase spending on AI x-risk mitigation by orders of magnitude and consider draconian measures such as stopping AI development. This is why it is so vital to understand that these estimates are not backed by any methodology. It would be incredibly unwise to make world-changing policy decisions based on so little evidence.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.normaltech.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">You&#8217;re reading AI Snake Oil, a blog about our <a href="https://www.aisnakeoil.com/p/ai-snake-oil-is-now-available-to">book</a>. Subscribe to get new posts.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><h3><strong>Forecasts of milestones suffer from outcome ambiguity</strong></h3><p>Is there a role for forecasting in AI policy? We think yes &#8212; just not forecasting existential risk. Forecasting AI milestones, such as performance on certain capability benchmarks or economic impacts, is more achievable and meaningful. If a forecaster has demonstrated skill in predicting when various AI milestones would be reached, it does give us evidence that they will do well in the future. We are no longer talking about unique or rare events. And when considering lower-stakes policy interventions &#8212; preparing for potential economic disruption rather than staving off killer robots &#8212; it is less critical that forecasts be justified to the satisfaction of every reasonable person.&nbsp;</p><p>The forecasting community devotes a lot of energy to milestone forecasting. On Metaculus, the question &#8220;Will there be Human-machine intelligence parity before 2040?&#8221; has an aggregate prediction of <a href="https://www.metaculus.com/questions/384/humanmachine-intelligence-parity-by-2040/">96%</a> based on over 1,300 forecasters. That&#8217;s remarkable! If we agreed with this forecast, we would be in favor of the position that managing the safe transition to AGI should be a global priority. Why don&#8217;t we?</p><p>The answer is in the fine print. There is no consensus on the definition of a fuzzy concept such as AGI. Even if we fix a definition, determining whether it has been achieved can be hard or impossible. For effective forecasting, it is extremely important to avoid ambiguous outcomes. The way the forecasting community gets around this is by defining it in terms of relatively narrow skills, such as exam performance.&nbsp;</p><p>The Metaculus intelligence parity question is defined in terms of the performance on graduate exams in math, physics, and computer science. Based on this definition, we do agree with the forecast of 96%. But we think the definition is so watered down that it doesn&#8217;t mean much for policy. Forget existential risk &#8212; as we&#8217;ve written before, AI performance on exams has so little <a href="https://www.aisnakeoil.com/p/gpt-4-and-professional-benchmarks">construct validity</a> that it doesn&#8217;t even let us predict whether AI will replace workers.</p><p>Other benchmarks <a href="https://www.aisnakeoil.com/p/new-paper-ai-agents-that-matter">aren&#8217;t much better</a>. In short, forecasting AI capability timelines is tricky because of the huge gap between benchmarks and real-world implications. Fortunately, <a href="https://www.openphilanthropy.org/rfp-llm-benchmarks/">better benchmarks</a> reflecting consequential real-world tasks are being developed. In addition to benchmarks, we need naturalistic evaluation, even if it is more costly. One type of naturalistic evaluation is to measure how people perform their jobs differently with AI assistance. Directly forecasting economic, social, or political impacts &#8212; such as labor market transformation or AI-related spending by militaries &#8212; could be even more useful, although harder to unambiguously define and measure.</p><h3><strong>Concluding thoughts</strong></h3><p>The responsibility for avoiding misuses of probability in policy lies with policymakers. We are not calling for forecasters to stop publishing forecasts in order to &#8220;protect&#8221; policymakers from being misled. That said, we think forecasts should be accompanied by a clear explanation of the process used and evidence considered. This would allow policymakers to make informed decisions about whether the justification presented meets the threshold that they are comfortable with. The XPT is a good example of transparency, as is <a href="https://arxiv.org/abs/2306.02519">this paper</a> (though it is not about x-risk). On the other hand, simply surveying a bunch of researchers and presenting aggregate numbers is misinformative and should be ignored by policymakers.</p><p>So what should governments do about AI x-risk? Our view isn&#8217;t that they should do nothing. But they should reject the kind of policies that might seem compelling if we view x-risk as urgent and serious, notably: restricting AI development. As we&#8217;ll argue in a future essay in this series, not only are such policies unnecessary, they are likely to <em>increase</em> x-risk. Instead, governments should adopt policies that are compatible with a range of possible estimates of AI risk, and are on balance helpful even if the risk is negligible. Fortunately, such policies exist. Governments should also change policymaking <em>processes</em> so that they are more responsive to new evidence. More on all that soon.</p><h4><strong>Further reading</strong></h4><ul><li><p>The XPT report is titled <a href="https://static1.squarespace.com/static/635693acf15a3e2a14a56a4a/t/64f0a7838ccbf43b6b5ee40c/1693493128111/XPT.pdf">Forecasting Existential Risks: Evidence from a Long-Run Forecasting Tournament</a>.</p></li><li><p>Tetlock and Gardner&#8217;s book <a href="https://www.amazon.com/Superforecasting-Science-Prediction-Philip-Tetlock/dp/0804136718">Superforecasting</a> summarizes research by Tetlock, Barbara Mellers, and others.</p></li><li><p>Scott Alexander refutes the claim that there is something wrong in principle with ascribing <a href="https://www.astralcodexten.com/p/in-continued-defense-of-non-frequentist">probabilities to unique events</a> (we largely agree). Our argument differs in two key ways from the position he addresses. We aren&#8217;t talking about forecasting in general; just its application to policymaking. And we don&#8217;t object to it in principle. Our argument is empirical: AI x-risk forecasts are extremely unreliable and lack justification. Theoretically this could change in the future, though we aren&#8217;t holding our breath.</p></li><li><p>A <a href="https://www.jstor.org/stable/45094450?seq=1">paper</a> by Friedman and Zuckhauser explains why probabilities aren&#8217;t the whole story: two forecasts that have the same risk estimate might have very different implications for policymakers. In our view, AI x-risk forecasts fare poorly on two of the three dimensions of confidence: a sound basis in evidence and a narrow range of reasonable opinion.</p></li><li><p>A paper by our Princeton CITP colleagues led by Shazeda Ahmed explains the <a href="https://firstmonday.org/ojs/index.php/fm/article/view/13626/11596">epistemic culture of AI safety</a>, which consists of &#8220;cohesive, interwoven social structures of knowledge-production and community-building&#8221;. It helps understand why practices such as forecasting have become pillars of how the AI safety community forms its beliefs, as opposed to the broader scientific community that centers practices such as peer review. Of course, we shouldn&#8217;t reject a view just because it doesn&#8217;t conform to scientific orthodoxy. But at the same time, we shouldn&#8217;t give any deference to the self-styled AI safety community&#8217;s views on AI safety. It is important to understand that the median member of the AI safety community holds one particular stance on AI safety &#8212; a stance that is highly contested and in our view rather alarmist.</p></li><li><p>We have written extensively about <a href="https://www.aisnakeoil.com/t/ai-safety">evidence-based AI safety</a>. Our best-known work includes the essay <a href="https://www.aisnakeoil.com/p/ai-safety-is-not-a-model-property">AI safety is not a model property</a> and the paper titled <a href="https://crfm.stanford.edu/open-fms/">On the Societal Impact of Open Foundation Models</a> which was the result of a large collaboration.</p></li></ul><p><strong>Acknowledgements.</strong> We are grateful to Benjamin Edelman, Ezra Karger, Matt Salganik, and Ollie Stephenson for feedback on a draft. This series of essays is based on an upcoming paper that benefited from feedback from many people, including Seth Lazar and members of the MINT lab at Australian National University, students in the Limits to Prediction course at Princeton, Shazeda Ahmed, and Zachary Siegel.</p><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>Specifically, by using data from nuclear weapons tests and making a few physics-informed assumptions, we can calculate what it would take to kick up enough dust to darken the skies for a prolonged period and lead to a collapse of global agriculture.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p>Suppose G, the conservative forecaster, has a floor of &#1013; for their forecasts (in our example, 0.01). There is only an &#1013; fraction of events where the difference between the two forecasters is even relevant. Even limiting ourselves to the set of events with a true probability less than &#1013;, some calculation shows that F&#8217;s expected log score is better than G&#8217;s by a tiny amount &#8212; O(&#1013;), and with the Brier score, O(&#1013;^2). So when looking at all events, the differences in expected scores are O(&#1013;^2) and O(&#1013;^3) respectively. Meanwhile, the variance in their scores comes out to about 0.1 for both scoring functions. To be able to confidently assert that the means of the two forecaster&#8217;s scores are different, we need the number of events N to be large enough that the standard deviation of the difference in mean scores, which is sqrt(0.1 / N), to be much less than the expected difference. So we need N ~ O(1/&#1013;^4) for the log score and N ~ O(1/&#1013;^6) for the Brier score.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-3" href="#footnote-anchor-3" class="footnote-number" contenteditable="false" target="_self">3</a><div class="footnote-content"><p>Of course, there are also reasons why forecasters might underestimate the risk. We find those reasons less persuasive, but more importantly, we consider them out of scope for this discussion. The reason is simple: the advocacy for governments to implement freedom-restricting policies is being justified by <em>high</em> x-risk estimates. The burden of evidence is thus asymmetric. Those who call for such policies must justify their probability estimates, including responding to concerns about upward biases. If, in some strange future world, some people advocate for restrictive policies based on <em>low</em> x-risk estimates, it will become important to subject potential <em>underestimation</em> to the same scrutiny.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-4" href="#footnote-anchor-4" class="footnote-number" contenteditable="false" target="_self">4</a><div class="footnote-content"><p>In the theory of forecasting, the logarithmic score and other so-called proper scoring rules are mathematically guaranteed to elicit the forecaster&#8217;s "true belief". But the definition of a proper scoring rule assumes that the true belief is a single value, whereas it tends to be a wide range. In such a scenario, the forecaster is incentivized to report the mean of their distribution, but what we intuitively mean by the &#8220;true belief&#8221; might better correspond to the median.</p><p></p></div></div>]]></content:encoded></item><item><title><![CDATA[New paper: AI agents that matter]]></title><description><![CDATA[Rethinking AI agent benchmarking and evaluation]]></description><link>https://www.normaltech.ai/p/new-paper-ai-agents-that-matter</link><guid isPermaLink="false">https://www.normaltech.ai/p/new-paper-ai-agents-that-matter</guid><dc:creator><![CDATA[Sayash Kapoor]]></dc:creator><pubDate>Wed, 03 Jul 2024 16:00:56 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/4de2d78f-afda-4151-b1ad-76986d007c29_1926x1494.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Some of the most exciting applications of large language models involve taking real-world action, such as booking flight tickets or finding and fixing software bugs. AI systems that carry out such tasks are called agents. They use LLMs in combination with other software to use tools such as web search and code terminals. </p><p>The North Star of this field is to build assistants like Siri or Alexa and get them to actually work &#8212; handle complex tasks, accurately interpret users&#8217; requests, and perform reliably. But this is far from a reality, and even the research direction is fairly new. To stimulate the development of agents and measure their effectiveness, researchers have created benchmark datasets. But as we&#8217;ve said before, <a href="https://www.cs.princeton.edu/~arvindn/talks/evaluating_llms_minefield/">LLM evaluation is a minefield</a>, and it turns out that agent evaluation has a bunch of additional pitfalls that affect today&#8217;s benchmarks and evaluation practices. This state of affairs encourages the development of agents that do well on benchmarks without being useful in practice.</p><p>We have released a new paper that identifies the challenges in evaluating agents and proposes ways to address them. <strong>Read the paper <a href="https://arxiv.org/abs/2407.01502">here</a>. </strong>The authors are Sayash Kapoor, <a href="https://citp.princeton.edu/citp-people/benedikt-strobl/">Benedikt Str&#246;bl</a>, <a href="https://www.zacharysiegel.org/">Zachary S. Siegel</a>, <a href="https://citp.princeton.edu/citp-people/nitya-nadgir/">Nitya Nadgir</a>, and Arvind Narayanan, all at Princeton University.&nbsp;</p><p>In this post, we offer thoughts on the definition of AI agents, why we are cautiously optimistic about the future of AI agent research, whether AI agents are more hype or substance, and give a brief overview of the paper.</p><h4><strong>What does the term agent mean? Is it just a buzzword?</strong></h4><p>The term agent has been used by AI researchers without a formal definition.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a> This has led to its being hijacked as a marketing term, and has generated a bit of pushback against its use. But the term isn&#8217;t meaningless. Many researchers have tried to formalize the community's intuitive understanding of what constitutes an agent in the context of language-model-based systems [<a href="https://cdn.openai.com/papers/practices-for-governing-agentic-ai-systems.pdf">1</a>, <a href="https://dl.acm.org/doi/10.1145/3593013.3594033">2</a>, <a href="https://arxiv.org/abs/2404.16244">3</a>, <a href="https://lilianweng.github.io/posts/2023-06-23-agent/">4</a>, <a href="https://blog.langchain.dev/what-is-an-agent/">5</a>]. Rather than a binary, it can be seen as a spectrum, sometimes denoted by the term <a href="https://www.deeplearning.ai/the-batch/welcoming-diverse-approaches-keeps-machine-learning-strong/">'agentic'</a>.&nbsp;</p><p>The five recent definitions of AI agents cited above are all distinct but with strong similarities to each other. Rather than propose a new definition, we identified three clusters of properties that cause an AI system to be considered more agentic according to existing definitions:</p><p><strong>Environment and goals.</strong> The more complex the environment, the more AI systems operating in that environment are agentic. Complex environments are those that have a range of tasks and domains, multiple stakeholders, a long time horizon to take action, and unexpected changes. Further, systems that pursue complex goals without being instructed on how to pursue the goal are more agentic.</p><p><strong>User interface and supervision.</strong> AI systems that can be instructed in natural language and act autonomously on the user's behalf are more agentic. In particular, systems that require less user supervision are more agentic. For example, chatbots cannot take real-world action, but adding plugins to chatbots (such as Zapier for ChatGPT) allows them to take some actions on behalf of users.</p><p><strong>System design.</strong> Systems that use tools (like web search or code terminal) or planning (like reflecting on previous outputs or decomposing goals into subgoals) are more agentic. Systems whose control flow is driven by an LLM, rather than LLMs being invoked by a static program, are more agentic.</p><h4><strong>Do agents even work?</strong></h4><p>While some agents such as ChatGPT&#8217;s code interpreter / data analysis mode have been useful, more ambitious agent-based products so far have failed. The two main product launches based on AI agents have been the <a href="https://www.rabbit.tech/">Rabbit R1</a> and <a href="https://humane.com/">Humane AI pin</a>. These devices promised to eliminate or reduce phone dependence, but turned out to be too slow and unreliable. Devin, an &#8220;AI software engineer&#8221;, was announced with great hype 4 months ago, but has been panned in a <a href="https://www.youtube.com/watch?v=tNmgmwEtoWE">video review</a> and remains in waitlist-only mode. It is clear that if AI agents are to be useful in real-world products, they have a long way to go.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!pOKK!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F483f2069-4a32-41ad-8bb4-94b2fb7793fc_1310x664.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!pOKK!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F483f2069-4a32-41ad-8bb4-94b2fb7793fc_1310x664.png 424w, https://substackcdn.com/image/fetch/$s_!pOKK!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F483f2069-4a32-41ad-8bb4-94b2fb7793fc_1310x664.png 848w, https://substackcdn.com/image/fetch/$s_!pOKK!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F483f2069-4a32-41ad-8bb4-94b2fb7793fc_1310x664.png 1272w, https://substackcdn.com/image/fetch/$s_!pOKK!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F483f2069-4a32-41ad-8bb4-94b2fb7793fc_1310x664.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!pOKK!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F483f2069-4a32-41ad-8bb4-94b2fb7793fc_1310x664.png" width="524" height="265.6" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/483f2069-4a32-41ad-8bb4-94b2fb7793fc_1310x664.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:664,&quot;width&quot;:1310,&quot;resizeWidth&quot;:524,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!pOKK!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F483f2069-4a32-41ad-8bb4-94b2fb7793fc_1310x664.png 424w, https://substackcdn.com/image/fetch/$s_!pOKK!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F483f2069-4a32-41ad-8bb4-94b2fb7793fc_1310x664.png 848w, https://substackcdn.com/image/fetch/$s_!pOKK!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F483f2069-4a32-41ad-8bb4-94b2fb7793fc_1310x664.png 1272w, https://substackcdn.com/image/fetch/$s_!pOKK!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F483f2069-4a32-41ad-8bb4-94b2fb7793fc_1310x664.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><a href="https://x.com/SnazzyLabs/status/1783220942812111220">Source</a></figcaption></figure></div><p>So are AI agents all hype? It&#8217;s too early to tell. We think there are research challenges to be solved before we can expect agents such as the ones above to work well enough to be widely adopted. The only way to find out is through more research, so we do think research on AI agents is worthwhile.</p><p>One major research challenge is reliability &#8212; LLMs are <a href="https://x.com/random_walker/status/1778770599340290103">already capable enough</a> to do many tasks that people want an assistant to handle, but not reliable enough that they can be successful products. To appreciate why, think of a flight-booking agent that needs to make dozens of calls to LLMs. If each of those went wrong independently with a probability of, say, just 2%, the overall system would be so unreliable as to be completely useless (this partly explains some of the product failures we&#8217;ve seen). So research on improving reliability might have many new applications <em>even if</em> the underlying language models don&#8217;t improve. And if <a href="https://www.aisnakeoil.com/p/ai-scaling-myths">scaling runs out</a>, agents are the most natural direction for further progress in AI.</p><p>Right now, however, research is itself contributing to hype and overoptimism because evaluation practices are not rigorous enough, much like the early days of machine learning research before the common task method took hold. That brings us to our paper.</p><h4><strong>Contributions of the paper</strong></h4><p>What changes must the AI community implement to help stimulate the development of AI agents that are useful in the real world, and not just on benchmarks? This is the paper&#8217;s central question. We make five recommendations:</p><p>1.<strong> Implement cost-controlled evaluations.</strong> The language models underlying most AI agents are stochastic. This means simply calling the underlying model multiple times can increase accuracy. We show that such simple tricks can outperform complex agent architectures on the HumanEval benchmark, while costing much less. We argue that all agent evaluation must control for cost. (We originally published this finding <a href="https://www.aisnakeoil.com/p/ai-leaderboards-are-no-longer-useful">here</a>. In the two months since we published this post, Pareto curves and joint optimization of cost and accuracy have become <a href="https://x.com/lmsysorg/status/1807812671238258931">increasingly common</a> in <a href="https://www.together.ai/blog/together-moa">agent</a> <a href="https://x.com/corbtt/status/1803813970018791845">evaluations</a>.)</p><p>2. <strong>Jointly optimize accuracy and cost.</strong> Visualizing evaluation results as a Pareto curve of accuracy and inference cost opens up a new space of agent design: jointly optimizing the two metrics. We show how we can lower cost while maintaining accuracy on HotPotQA by implementing a modification to the DSPy framework.</p><p>3. <strong>Distinguish model and downstream benchmarking. </strong>Through a case study of NovelQA, we show how benchmarks meant for model evaluation can be misleading when used for downstream evaluation. We argue that downstream evaluation should account for dollar costs, rather than proxies for cost such as the number of model parameters.</p><p>4. <strong>Prevent shortcuts in agent benchmarks.</strong> We show that many types of overfitting to agent benchmarks are possible. We identify 4 levels of generality of agents and argue that different types of hold-out samples are needed based on the desired level of generality. Without proper hold-outs, agent developers can take shortcuts, even unintentionally. We illustrate this with a case study of the WebArena benchmark.</p><p>5. <strong>Improve the standardization and reproducibility of agent benchmarks.</strong> We found pervasive shortcomings in the reproducibility of WebArena and HumanEval evaluations. These errors inflate accuracy estimates and lead to overoptimism about agent capabilities.</p><h4><strong>Concluding thoughts: reasons for cautious optimism</strong></h4><p>AI agent benchmarking is new and best practices haven't yet been established, making it hard to distinguish genuine advances from hype. We think agents are sufficiently different from models that benchmarking practices need to be rethought. In our paper, we take the first steps toward a principled approach to agent benchmarking. We hope these steps will raise the rigor of AI agent evaluation and provide a firm foundation for progress.</p><p>A different strand of our research concerns the reproducibility crisis in ML-based research in <a href="https://www.aisnakeoil.com/p/scientists-should-use-ai-as-a-tool">scientific fields</a> such as medicine or social science. At some level, our current paper is similar. In ML-based science, our outlook is that things will get worse before they get better. But in AI agents research, we are cautiously optimistic that practices will change quickly. One reason is that there is a stronger culture of sharing code and data alongside published papers, so errors are easier to spot. (This culture shift came about due to concerted efforts in the <a href="https://jmlr.org/papers/v22/20-303.html">last five years</a>.) Another reason is that overoptimistic research quickly gets a reality check when products based on misleading evaluations end up flopping. This is going to be an interesting space to watch over the next few years, both in terms of research and product releases.</p><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>In traditional AI, agents are defined entities that perceive and act upon their environment, but that definition is less useful in the LLM era &#8212; even a thermostat would qualify as an agent under that definition.</p><p></p></div></div>]]></content:encoded></item><item><title><![CDATA[AI scaling myths]]></title><description><![CDATA[Scaling will run out. The question is when.]]></description><link>https://www.normaltech.ai/p/ai-scaling-myths</link><guid isPermaLink="false">https://www.normaltech.ai/p/ai-scaling-myths</guid><dc:creator><![CDATA[Arvind Narayanan]]></dc:creator><pubDate>Thu, 27 Jun 2024 18:16:55 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/0ca1da67-ea35-4ded-aaf4-a0ff5d63e49b_1708x1310.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>So far, bigger and bigger language models have proven more and more capable. But does the past predict the future?</p><p>One popular view is that we should expect the trends that have held so far to continue for many more orders of magnitude, and that it will potentially get us to artificial general intelligence, or AGI. </p><p>This view rests on a series of myths and misconceptions. The seeming predictability of scaling is a misunderstanding of what research has shown. Besides, there are signs that LLM developers are already at the limit of high-quality training data. And the industry is seeing strong <em>downward</em> pressure on model size. While we can't predict exactly how far AI will advance through scaling, we think there&#8217;s virtually no chance that scaling alone will lead to AGI.&nbsp;</p><h4><strong>Scaling &#8220;laws&#8221; are often misunderstood</strong></h4><p>Research on <a href="https://arxiv.org/abs/2001.08361">scaling laws</a> shows that as we increase model size, training compute, and dataset size, language models get &#8220;better&#8221;. The improvement is truly striking in its predictability, and holds across many orders of magnitude. This is the main reason why many people believe that scaling will continue for the foreseeable future, with regular releases of larger, more powerful models from leading AI companies.</p><p>But this is a complete misinterpretation of scaling laws. What exactly is a &#8220;better&#8221; model? Scaling laws only quantify the decrease in perplexity, that is, improvement in how well models can predict the next word in a sequence. Of course, perplexity is more or less irrelevant to end users &#8212; what matters is &#8220;<a href="https://arxiv.org/abs/2206.07682">emergent abilities</a>&#8221;, that is, models&#8217; tendency to acquire new capabilities as size increases.</p><p>Emergence is not governed by any law-like behavior. It is true that so far, increases in scale have brought new capabilities. But there is no empirical regularity that gives us confidence that this will continue indefinitely.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a></p><p>Why might emergence not continue indefinitely? This gets at one of the core debates about LLM capabilities &#8212; are they capable of extrapolation or do they only learn tasks represented in the training data? The evidence is incomplete and there is a wide range of reasonable ways to interpret it. But we lean toward the skeptical view. On benchmarks designed to test the efficiency of acquiring skills to solve unseen tasks, LLMs tend to perform <a href="https://arcprize.org/arc">poorly</a>.&nbsp;</p><p>If LLMs can't do much beyond what's seen in training, at some point, having more data no longer helps because all the tasks that are ever going to be represented in it are already represented. Every traditional machine learning model eventually plateaus; maybe LLMs are no different.</p><h4><strong>Trend extrapolation is baseless speculation</strong></h4><p>Another barrier to continued scaling is obtaining training data. Companies are already using all the readily available data sources. Can they get more?</p><p>This is less likely than it might seem. People sometimes assume that new data sources, such as transcribing all of YouTube, will increase the available data volume by another order of magnitude or two. Indeed, YouTube has a remarkable <a href="https://journalqd.org/article/view/4066">150 billion minutes</a> of video. But considering that most of that has little or no usable audio (it is instead music, still images, video game footage, etc.), we end up with an estimate that is much <em>less</em> than the 15 trillion tokens that Llama 3 is already using &#8212; and that&#8217;s before deduplication and quality filtering of the transcribed YouTube audio, which is likely to knock off at least another order of magnitude.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a></p><p>People often discuss when companies will &#8220;run out&#8221; of training data. But this is not a meaningful question. There&#8217;s always more training data, but getting it will cost more and more. And now that copyright holders have <a href="https://reutersinstitute.politics.ox.ac.uk/how-many-news-websites-block-ai-crawlers">wised up</a> and want to be compensated, the cost might be especially steep. In addition to dollar costs, there could be reputational and regulatory costs because society might push back against data collection practices.</p><p>We can be certain that no exponential trend can continue indefinitely. But it can be hard to predict when a tech trend is about to plateau. This is especially so when the growth stops suddenly rather than gradually. The trendline itself contains no clue that it is about to plateau.&nbsp;</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!M59s!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffe6b2381-e80e-42b4-b326-814ab6422be7_756x468.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!M59s!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffe6b2381-e80e-42b4-b326-814ab6422be7_756x468.png 424w, https://substackcdn.com/image/fetch/$s_!M59s!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffe6b2381-e80e-42b4-b326-814ab6422be7_756x468.png 848w, https://substackcdn.com/image/fetch/$s_!M59s!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffe6b2381-e80e-42b4-b326-814ab6422be7_756x468.png 1272w, https://substackcdn.com/image/fetch/$s_!M59s!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffe6b2381-e80e-42b4-b326-814ab6422be7_756x468.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!M59s!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffe6b2381-e80e-42b4-b326-814ab6422be7_756x468.png" width="756" height="468" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/fe6b2381-e80e-42b4-b326-814ab6422be7_756x468.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:468,&quot;width&quot;:756,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!M59s!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffe6b2381-e80e-42b4-b326-814ab6422be7_756x468.png 424w, https://substackcdn.com/image/fetch/$s_!M59s!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffe6b2381-e80e-42b4-b326-814ab6422be7_756x468.png 848w, https://substackcdn.com/image/fetch/$s_!M59s!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffe6b2381-e80e-42b4-b326-814ab6422be7_756x468.png 1272w, https://substackcdn.com/image/fetch/$s_!M59s!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffe6b2381-e80e-42b4-b326-814ab6422be7_756x468.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">CPU clock speeds over time. The y-axis is logarithmic. [<a href="https://en.wikipedia.org/wiki/File:Clock_CPU_Scaling.jpg">Source</a>]</figcaption></figure></div><p>Two famous examples are CPU clock speeds in the 2000s and airplane speeds in the 1970s. CPU manufacturers decided that further increases to clock speed were too costly and mostly pointless (since CPU was no longer the bottleneck for overall performance), and simply decided to stop competing on this dimension, which suddenly removed the upward pressure on clock speed. With airplanes, the story is more complex but comes down to the market prioritizing <a href="https://theicct.org/sites/default/files/publications/Aircraft-fuel-burn-trends-sept2020.pdf">fuel</a> <a href="https://www.etw.de/uploads/pdfs/ATAG_Beginners_Guide_to_Aviation_Efficiency_web.pdf">efficiency</a> over speed.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-3" href="#footnote-3" target="_self">3</a></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!tGBk!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Facf7e082-b3c6-4ffb-82c6-85912217135e_2002x1316.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!tGBk!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Facf7e082-b3c6-4ffb-82c6-85912217135e_2002x1316.png 424w, https://substackcdn.com/image/fetch/$s_!tGBk!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Facf7e082-b3c6-4ffb-82c6-85912217135e_2002x1316.png 848w, https://substackcdn.com/image/fetch/$s_!tGBk!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Facf7e082-b3c6-4ffb-82c6-85912217135e_2002x1316.png 1272w, https://substackcdn.com/image/fetch/$s_!tGBk!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Facf7e082-b3c6-4ffb-82c6-85912217135e_2002x1316.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!tGBk!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Facf7e082-b3c6-4ffb-82c6-85912217135e_2002x1316.png" width="1456" height="957" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/acf7e082-b3c6-4ffb-82c6-85912217135e_2002x1316.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:957,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:279306,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!tGBk!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Facf7e082-b3c6-4ffb-82c6-85912217135e_2002x1316.png 424w, https://substackcdn.com/image/fetch/$s_!tGBk!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Facf7e082-b3c6-4ffb-82c6-85912217135e_2002x1316.png 848w, https://substackcdn.com/image/fetch/$s_!tGBk!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Facf7e082-b3c6-4ffb-82c6-85912217135e_2002x1316.png 1272w, https://substackcdn.com/image/fetch/$s_!tGBk!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Facf7e082-b3c6-4ffb-82c6-85912217135e_2002x1316.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Flight airspeed records over time. The SR-71 Blackbird record from 1976 still stands today. [<a href="https://en.wikipedia.org/wiki/Flight_airspeed_record">Source</a>]</figcaption></figure></div><p>With LLMs, we may have a couple of orders of magnitude of scaling left, or we may <em>already</em> be done. As with CPUs and airplanes, it is ultimately a business decision and fundamentally hard to predict in advance.&nbsp;</p><p>On the research front, the focus has shifted from compiling ever-larger datasets to improving the <a href="https://x.com/karpathy/status/1797313173449764933">quality</a> of training data. Careful data cleaning and filtering can allow building equally powerful models with <a href="https://www.microsoft.com/en-us/research/blog/phi-2-the-surprising-power-of-small-language-models/">much</a> <a href="https://arxiv.org/abs/2406.11794">smaller</a> datasets.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-4" href="#footnote-4" target="_self">4</a></p><h4><strong>Synthetic data is not magic</strong></h4><p>Synthetic data is often suggested as the path to continued scaling. In other words, maybe current models can be used to generate training data for the next generation of models.&nbsp;</p><p>But we think this rests on a misconception &#8212; we don't think developers are using (or can use) synthetic data to increase the <em>volume</em> of training data. <a href="https://arxiv.org/html/2404.07503v1">This paper</a> has a great list of uses for synthetic data for training, and it's all about fixing specific gaps and making domain-specific improvements like math, code, or low-resource languages. Similarly, Nvidia's recent <a href="https://developer.nvidia.com/blog/leverage-our-latest-open-models-for-synthetic-data-generation-with-nvidia-nemotron-4-340b/">Nemotron 340B</a> model, which is geared at synthetic data generation, targets alignment as the primary use case. There are a few secondary use cases, but replacing current sources of pre-training data is not one of them. In short, it's unlikely that mindless generation of synthetic training data will have the same effect as having more high-quality human data.</p><p>There are cases where synthetic training data has been spectacularly successful, such as <a href="https://www.nature.com/articles/nature24270">AlphaGo</a>, which beat the Go world champion in 2016, and its successors AlphaGo Zero and <a href="https://www.science.org/doi/10.1126/science.aar6404">AlphaZero</a>. These systems learned by playing games against themselves; the latter two did not use any human games as training data. They used a ton of calculation to generate somewhat high-quality games, used those games to train a neural network, which could then generate even higher-quality games when combined with calculation, resulting in an iterative improvement loop.&nbsp;</p><p>Self-play is the quintessential example of &#8220;System 2 --&gt; System 1 distillation&#8221;, in which a slow and expensive &#8220;System 2&#8221; process generates training data to train a fast and cheap &#8220;System 1&#8221; model. This works well for a game like Go which is a completely self-contained environment. Adapting self-play to domains beyond games is a valuable research direction. There are important domains like code generation where this strategy may be valuable. But we certainly can&#8217;t expect indefinite self-improvement for more open-ended tasks, say language translation. We should expect domains that admit significant improvement through self-play to be the exception rather than the rule.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.normaltech.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">You&#8217;re reading AI Snake Oil, a blog about our <a href="https://www.aisnakeoil.com/p/ai-snake-oil-is-now-available-to">book</a>. Subscribe to get new posts.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><h4><strong>Models have been getting smaller but are being trained for longer</strong></h4><p>Historically, the three axes of scaling &#8212; dataset size, model size, and training compute &#8212; have progressed in <a href="https://www.cnas.org/publications/reports/future-proofing-frontier-ai-regulation">tandem</a>, and this is known to be optimal. But what will happen if one of the axes (high-quality data) becomes a bottleneck? Will the other two axes, model size and training compute, continue to scale?</p><p>Based on current market trends, building bigger models does not seem like a wise business move, even if it would unlock new emergent capabilities. That&#8217;s because capability is no longer the barrier to adoption. In other words, there are many applications that are possible to build with <em>current</em> LLM capabilities but aren&#8217;t being built or adopted due to cost, among other reasons. This is especially true for &#8220;agentic&#8221; workflows which might invoke LLMs <a href="https://www.aisnakeoil.com/p/ai-leaderboards-are-no-longer-useful">tens or hundreds of times</a> to complete a task, such as <a href="https://www.youtube.com/watch?v=tNmgmwEtoWE">code generation</a>. </p><p>In the past year, much of the development effort has gone into producing <em>smaller</em> models at a given capability level.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-5" href="#footnote-5" target="_self">5</a> Frontier model developers no longer reveal model sizes, so we can&#8217;t be sure of this, but we can make educated guesses by using API pricing as a rough proxy for size. GPT-4o costs only 25% as much as GPT-4 does, while being similar or better in capabilities. We see the same pattern with Anthropic and Google. Claude 3 Opus is the most expensive (and presumably biggest) model in the Claude family, but the more recent Claude 3.5 Sonnet is both 5x cheaper and more capable. Similarly, Gemini 1.5 Pro is both cheaper and more capable than Gemini 1.0 Ultra. So with all three developers, the biggest model isn&#8217;t the most capable!</p><p>Training compute, on the other hand, will probably continue to scale for the time being. Paradoxically, smaller models require <em><a href="https://arxiv.org/abs/2203.15556">more</a></em><a href="https://arxiv.org/abs/2203.15556"> training</a> to reach the same level of performance. So the downward pressure on model size is putting upward pressure on training compute. In effect, developers are trading off training cost and inference cost. The earlier crop of models such as GPT-3.5 and GPT-4 was under-trained in the sense that inference costs over the model's lifetime are thought to dominate training cost. Ideally, the two should be roughly equal, given that it is always possible to <a href="https://epochai.org/blog/trading-off-compute-in-training-and-inference">trade off</a> training cost for inference cost and vice versa. In a notable example of this trend, Llama 3 used <em>20 times</em> as many training FLOPs for the 8 billion parameter model as the original Llama model did at roughly the same size (7 billion).</p><h4><strong>The ladder of generality</strong></h4><p>One sign consistent with the possibility that we won&#8217;t see much more capability improvement through scaling is that CEOs have been greatly <a href="https://www.cnbc.com/2024/01/16/openais-sam-altman-agi-coming-but-is-less-impactful-than-we-think.html">tamping down</a> AGI expectations. Unfortunately, instead of admitting they were wrong about their naive &#8220;AGI in 3 years&#8221; predictions, they've decided to save face by watering down what they mean by AGI so much that it's meaningless now. It helped that AGI was <a href="https://www.scientificamerican.com/article/what-does-artificial-general-intelligence-actually-mean/">never clearly defined</a> to begin with.</p><p>Instead of viewing generality as a binary, we can view it as a spectrum. Historically, the amount of effort it takes to get a computer to program a new task has decreased. We can view this as increasing generality. This trend began with the move from special-purpose computers to Turing machines. In this sense, the general-purpose nature of LLMs is not new.&nbsp;</p><p>This is the view we take in the <a href="https://www.amazon.com/Snake-Oil-Artificial-Intelligence-Difference/dp/069124913X">AI Snake Oil book</a>, which has a chapter dedicated to AGI. We conceptualize the history of AI as a punctuated equilibrium, which we call the ladder of generality (which isn&#8217;t meant to imply linear progress). Instruction-tuned LLMs are the latest step in the ladder. An unknown number of steps lie ahead before we can reach a level of generality where AI can perform any economically valuable job as effectively as any human (which is one definition of AGI).&nbsp;</p><p>Historically, standing on each step of the ladder, the AI research community has been terrible at predicting how much farther you can go with the current paradigm, what the next step will be, when it will arrive, what new applications it will enable, and what the implications for safety are. <em>That</em> is a trend we think will continue.</p><h4><strong>Further reading</strong></h4><p>A recent <a href="https://situational-awareness.ai/">essay</a> by Leopold Aschenbrenner made waves due to its claim that &#8220;AGI by 2027 is strikingly plausible&#8221;. We haven&#8217;t tried to give a point-by-point rebuttal here &#8212; most of this post was drafted before Aschenbrenner&#8217;s essay was released. His arguments for his timeline are entertaining and thought provoking, but fundamentally an exercise in trendline extrapolation. Also, like many AI boosters, he <a href="https://www.aisnakeoil.com/p/gpt-4-and-professional-benchmarks">conflates</a> benchmark performance with real-world usefulness.</p><p>Many AI researchers have made the skeptical case, including <a href="https://www.science.org/doi/10.1126/science.ado7069">Melanie Mitchell</a>, <a href="https://x.com/ylecun/status/1796982509567180927">Yann LeCun</a>, <a href="https://garymarcus.substack.com/p/breaking-news-scaling-will-never">Gary Marcus</a>, <a href="https://www.youtube.com/watch?v=UakqL6Pj9xo">Francois Chollet</a>, and <a href="https://arxiv.org/pdf/2402.01817">Subbarao Kambhampati and others</a>.</p><p>Dwarkesh Patel gives a nice <a href="https://www.dwarkeshpatel.com/p/will-scaling-work">overview</a> of both sides of the debate.</p><p><strong>Acknowledgements. </strong>We are grateful to Matt Salganik, Ollie Stephenson, and Benedikt Str&#246;bl for feedback on a draft.</p><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>Emergent abilities will be predictable if we can find a metric that changes <a href="https://arxiv.org/abs/2304.15004">smoothly</a> instead of discontinuously, but finding such a metric <a href="https://cset.georgetown.edu/article/emergent-abilities-in-large-language-models-an-explainer/">isn&#8217;t easy</a>, especially for tasks that require a <a href="https://windowsontheory.org/2023/12/22/emergent-abilities-and-grokking-fundamental-mirage-or-both/">combination</a> of skills. In practice, the question of whether and which new abilities will emerge at the next order of magnitude remains anyone&#8217;s guess.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p>AI companies do use <a href="https://www.nytimes.com/2024/04/06/technology/tech-giants-harvest-data-artificial-intelligence.html">transcribed YouTube data</a> for training, but the reason it is valuable is that it helps LLMs learn what spoken conversations look like, not because of its volume.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-3" href="#footnote-anchor-3" class="footnote-number" contenteditable="false" target="_self">3</a><div class="footnote-content"><p>Libertarian commentators predictably attribute the stagnation of airplane speeds entirely to <a href="https://www.mercatus.org/research/data-visualizations/airplane-speeds-have-stagnated-40-years">regulation</a>, but this is wrong or, at best, highly oversimplified. It&#8217;s true that the FAA essentially banned supersonic flight by civil aircraft over land in the U.S. in 1973. But the fastest aircraft are all military, so the ban doesn&#8217;t affect them. And civil aircraft cruise well below Mach 1 due to fuel efficiency and other considerations.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-4" href="#footnote-anchor-4" class="footnote-number" contenteditable="false" target="_self">4</a><div class="footnote-content"><p>There is a debate about whether LLM training can be made orders of magnitude more sample efficient. After all, children acquire language after being exposed to far fewer words than LLMs are. On the other hand, children are &#8220;<a href="https://www.amazon.com/Scientist-Crib-Early-Learning-Tells/dp/0688177883">scientists in the crib</a>&#8221;, developing world models and reasoning abilities early on, which might be what enables efficient language acquisition. This debate is orthogonal to our point. If task representation or difficulty of extrapolation is the bottleneck, it will represent an upper limit on LLM capabilities regardless of sample efficiency.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-5" href="#footnote-anchor-5" class="footnote-number" contenteditable="false" target="_self">5</a><div class="footnote-content"><p>Even when model developers have released larger models (in terms of parameter count), there is an increased focus on inference efficiency, such as in mixture-of-experts models like <a href="https://mistral.ai/news/mixtral-8x22b/">Mixtral 8x22B</a>, where the number of active parameters during inference is much lower than the total parameter count.</p></div></div>]]></content:encoded></item><item><title><![CDATA[Scientists should use AI as a tool, not an oracle]]></title><description><![CDATA[How AI hype leads to flawed research that fuels more hype]]></description><link>https://www.normaltech.ai/p/scientists-should-use-ai-as-a-tool</link><guid isPermaLink="false">https://www.normaltech.ai/p/scientists-should-use-ai-as-a-tool</guid><dc:creator><![CDATA[Arvind Narayanan]]></dc:creator><pubDate>Mon, 03 Jun 2024 18:34:13 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!8Au-!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F877ad57f-f2fc-4b70-9f03-a4bd52e490a7_1296x822.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Who produces AI hype? As we discuss in the AI Snake Oil <a href="https://www.aisnakeoil.com/p/ai-snake-oil-is-now-available-to">book</a>, it is not just companies and the media but also AI researchers. For example, a pair of widely-publicized papers in Nature in December 2023 claimed to have <a href="https://www.nature.com/articles/s41586-023-06735-9">discovered</a> over 2.2 million new materials using AI, and robotically <a href="https://www.nature.com/articles/s41586-023-06734-w">synthesized</a> 41 of them. Unfortunately, the claims were <a href="https://chemrxiv.org/engage/chemrxiv/article-details/65957d349138d231611ad8f7">quickly</a> <a href="https://x.com/Robert_Palgrave/status/1744383965270581615">debunked</a>: &#8220;Most of the [41] materials produced were misidentified, and the rest were already known&#8221;. As for the large dataset, examining a sample of 250 compounds showed that it was <a href="https://pubs.acs.org/doi/10.1021/acs.chemmater.4c00643">mostly junk</a>.</p><p>A core selling point of machine learning is discovery without understanding, which is why errors are particularly common in machine-learning-based science. Three years ago, we <a href="https://reproducible.cs.princeton.edu/">compiled evidence</a> revealing that an error called leakage &#8212; the machine learning version of teaching to the test &#8212; was pervasive, affecting hundreds of papers from 17 disciplines. Since then, we have been trying to understand the problem better and devise solutions.&nbsp;</p><p>This post presents an update. In short, we think things will get worse before they get better, although there are glimmers of hope on the horizon.&nbsp;</p><p><strong>The carnage continues</strong></p><p>In our most recent compilation, the number of disciplines where researchers have <a href="https://reproducible.cs.princeton.edu/#rep-failures">uncovered leakage</a> in published work has reached 30. The majority are medical fields, which we strongly suspect is due to the fact that since errors in medical research can be particularly consequential, medical fields seem to put much more effort into establishing best practices and critically reviewing previously published work. About 650 papers across all fields are affected, which we hypothesize is a vast underestimate &#8212; when researchers look for leakage systematically, in many fields they find that the <em>majority</em> of sampled studies commit the error of leakage.</p><p>Leakage is one of many reasons for reproducibility failures. There are widespread <a href="https://reforms.cs.princeton.edu/appendix3.html">shortcomings</a> in every step of ML-based science, from data collection to preprocessing and reporting results. Problems that might lead to irreproducibility include improper comparisons to baselines, unrepresentative samples, results being sensitive to specific modeling choices, and not reporting model uncertainties. There is also the basic problem of researchers failing to publish their code and data, precluding reproducibility. For example, Gabelica et al. <a href="https://pubmed.ncbi.nlm.nih.gov/35654271/">examined</a> 333 open-access journals indexed on BioMed Central in January 2019 and found that out of the 1,800 papers that pledged to share data upon request, 93% did not do so.</p><p><strong>The roots run deep</strong></p><p>Even before ML, many scientific fields have been facing reproducibility and replicability crises. The root causes include the publish-or-perish culture in science, the strong bias for publishing positive results (and the near-impossibility of publishing negative results), the lack of incentives for debunking faulty studies, and the lack of consequences for publishing shoddy work. For example, faulty papers are almost never retracted. Peers don&#8217;t even seem to notice replication failures &#8212; after a paper fails to replicate, <a href="https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10460510/">only 3%</a> of citing articles cited the replication attempt.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a> Science communicators love to claim that science self-corrects, but self-correction is practically nonexistent in our experience.</p><p>All of these cultural factors are also present in ML-based science. But ML introduces a bunch of additional reasons why we should be skeptical of published results. Performance evaluation is notoriously tricky and many aspects of it, such as uncertainty quantification, are unresolved research areas. Also, ML code tends to vastly more complex and less standardized than traditional statistical modeling. Since it is not peer reviewers&#8217; job to review code, coding errors are rarely discovered.</p><p>But we think the biggest reason for the poor quality of research is pervasive hype, resulting in the lack of a skeptical mindset among researchers, which is a cornerstone of good scientific practice. We&#8217;ve observed that when researchers have overoptimistic expectations, and their ML model performs poorly, they assume that they did something wrong and tweak the model, when in fact they should strongly consider the possibility that they have run up against inherent <a href="https://msalganik.github.io/cos597E-soc555_f2020/">limits to predictability</a>. Conversely, they tend to be credulous when their model performs well, when in fact they should be on high alert for leakage or other flaws. And if the model performs better than expected, they assume that it has discovered patterns in the data that no human could have thought of, and the myth of AI as an alien intelligence makes this explanation seem readily plausible.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!8Au-!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F877ad57f-f2fc-4b70-9f03-a4bd52e490a7_1296x822.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!8Au-!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F877ad57f-f2fc-4b70-9f03-a4bd52e490a7_1296x822.png 424w, https://substackcdn.com/image/fetch/$s_!8Au-!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F877ad57f-f2fc-4b70-9f03-a4bd52e490a7_1296x822.png 848w, https://substackcdn.com/image/fetch/$s_!8Au-!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F877ad57f-f2fc-4b70-9f03-a4bd52e490a7_1296x822.png 1272w, https://substackcdn.com/image/fetch/$s_!8Au-!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F877ad57f-f2fc-4b70-9f03-a4bd52e490a7_1296x822.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!8Au-!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F877ad57f-f2fc-4b70-9f03-a4bd52e490a7_1296x822.png" width="373" height="236.5787037037037" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/877ad57f-f2fc-4b70-9f03-a4bd52e490a7_1296x822.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:822,&quot;width&quot;:1296,&quot;resizeWidth&quot;:373,&quot;bytes&quot;:83688,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!8Au-!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F877ad57f-f2fc-4b70-9f03-a4bd52e490a7_1296x822.png 424w, https://substackcdn.com/image/fetch/$s_!8Au-!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F877ad57f-f2fc-4b70-9f03-a4bd52e490a7_1296x822.png 848w, https://substackcdn.com/image/fetch/$s_!8Au-!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F877ad57f-f2fc-4b70-9f03-a4bd52e490a7_1296x822.png 1272w, https://substackcdn.com/image/fetch/$s_!8Au-!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F877ad57f-f2fc-4b70-9f03-a4bd52e490a7_1296x822.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div><p>This is a feedback loop. Overoptimism fuels flawed research which further misleads other researchers in the field about what they should and shouldn&#8217;t expect AI to be able to do. In fact, we&#8217;ve encountered extreme versions of this in private correspondence with frustrated researchers: since flawed research goes uncorrected, it becomes literally impossible to publish good research since it will result in models that don&#8217;t beat the &#8220;state of the art&#8221;.</p><p>The more powerful and more black-box the tool, the more the potential for errors and overconfidence. The replication crises in psychology, medicine, etc. were the result of misapplication of plain old statistics. Given how relatively new ML is, our guess is that the reproducibility crisis in ML-based science will get worse for a while before it starts to get better. And now scientists are embracing large language models and generative AI, which open up many new pitfalls such as the <a href="https://www.nature.com/articles/s41586-024-07146-0">illusion of understanding</a>.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.normaltech.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">You&#8217;re reading AI Snake Oil, a blog about our <a href="https://www.aisnakeoil.com/p/ai-snake-oil-is-now-available-to">book</a>. Subscribe to get new posts.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p></p><p><strong>Glimmers of hope</strong></p><p>One good thing about ML-based science is that it usually involves only data analysis, not experimenting on people. So other researchers should in principle be able to download a paper&#8217;s code and data and check whether they can reproduce the reported results. They can also review the code for any errors or problematic choices. This is time consuming, but much less so than replicating a study in psychology or medicine, which is typically almost as costly as the original study.</p><p>Another good thing is that the vast majority of errors can be avoided if the researchers know what to look out for. In contrast, mitigations for the replication crisis in statistical science, such as pre-registration, have a much more spotty track record of effectiveness.</p><p>So we think that the problem can be greatly mitigated by a culture change where researchers systematically exercise more care in their work and reproducibility studies are incentivized. The ML <em>methods</em> community has already moved in this direction via the <a href="https://www.simonsfoundation.org/event/reproducible-research-and-the-common-task-method/">common task method</a> (which is decades old) and the <a href="https://arxiv.org/abs/2003.12206">reproducibility challenge</a> (which is more recent), but this has not yet happened in ML-based science, that is, in disciplines like medicine or psychology that use ML models to advance knowledge in their respective fields.</p><p>We have led a few efforts to change this. First, our leakage paper has had an impact. It has been used by researchers to clarify how they build models and document and demonstrate the <a href="https://sportrxiv.org/index.php/server/preprint/view/191/351">absence of leakage</a>. It has been used by researchers trying to find leakage in <a href="https://arxiv.org/pdf/2401.14497">published work</a>. It has also been used as a way to underscore the importance of studying leakage and coming up with <a href="https://www.nature.com/articles/s41559-023-02162-1">discipline-specific</a> <a href="https://www.sciencedirect.com/science/article/pii/S0928098723001926">guidelines</a>.</p><p>Beyond leakage, we led a group of 19 researchers across computer science, data science, social sciences, mathematics, and biomedical research to develop the <a href="https://reforms.cs.princeton.edu/">REFORMS</a> checklist for ML-based science. It is a 32-item checklist that can help researchers catch eight kinds of common pitfalls in ML-based science, of which leakage is only one. It was recently <a href="https://www.science.org/doi/epdf/10.1126/sciadv.adk3452">published</a> in Science Advances. Of course, checklists by themselves won&#8217;t help if there isn&#8217;t a culture change, but based on the reception so far, we are cautiously optimistic.</p><p><strong>Concluding thoughts</strong></p><p>Our point isn&#8217;t that AI is useless to scientists. We ourselves frequently use AI as a tool, even in our research that&#8217;s not about AI. The key word is tool. AI is not a revolution. It is not a replacement for human understanding &#8212; to think so is to miss the point of science. AI does not offer a shortcut to the hard work and frustration inherent to research. AI is not an oracle and cannot see the future.</p><p>Unfortunately, most scientific fields have succumbed to AI hype, leading to a suspension of common sense. For example, a line of research in political science claimed to predict the onset of civil war with an accuracy<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a> of well over 90%, a number that should sound facially impossible. (It <a href="https://reproducible.cs.princeton.edu/#civil-war">turned out</a> to be leakage, which is what got us interested in this whole line of research.)</p><p>We are at an interesting moment in the history of science. Look at these graphs showing the adoption of AI in various fields:<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-3" href="#footnote-3" target="_self">3</a></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Ik9I!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4527ec18-fa1f-49aa-95cd-f4395fed3dd6_1560x1234.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Ik9I!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4527ec18-fa1f-49aa-95cd-f4395fed3dd6_1560x1234.png 424w, https://substackcdn.com/image/fetch/$s_!Ik9I!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4527ec18-fa1f-49aa-95cd-f4395fed3dd6_1560x1234.png 848w, https://substackcdn.com/image/fetch/$s_!Ik9I!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4527ec18-fa1f-49aa-95cd-f4395fed3dd6_1560x1234.png 1272w, https://substackcdn.com/image/fetch/$s_!Ik9I!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4527ec18-fa1f-49aa-95cd-f4395fed3dd6_1560x1234.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Ik9I!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4527ec18-fa1f-49aa-95cd-f4395fed3dd6_1560x1234.png" width="1456" height="1152" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/4527ec18-fa1f-49aa-95cd-f4395fed3dd6_1560x1234.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1152,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Ik9I!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4527ec18-fa1f-49aa-95cd-f4395fed3dd6_1560x1234.png 424w, https://substackcdn.com/image/fetch/$s_!Ik9I!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4527ec18-fa1f-49aa-95cd-f4395fed3dd6_1560x1234.png 848w, https://substackcdn.com/image/fetch/$s_!Ik9I!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4527ec18-fa1f-49aa-95cd-f4395fed3dd6_1560x1234.png 1272w, https://substackcdn.com/image/fetch/$s_!Ik9I!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4527ec18-fa1f-49aa-95cd-f4395fed3dd6_1560x1234.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Percentage of AI-engaged papers by field, 1985&#8211;2023 by field. (<a href="https://arxiv.org/abs/2405.15828">Source: Duede et al. 2024</a>)</figcaption></figure></div><p>These hockey stick graphs are not good news. They should be terrifying. Adopting AI requires changes to scientific epistemology.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-4" href="#footnote-4" target="_self">4</a> No scientific field has the capacity to accomplish this on a timescale of a couple of years. This is not what happens when a tool or method is adopted organically. It happens when scientists jump on a trend to get funding. Given the level of hype, scientists don&#8217;t need <em>additional</em> incentives to adopt AI. That means AI-for-science funding programs are probably making things worse. We doubt the avalanche of flawed research can be stopped, but if at least a fraction of AI-for-science funding were diverted to better training, critical inquiry, meta-science, reproducibility, and other quality-control efforts, the havoc can be minimized.</p><p></p><p>Our book <em>AI Snake Oil</em> is now available to preorder. If you have enjoyed our blog and would like to support our work, please preorder via <a href="https://substack.com/redirect/f0945e82-44e2-4998-9535-cee6fe9f0fa5?j=eyJ1IjoiYmd4a3MifQ.EjRMsvQe8Xc2mF1xAwL5aBabUU37X2wfP2-gBTgHzJM">Amazon</a>, <a href="https://substack.com/redirect/fd84f1c5-276b-488f-b91b-3cc7f32821ba?j=eyJ1IjoiYmd4a3MifQ.EjRMsvQe8Xc2mF1xAwL5aBabUU37X2wfP2-gBTgHzJM">Bookshop</a>, or your favorite bookseller.</p><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>To be clear, replication failures don&#8217;t necessarily imply flaws in the original study. Our concern in this post is primarily about relatively clear-cut errors such as leakage.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p>Accuracy here refers to a metric called AUC; the baseline AUC is 50% even when one outcome (peace) is much more common than the other (war).</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-3" href="#footnote-anchor-3" class="footnote-number" contenteditable="false" target="_self">3</a><div class="footnote-content"><p>The paper clubs together different types of AI &#8220;engagement&#8221;: <em>Engagement could include (but is not limited to) the development of novel AI theory and approaches, technologies, or applications; the general use of AI models for domain-specific tasks; and critical engagement with AI, as typified by academic discourse in fields like philosophy and ethics.</em> This is unfortunate for our purposes, as our concern is solely about the second category, the use of AI for domain-specific tasks. We do think that outside of a few fields like computer science and philosophy, most AI engagement falls into this category.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-4" href="#footnote-anchor-4" class="footnote-number" contenteditable="false" target="_self">4</a><div class="footnote-content"><p>In particular, as the saying goes, &#8220;all models are wrong but some models are useful&#8221;. There is no straightforward answer to the question of when we can draw conclusions about the world based on a model, so validity has to be re-litigated in every field and for every type of model.</p><p></p></div></div>]]></content:encoded></item><item><title><![CDATA[AI leaderboards are no longer useful. It's time to switch to Pareto curves.]]></title><description><![CDATA[What spending $2,000 can tell us about evaluating AI agents]]></description><link>https://www.normaltech.ai/p/ai-leaderboards-are-no-longer-useful</link><guid isPermaLink="false">https://www.normaltech.ai/p/ai-leaderboards-are-no-longer-useful</guid><dc:creator><![CDATA[Sayash Kapoor]]></dc:creator><pubDate>Tue, 30 Apr 2024 14:03:42 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/eba33d2f-7779-4659-97e0-c55e5608d580_3404x1902.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>By Sayash Kapoor, <a href="https://citp.princeton.edu/citp-people/benedikt-strobl/">Benedikt Stroebl</a>, Arvind Narayanan</em></p><p>Which is the most accurate AI system for generating code? Surprisingly, there isn&#8217;t currently a good way to answer questions like these.&nbsp; </p><p>Based on <a href="https://paperswithcode.com/sota/code-generation-on-humaneval">HumanEval</a>, a widely used benchmark for code generation, the most accurate publicly available system is <a href="https://arxiv.org/abs/2402.16906">LDB</a> (short for LLM debugger).<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a> But there&#8217;s a catch. The most accurate generative AI systems, including LDB, tend to be agents,<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a> which repeatedly invoke language models like GPT-4. That means they can be orders of magnitude more costly to run than the models themselves (which are already pretty costly). If we eke out a 2% accuracy improvement for 100x the cost, is that really better?</p><p>In this post, we argue that:</p><ul><li><p>AI agent accuracy measurements that don&#8217;t control for cost aren&#8217;t useful.</p></li><li><p>Pareto curves can help visualize the accuracy-cost tradeoff.</p></li><li><p>Current state-of-the-art agent architectures are complex and costly but no more accurate than extremely simple baseline agents that cost 50x less in some cases.</p></li><li><p>Proxies for cost such as parameter count are misleading if the goal is to identify the best system for a given task. We should directly measure dollar costs instead.</p></li><li><p>Published agent evaluations are difficult to reproduce because of a lack of standardization and questionable, undocumented evaluation methods in some cases.</p><p></p></li></ul><h4><strong>Maximizing accuracy can lead to unbounded cost</strong></h4><p>LLMs are stochastic. Simply calling a model <a href="https://arxiv.org/abs/2402.05120">many</a> <a href="https://arxiv.org/pdf/2403.02419.pdf">times</a> and outputting the most common answer can increase accuracy.&nbsp;</p><p>On some tasks, there is seemingly no limit to the amount of inference compute that can improve accuracy.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-3" href="#footnote-3" target="_self">3</a> Google Deepmind's <a href="https://arxiv.org/pdf/2203.07814.pdf">AlphaCode</a>, which improved accuracy on automated coding evaluations, showed that this trend holds even when calling LLMs millions of times.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!2od5!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa22df54d-213b-44da-aacd-a55fb24f9089_688x544.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!2od5!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa22df54d-213b-44da-aacd-a55fb24f9089_688x544.jpeg 424w, https://substackcdn.com/image/fetch/$s_!2od5!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa22df54d-213b-44da-aacd-a55fb24f9089_688x544.jpeg 848w, https://substackcdn.com/image/fetch/$s_!2od5!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa22df54d-213b-44da-aacd-a55fb24f9089_688x544.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!2od5!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa22df54d-213b-44da-aacd-a55fb24f9089_688x544.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!2od5!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa22df54d-213b-44da-aacd-a55fb24f9089_688x544.jpeg" width="472" height="373.2093023255814" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/a22df54d-213b-44da-aacd-a55fb24f9089_688x544.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:544,&quot;width&quot;:688,&quot;resizeWidth&quot;:472,&quot;bytes&quot;:64235,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!2od5!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa22df54d-213b-44da-aacd-a55fb24f9089_688x544.jpeg 424w, https://substackcdn.com/image/fetch/$s_!2od5!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa22df54d-213b-44da-aacd-a55fb24f9089_688x544.jpeg 848w, https://substackcdn.com/image/fetch/$s_!2od5!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa22df54d-213b-44da-aacd-a55fb24f9089_688x544.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!2od5!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa22df54d-213b-44da-aacd-a55fb24f9089_688x544.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em>The accuracy of <a href="https://arxiv.org/pdf/2203.07814.pdf">AlphaCode</a> on coding tasks continues to improve even after making a million calls to the underlying model (the different curves represent varying parameter counts). Accuracy is measured by how often one of the top 10 answers generated by the model is correct.</em></figcaption></figure></div><p>A useful evaluation of agents must therefore ask: What did it cost? If we don&#8217;t do cost-controlled comparisons, it will encourage researchers to develop extremely costly agents just to claim they topped the leaderboard.</p><p>In fact, when we evaluate agents that have been proposed in the last year for solving coding tasks, we find that visualizing the tradeoff between cost and accuracy yields surprising insights.&nbsp;</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.normaltech.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Subscribe to AI Snake Oil receive new posts.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p></p><h4><strong>Visualizing the accuracy-cost tradeoff on HumanEval, with new baselines</strong></h4><p>We re-evaluated the accuracy of three agents that have been claimed to occupy top spots on the HumanEval leaderboard: <a href="https://arxiv.org/pdf/2402.16906v3.pdf">LDB</a>, <a href="https://arxiv.org/pdf/2310.04406v2.pdf">LATS</a>, and <a href="https://arxiv.org/pdf/2303.11366.pdf">Reflexion</a>.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-4" href="#footnote-4" target="_self">4</a> We also evaluated the cost and time requirements of running these agents.&nbsp;</p><p>These agents rely on running the code generated by the model, and if it fails the test cases provided with the problem description, they try to debug the code, look at alternative paths in the code generation process, or "reflect" on why the model's outputs were incorrect before generating another solution.&nbsp;</p><p>In addition, we calculated the accuracy, cost, and running time of&nbsp; a few simple baselines:&nbsp;</p><ul><li><p><strong>GPT-3.5</strong> and <strong>GPT-4 </strong>models (zero shot; no agent architecture<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-5" href="#footnote-5" target="_self">5</a>)</p></li><li><p><strong>Retry:</strong> We repeatedly invoke a model with the temperature set to zero, up to five times, if it fails the test cases provided with the problem description.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-6" href="#footnote-6" target="_self">6</a> Retrying makes sense because LLMs <a href="https://community.openai.com/t/run-same-query-many-times-different-results/140588#:~:text=Apr%202023-,OpenAI%20models%20are%20non%2Ddeterministic%2C%20meaning%20that%20identical%20inputs%20can%20yield,amount%20of%20variability%20may%20remain%20due%20to%20GPU%20floating%20point%20math.,-Solution">aren&#8217;t deterministic</a> even at temperature zero.</p></li><li><p><strong>Warming: </strong>This is the same as the retry strategy, but we gradually increase the temperature of the underlying model with each run, from 0 to 0.5. This increases the stochasticity of the model and, we hope, increases the likelihood that at least one of the retries will succeed.</p></li><li><p><strong>Escalation:</strong> We start with a cheap model (Llama-3 8B) and escalate to more expensive models (GPT-3.5, Llama-3 70B, GPT-4) if we encounter a test case failure.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-7" href="#footnote-7" target="_self">7</a></p></li></ul><p>Surprisingly, we are not aware of any papers that compare their proposed agent architectures with any of the latter three simple baselines.</p><p>Our most striking result is that <strong>agent architectures for HumanEval do not outperform our simpler baselines despite costing more</strong>. In fact, agents differ drastically in terms of cost: for substantially similar accuracy, the cost can differ by almost two orders of magnitude!<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-8" href="#footnote-8" target="_self">8</a> Yet, the cost of running these agents isn't a top-line metric reported in any of these papers.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-9" href="#footnote-9" target="_self">9</a></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!_AjE!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdcb16a9f-29bd-47a4-813e-929685ce3fd8_1029x705.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!_AjE!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdcb16a9f-29bd-47a4-813e-929685ce3fd8_1029x705.png 424w, https://substackcdn.com/image/fetch/$s_!_AjE!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdcb16a9f-29bd-47a4-813e-929685ce3fd8_1029x705.png 848w, https://substackcdn.com/image/fetch/$s_!_AjE!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdcb16a9f-29bd-47a4-813e-929685ce3fd8_1029x705.png 1272w, https://substackcdn.com/image/fetch/$s_!_AjE!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdcb16a9f-29bd-47a4-813e-929685ce3fd8_1029x705.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!_AjE!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdcb16a9f-29bd-47a4-813e-929685ce3fd8_1029x705.png" width="628" height="430.26239067055394" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/dcb16a9f-29bd-47a4-813e-929685ce3fd8_1029x705.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:705,&quot;width&quot;:1029,&quot;resizeWidth&quot;:628,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!_AjE!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdcb16a9f-29bd-47a4-813e-929685ce3fd8_1029x705.png 424w, https://substackcdn.com/image/fetch/$s_!_AjE!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdcb16a9f-29bd-47a4-813e-929685ce3fd8_1029x705.png 848w, https://substackcdn.com/image/fetch/$s_!_AjE!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdcb16a9f-29bd-47a4-813e-929685ce3fd8_1029x705.png 1272w, https://substackcdn.com/image/fetch/$s_!_AjE!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdcb16a9f-29bd-47a4-813e-929685ce3fd8_1029x705.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em>Our simple baselines offer Pareto improvements over existing agent architectures.&nbsp;We run each agent five times and report the mean accuracy and the mean total cost on the 164 HumanEval problems. Where results for LDB have two models/agents in parenthesis, they indicate the language model or agent used to generate the code, followed by the language model used to debug the code. Where they have just one, they indicate that the same model was used to both generate the code and debug it. Note that the y-axis is shown from 0.7 to 1; figures with the full axis (0 to 1) and error bars, robustness checks, and other details about our empirical results are included in the <a href="https://www.cs.princeton.edu/~sayashk/papers/ai-leaderboards-appendix.pdf">appendix</a>.&nbsp;</em></figcaption></figure></div><p>There is no significant accuracy difference between the warming strategy and the best-performing agent architecture. Yet, Reflexion and LDB cost over 50% more than the warming strategy,<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-10" href="#footnote-10" target="_self">10</a> and LATS over 50 times more (all these costs are entirely or predominantly from calls to GPT-4, so these ratios will be stable even if model costs change).&nbsp; Meanwhile, the escalation strategy strictly improves accuracy while costing less than half of LDB (GPT-3.5) at current inference prices.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-11" href="#footnote-11" target="_self">11</a>&nbsp;</p><p>Our results point to another underlying problem: papers making claims about the usefulness of agents have so far failed to test if simple agent baselines can lead to similar accuracy. This has led to widespread beliefs among AI researchers that complex ideas like planning, reflection, and debugging are responsible for accuracy gains. In fact, Lipton and Steinhardt noted a trend in the AI literature of failing to identify the sources of empirical gains <a href="https://arxiv.org/abs/1807.03341">back in 2018</a>.</p><p>Based on our findings, the question of whether debugging, reflection, and other such &#8220;System 2&#8221; approaches are useful for code generation remains open.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-12" href="#footnote-12" target="_self">12</a> It is possible that they will be useful on harder programming tasks than those represented in HumanEval. For now, the over-optimism about System 2 approaches is exacerbated by a lack of reproducibility and standardization that we report below.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-13" href="#footnote-13" target="_self">13</a></p><h4><strong>Proxies for cost are misleading</strong></h4><p>At first glance, reporting dollar costs is jarring. It breaks many properties of benchmarking that we take for granted: that measurements don&#8217;t change over time (whereas costs tend to come down) and that different models compete on a level playing field (whereas some developers may <a href="https://papers.nips.cc/paper_files/paper/2023/file/d1a14493e5f84d6c6129414f0cd1a7c6-Paper-Conference.pdf">benefit</a> from economies of scale, leading to lower inference costs). Because of this, researchers usually pick a different axis for the Pareto curve, such as parameter count.</p><p>The downsides of reporting costs are real, but we describe below how they can be mitigated. More importantly, we think using attributes like parameter count as a proxy for cost is a mistake and doesn&#8217;t solve the problem it&#8217;s intended to solve. To understand why, we need to introduce a conceptual distinction.</p><p>AI evaluations serve at least two distinct purposes. Model developers and AI researchers use them to identify which changes to the training data and architecture improve accuracy. We call this <a href="https://arxiv.org/pdf/2211.09110">model evaluation</a>. And downstream developers, such as programmers who use AI to build consumer-facing products, use evaluations to decide which AI systems to use in their products. We call this <a href="https://www.arthur.ai/product/bench">downstream evaluation</a>.&nbsp;</p><p>The difference between model evaluation and downstream evaluation is underappreciated. This has led to much confusion about how to factor in the cost of running AI.</p><p>Model evaluation is a scientific question of interest to researchers. So it makes sense to stay away from dollar costs for the aforementioned reasons. Instead, controlling for compute is a reasonable approach: if we normalize the amount of compute used to train a model, we can then understand if factors like architectural changes or changes in the data composition are responsible for improvements, as opposed to more compute. Notably, Nathan Lambert <a href="https://www.interconnects.ai/p/compute-efficient-open-llms">argues</a> that many of the accuracy gains in the last year (such as Meta's Llama 2) are simply consequences of using more compute.&nbsp;</p><p>On the other hand, downstream evaluation is an engineering question that helps inform a procurement decision. Here, cost is the actual construct of interest. The downsides of cost measurement aren&#8217;t downsides at all; they are exactly what&#8217;s needed. Inference costs do come down over time, and that greatly matters to downstream developers. It is unnecessary and counterproductive for the evaluation to stay frozen in time.</p><p>In this context, proxies for cost (such as the number of active parameters or amount of compute used) are misleading. For example, Mistral <a href="https://mistral.ai/news/mixtral-8x22b/">released</a> the figure below alongside their latest model, Mixtral 8x22B, to explain why developers should choose it over competitors.&nbsp;</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!yu-D!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F019ad4b8-e42b-4f9b-9d91-cb3ccc84fa3a_1030x640.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!yu-D!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F019ad4b8-e42b-4f9b-9d91-cb3ccc84fa3a_1030x640.jpeg 424w, https://substackcdn.com/image/fetch/$s_!yu-D!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F019ad4b8-e42b-4f9b-9d91-cb3ccc84fa3a_1030x640.jpeg 848w, https://substackcdn.com/image/fetch/$s_!yu-D!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F019ad4b8-e42b-4f9b-9d91-cb3ccc84fa3a_1030x640.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!yu-D!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F019ad4b8-e42b-4f9b-9d91-cb3ccc84fa3a_1030x640.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!yu-D!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F019ad4b8-e42b-4f9b-9d91-cb3ccc84fa3a_1030x640.jpeg" width="562" height="349.20388349514565" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/019ad4b8-e42b-4f9b-9d91-cb3ccc84fa3a_1030x640.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:640,&quot;width&quot;:1030,&quot;resizeWidth&quot;:562,&quot;bytes&quot;:73703,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!yu-D!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F019ad4b8-e42b-4f9b-9d91-cb3ccc84fa3a_1030x640.jpeg 424w, https://substackcdn.com/image/fetch/$s_!yu-D!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F019ad4b8-e42b-4f9b-9d91-cb3ccc84fa3a_1030x640.jpeg 848w, https://substackcdn.com/image/fetch/$s_!yu-D!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F019ad4b8-e42b-4f9b-9d91-cb3ccc84fa3a_1030x640.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!yu-D!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F019ad4b8-e42b-4f9b-9d91-cb3ccc84fa3a_1030x640.jpeg 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em>Substituting active parameters as a proxy for cost is misleading. Source: <a href="https://mistral.ai/news/mixtral-8x22b/">Mistral</a>.</em></figcaption></figure></div><p>In this figure, the number of active parameters is a poor proxy for cost. On <a href="https://docs.endpoints.anyscale.com/pricing/">Anyscale</a>, Mixtral 8x7B costs twice as much as Llama 2 13B, yet Mistral's figure shows it costs about the same, because they only consider the number of active parameters. Of course, downstream developers don't care about the number of active parameters when they're using an API. They simply care about the dollar cost relative to accuracy. Mistral chose &#8220;active parameters&#8221; as a proxy, presumably because it makes their models look better than dense models such as Meta&#8217;s Llama and Cohere&#8217;s Command R+. If we start using proxies for cost, every model developer can pick a proxy that makes their model look good.</p><p>Some hurdles to cost evaluation remain. Different providers can charge different amounts for <a href="https://docs.endpoints.anyscale.com/pricing/">the</a> <a href="https://www.together.ai/pricing">same</a> <a href="https://replicate.com/pricing">model</a>, the cost of an API call might change <a href="https://openai.com/blog/new-models-and-developer-products-announced-at-devday">overnight</a>, and cost might vary based on model developer decisions, such as whether <a href="https://twitter.com/OpenAIDevs/status/1779922566091522492?ref_src=twsrc%5Etfw%7Ctwcamp%5Etweetembed%7Ctwterm%5E1779922566091522492%7Ctwgr%5E100b538f87ce81b828a23634fa12d127c126b5ba%7Ctwcon%5Es1_c10&amp;ref_url=https%3A%2F%2Fwww.theverge.com%2F2024%2F4%2F15%2F24131401%2Fopenai-will-give-you-a-50-percent-discount-for-off-peak-gpt-use">bulk API</a> calls are charged differently. These downsides can be partly addressed by making the evaluation results customizable using mechanisms to adjust the cost of running models, i.e., providing users the option to adjust the cost of input and output tokens for their provider of choice to recalculate the tradeoff between cost and accuracy. In turn, downstream evaluations of agents should include input/output token counts in addition to dollar costs, so that anyone looking at the evaluation in the future can instantly recalculate the cost using current prices.</p><p>But ultimately, despite the hurdles, good measurement requires modeling the underlying construct of interest. For downstream evaluations, that underlying construct is cost. All other proxies are lacking.</p><h4><strong>Agent evaluations lack standardization and reproducibility</strong></h4><p>In the course of our evaluation, we found many shortcomings in the reproducibility and standardization of agent evaluations.</p><ul><li><p><strong>We were unable to reproduce the results of the LATS and LDB agents on HumanEval.</strong> In particular, across all 5 runs for LDB (Reflexion, GPT-3.5), the maximum accuracy was 91.5%, much lower than the 95.1% reported in the paper.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-14" href="#footnote-14" target="_self">14</a> The maximum accuracy of LATS across all five runs was similarly lower, at 91.5% instead of 94.4%.&nbsp;</p></li><li><p>Similarly, the accuracy for the baseline GPT-4 model reported in the LDB paper is drastically lower than our reproduction of the paper's code (75.0% vs. a mean of 89.6% across five runs). In fact, according to the paper, the GPT-3.5 and GPT-4 models perform very similarly (73.9% vs. 75.0%).<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-15" href="#footnote-15" target="_self">15</a> Weak baselines could give a false sense of the amount of improvement attributable to the agent architecture.</p></li><li><p>The LATS agent was evaluated on only a <a href="https://github.com/andyz245/LanguageAgentTreeSearch/blob/554886901183a9908183d2cb104c3088c493650a/programming/mcts.py#L138C1-L138C29">subset</a> of the test cases provided in the HumanEval benchmark. This exaggerated their accuracy numbers, since the code for a particular HumanEval problem might be incorrect, but if it passes only a portion of the test cases for that problem, it could still be marked as correct. In our analysis, this was responsible for a 3% difference in accuracy (mean across five runs), which explains a substantial part of the difference between the accuracy we found and the one reported in the paper. In addition, many details about the implementation, such as hyperparameter values, were not reported in the paper or GitHub repository (see <a href="https://www.cs.princeton.edu/~sayashk/papers/ai-leaderboards-appendix.pdf">appendix</a> for details).</p></li><li><p>To the best of our knowledge, this post is the first time the four agents with the highest accuracy&#8212;Retry, Warming, LDB (GPT-4), and LDB (GPT-4 + Reflexion)&#8212;have been tested on HumanEval.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-16" href="#footnote-16" target="_self">16</a>&nbsp;</p></li><li><p><strong>Reflexion, LDB, and LATS all use different subsets of HumanEval.</strong> Three (out of 164) coding problems in the original version of HumanEval lack example tests. Since these agents require example tests to debug or rerun their solutions, Reflexion removes the three problems that don't have example tests. LATS removes these three problems, plus another problem, for unreported reasons.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-17" href="#footnote-17" target="_self">17</a> LDB adds example tests for the three problems that are missing in the original benchmark. <strong>None of the three papers reports this. </strong>The paper introducing LATS claims (incorrectly): "We use all 164 problems for our experiments."<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-18" href="#footnote-18" target="_self">18</a> In our analysis, we conducted all evaluations on the version of the benchmark provided by LDB, since it contains example tests for all problems.</p></li><li><p>The LDB paper claims to use GPT-3.5 for code generation using Reflexion: "For Reflexion, we select the version based on GPT-3.5 and utilize the corresponding generated programs published in the official Github repository." However, the <a href="https://github.com/noahshinn/reflexion/blob/d15acda1c81d464d9a81648d7f29fb951e326c70/programming_runs/root/reflexion_humaneval_py_pass_at_1/reflexion_humaneval_py_pass_at_1.jsonl">generated program</a> they used from the Reflexion repository relies on GPT-4 for code generation, not GPT-3.5.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-19" href="#footnote-19" target="_self">19</a>&nbsp;</p></li></ul><p>These shortcomings in the empirical results have also led to errors of interpretation in broader discussions around the accuracy of AI agents. For example, a recent <a href="https://www.deeplearning.ai/the-batch/issue-241/">post</a> by Andrew Ng claimed that agents that use GPT-3.5 can outperform GPT-4. In particular, he claimed:&nbsp;</p><blockquote><p><em>[For HumanEval,] GPT-3.5 (zero shot) was 48.1% correct. GPT-4 (zero shot) does better at 67.0%. However, the improvement from GPT-3.5 to GPT-4 is dwarfed by incorporating an iterative agent workflow. Indeed, wrapped in an agent loop, GPT-3.5 achieves up to 95.1%.&nbsp;</em></p></blockquote><p>While this claim received a lot of attention, it is incorrect. The claim ("GPT-3.5 wrapped in an agent workflow achieves 95.1% accuracy") seems to be about the LDB agent. The <a href="https://paperswithcode.com/sota/code-generation-on-humaneval">Papers With Code leaderboard</a> for HumanEval makes the same claim. However, as we discussed above, for LDB, GPT-3.5 is only used to find bugs. The code is generated using GPT-4 (or the Reflexion agent that uses GPT-4), not GPT-3.5. Unfortunately, the error in the paper has led to much overoptimism about agents in the broader AI community.</p><p>Ng's post also makes the familiar error of repeating results from papers without verifying them or accounting for changes in prompts and model versions. For example, the zero-shot accuracy numbers of GPT-3.5 (48.1%) and GPT-4 (67.0%) seem to be copied from the <a href="https://arxiv.org/abs/2303.08774">GPT-4 technical report</a> from March 2023. However, the models have been updated many times since release. Indeed, in our comparison, we find that the base models perform much better compared to the claimed figures in Ng's post when we use them with the prompts provided with the LDB paper (GPT-3.5: 73.9%, GPT-4: 89.6%). As a result, the post drastically overestimates the improvement attributable to agent architectures.</p><p>Evaluation frameworks like Stanford's <a href="https://crfm.stanford.edu/helm/lite/latest/">HELM</a> and EleutherAI's <a href="https://www.eleuther.ai/projects/large-language-model-evaluation">LM Evaluation Harness</a> attempt to fix similar shortcomings for model evaluations, by providing standardized evaluation results. We are working on solutions to make agent evaluations standardized and reproducible, especially from the perspective of downstream evaluation of agents.</p><p>Finally, downstream developers should keep in mind that HumanEval or any other standardized benchmark is nothing more than a rough proxy for the specific tasks that arise in a particular downstream application. To understand how agents will perform in practice, it is necessary to evaluate them on a <a href="https://arxiv.org/abs/2404.12272">custom dataset</a> from the domain of interest &#8212; or even better, A/B test different agents in the production environment.</p><h4><strong>Further reading</strong></h4><ul><li><p><a href="https://bair.berkeley.edu/blog/2024/02/18/compound-ai-systems/">Zaharia et al.</a> observe that state-of-the-art accuracy on AI benchmarks is often attained by composite systems. If the adoption of agents continues, visualizing cost and accuracy as a Pareto curve would become even more necessary.</p></li><li><p><a href="https://arxiv.org/pdf/2212.01340.pdf">Santhanam et al.</a> point out the importance of evaluating cost alongside accuracy for information retrieval benchmarks.</p></li><li><p><a href="https://arxiv.org/html/2404.12387v1">Ozrmazabal et al.</a> highlight the <a href="https://arxiv.org/html/2404.12387v1#:~:text=Figure%201%3A,different%20LLM%20APIs.">accuracy vs. cost per output token tradeoffs</a> for various models (but not agents) on MMLU. While the cost of output tokens might not be a good indicator of the overall cost, given the varying input token costs as well as output lengths for different models, it is better than not reporting the tradeoffs at all.&nbsp;</p></li><li><p>The <a href="https://gorilla.cs.berkeley.edu/leaderboard.html">Berkeley Function Calling leaderboard</a> includes various metrics for language model evaluations of function calling, including cost and latency.</p></li><li><p><a href="https://arxiv.org/pdf/2404.07972.pdf">Xie et al.</a> develop OSWorld, a benchmark for evaluating agents in computer environments. In their <a href="https://github.com/xlang-ai/OSWorld">GitHub repository</a> (though not in the paper), they give a rough cost estimate for running various multimodal agents on their benchmark.</p></li><li><p>Unsurprisingly, the main impetus for cost vs. accuracy tradeoffs has come from the <a href="https://x.com/swyx/status/1772799201023557697">downstream developers</a> who <a href="https://x.com/jaredpalmer/status/1783899239140986884">use AI</a>.</p></li><li><p>In a previous talk, we discussed <a href="https://www.cs.princeton.edu/~arvindn/talks/evaluating_llms_minefield/">three major pitfalls</a> in LLM evaluation: prompt sensitivity, construct validity, and contamination. The current research is largely orthogonal: prompt sensitivity isn&#8217;t a concern for agent evaluation (as agents are allowed to define their own prompts); downstream developers can address contamination and construct validity by evaluating on custom datasets.</p></li></ul><p>The code for reproducing our analysis is available <a href="https://github.com/benediktstroebl/agent-eval">here</a>. The <a href="https://www.cs.princeton.edu/~sayashk/papers/ai-leaderboards-appendix.pdf">appendix</a> includes more details about our setup and results.</p><h4><strong>Acknowledgments</strong></h4><p>We thank Rishi Bommasani, Rumman Chowdhury, Percy Liang, Shayne Longpre, Yifan Mai, Nitya Nadgir, Matt Salganik, Hailey Schoelkopf, Zachary Siegel, and Venia Veselovsky for discussions and inputs that informed our analysis. We acknowledge Cunxiang Wang and Ruoxi Ning for their prompt responses to our questions about the NovelQA benchmark.&nbsp;</p><p>We are grateful to the authors of the papers we engage with in this post for their quick responses and for sharing their code, which makes such reproduction analysis possible in the first place. In particular, we are grateful to Zilong Wang (LDB), Andy Zhou (LATS), and Karthik Narasimhan (Reflexion), who gave us feedback in response to an earlier draft of this blog post.</p><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>The leaderboard on the linked page lists <a href="https://paperswithcode.com/paper/agentcoder-multi-agent-based-code-generation">AgentCoder</a> as the most accurate system. However, the code or data for reproducing the results of this agent are not available online, so we do not consider it in this blog post.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p>This post is about agents. Leaderboards are also becoming less useful for evaluating the <a href="https://www.arthur.ai/blog/whats-going-on-with-llm-leaderboards">underlying</a> <a href="https://aclanthology.org/2020.emnlp-main.393/">models</a>. There are many problems, including gameability. But controlling for inference cost isn&#8217;t the main problem, so our arguments don&#8217;t necessarily apply.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-3" href="#footnote-anchor-3" class="footnote-number" contenteditable="false" target="_self">3</a><div class="footnote-content"><p>Tasks where increased compute could help indefinitely are primarily those where verifying whether a solution is correct is easy. In the case of programming questions, this takes the form of test cases that are provided with each question to check if the answer is correct. Other examples include proving theorems, because verifying if a theorem is correct can be straightforward, as well as <a href="https://arxiv.org/pdf/2404.06474.pdf">some tasks on the internet</a> for agents that navigate the web. That said, even for tasks where there is no way to guess a solution and then verify, the costs of different agents can vary by orders of magnitude.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-4" href="#footnote-anchor-4" class="footnote-number" contenteditable="false" target="_self">4</a><div class="footnote-content"><p>We included agents from the HumanEval leaderboard on <a href="https://paperswithcode.com/sota/code-generation-on-humaneval">PapersWithCode</a> that share their code publicly. Reflexion is absent from the PapersWithCode list, but it has a reported accuracy of 91% (higher than any other agents with publicly available code apart from LDB and LATS), so we included it too.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-5" href="#footnote-anchor-5" class="footnote-number" contenteditable="false" target="_self">5</a><div class="footnote-content"><p>For the model evaluation, we only used the description of the coding problem as well as the example tests provided with the HumanEval dataset. Three of the 164 coding problems in HumanEval lack example tests. The authors of LDB include a modified version of HumanEval with example tests included for these three problems. We use this modified version for all experiments.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-6" href="#footnote-anchor-6" class="footnote-number" contenteditable="false" target="_self">6</a><div class="footnote-content"><p>In all of the baselines we provide, we don't use the test cases used to evaluate if the solution is correct when deciding to retry, only the ones in the problem description, to avoid <a href="https://reproducible.cs.princeton.edu/">leakage</a>.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-7" href="#footnote-anchor-7" class="footnote-number" contenteditable="false" target="_self">7</a><div class="footnote-content"><p>We evaluated Llama-3 using together.ai endpoints. The <a href="https://api.together.xyz/models">cost per million tokens on together.ai</a>, for both prompt and completion, is 0.20$ and 0.90$ for Llama-3-8B and Llama-3-70B, respectively.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-8" href="#footnote-anchor-8" class="footnote-number" contenteditable="false" target="_self">8</a><div class="footnote-content"><p>This is also true for other desired properties of agents, such as running time. We report results for time vs. accuracy tradeoffs in the <a href="https://www.cs.princeton.edu/~sayashk/papers/ai-leaderboards-appendix.pdf">appendix</a>.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-9" href="#footnote-anchor-9" class="footnote-number" contenteditable="false" target="_self">9</a><div class="footnote-content"><p>While some of the papers introducing these models discuss cost abstractly, such as the relationship between cost and number of times an agent retries, they don't report any concrete numbers on cost or compare token count to a baseline.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-10" href="#footnote-anchor-10" class="footnote-number" contenteditable="false" target="_self">10</a><div class="footnote-content"><p>The cost comparison is for LDB (Reflexion, GPT-3.5), since that is the top-performing agent reported by the authors of LDB.&nbsp;</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-11" href="#footnote-anchor-11" class="footnote-number" contenteditable="false" target="_self">11</a><div class="footnote-content"><p>In addition to HumanEval, we also ran experiments on the HotPotQA and NovelQA benchmarks for question answering. We found similar results for both benchmarks. In particular, we found that there can be large differences in cost underlying small improvements in accuracy for both benchmarks.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-12" href="#footnote-anchor-12" class="footnote-number" contenteditable="false" target="_self">12</a><div class="footnote-content"><p>One potential concern with our analysis is that while we relied on the April 2024 version of OpenAI models, many papers relied on older model versions for their results. To address this, we report results for an additional robustness check with the June 2023 version of OpenAI models in the <a href="https://www.cs.princeton.edu/~sayashk/papers/ai-leaderboards-appendix.pdf">appendix</a>; we find substantially similar results across model versions.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-13" href="#footnote-anchor-13" class="footnote-number" contenteditable="false" target="_self">13</a><div class="footnote-content"><p>While HumanEval is commonly used to evaluate how well AI can solve coding problems, it is limited due to its small size (only 164 questions), lack of difficult problems (none of the problems involve real-world tasks), and potential <a href="https://arxiv.org/pdf/2311.04850.pdf">contamination</a>, since language models have likely been trained on HumanEval problems, which might inflate the performance of the simple baselines we test. A more rigorous examination of hypotheses related to whether System 2 thinking helps will likely require the use of more comprehensive and robust benchmarks, such as <a href="https://www.swebench.com/">SWE-bench</a>.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-14" href="#footnote-anchor-14" class="footnote-number" contenteditable="false" target="_self">14</a><div class="footnote-content"><p>LDB uses already-existing solutions to improve them by debugging. The existing solutions can come from models like GPT-3.5 or GPT-4, or from agents like Reflexion. Since the authors of Reflexion provided all of the generated solutions in their Github repo, the authors of the LDB paper used code from the original Reflexion repository to run their analysis, rather than rerunning the Reflexion agent. The difference between the reported results and our reproduced results could be due to differences in the code generated by the Reflexion agent. Reusing Reflexion solutions is a reasonable choice for evaluating the usefulness of debugging (indeed, we see LDB increases the accuracy over using the models alone). The problem arises when their final accuracy is interpreted as a downstream evaluation, since it might give developers an inflated estimate of the accuracy of such techniques for coding.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-15" href="#footnote-anchor-15" class="footnote-number" contenteditable="false" target="_self">15</a><div class="footnote-content"><p>The authors acknowledge this and plan to update their results.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-16" href="#footnote-anchor-16" class="footnote-number" contenteditable="false" target="_self">16</a><div class="footnote-content"><p>The authors of LDB only tested the GPT-3.5 model as the debugger, which performed notably worse than the agent using GPT-4 as the debugger, with an accuracy of 88.9% for LDB (GPT-3.5 + Reflexion) vs. 92.9% for LDB (GPT-4 + Reflexion).</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-17" href="#footnote-anchor-17" class="footnote-number" contenteditable="false" target="_self">17</a><div class="footnote-content"><p>In correspondence with the authors of LATS, they clarified: "Originally, there was an execution error when evaluating some test cases for [one of the HumanEval test cases], so we opted to remove it from our setting."</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-18" href="#footnote-anchor-18" class="footnote-number" contenteditable="false" target="_self">18</a><div class="footnote-content"><p>The authors acknowledge this and plan to update the paper to address it.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-19" href="#footnote-anchor-19" class="footnote-number" contenteditable="false" target="_self">19</a><div class="footnote-content"><p>The authors acknowledge this and plan to update the paper to address it.</p><p></p></div></div>]]></content:encoded></item></channel></rss>