<?xml version="1.0" encoding="utf-8"?>
<feed xml:lang="en-us" xmlns="http://www.w3.org/2005/Atom"><title>Simon Willison's Weblog: predictions</title><link href="http://simonwillison.net/" rel="alternate"/><link href="http://simonwillison.net/tags/predictions.atom" rel="self"/><id>http://simonwillison.net/</id><updated>2026-01-27T16:58:08+00:00</updated><author><name>Simon Willison</name></author><entry><title>One Human + One Agent = One Browser From Scratch</title><link href="https://simonwillison.net/2026/Jan/27/one-human-one-agent-one-browser/#atom-tag" rel="alternate"/><published>2026-01-27T16:58:08+00:00</published><updated>2026-01-27T16:58:08+00:00</updated><id>https://simonwillison.net/2026/Jan/27/one-human-one-agent-one-browser/#atom-tag</id><summary type="html">
    
&lt;p&gt;&lt;strong&gt;&lt;a href="https://emsh.cat/one-human-one-agent-one-browser/"&gt;One Human + One Agent = One Browser From Scratch&lt;/a&gt;&lt;/strong&gt;&lt;/p&gt;
embedding-shapes was &lt;a href="https://emsh.cat/cursor-implied-success-without-evidence/"&gt;so infuriated&lt;/a&gt; by the hype around Cursor's &lt;a href="https://simonwillison.net/2026/Jan/23/fastrender/"&gt;FastRender browser project&lt;/a&gt; - thousands of parallel agents producing ~1.6 million lines of Rust - that they were inspired to take a go at building a web browser using coding agents themselves.&lt;/p&gt;
&lt;p&gt;The result is &lt;a href="https://github.com/embedding-shapes/one-agent-one-browser"&gt;one-agent-one-browser&lt;/a&gt; and it's &lt;em&gt;really&lt;/em&gt; impressive. Over three days they drove a single Codex CLI agent to build 20,000 lines of Rust that successfully renders HTML+CSS with no Rust crate dependencies at all - though it does (reasonably) use Windows, macOS and Linux system frameworks for image and text rendering.&lt;/p&gt;
&lt;p&gt;I installed the &lt;a href="https://github.com/embedding-shapes/one-agent-one-browser/releases/tag/0.1.0"&gt;1MB macOS binary release&lt;/a&gt; and ran it against my blog:&lt;/p&gt;
&lt;pre&gt;&lt;code&gt;chmod 755 ~/Downloads/one-agent-one-browser-macOS-ARM64 
~/Downloads/one-agent-one-browser-macOS-ARM64 https://simonwillison.net/
&lt;/code&gt;&lt;/pre&gt;
&lt;p&gt;Here's the result:&lt;/p&gt;
&lt;p&gt;&lt;img alt="My blog rendered in a window. Everything is in the right place, the CSS gradients look good, the feed subscribe SVG icon is rendered correctly but there's a missing PNG image." src="https://static.simonwillison.net/static/2026/one-agent-simonwillison.jpg" /&gt;&lt;/p&gt;
&lt;p&gt;It even rendered my SVG feed subscription icon! A PNG image is missing from the page, which looks like an intermittent bug (there's code to render PNGs).&lt;/p&gt;
&lt;p&gt;The code is pretty readable too - here's &lt;a href="https://github.com/embedding-shapes/one-agent-one-browser/blob/0.1.0/src/layout/flex.rs"&gt;the flexbox implementation&lt;/a&gt;.&lt;/p&gt;
&lt;p&gt;I had thought that "build a web browser" was the ideal prompt to really stretch the capabilities of coding agents - and that it would take sophisticated multi-agent harnesses (as seen in the Cursor project) and millions of lines of code to achieve.&lt;/p&gt;
&lt;p&gt;Turns out one agent driven by a talented engineer, three days and 20,000 lines of Rust is enough to get a very solid basic renderer working!&lt;/p&gt;
&lt;p&gt;I'm going to upgrade my &lt;a href="https://simonwillison.net/2026/Jan/8/llm-predictions-for-2026/#3-years-someone-will-build-a-new-browser-using-mainly-ai-assisted-coding-and-it-won-t-even-be-a-surprise"&gt;prediction for 2029&lt;/a&gt;: I think we're going to get a &lt;em&gt;production-grade&lt;/em&gt; web browser built by a small team using AI assistance by then.

    &lt;p&gt;&lt;small&gt;&lt;/small&gt;Via &lt;a href="https://news.ycombinator.com/item?id=46779522"&gt;Show Hacker News&lt;/a&gt;&lt;/small&gt;&lt;/p&gt;


    &lt;p&gt;Tags: &lt;a href="https://simonwillison.net/tags/browsers"&gt;browsers&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/predictions"&gt;predictions&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai"&gt;ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/rust"&gt;rust&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/generative-ai"&gt;generative-ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/llms"&gt;llms&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai-assisted-programming"&gt;ai-assisted-programming&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/coding-agents"&gt;coding-agents&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/codex-cli"&gt;codex-cli&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/browser-challenge"&gt;browser-challenge&lt;/a&gt;&lt;/p&gt;



</summary><category term="browsers"/><category term="predictions"/><category term="ai"/><category term="rust"/><category term="generative-ai"/><category term="llms"/><category term="ai-assisted-programming"/><category term="coding-agents"/><category term="codex-cli"/><category term="browser-challenge"/></entry><entry><title>LLM predictions for 2026, shared with Oxide and Friends</title><link href="https://simonwillison.net/2026/Jan/8/llm-predictions-for-2026/#atom-tag" rel="alternate"/><published>2026-01-08T19:42:13+00:00</published><updated>2026-01-08T19:42:13+00:00</updated><id>https://simonwillison.net/2026/Jan/8/llm-predictions-for-2026/#atom-tag</id><summary type="html">
    &lt;p&gt;I joined a recording of the Oxide and Friends podcast on Tuesday to talk about 1, 3 and 6 year predictions for the tech industry. This is my second appearance on their annual predictions episode, you can see &lt;a href="https://simonwillison.net/2025/Jan/10/ai-predictions/"&gt;my predictions from January 2025 here&lt;/a&gt;. Here's &lt;a href="https://oxide-and-friends.transistor.fm/episodes/predictions-2026"&gt;the page for this year's episode&lt;/a&gt;, with options to listen in all of your favorite podcast apps or &lt;a href="https://www.youtube.com/watch?v=lVDhQMiAbR8"&gt;directly on YouTube&lt;/a&gt;.&lt;/p&gt;
&lt;p&gt;Bryan Cantrill started the episode by declaring that he's never been so unsure about what's coming in the next year. I share that uncertainty - the significant advances in coding agents just in the last two months have left me certain that things will change significantly, but unclear as to what those changes will be.&lt;/p&gt;
&lt;p&gt;Here are the predictions I shared in the episode.&lt;/p&gt;
&lt;ul&gt;
  &lt;li&gt;&lt;a href="https://simonwillison.net/2026/Jan/8/llm-predictions-for-2026/#1-year-it-will-become-undeniable-that-llms-write-good-code"&gt;1 year: It will become undeniable that LLMs write good code&lt;/a&gt;&lt;/li&gt;
  &lt;li&gt;&lt;a href="https://simonwillison.net/2026/Jan/8/llm-predictions-for-2026/#1-year-we-re-finally-going-to-solve-sandboxing"&gt;1 year: We're finally going to solve sandboxing&lt;/a&gt;&lt;/li&gt;
  &lt;li&gt;&lt;a href="https://simonwillison.net/2026/Jan/8/llm-predictions-for-2026/#1-year-a-challenger-disaster-for-coding-agent-security"&gt;1 year: A "Challenger disaster" for coding agent security&lt;/a&gt;&lt;/li&gt;
  &lt;li&gt;&lt;a href="https://simonwillison.net/2026/Jan/8/llm-predictions-for-2026/#1-year-k-k-p-parrots-will-have-an-outstanding-breeding-season"&gt;1 year: Kākāpō parrots will have an outstanding breeding season&lt;/a&gt;&lt;/li&gt;
  &lt;li&gt;&lt;a href="https://simonwillison.net/2026/Jan/8/llm-predictions-for-2026/#3-years-the-coding-agents-jevons-paradox-for-software-engineering-will-resolve-one-way-or-the-other"&gt;3 years: the coding agents Jevons paradox for software engineering will resolve, one way or the other&lt;/a&gt;&lt;/li&gt;
  &lt;li&gt;&lt;a href="https://simonwillison.net/2026/Jan/8/llm-predictions-for-2026/#3-years-someone-will-build-a-new-browser-using-mainly-ai-assisted-coding-and-it-won-t-even-be-a-surprise"&gt;3 years: Someone will build a new browser using mainly AI-assisted coding and it won't even be a surprise&lt;/a&gt;&lt;/li&gt;
  &lt;li&gt;&lt;a href="https://simonwillison.net/2026/Jan/8/llm-predictions-for-2026/#6-years-typing-code-by-hand-will-go-the-way-of-punch-cards"&gt;6 years: Typing code by hand will go the way of punch cards&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;
&lt;h4 id="1-year-it-will-become-undeniable-that-llms-write-good-code"&gt;1 year: It will become undeniable that LLMs write good code &lt;a href="https://www.youtube.com/watch?v=lVDhQMiAbR8&amp;amp;t=1167s" class="predictions-video-link"&gt;▶ 19:27&lt;/a&gt;&lt;/h4&gt;
&lt;blockquote&gt;
&lt;p&gt;I think that there are still people out there who are convinced that LLMs cannot write good code. Those people are in for a very nasty shock in 2026. I do not think it will be possible to get to the end of even the next three months while still holding on to that idea that the code they write is all junk and it's it's likely any decent human programmer will write better code than they will.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;In 2023, saying that LLMs write garbage code was entirely correct. For most of 2024 that stayed true. In 2025 that changed, but you could be forgiven for continuing to hold out. In 2026 the quality of LLM-generated code will become impossible to deny.&lt;/p&gt;
&lt;p&gt;I base this on my own experience - I've spent more time exploring &lt;a href="https://simonwillison.net/tags/ai-assisted-programming/"&gt;AI-assisted programming&lt;/a&gt; than most.&lt;/p&gt;
&lt;p&gt;The key change in 2025 (see &lt;a href="https://simonwillison.net/2025/Dec/31/the-year-in-llms/#the-year-of-reasoning-"&gt;my overview for the year&lt;/a&gt;) was the introduction of "reasoning models" trained specifically against code using Reinforcement Learning. The major labs spent a full year competing with each other on who could get the best code capabilities from their models, and that problem turns out to be perfectly attuned to RL since code challenges come with built-in verifiable success conditions.&lt;/p&gt;
&lt;p&gt;Since Claude Opus 4.5 and GPT-5.2 came out in November and December respectively the amount of code I've written by hand has dropped to a single digit percentage of my overall output. The same is true for many other expert programmers I know.&lt;/p&gt;
&lt;p&gt;At this point if you continue to argue that LLMs write useless code you're damaging your own credibility.&lt;/p&gt;
&lt;h4 id="1-year-we-re-finally-going-to-solve-sandboxing"&gt;1 year: We're finally going to solve sandboxing &lt;a href="https://www.youtube.com/watch?v=lVDhQMiAbR8&amp;amp;t=1205s" class="predictions-video-link"&gt;▶ 20:05&lt;/a&gt;&lt;/h4&gt;
&lt;blockquote&gt;
&lt;p&gt;I think this year is the year we're going to solve sandboxing. I want to run code other people have written on my computing devices without it destroying my computing devices if it's malicious or has bugs. [...] It's crazy that it's 2026 and I still &lt;code&gt;pip install&lt;/code&gt; random code and then execute it in a way that it can steal all of my data and delete all my files. [...] I don't want to run a piece of code on any of my devices that somebody else wrote outside of sandbox ever again.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;This isn't just about LLMs, but it becomes even more important now there are so many more people writing code often without knowing what they're doing. Sandboxing is also a key part of the battle against prompt injection.&lt;/p&gt;
&lt;p&gt;We have a &lt;em&gt;lot&lt;/em&gt; of promising technologies in play already for this - containers and WebAssembly being the two I'm most optimistic about. There's real commercial value involved in solving this problem. The pieces are there, what's needed is UX work to reduce the friction in using them productively and securely.&lt;/p&gt;
&lt;h4 id="1-year-a-challenger-disaster-for-coding-agent-security"&gt;1 year: A "Challenger disaster" for coding agent security &lt;a href="https://www.youtube.com/watch?v=lVDhQMiAbR8&amp;amp;t=1281s" class="predictions-video-link"&gt;▶ 21:21&lt;/a&gt;&lt;/h4&gt;
&lt;blockquote&gt;
&lt;p&gt;I think we're due a Challenger disaster with respect to coding agent security[...] I think so many people, myself included, are running these coding agents practically as root, right? We're letting them do all of this stuff. And every time I do it, my computer doesn't get wiped. I'm like, "oh, it's fine". [...] The worst version of this is the worm - a prompt injection worm which infects people's computers and adds itself to the Python or NPM packages that person has access to.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;I used this as an opportunity to promote my favourite recent essay about AI security, &lt;a href="https://embracethered.com/blog/posts/2025/the-normalization-of-deviance-in-ai/"&gt;the Normalization of Deviance in AI&lt;/a&gt; by Johann Rehberger.&lt;/p&gt;
&lt;p&gt;The Normalization of Deviance describes the phenomenon where people and organizations get used to operating in an unsafe manner because nothing bad has happened to them yet, which can result in enormous problems (like the 1986 Challenger disaster) when their luck runs out.&lt;/p&gt;
&lt;p&gt;Every six months I predict that a headline-grabbing prompt injection attack is coming soon, and every six months it doesn't happen. This is my most recent version of that prediction!&lt;/p&gt;
&lt;h4 id="1-year-k-k-p-parrots-will-have-an-outstanding-breeding-season"&gt;1 year: Kākāpō parrots will have an outstanding breeding season &lt;a href="https://www.youtube.com/watch?v=lVDhQMiAbR8&amp;amp;t=3006s" class="predictions-video-link"&gt;▶ 50:06&lt;/a&gt;&lt;/h4&gt;

&lt;p&gt;(I dropped this one to lighten the mood after a discussion of the deep sense of existential dread that many programmers are feeling right now!)&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;I think that Kākāpō parrots in New Zealand are going to have an outstanding breeding season. The reason I think this is that the Rimu trees are in fruit right now. There's only 250 of them,  and they only breed if the Rimu trees have a good fruiting. The Rimu trees have been terrible since 2019, but this year the Rimu trees were all blooming. There are researchers saying that all 87 females of breeding age might lay an egg. And for a species with only 250 remaining parrots that's great news.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;(I just &lt;a href="https://en.wikipedia.org/wiki/K%C4%81k%C4%81p%C5%8D#Population_timeline"&gt;checked Wikipedia&lt;/a&gt; and I was right with the parrot numbers but wrong about the last good breeding season, apparently 2022 was a good year too.)&lt;/p&gt;
&lt;p&gt;In a year with precious little in the form of good news I am utterly delighted to share this story. Here's more:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;
&lt;a href="https://blog.doc.govt.nz/2025/06/27/kakapo-breeding-season-2026/"&gt;Kākāpō breeding season 2026&lt;/a&gt; introduction from the Department of Conservation from June 2025 .&lt;/li&gt;
&lt;li&gt;
&lt;a href="https://www.auckland.ac.nz/en/news/2025/12/03/bumper-breeding-season-for-kakapo-on-the-cards.html"&gt;Bumper breeding season for kākāpō on the cards&lt;/a&gt; - 3rd December 2025, University of Auckland.&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;I don't often use AI-generated images on this blog, but the Kākāpō image the Oxide team created for this episode is just &lt;em&gt;perfect&lt;/em&gt;:&lt;/p&gt;
&lt;p&gt;&lt;img src="https://static.simonwillison.net/static/2026/oxide-kakapo.jpg" alt="A beautiful green Kākāpō surrounded by candles gazes into a crystal ball" style="max-width: 100%;" /&gt;&lt;/p&gt;

&lt;h4 id="3-years-the-coding-agents-jevons-paradox-for-software-engineering-will-resolve-one-way-or-the-other"&gt;3 years: the coding agents Jevons paradox for software engineering will resolve, one way or the other &lt;a href="https://www.youtube.com/watch?v=lVDhQMiAbR8&amp;amp;t=3277s" class="predictions-video-link"&gt;▶ 54:37&lt;/a&gt;&lt;/h4&gt;

&lt;blockquote&gt;
&lt;p&gt;We will find out if the &lt;a href="https://en.wikipedia.org/wiki/Jevons_paradox"&gt;Jevons paradox&lt;/a&gt; saves our careers or not. This is a big question that anyone who's a software engineer has right now: we are driving the cost of actually producing working code down to a fraction of what it used to cost. Does that mean that our careers are completely devalued and we all have to learn to live on a tenth of our incomes, or does it mean that the demand for software, for custom software goes up by a factor of 10 and now our skills are even &lt;em&gt;more&lt;/em&gt; valuable because you can hire me and I can build you 10 times the software I used to be able to? I think by three years we will know for sure which way that one went.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;The quote says it all. There are two ways this coding agents thing could go: it could turn out software engineering skills are devalued, or it could turn out we're more valuable and effective than ever before.&lt;/p&gt;
&lt;p&gt;I'm crossing my fingers for the latter! So far it feels to me like it's working out that way.&lt;/p&gt;

&lt;h4 id="3-years-someone-will-build-a-new-browser-using-mainly-ai-assisted-coding-and-it-won-t-even-be-a-surprise"&gt;3 years: Someone will build a new browser using mainly AI-assisted coding and it won't even be a surprise &lt;a href="https://www.youtube.com/watch?v=lVDhQMiAbR8&amp;amp;t=3913s" class="predictions-video-link"&gt;▶ 65:13&lt;/a&gt;&lt;/h4&gt;

&lt;blockquote&gt;
&lt;p&gt;I think somebody will have built a full web browser mostly using AI assistance, and it won't even be surprising. Rolling a new web browser is one of the most complicated software projects I can imagine[...] the cheat code is the conformance suites. If there are existing tests that it'll get so much easier.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;A common complaint today from AI coding skeptics is that LLMs are fine for toy projects but can't be used for anything large and serious.&lt;/p&gt;
&lt;p&gt;I think within 3 years that will be comprehensively proven incorrect, to the point that it won't even be controversial anymore.&lt;/p&gt;
&lt;p&gt;I picked a web browser here because so much of the work building a browser involves writing code that has to conform to an enormous and daunting selection of both formal tests and informal websites-in-the-wild.&lt;/p&gt;
&lt;p&gt;Coding agents are &lt;em&gt;really good&lt;/em&gt; at tasks where you can define a concrete goal and then set them to work iterating in that direction.&lt;/p&gt;
&lt;p&gt;A web browser is the most ambitious project I can think of that leans into those capabilities.&lt;/p&gt;

&lt;h4 id="6-years-typing-code-by-hand-will-go-the-way-of-punch-cards"&gt;6 years: Typing code by hand will go the way of punch cards &lt;a href="https://www.youtube.com/watch?v=lVDhQMiAbR8&amp;amp;t=4839s" class="predictions-video-link"&gt;▶ 80:39&lt;/a&gt;&lt;/h4&gt;

&lt;blockquote&gt;
&lt;p&gt;I think the job of being paid money to type code into a computer will go the same way as punching punch cards [...] in six years time, I do not think anyone will be paid to just to do the thing where you type the code. I think software engineering will still be an enormous career. I just think the software engineers won't be spending multiple hours of their day in a text editor typing out syntax.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;The more time I spend on AI-assisted programming the less afraid I am for my job, because it turns out building software - especially at the rate it's now possible to build - still requires enormous skill, experience and depth of understanding.&lt;/p&gt;
&lt;p&gt;The skills are changing though! Being able to read a detailed specification and transform it into lines of code is the thing that's being automated away. What's left is everything else, and the more time I spend working with coding agents the larger that "everything else" becomes.&lt;/p&gt;
    
        &lt;p&gt;Tags: &lt;a href="https://simonwillison.net/tags/predictions"&gt;predictions&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/sandboxing"&gt;sandboxing&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai"&gt;ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/kakapo"&gt;kakapo&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/generative-ai"&gt;generative-ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/llms"&gt;llms&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai-assisted-programming"&gt;ai-assisted-programming&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/oxide"&gt;oxide&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/bryan-cantrill"&gt;bryan-cantrill&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/coding-agents"&gt;coding-agents&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/conformance-suites"&gt;conformance-suites&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/browser-challenge"&gt;browser-challenge&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/deep-blue"&gt;deep-blue&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/november-2025-inflection"&gt;november-2025-inflection&lt;/a&gt;&lt;/p&gt;
    

</summary><category term="predictions"/><category term="sandboxing"/><category term="ai"/><category term="kakapo"/><category term="generative-ai"/><category term="llms"/><category term="ai-assisted-programming"/><category term="oxide"/><category term="bryan-cantrill"/><category term="coding-agents"/><category term="conformance-suites"/><category term="browser-challenge"/><category term="deep-blue"/><category term="november-2025-inflection"/></entry><entry><title>Prediction: AI will make formal verification go mainstream</title><link href="https://simonwillison.net/2025/Dec/9/formal-verification/#atom-tag" rel="alternate"/><published>2025-12-09T03:11:19+00:00</published><updated>2025-12-09T03:11:19+00:00</updated><id>https://simonwillison.net/2025/Dec/9/formal-verification/#atom-tag</id><summary type="html">
    
&lt;p&gt;&lt;strong&gt;&lt;a href="https://martin.kleppmann.com/2025/12/08/ai-formal-verification.html"&gt;Prediction: AI will make formal verification go mainstream&lt;/a&gt;&lt;/strong&gt;&lt;/p&gt;
Martin Kleppmann makes the case for formal verification languages (things like &lt;a href="https://dafny.org/"&gt;Dafny&lt;/a&gt;, &lt;a href="https://github.com/marcoeilers/nagini"&gt;Nagini&lt;/a&gt;, and &lt;a href="https://github.com/verus-lang/verus"&gt;Verus&lt;/a&gt;) to finally start achieving more mainstream usage. Code generated by LLMs can benefit enormously from more robust verification, and LLMs themselves make these notoriously difficult systems easier to work with.&lt;/p&gt;
&lt;p&gt;The paper &lt;a href="https://arxiv.org/abs/2503.14183"&gt;Can LLMs Enable Verification in Mainstream Programming?&lt;/a&gt; by JetBrains Research in March 2025 found that Claude 3.5 Sonnet saw promising results for the three languages I listed above.

    &lt;p&gt;&lt;small&gt;&lt;/small&gt;Via &lt;a href="https://lobste.rs/s/zsgdbg/prediction_ai_will_make_formal"&gt;lobste.rs&lt;/a&gt;&lt;/small&gt;&lt;/p&gt;


    &lt;p&gt;Tags: &lt;a href="https://simonwillison.net/tags/predictions"&gt;predictions&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/programming-languages"&gt;programming-languages&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai"&gt;ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/generative-ai"&gt;generative-ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/llms"&gt;llms&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai-assisted-programming"&gt;ai-assisted-programming&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/martin-kleppmann"&gt;martin-kleppmann&lt;/a&gt;&lt;/p&gt;



</summary><category term="predictions"/><category term="programming-languages"/><category term="ai"/><category term="generative-ai"/><category term="llms"/><category term="ai-assisted-programming"/><category term="martin-kleppmann"/></entry><entry><title>My AI/LLM predictions for the next 1, 3 and 6 years, for Oxide and Friends</title><link href="https://simonwillison.net/2025/Jan/10/ai-predictions/#atom-tag" rel="alternate"/><published>2025-01-10T01:43:16+00:00</published><updated>2025-01-10T01:43:16+00:00</updated><id>https://simonwillison.net/2025/Jan/10/ai-predictions/#atom-tag</id><summary type="html">
    &lt;p&gt;The &lt;a href="https://oxide-and-friends.transistor.fm/"&gt;Oxide and Friends&lt;/a&gt; podcast has an annual tradition of asking guests to share their predictions for the next 1, 3 and 6 years. Here's &lt;a href="https://github.com/oxidecomputer/oxide-and-friends/blob/master/2022_01_03.md"&gt;2022&lt;/a&gt;, &lt;a href="https://github.com/oxidecomputer/oxide-and-friends/blob/master/2023_01_09.md"&gt;2023&lt;/a&gt; and &lt;a href="https://github.com/oxidecomputer/oxide-and-friends/blob/master/2024_01_08.md"&gt;2024&lt;/a&gt;. This year they invited me to participate. I've never been brave enough to share &lt;em&gt;any&lt;/em&gt; public predictions before, so this was a great opportunity to get outside my comfort zone!&lt;/p&gt;
&lt;p&gt;We recorded the episode live using Discord on Monday. It's now available &lt;a href="https://www.youtube.com/watch?v=-pk6VokHpGY"&gt;on YouTube&lt;/a&gt; and &lt;a href="https://oxide-and-friends.transistor.fm/"&gt;in podcast form&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;&lt;lite-youtube videoid="-pk6VokHpGY"
  title="Oxide and Friends 1/6/2025 -- Predictions 2025"
  playlabel="Play: Oxide and Friends 1/6/2025 -- Predictions 2025"
&gt; &lt;/lite-youtube&gt;&lt;/p&gt;

&lt;p&gt;Here are my predictions, written up here in a little more detail than the stream of consciousness I shared on the podcast.&lt;/p&gt;
&lt;p&gt;I should emphasize that I find the very idea of trying to predict AI/LLMs over a multi-year period to be completely absurd! I can't predict what's going to happen a week from now, six years is a different universe.&lt;/p&gt;
&lt;p&gt;With that disclaimer out of the way, here's an expanded version of what I said.&lt;/p&gt;
&lt;ul&gt;
  &lt;li&gt;&lt;a href="https://simonwillison.net/2025/Jan/10/ai-predictions/#one-year-agents-fail-to-happen-again"&gt;One year: Agents fail to happen, again&lt;/a&gt;&lt;/li&gt;
  &lt;li&gt;&lt;a href="https://simonwillison.net/2025/Jan/10/ai-predictions/#one-year-code-research-assistants"&gt;One year: ... except for code and research assistants&lt;/a&gt;&lt;/li&gt;
  &lt;li&gt;&lt;a href="https://simonwillison.net/2025/Jan/10/ai-predictions/#three-years-someone-wins-a-pulitzer-for-ai-assisted-investigative-reporting"&gt;Three years: Someone wins a Pulitzer for AI-assisted investigative reporting&lt;/a&gt;&lt;/li&gt;
  &lt;li&gt;&lt;a href="https://simonwillison.net/2025/Jan/10/ai-predictions/#three-years-part-two-privacy-laws-with-teeth"&gt;Three years part two: privacy laws with teeth&lt;/a&gt;&lt;/li&gt;
  &lt;li&gt;&lt;a href="https://simonwillison.net/2025/Jan/10/ai-predictions/#six-years-utopian-amazing-art"&gt;Six years utopian: amazing art&lt;/a&gt;&lt;/li&gt;
  &lt;li&gt;&lt;a href="https://simonwillison.net/2025/Jan/10/ai-predictions/#six-years-dystopian-agi-asi-causes-mass-civil-unrest"&gt;Six years dystopian: AGI/ASI causes mass civil unrest&lt;/a&gt;&lt;/li&gt;
  &lt;li&gt;&lt;a href="https://simonwillison.net/2025/Jan/10/ai-predictions/#my-total-lack-of-conviction"&gt;My total lack of conviction&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;
&lt;h4 id="one-year-agents-fail-to-happen-again"&gt;One year: Agents fail to happen, again&lt;/h4&gt;
&lt;p&gt;I wrote about how &lt;a href="https://simonwillison.net/2024/Dec/31/llms-in-2024/#-agents-still-haven-t-really-happened-yet"&gt;“Agents” still haven’t really happened yet&lt;/a&gt; in my review of Large Language Model developments  in 2024.&lt;/p&gt;
&lt;p&gt;I think we are going to see a &lt;em&gt;lot&lt;/em&gt; more froth about agents in 2025, but I expect the results will be a great disappointment to most of the people who are excited about this term. I expect a lot of money will be lost chasing after several different poorly defined dreams that share that name.&lt;/p&gt;
&lt;p&gt;What are agents anyway? Ask a dozen people and you'll get a dozen slightly different answers - I collected and &lt;a href="https://gist.github.com/simonw/beaa5f90133b30724c5cc1c4008d0654"&gt;then AI-summarized a bunch of those here&lt;/a&gt;.&lt;/p&gt;
&lt;p&gt;For the sake of argument, let's pick a definition that I can predict won't come to fruition: the idea of an AI assistant that can go out into the world and semi-autonomously act on your behalf. I think of this as the &lt;strong&gt;travel agent&lt;/strong&gt; definition of agents, because for some reason everyone always jumps straight to flight and hotel booking and itinerary planning when they describe this particular dream.&lt;/p&gt;
&lt;p&gt;Having the current generation of LLMs make material decisions on your behalf - like what to spend money on - is a &lt;em&gt;really bad idea&lt;/em&gt;. They're too unreliable, but more importantly they are too &lt;strong&gt;gullible&lt;/strong&gt;.&lt;/p&gt;
&lt;p&gt;If you're going to arm your AI assistant with a credit card and set it loose on the world, you need to be confident that it's not going to hit "buy" on the first website that claims to offer the best bargains!&lt;/p&gt;
&lt;p&gt;I'm confident that reliability is the reason we haven't seen LLM-powered agents that have taken off yet, despite the idea attracting a huge amount of buzz since right after ChatGPT first came out.&lt;/p&gt;
&lt;p&gt;I would be very surprised if any of the models released over the next twelve months had enough of a reliability improvement to make this work. Solving gullibility is an astonishingly difficult problem.&lt;/p&gt;
&lt;p&gt;(I had &lt;a href="https://www.youtube.com/watch?v=-pk6VokHpGY&amp;amp;t=1206s"&gt;a particularly spicy rant&lt;/a&gt; about how stupid the idea of sending a "digital twin" to a meeting on your behalf is.)&lt;/p&gt;
&lt;h4 id="one-year-code-research-assistants"&gt;One year: ... except for code and research assistants&lt;/h4&gt;
&lt;p&gt;There are two categories of "agent" that I do believe in, because they're proven to work already.&lt;/p&gt;
&lt;p&gt;The first is &lt;strong&gt;coding assistants&lt;/strong&gt; - where an LLM writes, executes and then refines computer code in a loop.&lt;/p&gt;
&lt;p&gt;I first saw this pattern demonstrated by OpenAI with their &lt;a href="https://simonwillison.net/tags/code-interpreter/"&gt;Code Interpreter&lt;/a&gt; feature for ChatGPT, released back in March/April of 2023.&lt;/p&gt;
&lt;p&gt;You can ask ChatGPT to solve a problem that can use Python code and it will write that Python, execute it in a secure sandbox (I think it's Kubernetes) and then use the output - or any error messages - to determine if the goal has been achieved.&lt;/p&gt;
&lt;p&gt;It's a beautiful pattern that worked great with early 2023 models (I believe it first shipped using original GPT-4), and continues to work today.&lt;/p&gt;
&lt;p&gt;Claude added their own version in October (&lt;a href="https://simonwillison.net/2024/Oct/24/claude-analysis-tool/"&gt;Claude analysis&lt;/a&gt;, using JavaScript that runs in the browser), Mistral have it, Gemini has a version and there are dozens of other implementations of the same pattern.&lt;/p&gt;
&lt;p&gt;The second category of agents that I believe in is &lt;strong&gt;research assistants&lt;/strong&gt; - where an LLM can run multiple searches, gather information and aggregate that into an answer to a question or write a report.&lt;/p&gt;
&lt;p&gt;&lt;a href="https://www.perplexity.ai/"&gt;Perplexity&lt;/a&gt; and &lt;a href="https://openai.com/index/introducing-chatgpt-search/"&gt;ChatGPT Search&lt;/a&gt; have both been operating in this space for a while, but by far the most impressive implementation I've seen is Google Gemini's &lt;a href="https://blog.google/products/gemini/google-gemini-deep-research/"&gt;Deep Research&lt;/a&gt; tool, which I've had access to for a few weeks.&lt;/p&gt;
&lt;p&gt;With Deep Research I can pose a question like this one:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;Pillar Point Harbor is one of the largest communal brown pelican roosts on the west coast of North America.&lt;/p&gt;
&lt;p&gt;find others&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;And Gemini will draft a plan, consult dozens of different websites via Google Search and then assemble a report (with all-important citations) describing what it found.&lt;/p&gt;
&lt;p&gt;Here's the plan it came up with:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;Pillar Point Harbor is one of the largest communal brown pelican roosts on the west coast of North America. Find other large communal brown pelican roosts on the west coast of North America.&lt;br /&gt;
(1) Find a list of brown pelican roosts on the west coast of North America.&lt;br /&gt;
(2) Find research papers or articles about brown pelican roosts and their size.&lt;br /&gt;
(3) Find information from birdwatching organizations or government agencies about brown pelican roosts.&lt;br /&gt;
(4) Compare the size of the roosts found in (3) to the size of the Pillar Point Harbor roost.&lt;br /&gt;
(5) Find any news articles or recent reports about brown pelican roosts and their populations.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;It dug up a whole bunch of details, but the one I cared most about was &lt;a href="https://birdallianceoregon.org/wp-content/uploads/2021/04/Brown-Pelican-survey_4-year_summary-infographic_2016-19_final.pdf"&gt;these PDF results for the 2016-2019 Pacific Brown Pelican Survey&lt;/a&gt; conducted by the West Coast Audubon network and partners - a PDF that included this delightful list:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;Top 10 Megaroosts (sites that traditionally host &amp;gt;500 pelicans) with average fall count numbers:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;Alameda Breakwater, CA (3,183)&lt;/li&gt;
&lt;li&gt;Pillar Point Harbor, CA (1,481)&lt;/li&gt;
&lt;li&gt;East Sand Island, OR (1,121)&lt;/li&gt;
&lt;li&gt;Ano Nuevo State Park, CA (1,068)&lt;/li&gt;
&lt;li&gt;Salinas River mouth, CA (762)&lt;/li&gt;
&lt;li&gt;Bolinas Lagoon, CA (755)&lt;/li&gt;
&lt;li&gt;Morro Rock, CA (725)&lt;/li&gt;
&lt;li&gt;Moss landing, CA (570)&lt;/li&gt;
&lt;li&gt;Crescent City Harbor, CA (514)&lt;/li&gt;
&lt;li&gt;Bird Rock Tomales, CA (514)&lt;/li&gt;
&lt;/ul&gt;
&lt;/blockquote&gt;
&lt;p&gt;My local harbor is the second biggest megaroost!&lt;/p&gt;
&lt;p&gt;It makes intuitive sense to me that this kind of research assistant can be built on our current generation of LLMs. They're competent at driving tools, they're capable of coming up with a relatively obvious research plan (look for newspaper articles and research papers) and they can synthesize sensible answers given the right collection of context gathered through search.&lt;/p&gt;
&lt;p&gt;Google are particularly well suited to solving this problem: they have the world's largest search index and their Gemini model has a 2 million token context. I expect Deep Research to get a whole lot better, and I expect it to attract plenty of competition.&lt;/p&gt;
&lt;h4 id="three-years-someone-wins-a-pulitzer-for-ai-assisted-investigative-reporting"&gt;Three years: Someone wins a Pulitzer for AI-assisted investigative reporting&lt;/h4&gt;
&lt;p&gt;I went for a bit of a self-serving prediction here: I think within three years someone is going to win a Pulitzer prize for a piece of investigative reporting that was aided by generative AI tools.&lt;/p&gt;

&lt;p&gt;&lt;em&gt;&lt;strong&gt;Update&lt;/strong&gt;: after publishing this piece I learned about this May 2024 story from Nieman Lab: &lt;a href="https://www.niemanlab.org/2024/05/for-the-first-time-two-pulitzer-winners-disclosed-using-ai-in-their-reporting/"&gt;For the first time, two Pulitzer winners disclosed using AI in their reporting&lt;/a&gt;. I think these were both examples of traditional machine learning as opposed to LLM-based generative AI, but this is yet another example of my predictions being less ambitious than I had thought!&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;I do &lt;em&gt;not&lt;/em&gt; mean that an LLM will write the article! I continue to think that having LLMs write on your behalf is one of the least interesting applications of these tools.&lt;/p&gt;
&lt;p&gt;I called this prediction self-serving because I want to help make this happen! My &lt;a href="https://datasette.io"&gt;Datasette&lt;/a&gt; suite of open source tools for data journalism has been growing AI features, like &lt;a href="https://simonwillison.net/2023/Dec/1/datasette-enrichments/"&gt;LLM-powered data enrichments&lt;/a&gt; and &lt;a href="https://www.datasette.cloud/blog/2024/datasette-extract/"&gt;extracting structured data&lt;/a&gt; into tables from unstructured text.&lt;/p&gt;
&lt;p&gt;My dream is for those tools - or tools like them - to be used for an award winning piece of investigative reporting.&lt;/p&gt;
&lt;p&gt;I picked three years for this because I think that's how long it will take for knowledge of how to responsibly and effectively use these tools to become widespread enough for that to happen.&lt;/p&gt;
&lt;p&gt;LLMs are not an obvious fit for journalism: journalists look for the truth, and LLMs are notoriously prone to hallucination and making things up. But journalists are also &lt;em&gt;really good&lt;/em&gt; at extracting useful information from potentially untrusted sources - that's a lot of what the craft of journalism is about.&lt;/p&gt;
&lt;p&gt;The two areas I think LLMs are particularly relevant to journalism are:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;Structured data extraction. If you have 10,000 PDFs from a successful Freedom of Information Act request, someone or something needs to kick off the process of reading through them to find the stories. LLMs are a fantastic way to take a vast amount of information and start making some element of sense from it. They can act as lead generators, helping identify the places to start looking more closely.&lt;/li&gt;
&lt;li&gt;Coding assistance. Writing code to help analyze data is a huge part of modern data journalism - from SQL queries through data cleanup scripts, custom web scrapers or visualizations to help find signal among the noise. Most newspapers don't have a team of programmers on staff: I think within three years we'll have robust enough tools built around this pattern that non-programmer journalists will be able to use them as part of their reporting process.&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;I hope to build some of these tools myself!&lt;/p&gt;
&lt;p&gt;So my concrete prediction for three years is that someone wins a Pulitzer with a small amount of assistance from LLMs.&lt;/p&gt;
&lt;p&gt;My more general prediction: within three years it won't be surprising at all to see most information professionals use LLMs as part of their daily workflow, in increasingly sophisticated ways. We'll know exactly what patterns work and how best to explain them to people. These skills will become widespread.&lt;/p&gt;
&lt;h4 id="three-years-part-two-privacy-laws-with-teeth"&gt;Three years part two: privacy laws with teeth&lt;/h4&gt;
&lt;p&gt;My other three year prediction concerned privacy legislation.&lt;/p&gt;
&lt;p&gt;The levels of (often justified) paranoia around both targeted advertising and what happens to the data people paste into these models is a constantly growing problem.&lt;/p&gt;
&lt;p&gt;I wrote recently about the &lt;a href="https://simonwillison.net/2025/Jan/2/they-spy-on-you-but-not-like-that/"&gt;inexterminable conspiracy theory that Apple target ads through spying through your phone's microphone&lt;/a&gt;. I've written in the past about &lt;a href="https://simonwillison.net/2023/Dec/14/ai-trust-crisis/"&gt;the AI trust crisis&lt;/a&gt;, where people refuse to believe that models are not being trained on their inputs no matter how emphatically the companies behind them deny it.&lt;/p&gt;
&lt;p&gt;I think the AI industry itself would benefit enormously from legislation that helps clarify what's going on with training on user-submitted data, and the wider tech industry could really do with harder rules around things like data retention and targeted advertising.&lt;/p&gt;
&lt;p&gt;I don't expect the next four years of US federal government to be effective at passing legislation, but I expect we'll see privacy legislation with sharper teeth emerging at the state level or internationally. Let's just hope we don't end up with a new generation of cookie-consent banners as a result!&lt;/p&gt;
&lt;h4 id="six-years-utopian-amazing-art"&gt;Six years utopian: amazing art&lt;/h4&gt;
&lt;p&gt;For six years I decided to go with two rival predictions, one optimistic and one pessimistic.&lt;/p&gt;
&lt;p&gt;I think six years is long enough that we'll figure out how to harness this stuff to make some &lt;strong&gt;really great art&lt;/strong&gt;.&lt;/p&gt;
&lt;p&gt;I don't think generative AI for art - images, video and music - deserves nearly the same level of respect as a useful tool as text-based LLMs. Generative art tools are a lot of fun to try out but the lack of fine-grained control over the output greatly limits its utility outside of personal amusement or generating &lt;a href="https://simonwillison.net/tags/slop/"&gt;slop&lt;/a&gt;.&lt;/p&gt;
&lt;p&gt;More importantly, they lack social acceptability. The vibes aren't good. Many talented artists have loudly rejected the idea of these tools, to the point that the very term "AI" is developing a distasteful connotation in society at large.&lt;/p&gt;
&lt;p&gt;Image and video models are also ground zero for the AI training data ethics debate, and for good reason: no artist wants to see a model trained on their work without their permission that then directly competes with them!&lt;/p&gt;
&lt;p&gt;I think six years is long enough for this whole thing to shake out - for society to figure out acceptable ways of using these tools to truly elevate human expression. What excites me is the idea of truly talented, visionary creative artists using whatever these tools have evolved into in six years to make meaningful art that could never have been achieved without them.&lt;/p&gt;
&lt;p&gt;On the podcast I talked about &lt;a href="https://en.wikipedia.org/wiki/Everything_Everywhere_All_at_Once"&gt;Everything Everywhere All at Once&lt;/a&gt;, a film that deserved every one of its seven Oscars. The core visual effects team on that film was just five people. Imagine what a team like that could do with the generative AI tools we'll have in six years time!&lt;/p&gt;
&lt;p id="since-recording"&gt;Since recording the podcast I learned from &lt;a href="https://www.swyx.io/"&gt;Swyx&lt;/a&gt; that Everything Everywhere All at Once &lt;a href="https://www.aboutamazon.com/news/aws/how-ai-tools-are-creating-new-possibilities-for-movies-and-visual-design-according-to-this-aws-powered-startup"&gt;used Runway ML as part of their toolset already&lt;/a&gt;:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;Evan Halleck was on this team, and he used Runway's AI tools to save time and automate tedious aspects of editing. Specifically in the film’s rock scene, he used Runway’s rotoscoping tool to get a quick, clean cut of the rocks as sand and dust were moving around the shot. This translated days of work to a matter of minutes.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;I said I thought a film that had used generative AI tools would win an Oscar within six years. Looks like I was eight years out on that one!&lt;/p&gt;
&lt;h4 id="six-years-dystopian-agi-asi-causes-mass-civil-unrest"&gt;Six years dystopian: AGI/ASI causes mass civil unrest&lt;/h4&gt;
&lt;p&gt;My pessimistic alternative take for 2031 concerns "AGI" - a term which, like "agents", is constantly being redefined. The Information &lt;a href="https://www.theinformation.com/articles/microsoft-and-openai-wrangle-over-terms-of-their-blockbuster-partnership"&gt;recently reported&lt;/a&gt; (see also &lt;a href="https://www.theverge.com/2025/1/6/24337106/sam-altman-says-openai-knows-how-to-build-agi-blog-post"&gt;The Verge&lt;/a&gt;) that Microsoft and OpenAI are now defining AGI as a system capable of generating $100bn in profit!&lt;/p&gt;
&lt;p&gt;If we assume AGI is the point at which AI systems are capable of performing almost any job currently reserved for a human being it's hard &lt;em&gt;not&lt;/em&gt; to see potentially negative consequences.&lt;/p&gt;
&lt;p&gt;Sam Altman may have &lt;a href="https://www.bloomberg.com/news/articles/2024-07-22/ubi-study-backed-by-openai-s-sam-altman-bolsters-support-for-basic-income"&gt;experimented with Universal Basic Income&lt;/a&gt;, but the USA is a country that can't even figure out universal healthcare! I have huge trouble imagining a future economy that works for the majority of people when the majority of jobs are being done by machines.&lt;/p&gt;
&lt;p&gt;So my dystopian prediction for 2031 is that if that form of AGI has come to pass it will be accompanied by extraordinarily bad economic outcomes and mass civil unrest.&lt;/p&gt;
&lt;p&gt;My version of an AI utopia is tools that augment existing humans. That's what we've had with LLMs so far, and my ideal is that those tools continue to improve and subsequently humans become able to take on &lt;a href="https://simonwillison.net/2023/Mar/27/ai-enhanced-development/"&gt;more ambitious work&lt;/a&gt;.&lt;/p&gt;
&lt;p&gt;If there's a version of AGI that results in that kind of utopia, I'm all for it.&lt;/p&gt;
&lt;h4 id="my-total-lack-of-conviction"&gt;My total lack of conviction&lt;/h4&gt;
&lt;p&gt;There's a reason I haven't made predictions like this before: my confidence in my ability to predict the future is almost non-existent. At least one of my predictions here &lt;a href="https://simonwillison.net/2025/Jan/10/ai-predictions/#since-recording"&gt;already proved to be eight years late&lt;/a&gt;!&lt;/p&gt;
&lt;p&gt;These predictions are in the public record now (I even &lt;a href="https://github.com/oxidecomputer/oxide-and-friends/pull/158"&gt;submitted a pull request&lt;/a&gt;).&lt;/p&gt;
&lt;p&gt;It's going to be interesting looking back at these in one, three and six years to see how I did.&lt;/p&gt;
    
        &lt;p&gt;Tags: &lt;a href="https://simonwillison.net/tags/data-journalism"&gt;data-journalism&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/predictions"&gt;predictions&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai"&gt;ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/openai"&gt;openai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/generative-ai"&gt;generative-ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/llms"&gt;llms&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai-assisted-programming"&gt;ai-assisted-programming&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/gemini"&gt;gemini&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/code-interpreter"&gt;code-interpreter&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/oxide"&gt;oxide&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai-agents"&gt;ai-agents&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/deep-research"&gt;deep-research&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai-assisted-search"&gt;ai-assisted-search&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/coding-agents"&gt;coding-agents&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/agent-definitions"&gt;agent-definitions&lt;/a&gt;&lt;/p&gt;
    

</summary><category term="data-journalism"/><category term="predictions"/><category term="ai"/><category term="openai"/><category term="generative-ai"/><category term="llms"/><category term="ai-assisted-programming"/><category term="gemini"/><category term="code-interpreter"/><category term="oxide"/><category term="ai-agents"/><category term="deep-research"/><category term="ai-assisted-search"/><category term="coding-agents"/><category term="agent-definitions"/></entry><entry><title>Quoting Sam Bowman</title><link href="https://simonwillison.net/2023/Apr/5/sam-bowman/#atom-tag" rel="alternate"/><published>2023-04-05T03:32:11+00:00</published><updated>2023-04-05T03:32:11+00:00</updated><id>https://simonwillison.net/2023/Apr/5/sam-bowman/#atom-tag</id><summary type="html">
    &lt;blockquote cite="https://cims.nyu.edu/~sbowman/eightthings.pdf"&gt;&lt;p&gt;Scaling laws allow us to precisely predict some coarse-but-useful measures of how capable future models will be as we scale them up along three dimensions: the amount of data they are fed, their size (measured in parameters), and the amount of computation used to train them (measured in FLOPs). [...] Our ability to make this kind of precise prediction is unusual in the history of software and unusual even in the history of modern AI research. It is also a powerful tool for driving investment since it allows R&amp;amp;D teams to propose model-training projects costing many millions of dollars, with reasonable confidence that these projects will succeed at producing economically valuable systems.&lt;/p&gt;&lt;/blockquote&gt;
&lt;p class="cite"&gt;&amp;mdash; &lt;a href="https://cims.nyu.edu/~sbowman/eightthings.pdf"&gt;Sam Bowman&lt;/a&gt;&lt;/p&gt;

    &lt;p&gt;Tags: &lt;a href="https://simonwillison.net/tags/predictions"&gt;predictions&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai"&gt;ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/generative-ai"&gt;generative-ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/llms"&gt;llms&lt;/a&gt;&lt;/p&gt;



</summary><category term="predictions"/><category term="ai"/><category term="generative-ai"/><category term="llms"/></entry><entry><title>Quoting Tim Bray</title><link href="https://simonwillison.net/2008/Jan/3/ongoing/#atom-tag" rel="alternate"/><published>2008-01-03T13:08:59+00:00</published><updated>2008-01-03T13:08:59+00:00</updated><id>https://simonwillison.net/2008/Jan/3/ongoing/#atom-tag</id><summary type="html">
    &lt;blockquote cite="http://www.tbray.org/ongoing/When/200x/2008/01/02/Prediction-Windows-OS-X-Linux"&gt;&lt;p&gt;The strain due to the fact that most business desktops are locked into the Microsoft platform, at a time when both the Apple and GNU/Linux alternatives are qualitatively safer, better, and cheaper to operate, will start to become impossible to ignore.&lt;/p&gt;&lt;/blockquote&gt;
&lt;p class="cite"&gt;&amp;mdash; &lt;a href="http://www.tbray.org/ongoing/When/200x/2008/01/02/Prediction-Windows-OS-X-Linux"&gt;Tim Bray&lt;/a&gt;&lt;/p&gt;

    &lt;p&gt;Tags: &lt;a href="https://simonwillison.net/tags/apple"&gt;apple&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/linux"&gt;linux&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/macos"&gt;macos&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/microsoft"&gt;microsoft&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/predictions"&gt;predictions&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/tim-bray"&gt;tim-bray&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/windows"&gt;windows&lt;/a&gt;&lt;/p&gt;



</summary><category term="apple"/><category term="linux"/><category term="macos"/><category term="microsoft"/><category term="predictions"/><category term="tim-bray"/><category term="windows"/></entry></feed>