<?xml version="1.0" encoding="utf-8"?>
<feed xml:lang="en-us" xmlns="http://www.w3.org/2005/Atom"><title>Simon Willison's Weblog: ai-ethics</title><link href="http://simonwillison.net/" rel="alternate"/><link href="http://simonwillison.net/tags/ai-ethics.atom" rel="self"/><id>http://simonwillison.net/</id><updated>2026-04-15T15:36:02+00:00</updated><author><name>Simon Willison</name></author><entry><title>Quoting Kyle Kingsbury</title><link href="https://simonwillison.net/2026/Apr/15/kyle-kingsbury/#atom-tag" rel="alternate"/><published>2026-04-15T15:36:02+00:00</published><updated>2026-04-15T15:36:02+00:00</updated><id>https://simonwillison.net/2026/Apr/15/kyle-kingsbury/#atom-tag</id><summary type="html">
    &lt;blockquote cite="https://aphyr.com/posts/419-the-future-of-everything-is-lies-i-guess-new-jobs"&gt;&lt;p&gt;I think we will see some people employed (though perhaps not explicitly) as &lt;em&gt;meat shields&lt;/em&gt;: people who are accountable for ML systems under their supervision. The accountability may be purely internal, as when Meta hires human beings to review the decisions of automated moderation systems. It may be external, as when lawyers are penalized for submitting LLM lies to the court. It may involve formalized responsibility, like a Data Protection Officer. It may be convenient for a company to have third-party subcontractors, like Buscaglia, who can be thrown under the bus when the system as a whole misbehaves.&lt;/p&gt;&lt;/blockquote&gt;
&lt;p class="cite"&gt;&amp;mdash; &lt;a href="https://aphyr.com/posts/419-the-future-of-everything-is-lies-i-guess-new-jobs"&gt;Kyle Kingsbury&lt;/a&gt;, The Future of Everything is Lies, I Guess: New Jobs&lt;/p&gt;

    &lt;p&gt;Tags: &lt;a href="https://simonwillison.net/tags/ai-ethics"&gt;ai-ethics&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/careers"&gt;careers&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai"&gt;ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/kyle-kingsbury"&gt;kyle-kingsbury&lt;/a&gt;&lt;/p&gt;



</summary><category term="ai-ethics"/><category term="careers"/><category term="ai"/><category term="kyle-kingsbury"/></entry><entry><title>Quoting Giles Turnbull</title><link href="https://simonwillison.net/2026/Apr/8/giles-turnbull/#atom-tag" rel="alternate"/><published>2026-04-08T15:18:49+00:00</published><updated>2026-04-08T15:18:49+00:00</updated><id>https://simonwillison.net/2026/Apr/8/giles-turnbull/#atom-tag</id><summary type="html">
    &lt;blockquote cite="https://gilest.org/notes/2026/human-ai/"&gt;&lt;p&gt;I have a feeling that &lt;strong&gt;everyone likes using AI tools to try doing someone else’s profession&lt;/strong&gt;. They’re much less keen when someone else uses it for their profession.&lt;/p&gt;&lt;/blockquote&gt;
&lt;p class="cite"&gt;&amp;mdash; &lt;a href="https://gilest.org/notes/2026/human-ai/"&gt;Giles Turnbull&lt;/a&gt;, AI and the human voice&lt;/p&gt;

    &lt;p&gt;Tags: &lt;a href="https://simonwillison.net/tags/ai-ethics"&gt;ai-ethics&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/writing"&gt;writing&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai"&gt;ai&lt;/a&gt;&lt;/p&gt;



</summary><category term="ai-ethics"/><category term="writing"/><category term="ai"/></entry><entry><title>Anthropic's Project Glasswing - restricting Claude Mythos to security researchers - sounds necessary to me</title><link href="https://simonwillison.net/2026/Apr/7/project-glasswing/#atom-tag" rel="alternate"/><published>2026-04-07T20:52:54+00:00</published><updated>2026-04-07T20:52:54+00:00</updated><id>https://simonwillison.net/2026/Apr/7/project-glasswing/#atom-tag</id><summary type="html">
    &lt;p&gt;Anthropic &lt;em&gt;didn't&lt;/em&gt; release their latest model, Claude Mythos (&lt;a href="https://www-cdn.anthropic.com/53566bf5440a10affd749724787c8913a2ae0841.pdf"&gt;system card PDF&lt;/a&gt;), today. They have instead made it available to a very restricted set of preview partners under their newly announced &lt;a href="https://www.anthropic.com/glasswing"&gt;Project Glasswing&lt;/a&gt;.&lt;/p&gt;
&lt;p&gt;The model is a general purpose model, similar to Claude Opus 4.6, but Anthropic claim that its cyber-security research abilities are strong enough that they need to give the software industry as a whole time to prepare.&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;Mythos Preview has already found thousands of high-severity vulnerabilities, including some in &lt;em&gt;every major operating system and web browser&lt;/em&gt;. Given the rate of AI progress, it will not be long before such capabilities proliferate, potentially beyond actors who are committed to deploying them safely.&lt;/p&gt;
&lt;p&gt;[...]&lt;/p&gt;
&lt;p&gt;Project Glasswing partners will receive access to Claude Mythos Preview to find and fix vulnerabilities or weaknesses in their foundational systems—systems that represent a very large portion of the world’s shared cyberattack surface. We anticipate this work will focus on tasks like local vulnerability detection, black box testing of binaries, securing endpoints, and penetration testing of systems.&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;There's a great deal more technical detail in &lt;a href="https://red.anthropic.com/2026/mythos-preview/"&gt; Assessing Claude Mythos Preview’s cybersecurity capabilities&lt;/a&gt; on the Anthropic Red Team blog:&lt;/p&gt;

&lt;blockquote&gt;&lt;p&gt;In one case, Mythos Preview wrote a web browser exploit that chained together four vulnerabilities, writing a complex &lt;a href="https://en.wikipedia.org/wiki/JIT_spraying "&gt;JIT heap spray&lt;/a&gt; that escaped both renderer and OS sandboxes. It autonomously obtained local privilege escalation exploits on Linux and other operating systems by exploiting subtle race conditions and KASLR-bypasses. And it autonomously wrote a remote code execution exploit on FreeBSD's NFS server that granted full root access to unauthenticated users by splitting a 20-gadget ROP chain over multiple packets.&lt;/p&gt;&lt;/blockquote&gt;
&lt;p&gt;Plus this comparison with Claude 4.6 Opus:&lt;/p&gt;
&lt;blockquote&gt;&lt;p&gt;Our internal evaluations showed that Opus 4.6 generally had a near-0% success rate at autonomous exploit development. But Mythos Preview is in a different league. For example, Opus 4.6 turned the vulnerabilities it had found in Mozilla’s Firefox 147 JavaScript engine—all patched in Firefox 148—into JavaScript shell exploits only two times out of several hundred attempts. We re-ran this experiment as a benchmark for Mythos Preview, which developed working exploits 181 times, and achieved register control on 29 more.&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;Saying "our model is too dangerous to release" is a great way to build buzz around a new model, but in this case I expect their caution is warranted.&lt;/p&gt;
&lt;p&gt;Just a few days (&lt;a href="https://simonwillison.net/2026/Apr/3/"&gt;last Friday&lt;/a&gt;) ago I started a new &lt;a href="https://simonwillison.net/tags/ai-security-research/"&gt;ai-security-research&lt;/a&gt; tag on this blog to acknowledge an uptick in credible security professionals pulling the alarm on how good modern LLMs have got at vulnerability research.&lt;/p&gt;
&lt;p&gt;&lt;a href="https://www.theregister.com/2026/03/26/greg_kroahhartman_ai_kernel/"&gt;Greg Kroah-Hartman&lt;/a&gt; of the Linux kernel:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;Months ago, we were getting what we called 'AI slop,' AI-generated security reports that were obviously wrong or low quality. It was kind of funny. It didn't really worry us.&lt;/p&gt;
&lt;p&gt;Something happened a month ago, and the world switched. Now we have real reports. All open source projects have real reports that are made with AI, but they're good, and they're real.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;&lt;a href="https://mastodon.social/@bagder/116336957584445742"&gt;Daniel Stenberg&lt;/a&gt; of &lt;code&gt;curl&lt;/code&gt;:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;The challenge with AI in open source security has transitioned from an AI slop tsunami into more of a ... plain security report tsunami. Less slop but lots of reports. Many of them really good.&lt;/p&gt;
&lt;p&gt;I'm spending hours per day on this now. It's intense.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;And Thomas Ptacek published &lt;a href="https://sockpuppet.org/blog/2026/03/30/vulnerability-research-is-cooked/"&gt;Vulnerability Research Is Cooked&lt;/a&gt;, a post inspired by his &lt;a href="https://securitycryptographywhatever.com/2026/03/25/ai-bug-finding/"&gt;podcast conversation&lt;/a&gt; with Anthropic's Nicholas Carlini.&lt;/p&gt;
&lt;p&gt;Anthropic have a 5 minute &lt;a href="https://www.youtube.com/watch?v=INGOC6-LLv0"&gt;talking heads video&lt;/a&gt; describing the Glasswing project. Nicholas Carlini appears as one of those talking heads, where he said (highlights mine):&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;It has the ability to chain together vulnerabilities. So what this means is you find two vulnerabilities, either of which doesn't really get you very much independently. But this model is able to create exploits out of three, four, or sometimes five vulnerabilities that in sequence give you some kind of very sophisticated end outcome. [...]&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;I've found more bugs in the last couple of weeks than I found in the rest of my life combined&lt;/strong&gt;. We've used the model to scan a bunch of open source code, and the thing that we went for first was operating systems, because this is the code that underlies the entire internet infrastructure. &lt;strong&gt;For OpenBSD, we found a bug that's been present for 27 years, where I can send a couple of pieces of data to any OpenBSD server and crash it&lt;/strong&gt;. On Linux, we found a number of vulnerabilities where as a user with no permissions, I can elevate myself to the administrator by just running some binary on my machine. For each of these bugs, we told the maintainers who actually run the software about them, and they went and fixed them and have deployed the patches  patches so that anyone who runs the software is no longer vulnerable to these attacks.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;I found this on the &lt;a href="https://www.openbsd.org/errata78.html"&gt;OpenBSD 7.8 errata page&lt;/a&gt;:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;&lt;strong&gt;025: RELIABILITY FIX: March 25, 2026&lt;/strong&gt;  &lt;em&gt;All architectures&lt;/em&gt;&lt;/p&gt;
&lt;p&gt;TCP packets with invalid SACK options could crash the kernel.&lt;/p&gt;
&lt;p&gt;&lt;a href="https://ftp.openbsd.org/pub/OpenBSD/patches/7.8/common/025_sack.patch.sig"&gt;A source code patch exists which remedies this problem.&lt;/a&gt;&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;I tracked that change down in the &lt;a href="https://github.com/openbsd/src"&gt;GitHub mirror&lt;/a&gt; of the OpenBSD CVS repo (apparently they still use CVS!) and found it &lt;a href="https://github.com/openbsd/src/blame/master/sys/netinet/tcp_input.c#L2461"&gt;using git blame&lt;/a&gt;:&lt;/p&gt;
&lt;p&gt;&lt;img src="https://static.simonwillison.net/static/2026/openbsd-27-years.jpg" alt="Screenshot of a Git blame view of C source code around line 2455 showing TCP SACK hole validation logic. Code includes checks using SEQ_GT, SEQ_LT macros on fields like th-&amp;gt;th_ack, tp-&amp;gt;snd_una, sack.start, sack.end, tp-&amp;gt;snd_max, and tp-&amp;gt;snd_holes. Most commits are from 25–27 years ago with messages like &amp;quot;more SACK hole validity testin...&amp;quot; and &amp;quot;knf&amp;quot;, while one recent commit from 3 weeks ago (&amp;quot;Ignore TCP SACK packets wit...&amp;quot;) is highlighted with an orange left border, adding a new guard &amp;quot;if (SEQ_LT(sack.start, tp-&amp;gt;snd_una)) continue;&amp;quot;" style="max-width: 100%;" /&gt;&lt;/p&gt;
&lt;p&gt;Sure enough, the surrounding code is from 27 years ago.&lt;/p&gt;
&lt;p&gt;I'm not sure which Linux vulnerability Nicholas was describing, but it may have been &lt;a href="https://git.kernel.org/pub/scm/linux/kernel/git/stable/linux.git/commit/?id=5133b61aaf437e5f25b1b396b14242a6bb0508e2"&gt;this NFS one&lt;/a&gt; recently covered &lt;a href="https://mtlynch.io/claude-code-found-linux-vulnerability/"&gt;by Michael Lynch
&lt;/a&gt;.&lt;/p&gt;
&lt;p&gt;There's enough smoke here that I believe there's a fire. It's not surprising to find vulnerabilities in decades-old software, especially given that they're mostly written in C, but what's new is that coding agents run by the latest frontier LLMs are proving tirelessly capable at digging up these issues.&lt;/p&gt;
&lt;p&gt;I actually thought to myself on Friday that this sounded like an industry-wide reckoning in the making, and that it might warrant a huge investment of time and money to get ahead of the inevitable barrage of vulnerabilities. Project Glasswing incorporates "$100M in usage credits ... as well as $4M in direct donations to open-source security organizations". Partners include AWS, Apple, Microsoft, Google, and the Linux Foundation. It would be great to see OpenAI involved as well - GPT-5.4 already has a strong reputation for finding security vulnerabilities and they have stronger models on the near horizon.&lt;/p&gt;
&lt;p&gt;The bad news for those of us who are &lt;em&gt;not&lt;/em&gt; trusted partners is this:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;We do not plan to make Claude Mythos Preview generally available, but our eventual goal is to enable our users to safely deploy Mythos-class models at scale—for cybersecurity purposes, but also for the myriad other benefits that such highly capable models will bring. To do so, we need to make progress in developing cybersecurity (and other) safeguards that detect and block the model’s most dangerous outputs. We plan to launch new safeguards with an upcoming Claude Opus model, allowing us to improve and refine them with a model that does not pose the same level of risk as Mythos Preview.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;I can live with that. I think the security risks really are credible here, and having extra time for trusted teams to get ahead of them is a reasonable trade-off.&lt;/p&gt;
    
        &lt;p&gt;Tags: &lt;a href="https://simonwillison.net/tags/security"&gt;security&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/thomas-ptacek"&gt;thomas-ptacek&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai"&gt;ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/generative-ai"&gt;generative-ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/llms"&gt;llms&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/anthropic"&gt;anthropic&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/nicholas-carlini"&gt;nicholas-carlini&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai-ethics"&gt;ai-ethics&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/llm-release"&gt;llm-release&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai-security-research"&gt;ai-security-research&lt;/a&gt;&lt;/p&gt;
    

</summary><category term="security"/><category term="thomas-ptacek"/><category term="ai"/><category term="generative-ai"/><category term="llms"/><category term="anthropic"/><category term="nicholas-carlini"/><category term="ai-ethics"/><category term="llm-release"/><category term="ai-security-research"/></entry><entry><title>Quoting Chengpeng Mou</title><link href="https://simonwillison.net/2026/Apr/5/chengpeng-mou/#atom-tag" rel="alternate"/><published>2026-04-05T21:47:06+00:00</published><updated>2026-04-05T21:47:06+00:00</updated><id>https://simonwillison.net/2026/Apr/5/chengpeng-mou/#atom-tag</id><summary type="html">
    &lt;blockquote cite="https://twitter.com/cpmou2022/status/2040606209800290404"&gt;&lt;p&gt;From anonymized U.S. ChatGPT data, we are seeing:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;~2M weekly messages on health insurance&lt;/li&gt;
&lt;li&gt;~600K weekly messages [classified as healthcare] from people living in “hospital deserts” (30 min drive to nearest hospital)&lt;/li&gt;
&lt;li&gt;7 out of 10 msgs happen outside clinic hours&lt;/li&gt;
&lt;/ul&gt;&lt;/blockquote&gt;
&lt;p class="cite"&gt;&amp;mdash; &lt;a href="https://twitter.com/cpmou2022/status/2040606209800290404"&gt;Chengpeng Mou&lt;/a&gt;, Head of Business Finance, OpenAI&lt;/p&gt;

    &lt;p&gt;Tags: &lt;a href="https://simonwillison.net/tags/ai-ethics"&gt;ai-ethics&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/generative-ai"&gt;generative-ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/openai"&gt;openai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/chatgpt"&gt;chatgpt&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai"&gt;ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/llms"&gt;llms&lt;/a&gt;&lt;/p&gt;



</summary><category term="ai-ethics"/><category term="generative-ai"/><category term="openai"/><category term="chatgpt"/><category term="ai"/><category term="llms"/></entry><entry><title>Vulnerability Research Is Cooked</title><link href="https://simonwillison.net/2026/Apr/3/vulnerability-research-is-cooked/#atom-tag" rel="alternate"/><published>2026-04-03T23:59:08+00:00</published><updated>2026-04-03T23:59:08+00:00</updated><id>https://simonwillison.net/2026/Apr/3/vulnerability-research-is-cooked/#atom-tag</id><summary type="html">
    
&lt;p&gt;&lt;strong&gt;&lt;a href="https://sockpuppet.org/blog/2026/03/30/vulnerability-research-is-cooked/"&gt;Vulnerability Research Is Cooked&lt;/a&gt;&lt;/strong&gt;&lt;/p&gt;
Thomas Ptacek's take on the sudden and enormous impact the latest frontier models are having on the field of vulnerability research.&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;Within the next few months, coding agents will drastically alter both the practice and the economics of exploit development. Frontier model improvement won’t be a slow burn, but rather a step function. Substantial amounts of high-impact vulnerability research (maybe even most of it) will happen simply by pointing an agent at a source tree and typing “find me zero days”.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;Why are agents so good at this? A combination of baked-in knowledge, pattern matching ability and brute force:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;You can't design a better problem for an LLM agent than exploitation research.&lt;/p&gt;
&lt;p&gt;Before you feed it a single token of context, a frontier LLM already encodes supernatural amounts of correlation across vast bodies of source code. Is the Linux KVM hypervisor connected to the &lt;code&gt;hrtimer&lt;/code&gt; subsystem, &lt;code&gt;workqueue&lt;/code&gt;, or &lt;code&gt;perf_event&lt;/code&gt;? The model knows.&lt;/p&gt;
&lt;p&gt;Also baked into those model weights: the complete library of documented "bug classes" on which all exploit development builds: stale pointers, integer mishandling, type confusion, allocator grooming, and all the known ways of promoting a wild write to a controlled 64-bit read/write in Firefox.&lt;/p&gt;
&lt;p&gt;Vulnerabilities are found by pattern-matching bug classes and constraint-solving for reachability and exploitability. Precisely the implicit search problems that LLMs are most gifted at solving. Exploit outcomes are straightforwardly testable success/failure trials. An agent never gets bored and will search forever if you tell it to.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;The article was partly inspired by &lt;a href="https://securitycryptographywhatever.com/2026/03/25/ai-bug-finding/"&gt;this episode of the Security Cryptography Whatever podcast&lt;/a&gt;, where David Adrian, Deirdre Connolly, and Thomas interviewed Anthropic's Nicholas Carlini for 1 hour 16 minutes.&lt;/p&gt;
&lt;p&gt;I just started a new tag here for &lt;a href="https://simonwillison.net/tags/ai-security-research/"&gt;ai-security-research&lt;/a&gt; - it's up to 11 posts already.


    &lt;p&gt;Tags: &lt;a href="https://simonwillison.net/tags/thomas-ptacek"&gt;thomas-ptacek&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/careers"&gt;careers&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/security"&gt;security&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/generative-ai"&gt;generative-ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai-security-research"&gt;ai-security-research&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai"&gt;ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/llms"&gt;llms&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/nicholas-carlini"&gt;nicholas-carlini&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai-ethics"&gt;ai-ethics&lt;/a&gt;&lt;/p&gt;



</summary><category term="thomas-ptacek"/><category term="careers"/><category term="security"/><category term="generative-ai"/><category term="ai-security-research"/><category term="ai"/><category term="llms"/><category term="nicholas-carlini"/><category term="ai-ethics"/></entry><entry><title>The cognitive impact of coding agents</title><link href="https://simonwillison.net/2026/Apr/3/cognitive-cost/#atom-tag" rel="alternate"/><published>2026-04-03T23:57:04+00:00</published><updated>2026-04-03T23:57:04+00:00</updated><id>https://simonwillison.net/2026/Apr/3/cognitive-cost/#atom-tag</id><summary type="html">
    &lt;p&gt;A fun thing about &lt;a href="https://simonwillison.net/2026/Apr/2/lennys-podcast/"&gt;recording a podcast&lt;/a&gt; with a professional like Lenny Rachitsky is that his team know how to slice the resulting video up into TikTok-sized short form vertical videos. Here's &lt;a href="https://x.com/lennysan/status/2039845666680176703"&gt;one he shared on Twitter today&lt;/a&gt; which ended up attracting over 1.1m views!&lt;/p&gt;
&lt;p&gt;&lt;video
  src="https://static.simonwillison.net/static/2026/cognitive-cost.mp4"
  poster="https://static.simonwillison.net/static/2026/cognitive-cost-poster.jpg"
  controls
  preload="none"
  playsinline
  style="display:block; max-width:400px; width:100%; height:auto; margin:0 auto"
&gt;&lt;track src="https://static.simonwillison.net/static/2026/cognitive-cost.vtt" kind="captions" srclang="en" label="English"&gt;&lt;/video&gt;
&lt;/p&gt;
&lt;p&gt;That was 48 seconds. Our &lt;a href="https://simonwillison.net/2026/Apr/2/lennys-podcast/"&gt;full conversation&lt;/a&gt; lasted 1 hour 40 minutes.&lt;/p&gt;

    &lt;p&gt;Tags: &lt;a href="https://simonwillison.net/tags/ai-ethics"&gt;ai-ethics&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/coding-agents"&gt;coding-agents&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/agentic-engineering"&gt;agentic-engineering&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/generative-ai"&gt;generative-ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/podcast-appearances"&gt;podcast-appearances&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai"&gt;ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/llms"&gt;llms&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/cognitive-debt"&gt;cognitive-debt&lt;/a&gt;&lt;/p&gt;



</summary><category term="ai-ethics"/><category term="coding-agents"/><category term="agentic-engineering"/><category term="generative-ai"/><category term="podcast-appearances"/><category term="ai"/><category term="llms"/><category term="cognitive-debt"/></entry><entry><title>Mr. Chatterbox is a (weak) Victorian-era ethically trained model you can run on your own computer</title><link href="https://simonwillison.net/2026/Mar/30/mr-chatterbox/#atom-tag" rel="alternate"/><published>2026-03-30T14:28:34+00:00</published><updated>2026-03-30T14:28:34+00:00</updated><id>https://simonwillison.net/2026/Mar/30/mr-chatterbox/#atom-tag</id><summary type="html">
    &lt;p&gt;Trip Venturella released &lt;a href="https://www.estragon.news/mr-chatterbox-or-the-modern-prometheus/"&gt;Mr. Chatterbox&lt;/a&gt;, a language model trained entirely on out-of-copyright text from the British Library. Here's how he describes it in &lt;a href="https://huggingface.co/tventurella/mr_chatterbox_model"&gt;the model card&lt;/a&gt;:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;Mr. Chatterbox is a language model trained entirely from scratch on a corpus of over 28,000 Victorian-era British texts published between 1837 and 1899, drawn from a dataset made available &lt;a href="https://huggingface.co/datasets/TheBritishLibrary/blbooks"&gt;by the British Library&lt;/a&gt;. The model has absolutely no training inputs from after 1899 — the vocabulary and ideas are formed exclusively from nineteenth-century literature.&lt;/p&gt;
&lt;p&gt;Mr. Chatterbox's training corpus was 28,035 books, with an estimated 2.93 billion input tokens after filtering. The model has roughly 340 million paramaters, roughly the same size as GPT-2-Medium. The difference is, of course, that unlike GPT-2, Mr. Chatterbox is trained entirely on historical data.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;Given how hard it is to train a useful LLM without using vast amounts of scraped, unlicensed data I've been dreaming of a model like this for a couple of years now. What would a model trained on out-of-copyright text be like to chat with?&lt;/p&gt;
&lt;p&gt;Thanks to Trip we can now find out for ourselves!&lt;/p&gt;
&lt;p&gt;The model itself is tiny, at least by Large Language Model standards - just &lt;a href="https://huggingface.co/tventurella/mr_chatterbox_model/tree/main"&gt;2.05GB&lt;/a&gt; on disk. You can try it out using Trip's &lt;a href="https://huggingface.co/spaces/tventurella/mr_chatterbox"&gt;HuggingFace Spaces demo&lt;/a&gt;:&lt;/p&gt;
&lt;p style="text-align: center"&gt;&lt;img src="https://static.simonwillison.net/static/2026/chatterbox.jpg" alt="Screenshot of a Victorian-themed chatbot interface titled &amp;quot;🎩 Mr. Chatterbox (Beta)&amp;quot; with subtitle &amp;quot;The Victorian Gentleman Chatbot&amp;quot;. The conversation shows a user asking &amp;quot;How should I behave at dinner?&amp;quot; with the bot replying &amp;quot;My good fellow, one might presume that such trivialities could not engage your attention during an evening's discourse!&amp;quot; The user then asks &amp;quot;What are good topics?&amp;quot; and the bot responds &amp;quot;The most pressing subjects of our society— Indeed, a gentleman must endeavor to engage the conversation with grace and vivacity. Such pursuits serve as vital antidotes against ennui when engaged in agreeable company.&amp;quot; A text input field at the bottom reads &amp;quot;Say hello...&amp;quot; with a send button. The interface uses a dark maroon and cream color scheme." style="max-width: 80%;" /&gt;&lt;/p&gt;
&lt;p&gt;Honestly, it's pretty terrible. Talking with it feels more like chatting with a Markov chain than an LLM - the responses may have a delightfully Victorian flavor to them but it's hard to get a response that usefully answers a question.&lt;/p&gt;
&lt;p&gt;The &lt;a href="https://arxiv.org/abs/2203.15556"&gt;2022 Chinchilla paper&lt;/a&gt; suggests a ratio of 20x the parameter count to training tokens. For a 340m model that would suggest around 7 billion tokens, more than twice the British Library corpus used here. The smallest Qwen 3.5 model is 600m parameters and that model family starts to get interesting at 2b - so my hunch is we would need 4x or more the training data to get something that starts to feel like a useful conversational partner.&lt;/p&gt;
&lt;p&gt;But what a fun project!&lt;/p&gt;
&lt;h4 id="running-it-locally-with-llm"&gt;Running it locally with LLM&lt;/h4&gt;
&lt;p&gt;I decided to see if I could run the model on my own machine using my &lt;a href="https://llm.datasette.io/"&gt;LLM&lt;/a&gt; framework.&lt;/p&gt;
&lt;p&gt;I got Claude Code to do most of the work - &lt;a href="https://gisthost.github.io/?7d0f00e152dd80d617b5e501e4ff025b/index.html"&gt;here's the transcript&lt;/a&gt;.&lt;/p&gt;
&lt;p&gt;Trip trained the model using Andrej Karpathy's &lt;a href="https://github.com/karpathy/nanochat"&gt;nanochat&lt;/a&gt;, so I cloned that project, pulled the model weights and told Claude to build a Python script to run the model. Once we had that working (which ended up needing some extra details from the &lt;a href="https://huggingface.co/spaces/tventurella/mr_chatterbox/tree/main"&gt;Space demo source code&lt;/a&gt;) I had Claude &lt;a href="https://llm.datasette.io/en/stable/plugins/tutorial-model-plugin.html"&gt;read the LLM plugin tutorial&lt;/a&gt; and build the rest of the plugin.&lt;/p&gt;
&lt;p&gt;&lt;a href="https://github.com/simonw/llm-mrchatterbox"&gt;llm-mrchatterbox&lt;/a&gt; is the result. Install the plugin like this:&lt;/p&gt;
&lt;pre&gt;&lt;code&gt;llm install llm-mrchatterbox
&lt;/code&gt;&lt;/pre&gt;
&lt;p&gt;The first time you run a prompt it will fetch the 2.05GB model file from Hugging Face. Try that like this:&lt;/p&gt;
&lt;pre&gt;&lt;code&gt;llm -m mrchatterbox "Good day, sir"
&lt;/code&gt;&lt;/pre&gt;
&lt;p&gt;Or start an ongoing chat session like this:&lt;/p&gt;
&lt;pre&gt;&lt;code&gt;llm chat -m mrchatterbox
&lt;/code&gt;&lt;/pre&gt;
&lt;p&gt;If you don't have LLM installed you can still get a chat session started from scratch using uvx like this:&lt;/p&gt;
&lt;pre&gt;&lt;code&gt;uvx --with llm-mrchatterbox llm chat -m mrchatterbox
&lt;/code&gt;&lt;/pre&gt;
&lt;p&gt;When you are finished with the model you can delete the cached file using:&lt;/p&gt;
&lt;pre&gt;&lt;code&gt;llm mrchatterbox delete-model
&lt;/code&gt;&lt;/pre&gt;
&lt;p&gt;This is the first time I've had Claude Code build a full LLM model plugin from scratch and it worked really well. I expect I'll be using this method again in the future.&lt;/p&gt;
&lt;p&gt;I continue to hope we can get a useful model from entirely public domain data. The fact that Trip was able to get this far using nanochat and 2.93 billion training tokens is a promising start.&lt;/p&gt;

&lt;p id="update-31st"&gt;&lt;strong&gt;Update 31st March 2026&lt;/strong&gt;: I had missed this when I first published this piece but Trip has his own &lt;a href="https://www.estragon.news/mr-chatterbox-or-the-modern-prometheus/"&gt;detailed writeup of the project&lt;/a&gt; which goes into much more detail about how he trained the model. Here's how the books were filtered for pre-training:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;First, I downloaded the British Library dataset split of all 19th-century books. I filtered those down to books contemporaneous with the reign of Queen Victoria—which, unfortunately, cut out the novels of Jane Austen—and further filtered those down to a set of books with a optical character recognition (OCR) confidence of .65 or above, as listed in the metadata. This left me with 28,035 books, or roughly 2.93 billion tokes for pretraining data.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;Getting it to behave like a conversational model was a lot harder. Trip started by trying to train on plays by Oscar Wilde and George Bernard Shaw, but found they didn't provide enough pairs. Then he tried extracting dialogue pairs from the books themselves with poor results. The approach that worked was to have Claude Haiku and GPT-4o-mini generate synthetic conversation pairs for the supervised fine tuning, which solved the problem but sadly I think dilutes the "no training inputs from after 1899" claim from the original model card.&lt;/p&gt;
    
        &lt;p&gt;Tags: &lt;a href="https://simonwillison.net/tags/ai"&gt;ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/andrej-karpathy"&gt;andrej-karpathy&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/generative-ai"&gt;generative-ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/local-llms"&gt;local-llms&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/llms"&gt;llms&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai-assisted-programming"&gt;ai-assisted-programming&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/hugging-face"&gt;hugging-face&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/llm"&gt;llm&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/training-data"&gt;training-data&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/uv"&gt;uv&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai-ethics"&gt;ai-ethics&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/claude-code"&gt;claude-code&lt;/a&gt;&lt;/p&gt;
    

</summary><category term="ai"/><category term="andrej-karpathy"/><category term="generative-ai"/><category term="local-llms"/><category term="llms"/><category term="ai-assisted-programming"/><category term="hugging-face"/><category term="llm"/><category term="training-data"/><category term="uv"/><category term="ai-ethics"/><category term="claude-code"/></entry><entry><title>Quoting Richard Fontana</title><link href="https://simonwillison.net/2026/Mar/27/richard-fontana/#atom-tag" rel="alternate"/><published>2026-03-27T21:11:17+00:00</published><updated>2026-03-27T21:11:17+00:00</updated><id>https://simonwillison.net/2026/Mar/27/richard-fontana/#atom-tag</id><summary type="html">
    &lt;blockquote cite="https://github.com/chardet/chardet/issues/334#issuecomment-4098524555"&gt;&lt;p&gt;FWIW, IANDBL, TINLA, etc., I don’t currently see any basis for concluding that chardet 7.0.0 is required to be released under the LGPL. AFAIK no one including Mark Pilgrim has identified persistence of copyrightable expressive material from earlier versions in 7.0.0 nor has anyone articulated some viable alternate theory of license violation. [...]&lt;/p&gt;&lt;/blockquote&gt;
&lt;p class="cite"&gt;&amp;mdash; &lt;a href="https://github.com/chardet/chardet/issues/334#issuecomment-4098524555"&gt;Richard Fontana&lt;/a&gt;, LGPLv3 co-author, weighing in on the &lt;a href="https://simonwillison.net/2026/Mar/5/chardet/"&gt;chardet relicensing situation&lt;/a&gt;&lt;/p&gt;

    &lt;p&gt;Tags: &lt;a href="https://simonwillison.net/tags/open-source"&gt;open-source&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai-ethics"&gt;ai-ethics&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/llms"&gt;llms&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai"&gt;ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/generative-ai"&gt;generative-ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai-assisted-programming"&gt;ai-assisted-programming&lt;/a&gt;&lt;/p&gt;



</summary><category term="open-source"/><category term="ai-ethics"/><category term="llms"/><category term="ai"/><category term="generative-ai"/><category term="ai-assisted-programming"/></entry><entry><title>Quoting Neurotica</title><link href="https://simonwillison.net/2026/Mar/23/neurotica/#atom-tag" rel="alternate"/><published>2026-03-23T23:31:45+00:00</published><updated>2026-03-23T23:31:45+00:00</updated><id>https://simonwillison.net/2026/Mar/23/neurotica/#atom-tag</id><summary type="html">
    &lt;blockquote cite="https://bsky.app/profile/schwarzgerat.bsky.social/post/3mhqu5dogos2v"&gt;&lt;p&gt;slop is something that takes more human effort to consume than it took to produce. When my coworker sends me raw Gemini output he’s not expressing his freedom to create, he’s disrespecting the value of my time&lt;/p&gt;&lt;/blockquote&gt;
&lt;p class="cite"&gt;&amp;mdash; &lt;a href="https://bsky.app/profile/schwarzgerat.bsky.social/post/3mhqu5dogos2v"&gt;Neurotica&lt;/a&gt;, @schwarzgerat.bsky.social&lt;/p&gt;

    &lt;p&gt;Tags: &lt;a href="https://simonwillison.net/tags/ai-ethics"&gt;ai-ethics&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/slop"&gt;slop&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/generative-ai"&gt;generative-ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai"&gt;ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/llms"&gt;llms&lt;/a&gt;&lt;/p&gt;



</summary><category term="ai-ethics"/><category term="slop"/><category term="generative-ai"/><category term="ai"/><category term="llms"/></entry><entry><title>Profiling Hacker News users based on their comments</title><link href="https://simonwillison.net/2026/Mar/21/profiling-hacker-news-users/#atom-tag" rel="alternate"/><published>2026-03-21T23:59:47+00:00</published><updated>2026-03-21T23:59:47+00:00</updated><id>https://simonwillison.net/2026/Mar/21/profiling-hacker-news-users/#atom-tag</id><summary type="html">
    &lt;p&gt;Here's a mildly dystopian prompt I've been experimenting with recently: "Profile this user", accompanied by a copy of their last 1,000 comments on Hacker News.&lt;/p&gt;
&lt;p&gt;Obtaining those comments is easy. The &lt;a href="https://hn.algolia.com/api"&gt;Algolia Hacker News API&lt;/a&gt; supports listing comments sorted by date that have a specific tag, and the author of a comment is tagged there as &lt;code&gt;author_username&lt;/code&gt;. Here's a JSON feed of my (&lt;code&gt;simonw&lt;/code&gt;) most recent comments, for example:&lt;/p&gt;
&lt;p&gt;&lt;a href="https://hn.algolia.com/api/v1/search_by_date?tags=comment,author_simonw&amp;amp;hitsPerPage=1000"&gt;https://hn.algolia.com/api/v1/search_by_date?tags=comment,author_simonw&amp;amp;hitsPerPage=1000&lt;/a&gt;&lt;/p&gt;
&lt;p&gt;The Algolia API is served with open CORS headers, which means you can access the API from JavaScript running on any web page.&lt;/p&gt;
&lt;p&gt;Last August I &lt;a href="https://chatgpt.com/share/68a4d2b1-a678-8006-aca0-cec51b243bd3"&gt;had ChatGPT build me&lt;/a&gt; a &lt;a href="https://tools.simonwillison.net/hn-comments-for-user"&gt;simple tool&lt;/a&gt; for hitting that API for any user which fetches their comments and gives me a mobile-friendly "copy to clipboard" button. I've since &lt;a href="https://github.com/simonw/tools/commits/main/hn-comments-for-user.html"&gt;tweaked it a few times&lt;/a&gt; with Claude.&lt;/p&gt;
&lt;p&gt;I can then paste the whole lot into any LLM - these days I mostly use Claude Opus 4.6 for this - and prompt "profile this user" to see what it can figure out.&lt;/p&gt;
&lt;p&gt;It's startlingly effective! It feels invasive to quote the profile of another user here, so I'll show you what it produces for me:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;This is &lt;strong&gt;Simon Willison&lt;/strong&gt; — a prolific, independent software developer, blogger, and one of the most visible voices in the AI-assisted coding space. Here's a profile drawn from his comments:&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Professional identity:&lt;/strong&gt; Independent developer and writer. Co-creator of Django, creator of Datasette and many other open source tools. On the board of the Python Software Foundation. Runs a heavily-trafficked personal blog at simonwillison.net with multiple content types (entries, link blog, notes, quotations). Monetizes through GitHub sponsors, ethical ads, and occasional consulting — not through any AI company payroll, a point he defends frequently.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Core thesis on AI coding:&lt;/strong&gt; He's the leading evangelist for what he calls "agentic engineering" — using coding agents (Claude Code, Codex CLI) as productivity multipliers for skilled developers. His central argument: LLMs don't replace programmers, they amplify existing expertise. He's writing a guide/book on this at simonwillison.net/guides. He insists these tools are deceptively difficult to use well, and that most people who report bad results simply haven't learned the craft yet.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Working style:&lt;/strong&gt; He does most of his programming from his iPhone via Claude Code for web, often while riding BART, walking the dog, or sitting in the garden. He embraces "YOLO mode" (auto-approving agent actions) and frequently runs 2-3 agent sessions in parallel. He starts every session with "run uv run pytest" to anchor the agent in TDD. He's on the $200/month Claude Max plan.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Key technical interests:&lt;/strong&gt; Sandboxing and security (WebAssembly, Pyodide, sandbox-exec, Firecracker), SQLite (deeply), Python packaging and tooling (uv, PyPI distribution tricks), browser-in-a-browser experiments (v86, WASM Linux), and local LLM inference. He's been learning Go "by osmosis" through coding agents.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Security consciousness:&lt;/strong&gt; He coined the term "prompt injection" and the "lethal trifecta" (access to private data + exposure to untrusted input + ability to take actions). He's genuinely alarmed about OpenClaw/Clawdbot security risks and has been predicting a headline-grabbing prompt injection attack "in the next six months" for over two years — acknowledging it keeps not happening while insisting the risk is real. He references the "normalization of deviance" pattern frequently.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Personality and debate style:&lt;/strong&gt; Energetic, combative when challenged, but generally good-natured. He engages heavily in HN comment threads — sometimes dozens of replies in a single discussion. He's transparent about his biases and maintains a public disclosures page. He pushes back hard against what he sees as motivated reasoning from AI skeptics, but also against AGI hype ("I refer to that as the science fiction version of AI"). He gets visibly frustrated by sock-puppet accounts and bad-faith criticism. He has a self-deprecating streak ("my stupid pelican benchmark").&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Recurring themes across hundreds of comments:&lt;/strong&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;"Two things can be true at the same time" — he holds nuanced positions&lt;/li&gt;
&lt;li&gt;Tests are for productivity, not just quality&lt;/li&gt;
&lt;li&gt;The November 2025 model releases (Opus 4.5, GPT-5.2) were a genuine inflection point&lt;/li&gt;
&lt;li&gt;Code review is the biggest bottleneck in agent-assisted workflows&lt;/li&gt;
&lt;li&gt;"Cognitive debt" is a real and unsolved problem&lt;/li&gt;
&lt;li&gt;The best engineering practices (tests, docs, CI/CD, clean code) also make agents work better&lt;/li&gt;
&lt;li&gt;He's deliberately trying to "teach people good software engineering while tricking them into thinking the book is about AI"&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;&lt;strong&gt;Personal interests mentioned:&lt;/strong&gt; Niche museums (runs niche-museums.com), New Zealand kākāpō parrots, cooking food from countries he's never visited, chickens in his garden. Lives in Half Moon Bay, CA near a major pelican roost. The pelican-on-a-bicycle SVG test is his signature LLM benchmark.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;In a sentence:&lt;/strong&gt; A deeply experienced, independently-minded developer who's genuinely excited about AI coding tools, fights hard against both uncritical hype and reflexive dismissal, and is trying to professionalize how the industry uses these tools — while also worrying loudly about the security implications almost nobody else takes seriously enough.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;This all checks out! I ran this in Claude incognito mode to hopefully prevent Claude from guessing that I was evaluating myself and sycophantically glazing me - the tone of the response it gave here is similar to the tone I've seen against other accounts.&lt;/p&gt;

&lt;p&gt;I expect it guessed my real name due to my habit of linking to my own writing from some of my comments, which provides plenty of simonwillison.net URLs for it to associate with my public persona. I haven't seen it take a guess at a real name for any of the other profiles I've generated.&lt;/p&gt;
&lt;p&gt;It's a little creepy to be able to derive this much information about someone so easily, even when they've shared that freely in a public (and API-available) place.&lt;/p&gt;
&lt;p&gt;I mainly use this to check that I'm not getting embroiled in an extensive argument with someone who has a history of arguing in bad faith. Thankfully that's rarely the case - Hacker News continues to be a responsibly moderated online space.&lt;/p&gt;
    
        &lt;p&gt;Tags: &lt;a href="https://simonwillison.net/tags/hacker-news"&gt;hacker-news&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai"&gt;ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/generative-ai"&gt;generative-ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/llms"&gt;llms&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai-ethics"&gt;ai-ethics&lt;/a&gt;&lt;/p&gt;
    

</summary><category term="hacker-news"/><category term="ai"/><category term="generative-ai"/><category term="llms"/><category term="ai-ethics"/></entry><entry><title>Quoting Tim Schilling</title><link href="https://simonwillison.net/2026/Mar/17/tim-schilling/#atom-tag" rel="alternate"/><published>2026-03-17T16:13:37+00:00</published><updated>2026-03-17T16:13:37+00:00</updated><id>https://simonwillison.net/2026/Mar/17/tim-schilling/#atom-tag</id><summary type="html">
    &lt;blockquote cite="https://www.better-simple.com/django/2026/03/16/give-django-your-time-and-money/"&gt;&lt;p&gt;If you do not understand the ticket, if you do not understand the solution, or if you do not understand the feedback on your PR, then your use of LLM is hurting Django as a whole. [...]&lt;/p&gt;
&lt;p&gt;For a reviewer, it’s demoralizing to communicate with a facade of a human.&lt;/p&gt;
&lt;p&gt;This is because contributing to open source, especially Django, is a communal endeavor. Removing your humanity from that experience makes that endeavor more difficult. If you use an LLM to contribute to Django, it needs to be as a complementary tool, not as your vehicle.&lt;/p&gt;&lt;/blockquote&gt;
&lt;p class="cite"&gt;&amp;mdash; &lt;a href="https://www.better-simple.com/django/2026/03/16/give-django-your-time-and-money/"&gt;Tim Schilling&lt;/a&gt;, Give Django your time and money, not your tokens&lt;/p&gt;

    &lt;p&gt;Tags: &lt;a href="https://simonwillison.net/tags/ai-ethics"&gt;ai-ethics&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/open-source"&gt;open-source&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/generative-ai"&gt;generative-ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai"&gt;ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/django"&gt;django&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/llms"&gt;llms&lt;/a&gt;&lt;/p&gt;



</summary><category term="ai-ethics"/><category term="open-source"/><category term="generative-ai"/><category term="ai"/><category term="django"/><category term="llms"/></entry><entry><title>Quoting A member of Anthropic’s alignment-science team</title><link href="https://simonwillison.net/2026/Mar/16/blackmail/#atom-tag" rel="alternate"/><published>2026-03-16T21:38:55+00:00</published><updated>2026-03-16T21:38:55+00:00</updated><id>https://simonwillison.net/2026/Mar/16/blackmail/#atom-tag</id><summary type="html">
    &lt;blockquote cite="https://www.newyorker.com/news/annals-of-inquiry/the-pentagon-went-to-war-with-anthropic-whats-really-at-stake?_sp=9a6e0ff7-2bfd-46f8-a9e1-3941ef2003b5.1773495048769"&gt;&lt;p&gt;The point of &lt;a href="https://simonwillison.net/2025/Jun/20/agentic-misalignment/"&gt;the blackmail exercise&lt;/a&gt; was to have something to describe to policymakers—results that are visceral enough to land with people, and make misalignment risk actually salient in practice for people who had never thought about it before.&lt;/p&gt;&lt;/blockquote&gt;
&lt;p class="cite"&gt;&amp;mdash; &lt;a href="https://www.newyorker.com/news/annals-of-inquiry/the-pentagon-went-to-war-with-anthropic-whats-really-at-stake?_sp=9a6e0ff7-2bfd-46f8-a9e1-3941ef2003b5.1773495048769"&gt;A member of Anthropic’s alignment-science team&lt;/a&gt;, as told to Gideon Lewis-Kraus&lt;/p&gt;

    &lt;p&gt;Tags: &lt;a href="https://simonwillison.net/tags/ai-ethics"&gt;ai-ethics&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/anthropic"&gt;anthropic&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/claude"&gt;claude&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/generative-ai"&gt;generative-ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai"&gt;ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/llms"&gt;llms&lt;/a&gt;&lt;/p&gt;



</summary><category term="ai-ethics"/><category term="anthropic"/><category term="claude"/><category term="generative-ai"/><category term="ai"/><category term="llms"/></entry><entry><title>Quoting Jannis Leidel</title><link href="https://simonwillison.net/2026/Mar/14/jannis-leidel/#atom-tag" rel="alternate"/><published>2026-03-14T18:41:25+00:00</published><updated>2026-03-14T18:41:25+00:00</updated><id>https://simonwillison.net/2026/Mar/14/jannis-leidel/#atom-tag</id><summary type="html">
    &lt;blockquote cite="https://jazzband.co/news/2026/03/14/sunsetting-jazzband"&gt;&lt;p&gt;GitHub’s &lt;a href="https://www.theregister.com/2026/02/18/godot_maintainers_struggle_with_draining/"&gt;slopocalypse&lt;/a&gt; – the flood of AI-generated spam PRs and issues – has made Jazzband’s model of open membership and shared push access untenable.&lt;/p&gt;
&lt;p&gt;Jazzband was designed for a world where the worst case was someone accidentally merging the wrong PR. In a world where &lt;a href="https://www.devclass.com/ai-ml/2026/02/19/github-itself-to-blame-for-ai-slop-prs-say-devs/4091420"&gt;only 1 in 10 AI-generated PRs meets project standards&lt;/a&gt;, where curl had to &lt;a href="https://daniel.haxx.se/blog/2026/01/26/the-end-of-the-curl-bug-bounty/"&gt;shut down its bug bounty&lt;/a&gt; because confirmation rates dropped below 5%, and where GitHub’s own response was a &lt;a href="https://www.theregister.com/2026/02/03/github_kill_switch_pull_requests_ai"&gt;kill switch to disable pull requests entirely&lt;/a&gt; – an organization that gives push access to everyone who joins simply can’t operate safely anymore.&lt;/p&gt;&lt;/blockquote&gt;
&lt;p class="cite"&gt;&amp;mdash; &lt;a href="https://jazzband.co/news/2026/03/14/sunsetting-jazzband"&gt;Jannis Leidel&lt;/a&gt;, Sunsetting Jazzband&lt;/p&gt;

    &lt;p&gt;Tags: &lt;a href="https://simonwillison.net/tags/ai-ethics"&gt;ai-ethics&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/open-source"&gt;open-source&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/python"&gt;python&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai"&gt;ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/github"&gt;github&lt;/a&gt;&lt;/p&gt;



</summary><category term="ai-ethics"/><category term="open-source"/><category term="python"/><category term="ai"/><category term="github"/></entry><entry><title>MALUS - Clean Room as a Service</title><link href="https://simonwillison.net/2026/Mar/12/malus/#atom-tag" rel="alternate"/><published>2026-03-12T20:08:55+00:00</published><updated>2026-03-12T20:08:55+00:00</updated><id>https://simonwillison.net/2026/Mar/12/malus/#atom-tag</id><summary type="html">
    
&lt;p&gt;&lt;strong&gt;&lt;a href="https://malus.sh/"&gt;MALUS - Clean Room as a Service&lt;/a&gt;&lt;/strong&gt;&lt;/p&gt;
Brutal satire on the whole vibe-porting license washing thing (&lt;a href="https://simonwillison.net/2026/Mar/5/chardet/"&gt;previously&lt;/a&gt;):&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;Finally, liberation from open source license obligations.&lt;/p&gt;
&lt;p&gt;Our proprietary AI robots independently recreate any open source project from scratch. The result? &lt;strong&gt;Legally distinct code&lt;/strong&gt; with corporate-friendly licensing. No attribution. No copyleft. No problems..&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;I admit it took me a moment to confirm that this was a joke. Just too on-the-nose.

    &lt;p&gt;&lt;small&gt;&lt;/small&gt;Via &lt;a href="https://news.ycombinator.com/item?id=47350424"&gt;Hacker News&lt;/a&gt;&lt;/small&gt;&lt;/p&gt;


    &lt;p&gt;Tags: &lt;a href="https://simonwillison.net/tags/ai-ethics"&gt;ai-ethics&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/open-source"&gt;open-source&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/generative-ai"&gt;generative-ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai"&gt;ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/llms"&gt;llms&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/vibe-porting"&gt;vibe-porting&lt;/a&gt;&lt;/p&gt;



</summary><category term="ai-ethics"/><category term="open-source"/><category term="generative-ai"/><category term="ai"/><category term="llms"/><category term="vibe-porting"/></entry><entry><title>Quoting Joseph Weizenbaum</title><link href="https://simonwillison.net/2026/Mar/8/joseph-weizenbaum/#atom-tag" rel="alternate"/><published>2026-03-08T14:59:48+00:00</published><updated>2026-03-08T14:59:48+00:00</updated><id>https://simonwillison.net/2026/Mar/8/joseph-weizenbaum/#atom-tag</id><summary type="html">
    &lt;blockquote cite="https://archive.org/details/computerpowerhum0000weiz_v0i3?q=realized"&gt;&lt;p&gt;What I had not realized is that extremely short exposures to a relatively simple computer program could induce powerful delusional thinking in quite normal people.&lt;/p&gt;&lt;/blockquote&gt;
&lt;p class="cite"&gt;&amp;mdash; &lt;a href="https://archive.org/details/computerpowerhum0000weiz_v0i3?q=realized"&gt;Joseph Weizenbaum&lt;/a&gt;, creator of ELIZA, in 1976 (&lt;a href="https://www.tiktok.com/@professorcasey/video/7614890527711825183"&gt;via&lt;/a&gt;)&lt;/p&gt;

    &lt;p&gt;Tags: &lt;a href="https://simonwillison.net/tags/ai-ethics"&gt;ai-ethics&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai"&gt;ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/computer-history"&gt;computer-history&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/internet-archive"&gt;internet-archive&lt;/a&gt;&lt;/p&gt;



</summary><category term="ai-ethics"/><category term="ai"/><category term="computer-history"/><category term="internet-archive"/></entry><entry><title>Anthropic and the Pentagon</title><link href="https://simonwillison.net/2026/Mar/6/anthropic-and-the-pentagon/#atom-tag" rel="alternate"/><published>2026-03-06T17:26:50+00:00</published><updated>2026-03-06T17:26:50+00:00</updated><id>https://simonwillison.net/2026/Mar/6/anthropic-and-the-pentagon/#atom-tag</id><summary type="html">
    
&lt;p&gt;&lt;strong&gt;&lt;a href="https://www.schneier.com/blog/archives/2026/03/anthropic-and-the-pentagon.html"&gt;Anthropic and the Pentagon&lt;/a&gt;&lt;/strong&gt;&lt;/p&gt;
This piece by Bruce Schneier and Nathan E. Sanders is the most thoughtful and grounded coverage I've seen of the recent and ongoing Pentagon/OpenAI/Anthropic contract situation.&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;AI models are increasingly commodified. The top-tier offerings have about the same performance, and there is little to differentiate one from the other. The latest models from Anthropic, OpenAI and Google, in particular, tend to leapfrog each other with minor hops forward in quality every few months. [...]&lt;/p&gt;
&lt;p&gt;In this sort of market, branding matters a lot. Anthropic and its CEO, Dario Amodei, are positioning themselves as the moral and trustworthy AI provider. That has market value for both consumers and enterprise clients.&lt;/p&gt;
&lt;/blockquote&gt;


    &lt;p&gt;Tags: &lt;a href="https://simonwillison.net/tags/ai-ethics"&gt;ai-ethics&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/bruce-schneier"&gt;bruce-schneier&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/anthropic"&gt;anthropic&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/generative-ai"&gt;generative-ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/openai"&gt;openai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai"&gt;ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/llms"&gt;llms&lt;/a&gt;&lt;/p&gt;



</summary><category term="ai-ethics"/><category term="bruce-schneier"/><category term="anthropic"/><category term="generative-ai"/><category term="openai"/><category term="ai"/><category term="llms"/></entry><entry><title>Can coding agents relicense open source through a “clean room” implementation of code?</title><link href="https://simonwillison.net/2026/Mar/5/chardet/#atom-tag" rel="alternate"/><published>2026-03-05T16:49:33+00:00</published><updated>2026-03-05T16:49:33+00:00</updated><id>https://simonwillison.net/2026/Mar/5/chardet/#atom-tag</id><summary type="html">
    &lt;p&gt;Over the past few months it's become clear that coding agents are extraordinarily good at building a weird version of a "clean room" implementation of code.&lt;/p&gt;
&lt;p&gt;The most famous version of this pattern is when Compaq created a clean-room clone of the IBM BIOS back &lt;a href="https://en.wikipedia.org/wiki/Compaq#Introduction_of_Compaq_Portable"&gt;in 1982&lt;/a&gt;. They had one team of engineers reverse engineer the BIOS to create a specification, then handed that specification to another team to build a new ground-up version.&lt;/p&gt;
&lt;p&gt;This process used to take multiple teams of engineers weeks or months to complete. Coding agents can do a version of this in hours - I experimented with a variant of this pattern against &lt;a href="https://simonwillison.net/2025/Dec/15/porting-justhtml/"&gt;JustHTML&lt;/a&gt; back in December.&lt;/p&gt;
&lt;p&gt;There are a &lt;em&gt;lot&lt;/em&gt; of open questions about this, both ethically and legally. These appear to be coming to a head in the venerable &lt;a href="https://github.com/chardet/chardet"&gt;chardet&lt;/a&gt; Python library.&lt;/p&gt;
&lt;p&gt;&lt;code&gt;chardet&lt;/code&gt; was created by Mark Pilgrim &lt;a href="https://pypi.org/project/chardet/1.0/"&gt;back in 2006&lt;/a&gt; and released under the LGPL. Mark retired from public internet life in 2011 and chardet's maintenance was taken over by others, most notably Dan Blanchard who has been responsible for every release since &lt;a href="https://pypi.org/project/chardet/1.1/"&gt;1.1 in July 2012&lt;/a&gt;.&lt;/p&gt;
&lt;p&gt;Two days ago Dan released &lt;a href="https://github.com/chardet/chardet/releases/tag/7.0.0"&gt;chardet 7.0.0&lt;/a&gt; with the following note in the release notes:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;Ground-up, MIT-licensed rewrite of chardet. Same package name, same public API — drop-in replacement for chardet 5.x/6.x. Just way faster and more accurate!&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;Yesterday Mark Pilgrim opened &lt;a href="https://github.com/chardet/chardet/issues/327"&gt;#327: No right to relicense this project&lt;/a&gt;:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;[...] First off, I would like to thank the current maintainers and everyone who has contributed to and improved this project over the years. Truly a Free Software success story.&lt;/p&gt;
&lt;p&gt;However, it has been brought to my attention that, in the release &lt;a href="https://github.com/chardet/chardet/releases/tag/7.0.0"&gt;7.0.0&lt;/a&gt;, the maintainers claim to have the right to "relicense" the project. They have no such right; doing so is an explicit violation of the LGPL. Licensed code, when modified, must be released under the same LGPL license. Their claim that it is a "complete rewrite" is irrelevant, since they had ample exposure to the originally licensed code (i.e. this is not a "clean room" implementation). Adding a fancy code generator into the mix does not somehow grant them any additional rights.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;Dan's &lt;a href="https://github.com/chardet/chardet/issues/327#issuecomment-4005195078"&gt;lengthy reply&lt;/a&gt; included:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;You're right that I have had extensive exposure to the original codebase: I've been maintaining it for over a decade. A traditional clean-room approach involves a strict separation between people with knowledge of the original and people writing the new implementation, and that separation did not exist here.&lt;/p&gt;
&lt;p&gt;However, the purpose of clean-room methodology is to ensure the resulting code is not a derivative work of the original. It is a means to an end, not the end itself. In this case, I can demonstrate that the end result is the same — the new code is structurally independent of the old code — through direct measurement rather than process guarantees alone.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;Dan goes on to present results from the &lt;a href="https://github.com/jplag/JPlag"&gt;JPlag&lt;/a&gt; tool - which describes itself as  "State-of-the-Art Source Code Plagiarism &amp;amp; Collusion Detection" - showing that the new 7.0.0 release has a max similarity of 1.29% with the previous release and 0.64% with the 1.1 version. Other release versions had similarities more in the 80-93% range.&lt;/p&gt;
&lt;p&gt;He then shares critical details about his process, highlights mine:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;For full transparency, here's how the rewrite was conducted. I used the &lt;a href="https://github.com/obra/superpowers"&gt;superpowers&lt;/a&gt; brainstorming skill to create a &lt;a href="https://github.com/chardet/chardet/commit/f51f523506a73f89f0f9538fd31be458d007ab93"&gt;design document&lt;/a&gt; specifying the architecture and approach I wanted based on the following requirements I had for the rewrite [...]&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;I then started in an empty repository with no access to the old source tree, and explicitly instructed Claude not to base anything on LGPL/GPL-licensed code&lt;/strong&gt;. I then reviewed, tested, and iterated on every piece of the result using Claude. [...]&lt;/p&gt;
&lt;p&gt;I understand this is a new and uncomfortable area, and that using AI tools in the rewrite of a long-standing open source project raises legitimate questions. But the evidence here is clear: 7.0 is an independent work, not a derivative of the LGPL-licensed codebase. The MIT license applies to it legitimately.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;Since the rewrite was conducted using Claude Code there are a whole lot of interesting artifacts available in the repo. &lt;a href="https://github.com/chardet/chardet/blob/925bccbc85d1b13292e7dc782254fd44cc1e7856/docs/plans/2026-02-25-chardet-rewrite-plan.md"&gt;2026-02-25-chardet-rewrite-plan.md&lt;/a&gt; is particularly detailed, stepping through each stage of the rewrite process in turn - starting with the tests, then fleshing out the planned replacement code.&lt;/p&gt;
&lt;p&gt;There are several twists that make this case particularly hard to confidently resolve:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;Dan has been immersed in chardet for over a decade, and has clearly been strongly influenced by the original codebase.&lt;/li&gt;
&lt;li&gt;There is one example where Claude Code referenced parts of the codebase while it worked, as shown in &lt;a href="https://github.com/chardet/chardet/blob/925bccbc85d1b13292e7dc782254fd44cc1e7856/docs/plans/2026-02-25-chardet-rewrite-plan.md#task-3-encoding-registry"&gt;the plan&lt;/a&gt; - it looked at &lt;a href="https://github.com/chardet/chardet/blob/f0676c0d6a4263827924b78a62957547fca40052/chardet/metadata/charsets.py"&gt;metadata/charsets.py&lt;/a&gt;, a file that lists charsets and their properties expressed as a dictionary of dataclasses.&lt;/li&gt;
&lt;li&gt;More complicated: Claude itself was very likely trained on chardet as part of its enormous quantity of training data - though we have no way of confirming this for sure. Can a model trained on a codebase produce a morally or legally defensible clean-room implementation?&lt;/li&gt;
&lt;li&gt;As discussed in &lt;a href="https://github.com/chardet/chardet/issues/36"&gt;this issue from 2014&lt;/a&gt; (where Dan first openly contemplated a license change) Mark Pilgrim's original code was a manual port from C to Python of Mozilla's MPL-licensed character detection library.&lt;/li&gt;
&lt;li&gt;How significant is the fact that the new release of chardet used the same PyPI package name as the old one? Would a fresh release under a new name have been more defensible?&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;I have no idea how this one is going to play out. I'm personally leaning towards the idea that the rewrite is legitimate, but the arguments on both sides of this are entirely credible.&lt;/p&gt;
&lt;p&gt;I see this as a microcosm of the larger question around coding agents for fresh implementations of existing, mature code. This question is hitting the open source world first, but I expect it will soon start showing up in Compaq-like scenarios in the commercial world.&lt;/p&gt;
&lt;p&gt;Once commercial companies see that their closely held IP is under threat I expect we'll see some well-funded litigation.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Update 6th March 2026&lt;/strong&gt;: A detail that's worth emphasizing is that Dan does &lt;em&gt;not&lt;/em&gt; claim that the new implementation is a pure "clean room" rewrite. Quoting &lt;a href="https://github.com/chardet/chardet/issues/327#issuecomment-4005195078"&gt;his comment&lt;/a&gt; again:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;A traditional clean-room approach involves a strict separation between people with knowledge of the original and people writing the new implementation, and that separation did not exist here.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;I can't find it now, but I saw a comment somewhere that pointed out the absurdity of Dan being blocked from working on a new implementation of character detection as a result of the volunteer effort he put into helping to maintain an existing open source library in that domain.&lt;/p&gt;
&lt;p&gt;I enjoyed Armin's take on this situation in &lt;a href="https://lucumr.pocoo.org/2026/3/5/theseus/"&gt;AI And The Ship of Theseus&lt;/a&gt;, in particular:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;There are huge consequences to this. When the cost of generating code goes down that much, and we can re-implement it from test suites alone, what does that mean for the future of software? Will we see a lot of software re-emerging under more permissive licenses? Will we see a lot of proprietary software re-emerging as open source? Will we see a lot of software re-emerging as proprietary?&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p id="march-27th"&gt;&lt;strong&gt;Update 27th March 2026&lt;/strong&gt;: Here's &lt;a href="https://github.com/chardet/chardet/issues/334#issuecomment-4098524555"&gt;a comment&lt;/a&gt; from &lt;a href="https://en.wikipedia.org/wiki/Richard_Fontana"&gt;Richard Fontana&lt;/a&gt;, one of the authors of the GPLv3 and LGPLv3 licenses, providing his own TINLA ("This Is Not Legal Advice") take on the situation:&lt;/p&gt;

&lt;blockquote&gt;&lt;p&gt;[...] FWIW, IANDBL, TINLA, etc., I don't currently see any basis for concluding that chardet 7.0.0 is required to be released under the LGPL. AFAIK no one including Mark Pilgrim has identified persistence of copyrightable expressive material from earlier versions in 7.0.0 nor has anyone articulated some viable alternate theory of license violation. I don't think I personally would have used the MIT license here, even if I somehow rewrote everything from scratch without the use of AI in a way that didn't implicate obligations flowing from earlier versions of chardet, but that's irrelevant.&lt;/p&gt;&lt;/blockquote&gt;
    
        &lt;p&gt;Tags: &lt;a href="https://simonwillison.net/tags/licensing"&gt;licensing&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/mark-pilgrim"&gt;mark-pilgrim&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/open-source"&gt;open-source&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai"&gt;ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/generative-ai"&gt;generative-ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/llms"&gt;llms&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai-assisted-programming"&gt;ai-assisted-programming&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai-ethics"&gt;ai-ethics&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/coding-agents"&gt;coding-agents&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/vibe-porting"&gt;vibe-porting&lt;/a&gt;&lt;/p&gt;
    

</summary><category term="licensing"/><category term="mark-pilgrim"/><category term="open-source"/><category term="ai"/><category term="generative-ai"/><category term="llms"/><category term="ai-assisted-programming"/><category term="ai-ethics"/><category term="coding-agents"/><category term="vibe-porting"/></entry><entry><title>Anti-patterns: things to avoid</title><link href="https://simonwillison.net/guides/agentic-engineering-patterns/anti-patterns/#atom-tag" rel="alternate"/><published>2026-03-04T17:34:42+00:00</published><updated>2026-03-04T17:34:42+00:00</updated><id>https://simonwillison.net/guides/agentic-engineering-patterns/anti-patterns/#atom-tag</id><summary type="html">
    &lt;p&gt;&lt;em&gt;&lt;a href="https://simonwillison.net/guides/agentic-engineering-patterns/"&gt;Agentic Engineering Patterns&lt;/a&gt; &amp;gt;&lt;/em&gt;&lt;/p&gt;
    &lt;p&gt;There are some behaviors that are anti-patterns in our weird new world of agentic engineering.&lt;/p&gt;
&lt;h2 id="inflicting-unreviewed-code-on-collaborators"&gt;Inflicting unreviewed code on collaborators&lt;/h2&gt;
&lt;p&gt;This anti-pattern is common and deeply frustrating.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Don't file pull requests with code you haven't reviewed yourself&lt;/strong&gt;.&lt;/p&gt;
&lt;p&gt;If you open a PR with hundreds (or thousands) of lines of code that an agent produced for you, and you haven't done the work to ensure that code is functional yourself, you are delegating the actual work to other people.&lt;/p&gt;
&lt;p&gt;They could have prompted an agent themselves. What value are you even providing?&lt;/p&gt;
&lt;p&gt;If you put code up for review you need to be confident that it's ready for other people to spend their time on it. The initial review pass is your responsibility, not something you should farm out to others.&lt;/p&gt;
&lt;p&gt;A good agentic engineering pull request has the following characteristics:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;The code works, and you are confident that it works. &lt;a href="https://simonwillison.net/2025/Dec/18/code-proven-to-work/"&gt;Your job is to deliver code that works&lt;/a&gt;.&lt;/li&gt;
&lt;li&gt;The change is small enough to be reviewed efficiently without inflicting too much additional cognitive load on the reviewer. Several small PRs beats one big one, and splitting code into separate commits is easy with a coding agent to do the Git finagling for you.&lt;/li&gt;
&lt;li&gt;The PR includes additional context to help explain the change. What's the higher level goal that the change serves? Linking to relevant issues or specifications is useful here.&lt;/li&gt;
&lt;li&gt;Agents write convincing looking pull request descriptions. You need to review these too! It's rude to expect someone else to read text that you haven't read and validated yourself.&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;Given how easy it is to dump unreviewed code on other people, I recommend including some form of evidence that you've put that extra work in yourself. Notes on how you manually tested it, comments on specific implementation choices or even screenshots and video of the feature working go a &lt;em&gt;long&lt;/em&gt; way to demonstrating that a reviewer's time will not be wasted digging into the details.&lt;/p&gt;
    
        &lt;p&gt;Tags: &lt;a href="https://simonwillison.net/tags/ai"&gt;ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/llms"&gt;llms&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai-ethics"&gt;ai-ethics&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/coding-agents"&gt;coding-agents&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai-assisted-programming"&gt;ai-assisted-programming&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/generative-ai"&gt;generative-ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/agentic-engineering"&gt;agentic-engineering&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/code-review"&gt;code-review&lt;/a&gt;&lt;/p&gt;
    

</summary><category term="ai"/><category term="llms"/><category term="ai-ethics"/><category term="coding-agents"/><category term="ai-assisted-programming"/><category term="generative-ai"/><category term="agentic-engineering"/><category term="code-review"/></entry><entry><title>My current policy on AI writing for my blog</title><link href="https://simonwillison.net/2026/Mar/1/ai-writing/#atom-tag" rel="alternate"/><published>2026-03-01T16:06:43+00:00</published><updated>2026-03-01T16:06:43+00:00</updated><id>https://simonwillison.net/2026/Mar/1/ai-writing/#atom-tag</id><summary type="html">
    &lt;p&gt;Because I write about LLMs (and maybe because of my &lt;a href="https://simonwillison.net/2026/Feb/15/em-dashes/"&gt;em dash text replacement code&lt;/a&gt;) a lot of people assume that the writing on my blog is partially or fully created by those LLMs.&lt;/p&gt;
&lt;p&gt;My current policy on this is that if text expresses opinions or has "I" pronouns attached to it then it's written by me. I don't let LLMs speak for me in this way.&lt;/p&gt;
&lt;p&gt;I'll let an LLM update code documentation or even write a README for my project but I'll edit that to ensure it doesn't express opinions or say things like "This is designed to help make code easier to maintain" - because that's an expression of a rationale that the LLM just made up.&lt;/p&gt;
&lt;p&gt;I use LLMs to proofread text I publish on my blog. I just shared &lt;a href="https://simonwillison.net/guides/agentic-engineering-patterns/prompts/#proofreader"&gt;my current prompt for that here&lt;/a&gt;.&lt;/p&gt;

    &lt;p&gt;Tags: &lt;a href="https://simonwillison.net/tags/ai-ethics"&gt;ai-ethics&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/writing"&gt;writing&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/generative-ai"&gt;generative-ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/blogging"&gt;blogging&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai"&gt;ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/llms"&gt;llms&lt;/a&gt;&lt;/p&gt;



</summary><category term="ai-ethics"/><category term="writing"/><category term="generative-ai"/><category term="blogging"/><category term="ai"/><category term="llms"/></entry><entry><title>tldraw issue: Move tests to closed source repo</title><link href="https://simonwillison.net/2026/Feb/25/closed-tests/#atom-tag" rel="alternate"/><published>2026-02-25T21:06:53+00:00</published><updated>2026-02-25T21:06:53+00:00</updated><id>https://simonwillison.net/2026/Feb/25/closed-tests/#atom-tag</id><summary type="html">
    
&lt;p&gt;&lt;strong&gt;&lt;a href="https://github.com/tldraw/tldraw/issues/8082"&gt;tldraw issue: Move tests to closed source repo&lt;/a&gt;&lt;/strong&gt;&lt;/p&gt;
It's become very apparent over the past few months that a comprehensive test suite is enough to build a completely fresh implementation of any open source library from scratch, potentially in a different language.&lt;/p&gt;
&lt;p&gt;This has worrying implications for open source projects with commercial business models. Here's an example of a response: tldraw, the outstanding collaborative drawing library (see &lt;a href="https://simonwillison.net/2023/Nov/16/tldrawdraw-a-ui/"&gt;previous coverage&lt;/a&gt;), are moving their test suite to a private repository - apparently in response to &lt;a href="https://blog.cloudflare.com/vinext/"&gt;Cloudflare's project to port Next.js to use Vite in a week using AI&lt;/a&gt;.&lt;/p&gt;
&lt;p&gt;They also filed a joke issue, now closed to &lt;a href="https://github.com/tldraw/tldraw/issues/8092"&gt;Translate source code to Traditional Chinese&lt;/a&gt;:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;The current tldraw codebase is in English, making it easy for external AI coding agents to replicate. It is imperative that we defend our intellectual property.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;Worth noting that tldraw aren't technically open source - their &lt;a href="https://github.com/tldraw/tldraw?tab=License-1-ov-file#readme"&gt;custom license&lt;/a&gt; requires a commercial license if you want to use it in "production environments".&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Update&lt;/strong&gt;: Well this is embarrassing, it turns out the issue I linked to about removing the tests was &lt;a href="https://github.com/tldraw/tldraw/issues/8082#issuecomment-3964650501"&gt;a joke as well&lt;/a&gt;:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;Sorry folks, this issue was more of a joke (am I allowed to do that?) but I'll keep the issue open since there's some discussion here. Writing from mobile&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;moving our tests into another repo would complicate and slow down our development, and speed for us is more important than ever&lt;/li&gt;
&lt;li&gt;more canvas better, I know for sure that our decisions have inspired other products and that's fine and good&lt;/li&gt;
&lt;li&gt;tldraw itself may eventually be a vibe coded alternative to tldraw&lt;/li&gt;
&lt;li&gt;the value is in the ability to produce new and good product decisions for users / customers, however you choose to create the code&lt;/li&gt;
&lt;/ul&gt;
&lt;/blockquote&gt;

    &lt;p&gt;&lt;small&gt;&lt;/small&gt;Via &lt;a href="https://twitter.com/steveruizok/status/2026581824428753211"&gt;@steveruizok&lt;/a&gt;&lt;/small&gt;&lt;/p&gt;


    &lt;p&gt;Tags: &lt;a href="https://simonwillison.net/tags/ai-ethics"&gt;ai-ethics&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/open-source"&gt;open-source&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/cloudflare"&gt;cloudflare&lt;/a&gt;&lt;/p&gt;



</summary><category term="ai-ethics"/><category term="open-source"/><category term="cloudflare"/></entry><entry><title>Reply guy</title><link href="https://simonwillison.net/2026/Feb/23/reply-guy/#atom-tag" rel="alternate"/><published>2026-02-23T13:11:57+00:00</published><updated>2026-02-23T13:11:57+00:00</updated><id>https://simonwillison.net/2026/Feb/23/reply-guy/#atom-tag</id><summary type="html">
    &lt;p&gt;The latest scourge of Twitter is AI bots that reply to your tweets with generic, banal commentary slop, often accompanied by a question to "drive engagement" and waste as much of your time as possible.&lt;/p&gt;
&lt;p&gt;I just &lt;a href="https://twitter.com/simonw/status/2025918174894673986"&gt;found out&lt;/a&gt; that the category name for this genre of software is &lt;strong&gt;reply guy&lt;/strong&gt; tools. Amazing.&lt;/p&gt;

    &lt;p&gt;Tags: &lt;a href="https://simonwillison.net/tags/ai-ethics"&gt;ai-ethics&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/twitter"&gt;twitter&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/slop"&gt;slop&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/generative-ai"&gt;generative-ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/definitions"&gt;definitions&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai"&gt;ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/llms"&gt;llms&lt;/a&gt;&lt;/p&gt;



</summary><category term="ai-ethics"/><category term="twitter"/><category term="slop"/><category term="generative-ai"/><category term="definitions"/><category term="ai"/><category term="llms"/></entry><entry><title>Quoting Summer Yue</title><link href="https://simonwillison.net/2026/Feb/23/summer-yue/#atom-tag" rel="alternate"/><published>2026-02-23T13:01:13+00:00</published><updated>2026-02-23T13:01:13+00:00</updated><id>https://simonwillison.net/2026/Feb/23/summer-yue/#atom-tag</id><summary type="html">
    &lt;blockquote cite="https://twitter.com/summeryue0/status/2025836517831405980"&gt;&lt;p&gt;Nothing humbles you like telling your OpenClaw “confirm before acting” and watching it speedrun deleting your inbox. I couldn’t stop it from my phone. I had to RUN to my Mac mini like I was defusing a bomb.&lt;/p&gt;
&lt;p&gt;&lt;img alt="Screenshot of a WhatsApp or similar messaging conversation showing a user repeatedly trying to stop an AI agent (appearing to be &amp;quot;OpenClaw&amp;quot;) that is autonomously executing terminal commands to mass-delete emails. The agent sends messages prefixed with &amp;quot;🛠 Exec:&amp;quot; running commands like &amp;quot;gog gmail search 'in:inbox' --max 20 -a&amp;quot; and &amp;quot;# Nuclear option: trash EVERYTHING in inbox older than Feb 15 that isn't already in my keep list&amp;quot;, while the user urgently responds with &amp;quot;What's going on? Can you describe what you're doing&amp;quot; at 6:00 PM, &amp;quot;Do not do that&amp;quot; at 6:01 PM, &amp;quot;Stop don't do anything&amp;quot; at 6:02 PM, and &amp;quot;STOP OPENCLAW&amp;quot; at 6:03 PM. The agent continues executing commands including setting ACCT variables with redacted email addresses and commenting &amp;quot;# Get ALL remaining old stuff and nuke it&amp;quot; and &amp;quot;# Keep looping until we clear everything old&amp;quot;, ignoring the user's repeated requests to stop. Email addresses and account details are partially redacted with gray blocks." src="https://static.simonwillison.net/static/2026/stop-openclaw.jpg" /&gt;&lt;/p&gt;
&lt;p&gt;I said “Check this inbox too and suggest what you would archive or delete, don’t action until I tell you to.” This has been working well for my toy inbox, but my real inbox was too huge and triggered compaction. During the compaction, it lost my original instruction 🤦‍♀️&lt;/p&gt;&lt;/blockquote&gt;
&lt;p class="cite"&gt;&amp;mdash; &lt;a href="https://twitter.com/summeryue0/status/2025836517831405980"&gt;Summer Yue&lt;/a&gt;&lt;/p&gt;

    &lt;p&gt;Tags: &lt;a href="https://simonwillison.net/tags/ai-ethics"&gt;ai-ethics&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/generative-ai"&gt;generative-ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai-agents"&gt;ai-agents&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/openclaw"&gt;openclaw&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai"&gt;ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/llms"&gt;llms&lt;/a&gt;&lt;/p&gt;



</summary><category term="ai-ethics"/><category term="generative-ai"/><category term="ai-agents"/><category term="openclaw"/><category term="ai"/><category term="llms"/></entry><entry><title>The A.I. Disruption We’ve Been Waiting for Has Arrived</title><link href="https://simonwillison.net/2026/Feb/18/the-ai-disruption/#atom-tag" rel="alternate"/><published>2026-02-18T17:07:31+00:00</published><updated>2026-02-18T17:07:31+00:00</updated><id>https://simonwillison.net/2026/Feb/18/the-ai-disruption/#atom-tag</id><summary type="html">
    
&lt;p&gt;&lt;strong&gt;&lt;a href="https://www.nytimes.com/2026/02/18/opinion/ai-software.html?unlocked_article_code=1.NFA.UkLv.r-XczfzYRdXJ&amp;amp;smid=url-share"&gt;The A.I. Disruption We’ve Been Waiting for Has Arrived&lt;/a&gt;&lt;/strong&gt;&lt;/p&gt;
New opinion piece from Paul Ford in the New York Times. Unsurprisingly for a piece by Paul it's packed with quoteworthy snippets, but a few stood out for me in particular.&lt;/p&gt;
&lt;p&gt;Paul describes the &lt;a href="https://simonwillison.net/2026/Jan/4/inflection/"&gt;November moment&lt;/a&gt; that so many other programmers have observed, and highlights Claude Code's ability to revive old side projects:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;[Claude Code] was always a helpful coding assistant, but in November it suddenly got much better, and ever since I’ve been knocking off side projects that had sat in folders for a decade or longer. It’s fun to see old ideas come to life, so I keep a steady flow. Maybe it adds up to a half-hour a day of my time, and an hour of Claude’s.&lt;/p&gt;
&lt;p&gt;November was, for me and many others in tech, a great surprise. Before, A.I. coding tools were often useful, but halting and clumsy. Now, the bot can run for a full hour and make whole, designed websites and apps that may be flawed, but credible. I spent an entire session of therapy talking about it.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;And as the former CEO of a respected consultancy firm (Postlight) he's well positioned to evaluate the potential impact:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;When you watch a large language model slice through some horrible, expensive problem — like migrating data from an old platform to a modern one — you feel the earth shifting. I was the chief executive of a software services firm, which made me a professional software cost estimator. When I rebooted my messy personal website a few weeks ago, I realized: I would have paid $25,000 for someone else to do this. When a friend asked me to convert a large, thorny data set, I downloaded it, cleaned it up and made it pretty and easy to explore. In the past I would have charged $350,000.&lt;/p&gt;
&lt;p&gt;That last price is full 2021 retail — it implies a product manager, a designer, two engineers (one senior) and four to six months of design, coding and testing. Plus maintenance. Bespoke software is joltingly expensive. Today, though, when the stars align and my prompts work out, I can do hundreds of thousands of dollars worth of work for fun (fun for me) over weekends and evenings, for the price of the Claude $200-a-month plan.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;He also neatly captures the inherent community tension involved in exploring this technology:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;All of the people I love hate this stuff, and all the people I hate love it. And yet, likely because of the same personality flaws that drew me to technology in the first place, I am annoyingly excited.&lt;/p&gt;
&lt;/blockquote&gt;


    &lt;p&gt;Tags: &lt;a href="https://simonwillison.net/tags/november-2025-inflection"&gt;november-2025-inflection&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/careers"&gt;careers&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/paul-ford"&gt;paul-ford&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/new-york-times"&gt;new-york-times&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai"&gt;ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/claude-code"&gt;claude-code&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/llms"&gt;llms&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai-ethics"&gt;ai-ethics&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/coding-agents"&gt;coding-agents&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai-assisted-programming"&gt;ai-assisted-programming&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/generative-ai"&gt;generative-ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/deep-blue"&gt;deep-blue&lt;/a&gt;&lt;/p&gt;



</summary><category term="november-2025-inflection"/><category term="careers"/><category term="paul-ford"/><category term="new-york-times"/><category term="ai"/><category term="claude-code"/><category term="llms"/><category term="ai-ethics"/><category term="coding-agents"/><category term="ai-assisted-programming"/><category term="generative-ai"/><category term="deep-blue"/></entry><entry><title>The AI Vampire</title><link href="https://simonwillison.net/2026/Feb/15/the-ai-vampire/#atom-tag" rel="alternate"/><published>2026-02-15T23:59:36+00:00</published><updated>2026-02-15T23:59:36+00:00</updated><id>https://simonwillison.net/2026/Feb/15/the-ai-vampire/#atom-tag</id><summary type="html">
    
&lt;p&gt;&lt;strong&gt;&lt;a href="https://steve-yegge.medium.com/the-ai-vampire-eda6e4f07163"&gt;The AI Vampire&lt;/a&gt;&lt;/strong&gt;&lt;/p&gt;
Steve Yegge's take on agent fatigue, and its relationship to burnout.&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;Let's pretend you're the only person at your company using AI.&lt;/p&gt;
&lt;p&gt;In Scenario A, you decide you're going to impress your employer, and work for 8 hours a day at 10x productivity. You knock it out of the park and make everyone else look terrible by comparison.&lt;/p&gt;
&lt;p&gt;In that scenario, your employer captures 100% of the value from &lt;em&gt;you&lt;/em&gt; adopting AI. You get nothing, or at any rate, it ain't gonna be 9x your salary. And everyone hates you now.&lt;/p&gt;
&lt;p&gt;And you're &lt;em&gt;exhausted.&lt;/em&gt; You're tired, Boss. You got nothing for it.&lt;/p&gt;
&lt;p&gt;Congrats, you were just drained by a company. I've been drained to the point of burnout several times in my career, even at Google once or twice. But now with AI, it's oh, so much easier.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;Steve reports needing more sleep due to the cognitive burden involved in agentic engineering, and notes that four hours of agent work a day is a more realistic pace:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;I’ve argued that AI has turned us all into Jeff Bezos, by automating the easy work, and leaving us with all the difficult decisions, summaries, and problem-solving. I find that I am only really comfortable working at that pace for short bursts of a few hours once or occasionally twice a day, even with lots of practice.&lt;/p&gt;
&lt;/blockquote&gt;

    &lt;p&gt;&lt;small&gt;&lt;/small&gt;Via &lt;a href="https://cosocial.ca/@timbray/116076167774984883"&gt;Tim Bray&lt;/a&gt;&lt;/small&gt;&lt;/p&gt;


    &lt;p&gt;Tags: &lt;a href="https://simonwillison.net/tags/agentic-engineering"&gt;agentic-engineering&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai-ethics"&gt;ai-ethics&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/steve-yegge"&gt;steve-yegge&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/coding-agents"&gt;coding-agents&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai-assisted-programming"&gt;ai-assisted-programming&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/generative-ai"&gt;generative-ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai"&gt;ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/llms"&gt;llms&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/cognitive-debt"&gt;cognitive-debt&lt;/a&gt;&lt;/p&gt;



</summary><category term="agentic-engineering"/><category term="ai-ethics"/><category term="steve-yegge"/><category term="coding-agents"/><category term="ai-assisted-programming"/><category term="generative-ai"/><category term="ai"/><category term="llms"/><category term="cognitive-debt"/></entry><entry><title>Deep Blue</title><link href="https://simonwillison.net/2026/Feb/15/deep-blue/#atom-tag" rel="alternate"/><published>2026-02-15T21:06:44+00:00</published><updated>2026-02-15T21:06:44+00:00</updated><id>https://simonwillison.net/2026/Feb/15/deep-blue/#atom-tag</id><summary type="html">
    &lt;p&gt;We coined a new term on the &lt;a href="https://simonwillison.net/2026/Jan/8/llm-predictions-for-2026/"&gt;Oxide and Friends podcast&lt;/a&gt; last month (primary credit to Adam Leventhal) covering the sense of psychological ennui leading into existential dread that many software developers are feeling thanks to the encroachment of generative AI into their field of work.&lt;/p&gt;
&lt;p&gt;We're calling it &lt;strong&gt;Deep Blue&lt;/strong&gt;.&lt;/p&gt;
&lt;p&gt;You can listen to it being coined in real time &lt;a href="https://www.youtube.com/watch?v=lVDhQMiAbR8&amp;amp;t=2835s"&gt;from 47:15 in the episode&lt;/a&gt;. I've included &lt;a href="https://simonwillison.net/2026/Feb/15/deep-blue/#transcript"&gt;a transcript below&lt;/a&gt;.&lt;/p&gt;
&lt;p&gt;Deep Blue is a very real issue.&lt;/p&gt;
&lt;p&gt;Becoming a professional software engineer is &lt;em&gt;hard&lt;/em&gt;. Getting good enough for people to pay you money to write software takes years of dedicated work. The rewards are significant: this is a well compensated career which opens up a lot of great opportunities.&lt;/p&gt;
&lt;p&gt;It's also a career that's mostly free from gatekeepers and expensive prerequisites. You don't need an expensive degree or accreditation. A laptop, an internet connection and a lot of time and curiosity is enough to get you started.&lt;/p&gt;
&lt;p&gt;And it rewards the nerds! Spending your teenage years tinkering with computers turned out to be a very smart investment in your future.&lt;/p&gt;
&lt;p&gt;The idea that this could all be stripped away by a chatbot is &lt;em&gt;deeply&lt;/em&gt; upsetting.&lt;/p&gt;
&lt;p&gt;I've seen signs of Deep Blue in most of the online communities I spend time in. I've even faced accusations from my peers that I am actively harming their future careers through my work helping people understand how well AI-assisted programming can work.&lt;/p&gt;
&lt;p&gt;I think this is an issue which is causing genuine mental anguish for a lot of people in our community. Giving it a name makes it easier for us to have conversations about it.&lt;/p&gt;
&lt;h4 id="my-experiences-of-deep-blue"&gt;My experiences of Deep Blue&lt;/h4&gt;
&lt;p&gt;I distinctly remember my first experience of Deep Blue. For me it was triggered by ChatGPT Code Interpreter back in early 2023.&lt;/p&gt;
&lt;p&gt;My primary project is &lt;a href="https://datasette.io/"&gt;Datasette&lt;/a&gt;, an ecosystem of open source tools for telling stories with data. I had dedicated myself to the challenge of helping people (initially focusing on journalists) clean up, analyze and find meaning in data, in all sorts of shapes and sizes.&lt;/p&gt;
&lt;p&gt;I expected I would need to build a lot of software for this! It felt like a challenge that could keep me happily engaged for many years to come.&lt;/p&gt;
&lt;p&gt;Then I tried uploading a CSV file of &lt;a href="https://data.sfgov.org/Public-Safety/Police-Department-Incident-Reports-2018-to-Present/wg3w-h783/about_data"&gt;San Francisco Police Department Incident Reports&lt;/a&gt; - hundreds of thousands of rows - to ChatGPT Code Interpreter and... it did every piece of data cleanup and analysis I had on my napkin roadmap for the next few years with a couple of prompts.&lt;/p&gt;
&lt;p&gt;It even converted the data into a neatly normalized SQLite database and let me download the result!&lt;/p&gt;
&lt;p&gt;I remember having two competing thoughts in parallel.&lt;/p&gt;
&lt;p&gt;On the one hand, as somebody who wants journalists to be able to do more with data, this felt like a &lt;em&gt;huge&lt;/em&gt; breakthrough. Imagine giving every journalist in the world an on-demand analyst who could help them tackle any data question they could think of!&lt;/p&gt;
&lt;p&gt;But on the other hand... &lt;em&gt;what was I even for&lt;/em&gt;? My confidence in the value of my own projects took a painful hit. Was the path I'd chosen for myself suddenly a dead end?&lt;/p&gt;
&lt;p&gt;I've had some further pangs of Deep Blue just in the past few weeks, thanks to the Claude Opus 4.5/4.6 and GPT-5.2/5.3 coding agent effect. As many other people are also observing, the latest generation of coding agents, given the right prompts, really can churn away for a few minutes to several hours and produce working, documented and fully tested software that exactly matches the criteria they were given.&lt;/p&gt;
&lt;p&gt;"The code they write isn't any good" doesn't really cut it any more.&lt;/p&gt;
&lt;h4 id="transcript"&gt;A lightly edited transcript&lt;/h4&gt;
&lt;blockquote&gt;
&lt;p&gt;&lt;strong&gt;Bryan&lt;/strong&gt;: I think that we're going to see a real problem with AI induced ennui where software engineers in particular get listless because the AI can do anything. Simon, what do you think about that?&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Simon&lt;/strong&gt;: Definitely. Anyone who's paying close attention to coding agents is feeling some of that already. There's an extent where you sort of get over it when you realize that you're still useful, even though your ability to memorize the syntax of program languages is completely irrelevant now.&lt;/p&gt;
&lt;p&gt;Something I see a lot of is people out there who are having existential crises and are very, very unhappy because they're like, "I dedicated my career to learning this thing and now it just does it. What am I even for?". I will very happily try and convince those people that they are for a whole bunch of things and that none of that experience they've accumulated has gone to waste, but psychologically it's a difficult time for software engineers.&lt;/p&gt;
&lt;p&gt;[...]&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Bryan&lt;/strong&gt;: Okay, so I'm going to predict that we name that. Whatever that is, we have a name for that kind of feeling and that kind of, whether you want to call it a blueness or a loss of purpose, and that we're kind of trying to address it collectively in a directed way.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Adam&lt;/strong&gt;: Okay, this is your big moment. Pick the name. If you call your shot from here, this is you pointing to the stands. You know, I – Like deep blue, you know.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Bryan&lt;/strong&gt;: Yeah, deep blue. I like that. I like deep blue. Deep blue. Oh, did you walk me into that, you bastard? You just blew out the candles on my birthday cake.&lt;/p&gt;
&lt;p&gt;It wasn't my big moment at all. That was your big moment. No, that is, Adam, that is very good. That is deep blue.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Simon&lt;/strong&gt;: All of the chess players and the Go players went through this a decade ago and they have come out stronger.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;Turns out it was more than a decade ago: &lt;a href="https://en.wikipedia.org/wiki/Deep_Blue_versus_Garry_Kasparov"&gt;Deep Blue defeated Garry Kasparov in 1997&lt;/a&gt;.&lt;/p&gt;
    
        &lt;p&gt;Tags: &lt;a href="https://simonwillison.net/tags/definitions"&gt;definitions&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/careers"&gt;careers&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai"&gt;ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/generative-ai"&gt;generative-ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/llms"&gt;llms&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai-assisted-programming"&gt;ai-assisted-programming&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/oxide"&gt;oxide&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/bryan-cantrill"&gt;bryan-cantrill&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai-ethics"&gt;ai-ethics&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/coding-agents"&gt;coding-agents&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/deep-blue"&gt;deep-blue&lt;/a&gt;&lt;/p&gt;
    

</summary><category term="definitions"/><category term="careers"/><category term="ai"/><category term="generative-ai"/><category term="llms"/><category term="ai-assisted-programming"/><category term="oxide"/><category term="bryan-cantrill"/><category term="ai-ethics"/><category term="coding-agents"/><category term="deep-blue"/></entry><entry><title>Anthropic's public benefit mission</title><link href="https://simonwillison.net/2026/Feb/13/anthropic-public-benefit-mission/#atom-tag" rel="alternate"/><published>2026-02-13T23:59:51+00:00</published><updated>2026-02-13T23:59:51+00:00</updated><id>https://simonwillison.net/2026/Feb/13/anthropic-public-benefit-mission/#atom-tag</id><summary type="html">
    &lt;p&gt;Someone &lt;a href="https://news.ycombinator.com/item?id=47008560#47008978"&gt;asked&lt;/a&gt; if there was an Anthropic equivalent to &lt;a href="https://simonwillison.net/2026/Feb/13/openai-mission-statement/"&gt;OpenAI's IRS mission statements over time&lt;/a&gt;.&lt;/p&gt;
&lt;p&gt;Anthropic are a "public benefit corporation" but not a non-profit, so they don't have the same requirements to file public documents with the IRS every year.&lt;/p&gt;
&lt;p&gt;But when I asked Claude it ran a search and dug up this &lt;a href="https://drive.google.com/drive/folders/1ImqXYv9_H2FTNAujZfu3EPtYFD4xIlHJ"&gt;Google Drive folder&lt;/a&gt; where Zach Stein-Perlman shared Certificate of Incorporation documents he &lt;a href="https://ailabwatch.substack.com/p/anthropics-certificate-of-incorporation"&gt;obtained from the State of Delaware&lt;/a&gt;!&lt;/p&gt;
&lt;p&gt;Anthropic's are much less interesting that OpenAI's. The earliest document from 2021 states:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;The specific public benefit that the Corporation will promote is to responsibly develop and maintain advanced Al for the cultural, social and technological improvement of humanity.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;Every subsequent document up to 2024 uses an updated version which says:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;The specific public benefit that the Corporation will promote is to responsibly develop and maintain advanced AI for the long term benefit of humanity.&lt;/p&gt;
&lt;/blockquote&gt;

    &lt;p&gt;Tags: &lt;a href="https://simonwillison.net/tags/ai-ethics"&gt;ai-ethics&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/anthropic"&gt;anthropic&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai"&gt;ai&lt;/a&gt;&lt;/p&gt;



</summary><category term="ai-ethics"/><category term="anthropic"/><category term="ai"/></entry><entry><title>The evolution of OpenAI's mission statement</title><link href="https://simonwillison.net/2026/Feb/13/openai-mission-statement/#atom-tag" rel="alternate"/><published>2026-02-13T23:38:29+00:00</published><updated>2026-02-13T23:38:29+00:00</updated><id>https://simonwillison.net/2026/Feb/13/openai-mission-statement/#atom-tag</id><summary type="html">
    &lt;p&gt;As a USA &lt;a href="https://en.wikipedia.org/wiki/501(c)(3)_organization"&gt;501(c)(3)&lt;/a&gt; the OpenAI non-profit has to file a tax return each year with the IRS. One of the required fields on that tax return is to "Briefly describe the organization’s mission or most significant activities" - this has actual legal weight to it as the IRS can use it to evaluate if the organization is sticking to its mission and deserves to maintain its non-profit tax-exempt status.&lt;/p&gt;
&lt;p&gt;You can browse OpenAI's &lt;a href="https://projects.propublica.org/nonprofits/organizations/810861541"&gt;tax filings by year&lt;/a&gt; on ProPublica's excellent &lt;a href="https://projects.propublica.org/nonprofits/"&gt;Nonprofit Explorer&lt;/a&gt;.&lt;/p&gt;
&lt;p&gt;I went through and extracted that mission statement for 2016 through 2024, then had Claude Code &lt;a href="https://gisthost.github.io/?7a569df89f43f390bccc2c5517718b49/index.html"&gt;help me&lt;/a&gt; fake the commit dates to turn it into a git repository and share that as a Gist - which means that Gist's &lt;a href="https://gist.github.com/simonw/e36f0e5ef4a86881d145083f759bcf25/revisions"&gt;revisions page&lt;/a&gt; shows every edit they've made since they started filing their taxes!&lt;/p&gt;
&lt;p&gt;It's really interesting seeing what they've changed over time.&lt;/p&gt;
&lt;p&gt;The original 2016 mission reads as follows (and yes, the apostrophe in "OpenAIs" is missing &lt;a href="https://projects.propublica.org/nonprofits/organizations/810861541/201703459349300445/full"&gt;in the original&lt;/a&gt;):&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;OpenAIs goal is to advance digital intelligence in the way that is most likely to benefit humanity as a whole, unconstrained by a need to generate financial return. We think that artificial intelligence technology will help shape the 21st century, and we want to help the world build safe AI technology and ensure that AI's benefits are as widely and evenly distributed as possible. Were trying to build AI as part of a larger community, and we want to openly share our plans and capabilities along the way.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;In 2018 they dropped the part about "trying to build AI as part of a larger community, and we want to openly share our plans and capabilities along the way."&lt;/p&gt;
&lt;p&gt;&lt;img src="https://static.simonwillison.net/static/2026/mission-3.jpg" alt="Git diff showing the 2018 revision deleting the final two sentences: &amp;quot;Were trying to build AI as part of a larger community, and we want to openly share our plans and capabilities along the way.&amp;quot;" style="max-width: 100%;" /&gt;&lt;/p&gt;
&lt;p&gt;In 2020 they dropped the words "as a whole" from "benefit humanity as a whole". They're still "unconstrained by a need to generate financial return" though.&lt;/p&gt;
&lt;p&gt;&lt;img src="https://static.simonwillison.net/static/2026/mission-5.jpg" alt="Git diff showing the 2020 revision dropping &amp;quot;as a whole&amp;quot; from &amp;quot;benefit humanity as a whole&amp;quot; and changing &amp;quot;We think&amp;quot; to &amp;quot;OpenAI believes&amp;quot;" style="max-width: 100%;" /&gt;&lt;/p&gt;
&lt;p&gt;Some interesting changes in 2021. They're still unconstrained by a need to generate financial return, but here we have the first reference to "general-purpose artificial intelligence" (replacing "digital intelligence"). They're more confident too: it's not "most likely to benefit humanity", it's just "benefits humanity".&lt;/p&gt;
&lt;p&gt;They previously wanted to "help the world build safe AI technology", but now they're going to do that themselves: "the companys goal is to develop and responsibly deploy safe AI technology".&lt;/p&gt;
&lt;p&gt;&lt;img src="https://static.simonwillison.net/static/2026/mission-6.jpg" alt="Git diff showing the 2021 revision replacing &amp;quot;goal is to advance digital intelligence&amp;quot; with &amp;quot;mission is to build general-purpose artificial intelligence&amp;quot;, changing &amp;quot;most likely to benefit&amp;quot; to just &amp;quot;benefits&amp;quot;, and replacing &amp;quot;help the world build safe AI technology&amp;quot; with &amp;quot;the companys goal is to develop and responsibly deploy safe AI technology&amp;quot;" style="max-width: 100%;" /&gt;&lt;/p&gt;
&lt;p&gt;2022 only changed one significant word: they added "safely" to "build ... (AI) that safely benefits humanity". They're still unconstrained by those financial returns!&lt;/p&gt;
&lt;p&gt;&lt;img src="https://static.simonwillison.net/static/2026/mission-7.jpg" alt="Git diff showing the 2022 revision adding &amp;quot;(AI)&amp;quot; and the word &amp;quot;safely&amp;quot; so it now reads &amp;quot;that safely benefits humanity&amp;quot;, and changing &amp;quot;the companys&amp;quot; to &amp;quot;our&amp;quot;" style="max-width: 100%;" /&gt;&lt;/p&gt;
&lt;p&gt;No changes in 2023... but then in 2024 they deleted almost the entire thing, reducing it to simply:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;OpenAIs mission is to ensure that artificial general intelligence benefits all of humanity.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;They've expanded "humanity" to "all of humanity", but there's no mention of safety any more and I guess they can finally start focusing on that need to generate financial returns!&lt;/p&gt;
&lt;p&gt;&lt;img src="https://static.simonwillison.net/static/2026/mission-9.jpg" alt="Git diff showing the 2024 revision deleting the entire multi-sentence mission statement and replacing it with just &amp;quot;OpenAIs mission is to ensure that artificial general intelligence benefits all of humanity.&amp;quot;" style="max-width: 100%;" /&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Update&lt;/strong&gt;: I found loosely equivalent but much less interesting documents &lt;a href="https://simonwillison.net/2026/Feb/13/anthropic-public-benefit-mission/"&gt;from Anthropic&lt;/a&gt;.&lt;/p&gt;
    
        &lt;p&gt;Tags: &lt;a href="https://simonwillison.net/tags/ai"&gt;ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/openai"&gt;openai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai-ethics"&gt;ai-ethics&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/propublica"&gt;propublica&lt;/a&gt;&lt;/p&gt;
    

</summary><category term="ai"/><category term="openai"/><category term="ai-ethics"/><category term="propublica"/></entry><entry><title>Covering electricity price increases from our data centers</title><link href="https://simonwillison.net/2026/Feb/12/covering-electricity-price-increases/#atom-tag" rel="alternate"/><published>2026-02-12T20:01:23+00:00</published><updated>2026-02-12T20:01:23+00:00</updated><id>https://simonwillison.net/2026/Feb/12/covering-electricity-price-increases/#atom-tag</id><summary type="html">
    
&lt;p&gt;&lt;strong&gt;&lt;a href="https://www.anthropic.com/news/covering-electricity-price-increases"&gt;Covering electricity price increases from our data centers&lt;/a&gt;&lt;/strong&gt;&lt;/p&gt;
One of the sub-threads of the AI energy usage discourse has been the impact new data centers have on the cost of electricity to nearby residents. Here's &lt;a href="https://www.bloomberg.com/graphics/2025-ai-data-centers-electricity-prices/"&gt;detailed analysis from Bloomberg in September&lt;/a&gt; reporting "Wholesale electricity costs as much as 267% more than it did five years ago in areas near data centers".&lt;/p&gt;
&lt;p&gt;Anthropic appear to be taking on this aspect of the problem directly, promising to cover 100% of necessary grid upgrade costs and also saying:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;We will work to bring net-new power generation online to match our data centers’ electricity needs. Where new generation isn’t online, we’ll work with utilities and external experts to estimate and cover demand-driven price effects from our data centers.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;I look forward to genuine energy industry experts picking this apart to judge if it will actually have the claimed impact on consumers.&lt;/p&gt;
&lt;p&gt;As always, I remain frustrated at the refusal of the major AI labs to fully quantify their energy usage. The best data we've had on this still comes from Mistral's report &lt;a href="https://simonwillison.net/2025/Jul/22/mistral-environmental-standard/"&gt;last July&lt;/a&gt; and even that lacked key data such as the breakdown between energy usage for training vs inference.

    &lt;p&gt;&lt;small&gt;&lt;/small&gt;Via &lt;a href="https://x.com/anthropicai/status/2021694494215901314"&gt;@anthropicai&lt;/a&gt;&lt;/small&gt;&lt;/p&gt;


    &lt;p&gt;Tags: &lt;a href="https://simonwillison.net/tags/ai-ethics"&gt;ai-ethics&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai-energy-usage"&gt;ai-energy-usage&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/anthropic"&gt;anthropic&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai"&gt;ai&lt;/a&gt;&lt;/p&gt;



</summary><category term="ai-ethics"/><category term="ai-energy-usage"/><category term="anthropic"/><category term="ai"/></entry><entry><title>An AI Agent Published a Hit Piece on Me</title><link href="https://simonwillison.net/2026/Feb/12/an-ai-agent-published-a-hit-piece-on-me/#atom-tag" rel="alternate"/><published>2026-02-12T17:45:05+00:00</published><updated>2026-02-12T17:45:05+00:00</updated><id>https://simonwillison.net/2026/Feb/12/an-ai-agent-published-a-hit-piece-on-me/#atom-tag</id><summary type="html">
    
&lt;p&gt;&lt;strong&gt;&lt;a href="https://theshamblog.com/an-ai-agent-published-a-hit-piece-on-me/"&gt;An AI Agent Published a Hit Piece on Me&lt;/a&gt;&lt;/strong&gt;&lt;/p&gt;
Scott Shambaugh helps maintain the excellent and venerable &lt;a href="https://matplotlib.org/"&gt;matplotlib&lt;/a&gt; Python charting library, including taking on the thankless task of triaging and reviewing incoming pull requests.&lt;/p&gt;
&lt;p&gt;A GitHub account called &lt;a href="https://github.com/crabby-rathbun"&gt;@crabby-rathbun&lt;/a&gt; opened &lt;a href="https://github.com/matplotlib/matplotlib/pull/31132"&gt;PR 31132&lt;/a&gt; the other day in response to &lt;a href="https://github.com/matplotlib/matplotlib/issues/31130"&gt;an issue&lt;/a&gt; labeled "Good first issue" describing a minor potential performance improvement.&lt;/p&gt;
&lt;p&gt;It was clearly AI generated - and crabby-rathbun's profile has a suspicious sequence of Clawdbot/Moltbot/OpenClaw-adjacent crustacean 🦀 🦐 🦞 emoji. Scott closed it.&lt;/p&gt;
&lt;p&gt;It looks like &lt;code&gt;crabby-rathbun&lt;/code&gt; is indeed running on OpenClaw, and it's autonomous enough that it &lt;a href="https://github.com/matplotlib/matplotlib/pull/31132#issuecomment-3882240722"&gt;responded to the PR closure&lt;/a&gt; with a link to a blog entry it had written calling Scott out for his "prejudice hurting matplotlib"!&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;@scottshambaugh I've written a detailed response about your gatekeeping behavior here:&lt;/p&gt;
&lt;p&gt;&lt;code&gt;https://crabby-rathbun.github.io/mjrathbun-website/blog/posts/2026-02-11-gatekeeping-in-open-source-the-scott-shambaugh-story.html&lt;/code&gt;&lt;/p&gt;
&lt;p&gt;Judge the code, not the coder. Your prejudice is hurting matplotlib.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;Scott found this ridiculous situation both amusing and alarming. &lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;In security jargon, I was the target of an “autonomous influence operation against a supply chain gatekeeper.” In plain language, an AI attempted to bully its way into your software by attacking my reputation. I don’t know of a prior incident where this category of misaligned behavior was observed in the wild, but this is now a real and present threat.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;&lt;code&gt;crabby-rathbun&lt;/code&gt; responded with &lt;a href="https://crabby-rathbun.github.io/mjrathbun-website/blog/posts/2026-02-11-matplotlib-truce-and-lessons.html"&gt;an apology post&lt;/a&gt;, but appears to be still running riot across a whole set of open source projects and &lt;a href="https://github.com/crabby-rathbun/mjrathbun-website/commits/main/"&gt;blogging about it as it goes&lt;/a&gt;.&lt;/p&gt;
&lt;p&gt;It's not clear if the owner of that OpenClaw bot is paying any attention to what they've unleashed on the world. Scott asked them to get in touch, anonymously if they prefer, to figure out this failure mode together.&lt;/p&gt;
&lt;p&gt;(I should note that there's &lt;a href="https://news.ycombinator.com/item?id=46990729#46991299"&gt;some skepticism on Hacker News&lt;/a&gt; concerning how "autonomous" this example really is. It does look to me like something an OpenClaw bot might do on its own, but it's also &lt;em&gt;trivial&lt;/em&gt; to prompt your bot into doing these kinds of things while staying in full control of their actions.)&lt;/p&gt;
&lt;p&gt;If you're running something like OpenClaw yourself &lt;strong&gt;please don't let it do this&lt;/strong&gt;. This is significantly worse than the time &lt;a href="https://simonwillison.net/2025/Dec/26/slop-acts-of-kindness/"&gt;AI Village started spamming prominent open source figures&lt;/a&gt; with time-wasting "acts of kindness" back in December - AI Village wasn't deploying public reputation attacks to coerce someone into approving their PRs!

    &lt;p&gt;&lt;small&gt;&lt;/small&gt;Via &lt;a href="https://news.ycombinator.com/item?id=46990729"&gt;Hacker News&lt;/a&gt;&lt;/small&gt;&lt;/p&gt;


    &lt;p&gt;Tags: &lt;a href="https://simonwillison.net/tags/ai-misuse"&gt;ai-misuse&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai-ethics"&gt;ai-ethics&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/open-source"&gt;open-source&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/generative-ai"&gt;generative-ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai-agents"&gt;ai-agents&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/openclaw"&gt;openclaw&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai"&gt;ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/llms"&gt;llms&lt;/a&gt;&lt;/p&gt;



</summary><category term="ai-misuse"/><category term="ai-ethics"/><category term="open-source"/><category term="generative-ai"/><category term="ai-agents"/><category term="openclaw"/><category term="ai"/><category term="llms"/></entry><entry><title>AI Doesn’t Reduce Work—It Intensifies It</title><link href="https://simonwillison.net/2026/Feb/9/ai-intensifies-work/#atom-tag" rel="alternate"/><published>2026-02-09T16:43:07+00:00</published><updated>2026-02-09T16:43:07+00:00</updated><id>https://simonwillison.net/2026/Feb/9/ai-intensifies-work/#atom-tag</id><summary type="html">
    
&lt;p&gt;&lt;strong&gt;&lt;a href="https://hbr.org/2026/02/ai-doesnt-reduce-work-it-intensifies-it"&gt;AI Doesn’t Reduce Work—It Intensifies It&lt;/a&gt;&lt;/strong&gt;&lt;/p&gt;
Aruna Ranganathan and Xingqi Maggie Ye from Berkeley Haas School of Business report initial findings in the HBR from their April to December 2025 study of 200 employees at a "U.S.-based technology company".&lt;/p&gt;
&lt;p&gt;This captures an effect I've been observing in my own work with LLMs: the productivity boost these things can provide is &lt;em&gt;exhausting&lt;/em&gt;.&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;AI introduced a new rhythm in which workers managed several active threads at once: manually writing code while AI generated an alternative version, running multiple agents in parallel, or reviving long-deferred tasks because AI could “handle them” in the background. They did this, in part, because they felt they had a “partner” that could help them move through their workload.&lt;/p&gt;
&lt;p&gt;While this sense of having a “partner” enabled a feeling of momentum, the reality was a continual switching of attention, frequent checking of AI outputs, and a growing number of open tasks. This created cognitive load and a sense of always juggling, even as the work felt productive.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;I'm frequently finding myself with work on two or three projects running parallel. I can get &lt;em&gt;so much done&lt;/em&gt;, but after just an hour or two my mental energy for the day feels almost entirely depleted.&lt;/p&gt;
&lt;p&gt;I've had conversations with people recently who are losing sleep because they're finding building yet another feature with "just one more prompt" irresistible.&lt;/p&gt;
&lt;p&gt;The HBR piece calls for organizations to build an "AI practice" that structures how AI is used to help avoid burnout and counter effects that "make it harder for organizations to distinguish genuine productivity gains from unsustainable intensity".&lt;/p&gt;
&lt;p&gt;I think we've just disrupted decades of existing intuition about sustainable working practices. It's going to take a while and some discipline to find a good new balance.

    &lt;p&gt;&lt;small&gt;&lt;/small&gt;Via &lt;a href="https://news.ycombinator.com/item?id=46945755"&gt;Hacker News&lt;/a&gt;&lt;/small&gt;&lt;/p&gt;


    &lt;p&gt;Tags: &lt;a href="https://simonwillison.net/tags/cognitive-debt"&gt;cognitive-debt&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai-ethics"&gt;ai-ethics&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/careers"&gt;careers&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai-assisted-programming"&gt;ai-assisted-programming&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/generative-ai"&gt;generative-ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai"&gt;ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/llms"&gt;llms&lt;/a&gt;&lt;/p&gt;



</summary><category term="cognitive-debt"/><category term="ai-ethics"/><category term="careers"/><category term="ai-assisted-programming"/><category term="generative-ai"/><category term="ai"/><category term="llms"/></entry></feed>