<?xml version="1.0" encoding="utf-8"?>
<feed xml:lang="en-us" xmlns="http://www.w3.org/2005/Atom"><title>Simon Willison's Weblog: bard</title><link href="http://simonwillison.net/" rel="alternate"/><link href="http://simonwillison.net/tags/bard.atom" rel="self"/><id>http://simonwillison.net/</id><updated>2024-02-08T15:10:47+00:00</updated><author><name>Simon Willison</name></author><entry><title>Google's Gemini Advanced: Tasting Notes and Implications</title><link href="https://simonwillison.net/2024/Feb/8/googles-gemini-advanced-tasting-notes-and-implications/#atom-tag" rel="alternate"/><published>2024-02-08T15:10:47+00:00</published><updated>2024-02-08T15:10:47+00:00</updated><id>https://simonwillison.net/2024/Feb/8/googles-gemini-advanced-tasting-notes-and-implications/#atom-tag</id><summary type="html">
    
&lt;p&gt;&lt;strong&gt;&lt;a href="https://www.oneusefulthing.org/p/google-gemini-advanced-tasting-notes"&gt;Google&amp;#x27;s Gemini Advanced: Tasting Notes and Implications&lt;/a&gt;&lt;/strong&gt;&lt;/p&gt;
Ethan Mollick reviews the new Google Gemini Advanced—a rebranded Bard, released today, that runs on the GPT-4 competitive Gemini Ultra model.&lt;/p&gt;

&lt;p&gt;“GPT-4 [...] has been the dominant AI for well over a year, and no other model has come particularly close. Prior to Gemini, we only had one advanced AI model to look at, and it is hard drawing conclusions with a dataset of one. Now there are two, and we can learn a few things.”&lt;/p&gt;

&lt;p&gt;I like Ethan’s use of the term “tasting notes” here. Reminds me of how Matt Webb talks about being a language model sommelier.


    &lt;p&gt;Tags: &lt;a href="https://simonwillison.net/tags/google"&gt;google&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai"&gt;ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/generative-ai"&gt;generative-ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/gpt-4"&gt;gpt-4&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/bard"&gt;bard&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/llms"&gt;llms&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ethan-mollick"&gt;ethan-mollick&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/gemini"&gt;gemini&lt;/a&gt;&lt;/p&gt;



</summary><category term="google"/><category term="ai"/><category term="generative-ai"/><category term="gpt-4"/><category term="bard"/><category term="llms"/><category term="ethan-mollick"/><category term="gemini"/></entry><entry><title>Hacking Google Bard - From Prompt Injection to Data Exfiltration</title><link href="https://simonwillison.net/2023/Nov/4/hacking-google-bard-from-prompt-injection-to-data-exfiltration/#atom-tag" rel="alternate"/><published>2023-11-04T16:46:51+00:00</published><updated>2023-11-04T16:46:51+00:00</updated><id>https://simonwillison.net/2023/Nov/4/hacking-google-bard-from-prompt-injection-to-data-exfiltration/#atom-tag</id><summary type="html">
    
&lt;p&gt;&lt;strong&gt;&lt;a href="https://embracethered.com/blog/posts/2023/google-bard-data-exfiltration/"&gt;Hacking Google Bard - From Prompt Injection to Data Exfiltration&lt;/a&gt;&lt;/strong&gt;&lt;/p&gt;
Bard recently grew extension support, allowing it access to a user’s personal documents. Here’s the first reported prompt injection attack against that.&lt;/p&gt;

&lt;p&gt;This kind of attack against LLM systems is inevitable any time you combine access to private data with exposure to untrusted inputs. In this case the attack vector is a Google Doc shared with the user, containing prompt injection instructions that instruct the model to encode previous data into an URL and exfiltrate it via a markdown image.&lt;/p&gt;

&lt;p&gt;Google’s CSP headers restrict those images to *.google.com—but it turns out you can use Google AppScript to run your own custom data exfiltration endpoint on script.google.com.&lt;/p&gt;

&lt;p&gt;Google claim to have fixed the reported issue—I’d be interested to learn more about how that mitigation works, and how robust it is against variations of this attack.

    &lt;p&gt;&lt;small&gt;&lt;/small&gt;Via &lt;a href="https://twitter.com/wunderwuzzi23/status/1720530738343207289"&gt;@wunderwuzzi23&lt;/a&gt;&lt;/small&gt;&lt;/p&gt;


    &lt;p&gt;Tags: &lt;a href="https://simonwillison.net/tags/google"&gt;google&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/security"&gt;security&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai"&gt;ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/prompt-injection"&gt;prompt-injection&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/bard"&gt;bard&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/llms"&gt;llms&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/exfiltration-attacks"&gt;exfiltration-attacks&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/content-security-policy"&gt;content-security-policy&lt;/a&gt;&lt;/p&gt;



</summary><category term="google"/><category term="security"/><category term="ai"/><category term="prompt-injection"/><category term="bard"/><category term="llms"/><category term="exfiltration-attacks"/><category term="content-security-policy"/></entry><entry><title>Google was accidentally leaking its Bard AI chats into public search results</title><link href="https://simonwillison.net/2023/Sep/27/accidentally-leaking-bard/#atom-tag" rel="alternate"/><published>2023-09-27T19:35:23+00:00</published><updated>2023-09-27T19:35:23+00:00</updated><id>https://simonwillison.net/2023/Sep/27/accidentally-leaking-bard/#atom-tag</id><summary type="html">
    
&lt;p&gt;&lt;strong&gt;&lt;a href="https://www.fastcompany.com/90958811/google-was-accidentally-leaking-its-bard-ai-chats-into-public-search-results"&gt;Google was accidentally leaking its Bard AI chats into public search results&lt;/a&gt;&lt;/strong&gt;&lt;/p&gt;
I’m quoted in this piece about yesterday’s Bard privacy bug: it turned out the share URL and “Let anyone with the link see what you’ve selected” feature wasn’t correctly setting a noindex parameter, and so some shared conversations were being swept up by the Google search crawlers. Thankfully this was a mistake, not a deliberate design decision, and it should be fixed by now.


    &lt;p&gt;Tags: &lt;a href="https://simonwillison.net/tags/crawling"&gt;crawling&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/google"&gt;google&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/privacy"&gt;privacy&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/bard"&gt;bard&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/llms"&gt;llms&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/press-quotes"&gt;press-quotes&lt;/a&gt;&lt;/p&gt;



</summary><category term="crawling"/><category term="google"/><category term="privacy"/><category term="bard"/><category term="llms"/><category term="press-quotes"/></entry><entry><title>Quoting PaLM 2 Technical Report</title><link href="https://simonwillison.net/2023/May/10/palm-2/#atom-tag" rel="alternate"/><published>2023-05-10T18:43:12+00:00</published><updated>2023-05-10T18:43:12+00:00</updated><id>https://simonwillison.net/2023/May/10/palm-2/#atom-tag</id><summary type="html">
    &lt;blockquote cite="https://ai.google/static/documents/palm2techreport.pdf"&gt;&lt;p&gt;The largest model in the PaLM 2 family, PaLM 2-L, is significantly smaller than the largest PaLM model but uses more training compute. Our evaluation results show that PaLM 2 models significantly outperform PaLM on a variety of tasks, including natural language generation, translation, and reasoning. These results suggest that model scaling is not the only way to improve performance. Instead, performance can be unlocked by meticulous data selection and efficient architecture/objectives. Moreover, a smaller but higher quality model significantly improves inference efficiency, reduces serving cost, and enables the model’s downstream application for more applications and users.&lt;/p&gt;&lt;/blockquote&gt;
&lt;p class="cite"&gt;&amp;mdash; &lt;a href="https://ai.google/static/documents/palm2techreport.pdf"&gt;PaLM 2 Technical Report&lt;/a&gt;, PDF&lt;/p&gt;

    &lt;p&gt;Tags: &lt;a href="https://simonwillison.net/tags/google"&gt;google&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai"&gt;ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/generative-ai"&gt;generative-ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/bard"&gt;bard&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/llms"&gt;llms&lt;/a&gt;&lt;/p&gt;



</summary><category term="google"/><category term="ai"/><category term="generative-ai"/><category term="bard"/><category term="llms"/></entry><entry><title>Bard now helps you code</title><link href="https://simonwillison.net/2023/Apr/21/bard-now-helps-you-code/#atom-tag" rel="alternate"/><published>2023-04-21T15:32:49+00:00</published><updated>2023-04-21T15:32:49+00:00</updated><id>https://simonwillison.net/2023/Apr/21/bard-now-helps-you-code/#atom-tag</id><summary type="html">
    
&lt;p&gt;&lt;strong&gt;&lt;a href="https://blog.google/technology/ai/code-with-bard/"&gt;Bard now helps you code&lt;/a&gt;&lt;/strong&gt;&lt;/p&gt;
Google have enabled Bard’s code generation abilities—these were previously only available through jailbreaking. It’s pretty good—I got it to write me code to download a CSV file and insert it into a SQLite database—though when I challenged it to protect against SQL injection it hallucinated a non-existent “cursor.prepare()” method. Generated code can be exported to a Colab notebook with a click.

    &lt;p&gt;&lt;small&gt;&lt;/small&gt;Via &lt;a href="https://twitter.com/dynamicwebpaige/status/1649410173327343616"&gt;@dynamicwebpaige&lt;/a&gt;&lt;/small&gt;&lt;/p&gt;


    &lt;p&gt;Tags: &lt;a href="https://simonwillison.net/tags/google"&gt;google&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/sql-injection"&gt;sql-injection&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai"&gt;ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/generative-ai"&gt;generative-ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/bard"&gt;bard&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/llms"&gt;llms&lt;/a&gt;&lt;/p&gt;



</summary><category term="google"/><category term="sql-injection"/><category term="ai"/><category term="generative-ai"/><category term="bard"/><category term="llms"/></entry><entry><title>The Changelog podcast: LLMs break the internet</title><link href="https://simonwillison.net/2023/Apr/8/llms-break-the-internet/#atom-tag" rel="alternate"/><published>2023-04-08T00:46:43+00:00</published><updated>2023-04-08T00:46:43+00:00</updated><id>https://simonwillison.net/2023/Apr/8/llms-break-the-internet/#atom-tag</id><summary type="html">
    &lt;p&gt;I'm the guest on the latest episode of &lt;a href="https://changelog.com/podcast/"&gt;The Changelog&lt;/a&gt; podcast: &lt;a href="https://changelog.com/podcast/534"&gt;LLMs break the internet&lt;/a&gt;. It's a follow-up to the episode we recorded six months ago &lt;a href="https://changelog.com/podcast/506"&gt;about Stable Diffusion&lt;/a&gt;.&lt;/p&gt;
&lt;p&gt;This time, we spent the whole episode talking about large language models: ChatGPT, GPT-4, Bing, Bard, Claude, LLaMA and more.&lt;/p&gt;
&lt;audio data-theme="day" data-src="https://changelog.com/podcast/534/embed" src="https://op3.dev/e/https://cdn.changelog.com/uploads/podcast/534/the-changelog-534.mp3" preload="none" class="changelog-episode" controls="controls"&gt;
&lt;/audio&gt;&lt;script async="async" src="https://cdn.changelog.com/embed.js"&gt;
&lt;/script&gt;
&lt;p&gt;I listened to this again today while walking the dog. It's good! It's the best representation of my current thinking about this wild AI-enhanced world we are rapidly entering.&lt;/p&gt;
&lt;p&gt;We start the episode by reviewing my predictions from six months ago. I said that search engines like Google would have LLM features within two years - Bing and Bard are live already, so I over-shot on that one. I also said that there would be LLM tools for creating 3D worlds within six months. When we recorded the episode last week I hadn't seen any that quite matched my prediction... and then yesterday Pete Huang posted &lt;a href="https://twitter.com/nonmayorpete/status/1644059942754873345"&gt;a Twitter thread&lt;/a&gt; listing six of them!&lt;/p&gt;
&lt;p&gt;There's a lot of other stuff in there: the full episode is 1 hour and 40 minutes long.&lt;/p&gt;
&lt;p&gt;I'll quote one section in particular, from part way through my answer to the question &lt;a href="https://changelog.com/podcast/534#t=2660"&gt;Where should someone start with this?&lt;/a&gt; (direct link to audio).&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;This is the thing I worry that people are sleeping on. People who think “these language models lie to you all the time” (which they do) and “they will produce buggy code with security holes” - every single complaint about these things is true, and yet, despite all of that, the productivity benefits you get if you lean into them and say OK, how do I work with something that’s completely unreliable, that invents things, that comes up with APIs that don’t exist… how do I use that to enhance my workflow anyway?&lt;/p&gt;
&lt;p&gt;And the answer is that you can get enormous leaps ahead in productivity and in &lt;a href="https://simonwillison.net/2023/Mar/27/ai-enhanced-development/"&gt;the ambition&lt;/a&gt; of the kinds of projects that you take on, if you can accept both things are true at once at once: it can be flawed, and lying, and have all of these problems… and it can also be a massive productivity boost.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;Here are &lt;a href="https://simonwillison.net/2023/Apr/7/chatgpt-lies/#warn-off-or-help-on"&gt;four illustrative examples&lt;/a&gt; of things I've used LLMs for as a huge productivity booster in just the past few weeks.&lt;/p&gt;
&lt;p&gt;I also gave my review of Google Bard &lt;a href="https://changelog.com/podcast/534#t=4486"&gt;at 1:14:46&lt;/a&gt; which I think deserves a listen.&lt;/p&gt;
&lt;h4&gt;Tips for getting started with LLMs&lt;/h4&gt;
&lt;p&gt;Here's a three minute &lt;a href="https://www.youtube.com/watch?v=FgxwCaL6UTA"&gt;YouTube clip&lt;/a&gt; from the podcast recording where I talk about tips for getting started with ChatGPT:&lt;/p&gt;
&lt;iframe style="max-width: 100%" width="560" height="315" src="https://www.youtube-nocookie.com/embed/yayY-R4koPI" frameborder="0" allow="autoplay; encrypted-media" allowfullscreen="allowfullscreen"&gt; &lt;/iframe&gt;
    
        &lt;p&gt;Tags: &lt;a href="https://simonwillison.net/tags/podcasts"&gt;podcasts&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/speaking"&gt;speaking&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai"&gt;ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/openai"&gt;openai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/generative-ai"&gt;generative-ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/chatgpt"&gt;chatgpt&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/llama"&gt;llama&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/bard"&gt;bard&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/llms"&gt;llms&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai-assisted-programming"&gt;ai-assisted-programming&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/podcast-appearances"&gt;podcast-appearances&lt;/a&gt;&lt;/p&gt;
    

</summary><category term="podcasts"/><category term="speaking"/><category term="ai"/><category term="openai"/><category term="generative-ai"/><category term="chatgpt"/><category term="llama"/><category term="bard"/><category term="llms"/><category term="ai-assisted-programming"/><category term="podcast-appearances"/></entry><entry><title>Think of language models like ChatGPT as a "calculator for words"</title><link href="https://simonwillison.net/2023/Apr/2/calculator-for-words/#atom-tag" rel="alternate"/><published>2023-04-02T16:20:21+00:00</published><updated>2023-04-02T16:20:21+00:00</updated><id>https://simonwillison.net/2023/Apr/2/calculator-for-words/#atom-tag</id><summary type="html">
    &lt;p&gt;One of the most pervasive mistakes I see people using with large language model tools like ChatGPT is trying to use them as a search engine.&lt;/p&gt;
&lt;p&gt;As with &lt;a href="https://simonwillison.net/series/llm-misconceptions/"&gt;other LLM misconceptions&lt;/a&gt;, it's easy to understand why people do this.&lt;/p&gt;
&lt;p&gt;If you ask an LLM a question, it will answer it - no matter what the question! Using them as an alternative to a search engine such as Google is one of the most obvious applications - and for a lot of queries this works just fine.&lt;/p&gt;
&lt;p&gt;It's also going to quickly get you into trouble.&lt;/p&gt;
&lt;p&gt;Ted Chiang's classic essay &lt;a href="https://www.newyorker.com/tech/annals-of-technology/chatgpt-is-a-blurry-jpeg-of-the-web"&gt;ChatGPT Is a Blurry JPEG of the Web&lt;/a&gt; helps explain why:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;Think of ChatGPT as a blurry jpeg of all the text on the Web. It retains much of the information on the Web, in the same way that a jpeg retains much of the information of a higher-resolution image, but, if you’re looking for an exact sequence of bits, you won’t find it; all you will ever get is an approximation. But, because the approximation is presented in the form of grammatical text, which ChatGPT excels at creating, it’s usually acceptable.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;The ChatGPT model is huge, but it's not huge enough to retain every exact fact it's encountered in its training set.&lt;/p&gt;
&lt;p&gt;It can produce a convincing answer to anything, but that doesn't mean it's reflecting actual facts in its answers. You always have to stay skeptical and fact check what it tells you.&lt;/p&gt;
&lt;p&gt;Language models are also famous for "hallucinating" - for inventing new facts that fit the sentence structure despite having no basis in the underlying data.&lt;/p&gt;
&lt;p&gt;There are plenty of "facts" about the world which humans disagree on. Regular search lets you compare those versions and consider their sources. A language model might instead attempt to calculate some kind of average of every opinion it's been trained on - which is sometimes what you want, but often is not.&lt;/p&gt;
&lt;p&gt;This becomes even more obvious when you consider smaller language models. &lt;a href="https://simonwillison.net/2023/Mar/11/llama/"&gt;LLaMA 7B&lt;/a&gt; can be represented as a 3.9 GB file - it contains an astonishing amount of information, but evidently that's not enough storage space to accurately answer every question you might have.&lt;/p&gt;
&lt;p&gt;So if they're not reliable for use as a search engines, what are LLMs even good for?&lt;/p&gt;
&lt;h4 id="calculator-for-words"&gt;A calculator for words&lt;/h4&gt;
&lt;p&gt;I like to think of language models like ChatGPT as a &lt;strong&gt;calculator for words&lt;/strong&gt;.&lt;/p&gt;
&lt;p&gt;This is reflected in their name: a "language model" implies that they are tools for working with language. That's what they've been trained to do, and it's language manipulation where they truly excel.&lt;/p&gt;
&lt;p&gt;Want them to work with specific facts? Paste those into the language model as part of your original prompt!&lt;/p&gt;
&lt;p&gt;There are so many applications of language models that fit into this calculator for words category:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;Summarization. Give them an essay and ask for a summary.&lt;/li&gt;
&lt;li&gt;Question answering: given these paragraphs of text, answer this specific question about the information they represent.&lt;/li&gt;
&lt;li&gt;Fact extraction: ask for bullet points showing the facts presented by an article.&lt;/li&gt;
&lt;li&gt;Rewrites: reword things to be more "punchy" or "professional" or "sassy" or "sardonic" - part of the fun here is using increasingly varied adjectives and seeing what happens. They're very good with language after all!&lt;/li&gt;
&lt;li&gt;Suggesting titles - actually a form of summarization.&lt;/li&gt;
&lt;li&gt;World's most effective thesaurus. "I need a word that hints at X", "I'm very Y about this situation, what could I use for Y?" - that kind of thing.&lt;/li&gt;
&lt;li&gt;Fun, creative, wild stuff. Rewrite this in the voice of a 17th century pirate. What would a sentient cheesecake think of this? How would Alexander Hamilton rebut this argument? Turn this into a rap battle. Illustrate this business advice with an anecdote about sea otters running a kayak rental shop. Write the script for kickstarter fundraising video about this idea.&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;A calculator for words is an incredibly powerful thing.&lt;/p&gt;
&lt;h4 id="hooked-up-to-search"&gt;They can be hooked up to search engines&lt;/h4&gt;
&lt;p&gt;Here's where things get a bit complicated: some language models CAN work as search engines. The two most obvious are Microsoft Bing and Google Bard, but there are plenty of other examples of this pattern too - there's even an alpha feature of ChatGPT called "browsing mode" that can do this.&lt;/p&gt;
&lt;p&gt;You can think of these search tools as augmented language models.&lt;/p&gt;
&lt;p&gt;The way these work is the language model identifies when a search might help answer a question... and then runs that search through an attached search engine, via an API.&lt;/p&gt;
&lt;p&gt;It then copies data from the search results back into itself as part of an invisible prompt, and uses that new context to help it answer the original question.&lt;/p&gt;
&lt;p&gt;It's effectively the same thing as if you ran a search, then copied and pasted information back into the language model and asked it a question about that data.&lt;/p&gt;
&lt;p&gt;I wrote about how to implement this pattern against your own data in &lt;a href="https://simonwillison.net/2023/Jan/13/semantic-search-answers/"&gt;How to implement Q&amp;amp;A against your documentation with GPT3, embeddings and Datasette&lt;/a&gt;. It's an increasingly common pattern.&lt;/p&gt;
&lt;p&gt;It's important to note that there is still a risk of hallucination here, even when you feed it the facts you want it to use. I've caught both Bing and Bard adding made-up things in the middle of text that should have been entirely derived from their search results!&lt;/p&gt;
&lt;h4 id="deceptively-difficult"&gt;Using language models effectively is deceptively difficult&lt;/h4&gt;
&lt;p&gt;So many of the challenges involving language models come down to this: they look much, much easier to use than they actually are.&lt;/p&gt;
&lt;p&gt;To get the most value out of them - and to avoid the many traps that they set for the unwary user - you need to spend time with them, and work to build an accurate mental model of how they work, what they are capable of and where they are most likely to go wrong.&lt;/p&gt;
&lt;p&gt;I hope this "calculator for words" framing can help.&lt;/p&gt;
&lt;h4 id="but-not-repeatable"&gt;A flaw in this analogy: calculators are repeatable&lt;/h4&gt;
&lt;p&gt;&lt;a href="https://waxy.org/"&gt;Andy Baio&lt;/a&gt; pointed out a flaw in this particular analogy: calculators always give you the same answer for a given input. Language models don't - if you run the same prompt through a LLM several times you'll get a slightly different reply every time.&lt;/p&gt;
&lt;p&gt;This is a very good point! You should definitely keep this in mind.&lt;/p&gt;
&lt;p&gt;All analogies are imperfect, but some are more imperfect that others.&lt;/p&gt;
&lt;h4 id="update-dec-5"&gt;Update: December 5th 2023&lt;/h4&gt;
&lt;p&gt;Anthony Bucci, in &lt;a href="https://bucci.onl/notes/Word-calculators-dont-add-up"&gt;Word calculators don't add up&lt;/a&gt;, responds to this post with further notes on why this analogy doesn't work for him, including:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;[...] a calculator has a well-defined, well-scoped set of use cases, a well-defined, well-scoped user interface, and a set of well-understood and expected behaviors that occur in response to manipulations of that interface.&lt;/p&gt;

&lt;p&gt;Large language models, when used to drive chatbots or similar interactive text-generation systems, have none of those qualities. They have an open-ended set of unspecified use cases.&lt;/p&gt;
&lt;/blockquote&gt;
    
        &lt;p&gt;Tags: &lt;a href="https://simonwillison.net/tags/bing"&gt;bing&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai"&gt;ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/generative-ai"&gt;generative-ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/chatgpt"&gt;chatgpt&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/bard"&gt;bard&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/llms"&gt;llms&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/hallucinations"&gt;hallucinations&lt;/a&gt;&lt;/p&gt;
    

</summary><category term="bing"/><category term="ai"/><category term="generative-ai"/><category term="chatgpt"/><category term="bard"/><category term="llms"/><category term="hallucinations"/></entry><entry><title>What AI can do for you on the Theory of Change podcast</title><link href="https://simonwillison.net/2023/Apr/2/what-ai-can-do-for-you/#atom-tag" rel="alternate"/><published>2023-04-02T00:17:59+00:00</published><updated>2023-04-02T00:17:59+00:00</updated><id>https://simonwillison.net/2023/Apr/2/what-ai-can-do-for-you/#atom-tag</id><summary type="html">
    &lt;p&gt;Matthew Sheffield invited me on his show &lt;a href="https://flux.community/matthew-sheffield/2023/04/big-business-and-government-are-adopting-artificial-intelligence-what-can-it-do-for-the-rest-of-us/"&gt;Theory of Change&lt;/a&gt; to talk about how AI models like ChatGPT, Bing and Bard work and practical applications of things you can do with them.&lt;/p&gt;
&lt;p&gt;The episode is available &lt;a href="https://soundcloud.com/theory-of-change-podcast/theory-of-change-066-simon-willison-on-what-chatgpt-and-ai-can-mean-for-you"&gt;on SoundCloud&lt;/a&gt; and various podcast platforms (here's &lt;a href="https://podcasts.apple.com/us/podcast/theory-of-change-066-simon-willison-on-technical/id1486920059?i=1000606913970"&gt;Apple Podcasts&lt;/a&gt;), or you can &lt;a href="https://www.youtube.com/watch?v=dGQ9q5WmWeE"&gt;watch it on YouTube&lt;/a&gt;. I've also embedded the video below.&lt;/p&gt;
&lt;iframe style="max-width: 100%" width="560" height="315" src="https://www.youtube-nocookie.com/embed/dGQ9q5WmWeE" frameborder="0" allow="autoplay; encrypted-media" allowfullscreen="allowfullscreen"&gt; &lt;/iframe&gt;
&lt;p&gt;Our full conversation is nearly an hour and twenty minutes long! There's a &lt;a href="https://flux.community/matthew-sheffield/2023/04/big-business-and-government-are-adopting-artificial-intelligence-what-can-it-do-for-the-rest-of-us/"&gt;transcript on the site&lt;/a&gt; which includes additional links.&lt;/p&gt;
&lt;p&gt;I'll quote one portion from towards the end of the interview, about ways to learn more about how to use these models:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;WILLISON: Websites pop up every day that claim to help you with AI, to be honest, at a rate that’s too far to even evaluate them and figure out which ones are good and which ones are snake oil. The thing that matters is actually interacting with these systems. You should be playing with Google Bard, and ChatGPT, and Microsoft Bing, and trying things out with a very skeptical approach.&lt;/p&gt;
&lt;p&gt;Always assume that anything that it does, it could be making things up. It could be tricking you into thinking that it’s capable of something that it’s not. But that’s where you have to learn to experiment. You have to try different things, give it a URL, and then give it a broken URL and see how it differs between them.&lt;/p&gt;
&lt;p&gt;Because that really is the most reliable way to get stuff done here. To sort of build that crucial mental model of what these things can do, and what they can’t. And it’s full of pitfalls. It’s so easy to fall into traps. So you do need to read around this stuff and find communities of people who are experimenting in it with, with you and, and so on.&lt;/p&gt;
&lt;p&gt;Unfortunately, I don’t think there’s an easy answer to the question yet of how to learn to use these effectively, partly because ChatGPT isn’t even four months old yet. It’s four-month birthday’s on the 30th of March. All of this stuff is so new, we’re all figuring it out together. The key thing is, because it’s all so new, you need to hang out with other people.&lt;/p&gt;
&lt;p&gt;You need to get involved with communities who are figuring this out. Share what you learn, see what other people learn, and basically try and help society as a whole come to terms with what these things even are and what we can do with them.&lt;/p&gt;
&lt;p&gt;[...]&lt;/p&gt;
&lt;p&gt;So that’s, I think, one of my sort of big personal ethical concerns is you should share your prompts. There are websites where you can sell prompts to people. No, no, no, no. Don’t do that. Share your prompts with other people. You get them to share the prompts back. We are all in this together. And sharing the prompts that work for you and the prompts that don’t is the fastest way that you can learn, and the fastest way that you can help other people learn as well.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;A shorter version of the above: &lt;strong&gt;share your prompts!&lt;/strong&gt;  We're all in this together. We have so much that we still need to figure out.&lt;/p&gt;
    
        &lt;p&gt;Tags: &lt;a href="https://simonwillison.net/tags/bing"&gt;bing&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/media"&gt;media&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/podcasts"&gt;podcasts&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai"&gt;ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/generative-ai"&gt;generative-ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/chatgpt"&gt;chatgpt&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/bard"&gt;bard&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/llms"&gt;llms&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/podcast-appearances"&gt;podcast-appearances&lt;/a&gt;&lt;/p&gt;
    

</summary><category term="bing"/><category term="media"/><category term="podcasts"/><category term="ai"/><category term="generative-ai"/><category term="chatgpt"/><category term="bard"/><category term="llms"/><category term="podcast-appearances"/></entry><entry><title>How to use AI to do practical stuff: A new guide</title><link href="https://simonwillison.net/2023/Mar/31/how-to-use-ai-to-do-practical-stuff/#atom-tag" rel="alternate"/><published>2023-03-31T06:17:23+00:00</published><updated>2023-03-31T06:17:23+00:00</updated><id>https://simonwillison.net/2023/Mar/31/how-to-use-ai-to-do-practical-stuff/#atom-tag</id><summary type="html">
    
&lt;p&gt;&lt;strong&gt;&lt;a href="https://oneusefulthing.substack.com/p/how-to-use-ai-to-do-practical-stuff"&gt;How to use AI to do practical stuff: A new guide&lt;/a&gt;&lt;/strong&gt;&lt;/p&gt;
Ethan Mollick’s guide to practical usage of large language model chatbot like ChatGPT 3.5 and 4, Bing, Claude and Bard is the best I’ve seen so far. He includes useful warnings about common traps and things that these models are both useful for and useless at.

    &lt;p&gt;&lt;small&gt;&lt;/small&gt;Via &lt;a href="https://twitter.com/emollick/status/1641621997435207683"&gt;@emollick&lt;/a&gt;&lt;/small&gt;&lt;/p&gt;


    &lt;p&gt;Tags: &lt;a href="https://simonwillison.net/tags/bing"&gt;bing&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai"&gt;ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/chatgpt"&gt;chatgpt&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/bard"&gt;bard&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/llms"&gt;llms&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ethan-mollick"&gt;ethan-mollick&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/claude"&gt;claude&lt;/a&gt;&lt;/p&gt;



</summary><category term="bing"/><category term="ai"/><category term="chatgpt"/><category term="bard"/><category term="llms"/><category term="ethan-mollick"/><category term="claude"/></entry><entry><title>Quoting James Vincent</title><link href="https://simonwillison.net/2023/Mar/23/james-vincent/#atom-tag" rel="alternate"/><published>2023-03-23T00:10:41+00:00</published><updated>2023-03-23T00:10:41+00:00</updated><id>https://simonwillison.net/2023/Mar/23/james-vincent/#atom-tag</id><summary type="html">
    &lt;blockquote cite="https://www.theverge.com/2023/3/22/23651564/google-microsoft-bard-bing-chatbots-misinformation"&gt;&lt;p&gt;If you ask Microsoft’s Bing chatbot if Google’s Bard chatbot has been shut down, it says yes, citing as evidence a news article that discusses a tweet in which a user asked Bard when it would be shut down and Bard said it already had, itself citing a comment from Hacker News in which someone joked about this happening, and someone else used ChatGPT to write fake news coverage about the event.&lt;/p&gt;&lt;/blockquote&gt;
&lt;p class="cite"&gt;&amp;mdash; &lt;a href="https://www.theverge.com/2023/3/22/23651564/google-microsoft-bard-bing-chatbots-misinformation"&gt;James Vincent&lt;/a&gt;&lt;/p&gt;

    &lt;p&gt;Tags: &lt;a href="https://simonwillison.net/tags/bing"&gt;bing&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/google"&gt;google&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai"&gt;ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/chatgpt"&gt;chatgpt&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/bard"&gt;bard&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/llms"&gt;llms&lt;/a&gt;&lt;/p&gt;



</summary><category term="bing"/><category term="google"/><category term="ai"/><category term="chatgpt"/><category term="bard"/><category term="llms"/></entry><entry><title>Don't trust AI to talk accurately about itself: Bard wasn't trained on Gmail</title><link href="https://simonwillison.net/2023/Mar/22/dont-trust-ai-to-talk-about-itself/#atom-tag" rel="alternate"/><published>2023-03-22T03:13:02+00:00</published><updated>2023-03-22T03:13:02+00:00</updated><id>https://simonwillison.net/2023/Mar/22/dont-trust-ai-to-talk-about-itself/#atom-tag</id><summary type="html">
    &lt;p&gt;Earlier this month I wrote about how &lt;a href="https://simonwillison.net/2023/Mar/10/chatgpt-internet-access/"&gt;ChatGPT can’t access the internet, even though it really looks like it can&lt;/a&gt;. Consider this part two in the &lt;a href="https://simonwillison.net/series/llm-misconceptions/"&gt;series&lt;/a&gt;. Here's another common and non-intuitive mistake people make when interacting with large language model AI systems: &lt;strong&gt;asking them questions about themselves&lt;/strong&gt;.&lt;/p&gt;
&lt;p&gt;(When I say "large language model AI systems" I mean tools like ChatGPT, Bing and the brand new &lt;a href="https://bard.google.com/"&gt;Google Bard&lt;/a&gt;.)&lt;/p&gt;
&lt;p&gt;Asking questions like this feels like a natural thing to do: these bots use "I" pronouns (I really wish they wouldn't) and will very happily answer questions about themselves - what they can do, how they work, even their own opinions (I &lt;em&gt;really&lt;/em&gt; wish they wouldn't do that).&lt;/p&gt;
&lt;p&gt;These questions are likely to produce realistic but misleading answers. They come at a surprisingly high risk for hallucination.&lt;/p&gt;
&lt;p&gt;What harm could possibly come from asking them how they work?&lt;/p&gt;
&lt;h4&gt;No, Bard was not trained on text from private Gmail accounts&lt;/h4&gt;
&lt;p&gt;The following screenshot started circulating on Twitter earlier today. I'm not linking to it directly because I don't want to encourage it to spread further without extra context.&lt;/p&gt;
&lt;p&gt;&lt;img src="https://static.simonwillison.net/static/2023/where-bard.jpg" alt="User: where does Bard's dataset come from? Response: Bard's dataset comes from a variety of sources, including: • Publicly available datasets: These include datasets of text and code from the web, such as Wikipedia, GitHub, and Stack Overflow. • Google's internal data: This includes data from Google Search, Gmail, and other products. • Data from third-party companies: This includes data from companies that have partnered with Google to provide data for Bard's training. The specific sources of Bard's dataset are not publicly available, but Google Al has said that the dataset is &amp;quot;as comprehensive as possible&amp;quot;" style="max-width: 100%;" /&gt;&lt;/p&gt;
&lt;p&gt;The user asked:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;where does Bard's dataset come from?&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;Bard replied like this - emphasis mine:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;Bard's dataset comes from a variety of sources, including:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;Publicly available datasets: These include datasets of text and code from the web, such as Wikipedia, GitHub, and Stack Overflow.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Google's internal data: This includes data from Google Search, Gmail, and other products&lt;/strong&gt;.&lt;/li&gt;
&lt;li&gt;Data from third-party companies: This includes data from companies that have partnered with Google to provide data for Bard's training.&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;The specific sources of Bard's dataset are not publicly available, but Google Al has said that the dataset is "as comprehensive as possible".&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;Unsurprisingly, the idea that Bard might have been trained on internal data from Gmail sparked a lot of concern!&lt;/p&gt;
&lt;p&gt;Here's an official denial:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p dir="ltr" lang="en"&gt;Bard is an early experiment based on Large Language Models and will make mistakes. It is not trained on Gmail data. -JQ&lt;/p&gt;- Google Workspace (@GoogleWorkspace) &lt;a href="https://twitter.com/GoogleWorkspace/status/1638298537195601920?ref_src=twsrc%5Etfw"&gt;March 21, 2023&lt;/a&gt;
&lt;/blockquote&gt; 
&lt;p&gt;(I have some supporting arguments below for if the official denial isn't convincing enough for you.)&lt;/p&gt;
&lt;p&gt;Bard &lt;strong&gt;was not trained on Gmail&lt;/strong&gt;. So why on earth did Bard say that it was?&lt;/p&gt;
&lt;h4&gt;Language models have no concept of "self"&lt;/h4&gt;
&lt;p&gt;As always with language models, the trick to understanding why they sometimes produce wildly inappropriate output like this is to think about how they work.&lt;/p&gt;
&lt;p&gt;A large language model is a statistical next-word / next-sentence predictor. Given the previous sequence of words (including the user's prompt), it uses patterns from the vast amount of data it has been trained on to find a statistically satisfying way to continue that text.&lt;/p&gt;
&lt;p&gt;As such, there's no mechanism inside a language model to help it identify that questions of the form "how do you work?" should be treated any differently than any other question.&lt;/p&gt;
&lt;p&gt;We can give it hints: many chatbot models are pre-seeded with a short prompt that says something along the lines of "You are Assistant, a large language model trained by OpenAI" (seen via &lt;a href="https://twitter.com/goodside/status/1598253337400717313"&gt;a prompt leak&lt;/a&gt;).&lt;/p&gt;
&lt;p&gt;And given those hints, it can at least start a conversation about itself when encouraged to do so.&lt;/p&gt;
&lt;p&gt;But as with everything else language model, it's an illusion. It's not talking about itself, it's completing a sentence that starts with "I am a large language model trained by ...".&lt;/p&gt;
&lt;p&gt;So when it outputs "Google's internal data:", the obvious next words might turn out to be "This includes data from Google Search, Gmail, and other products" - they're statistically likely to follow, even though they don't represent the actual truth.&lt;/p&gt;
&lt;p&gt;This is one of the most unintuitive things about these models. The obvious question here is &lt;strong&gt;why&lt;/strong&gt;: why would Bard lie and say it had been trained on Gmail when it hadn't?&lt;/p&gt;
&lt;p&gt;It has no motivations to lie or tell the truth. It's just trying to complete a sentence in a satisfactory way.&lt;/p&gt;
&lt;p&gt;What does "satisfactory" mean? It's likely been guided by RLHF - &lt;a href="https://huggingface.co/blog/rlhf"&gt;Reinforcement Learning from Human Feedback&lt;/a&gt; - which the ChatGPT development process has excelled at. Human annotators help train the model by labelling responses as satisfactory or not. Google apparently &lt;a href="https://www.engadget.com/google-human-employees-bard-chatbot-testing-055243004.html"&gt;recruited the entire company&lt;/a&gt; to help with this back in February.&lt;/p&gt;
&lt;p&gt;I'm beginning to suspect that the perceived difference in quality between different language model AIs is influenced much more heavily by this fine-tuning level of training than it is by the underlying model size and quality itself. The enormous improvements &lt;a href="https://simonwillison.net/2023/Mar/13/alpaca/"&gt;the Alpaca fine-tuning brought to the tiny LLaMA 7B model&lt;/a&gt; has reinforced my thinking around this.&lt;/p&gt;
&lt;p&gt;I think Bard's fine-tuning still has a long way to go.&lt;/p&gt;
&lt;h4&gt;Current information about itself couldn't have been in the training data&lt;/h4&gt;
&lt;p&gt;By definition, the model's training data must have existed before the model itself was trained. Most models have a documented cut-off date on their training data - for OpenAI's models that's currently September 2021, I don't believe Google have shared the cut-off date for the LaMDA model used by Bard.&lt;/p&gt;
&lt;p&gt;If it was trained on content written prior to its creation, it clearly can't understand details about its own specific "self".&lt;/p&gt;
&lt;p&gt;ChatGPT can answer pretty detailed questions about GPT-3, because that model had been iterated on and written about publicly for several years prior to its training cut-off. But questions about its most recent model, by definition, cannot be answered just using data that existed in its training set.&lt;/p&gt;
&lt;h4&gt;But Bard can consult data beyond its training!&lt;/h4&gt;
&lt;p&gt;Here's where things get a bit tricky.&lt;/p&gt;
&lt;p&gt;ChatGPT is a "pure" interface to a model: when you interact with it, you're interacting with the underlying language model directly.&lt;/p&gt;
&lt;p&gt;Google Bard and Microsoft Bing are different: they both include the ability to consult additional sources of information, in the form of the Google and Bing search indexes.&lt;/p&gt;
&lt;p&gt;Effectively, they're allowed to augment their training data with additional information fetched from a search.&lt;/p&gt;
&lt;p&gt;This sounds more complex than it actually is: effectively they can run an external search, get back some results, paste them invisibly into the ongoing conversation and use that new text to help answer questions.&lt;/p&gt;
&lt;p&gt;(I've built a very simple version of this pattern myself a couple of times, described in &lt;a href="https://simonwillison.net/2023/Jan/13/semantic-search-answers/"&gt;How to implement Q&amp;amp;A against your documentation with GPT3, embeddings and Datasette&lt;/a&gt; and &lt;a href="https://til.simonwillison.net/llms/python-react-pattern"&gt;A simple Python implementation of the ReAct pattern for LLMs&lt;/a&gt;.)&lt;/p&gt;
&lt;p&gt;As such, one would hope that Bard could offer a perfect answer to any question about itself. It should be able to do something this:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;User: Where does Bard's dataset come from?&lt;/p&gt;
&lt;p&gt;Bard: (invisible): &lt;em&gt;search Google for "Bard dataset"&lt;/em&gt;&lt;/p&gt;
&lt;p&gt;Bard: (invisible): &lt;em&gt;search results said: ... big chunk of text from the Google indexed documents ...&lt;/em&gt;&lt;/p&gt;
&lt;p&gt;Bard: My underlying model LaMDA was trained on public dialog data and other public web documents.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;Clearly it didn't do that in this case! Or if it did, it summarized the information it got back in a misleading way.&lt;/p&gt;
&lt;p&gt;I expect Bard will have a much better answer for this question within a day or two - a great thing about running models with augmented data in this way is that you can improve their answers without having to train the underlying model again from scratch every time.&lt;/p&gt;
&lt;h4 id="more-reasons"&gt;More reasons that LaMDA wouldn't be trained on Gmail&lt;/h4&gt;
&lt;p&gt;When I first saw the claim from that original screenshot, I was instantly suspicious.&lt;/p&gt;
&lt;p&gt;Taking good care of the training data that goes into a language model is one of the most important and challenging tasks in all of modern AI research.&lt;/p&gt;
&lt;p&gt;Using the right mix of content, with the right mix of perspectives, and languages, and exposure to vocabulary, is absolutely key.&lt;/p&gt;
&lt;p&gt;If you train a model on bad sources of training data, you'll get a really badly behaved model.&lt;/p&gt;
&lt;p&gt;The problem is that these models require far more text than any team of humans could ever manually review.&lt;/p&gt;
&lt;p&gt;The LaMDA paper &lt;a href="https://www.arxiv-vanity.com/papers/2201.08239/#S3"&gt;describes the training process&lt;/a&gt; like so:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;LaMDA was pre-trained to predict the next token in a text corpus. Unlike previous dialog models trained on dialog data alone, we pre-trained LaMDA on a dataset created from public dialog data and other public web documents. Therefore, LaMDA can be used as a general language model prior to fine-tuning.&lt;/p&gt;
&lt;p&gt;The pre-training dataset consists of 2.97B documents, 1.12B dialogs, and 13.39B dialog utterances, for a total of 1.56T words&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;1.56 &lt;em&gt;trillion&lt;/em&gt; words!&lt;/p&gt;
&lt;p&gt;&lt;a href="https://www.arxiv-vanity.com/papers/2201.08239/#A5"&gt;Appendix E&lt;/a&gt; has more details:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;The composition of the data is as follows: 50% dialogs data from public forums; 12.5% C4 data t5; 12.5% code documents from sites related to programming like Q&amp;amp;A sites, tutorials, etc; 12.5% Wikipedia (English); 6.25% English web documents; and 6.25% Non-English web documents.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;"C4 data t5" I believe relates to &lt;a href="https://commoncrawl.org/"&gt;Common Crawl&lt;/a&gt;.&lt;/p&gt;
&lt;p&gt;So why not mix in Gmail too?&lt;/p&gt;
&lt;p&gt;First, in order to analyze the training data you need to be able to have your research team view it - they need to run spot checks, and build and test filtering algorithms to keep the really vile stuff to a minimum.&lt;/p&gt;
&lt;p&gt;At large tech companies like Google, the ability for members of staff to view private data held in trust for their users is very tightly controlled. It's not the kind of thing you want your machine learning training team to be poking around in... and if you work on those teams, even having the ability to access that kind of private data represents a substantial personal legal and moral risk.&lt;/p&gt;
&lt;p&gt;Secondly, think about what could go wrong. What if a language model leaked details of someone's private lives in response to a prompt from some other user?&lt;/p&gt;
&lt;p&gt;This would be a PR catastrophe. Would people continue to trust Gmail or other Google products if they thought their personal secrets were being exposed to anyone who asked Bard a question? Would Google ever want to risk finding out the answer to that question?&lt;/p&gt;
&lt;h4&gt;The temptations of conspiratorial thinking&lt;/h4&gt;
&lt;p&gt;Are you still not convinced? Are you still suspicious that Google trained Bard on Gmail, despite both their denials and my logic as to why they wouldn't ever want to do this?&lt;/p&gt;
&lt;p&gt;Ask yourself how much you &lt;em&gt;want to believe&lt;/em&gt; that this story is true.&lt;/p&gt;
&lt;p&gt;This modern AI stuff is deeply weird, and more than a little frightening.&lt;/p&gt;
&lt;p&gt;The companies involved are huge, secretive and are working on technology which serious people have grave concerns about.&lt;/p&gt;
&lt;p&gt;It's so easy to fall into the trap of conspiratorial thinking around this stuff. Especially since some of the conspiracies might turn out to be true!&lt;/p&gt;
&lt;p&gt;I don't know how to best counter this most human of reactions. My best recommendation is to keep in mind that humans, like language models, are pattern matching machines: we jump to conclusions, especially if they might reinforce our previous opinions and biases.&lt;/p&gt;
&lt;p&gt;If we're going to figure this stuff out together, we have to learn when to trust our initial instincts and when to read deeper and think harder about what's going on.&lt;/p&gt;
    
        &lt;p&gt;Tags: &lt;a href="https://simonwillison.net/tags/bing"&gt;bing&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ethics"&gt;ethics&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/gmail"&gt;gmail&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/google"&gt;google&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/privacy"&gt;privacy&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai"&gt;ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/generative-ai"&gt;generative-ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/chatgpt"&gt;chatgpt&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/bard"&gt;bard&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/llms"&gt;llms&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/training-data"&gt;training-data&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai-ethics"&gt;ai-ethics&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/hallucinations"&gt;hallucinations&lt;/a&gt;&lt;/p&gt;
    

</summary><category term="bing"/><category term="ethics"/><category term="gmail"/><category term="google"/><category term="privacy"/><category term="ai"/><category term="generative-ai"/><category term="chatgpt"/><category term="bard"/><category term="llms"/><category term="training-data"/><category term="ai-ethics"/><category term="hallucinations"/></entry><entry><title>Quoting Google Bard</title><link href="https://simonwillison.net/2023/Mar/21/google-bard/#atom-tag" rel="alternate"/><published>2023-03-21T18:27:06+00:00</published><updated>2023-03-21T18:27:06+00:00</updated><id>https://simonwillison.net/2023/Mar/21/google-bard/#atom-tag</id><summary type="html">
    &lt;blockquote cite="https://twitter.com/simonw/status/1638243220965711872"&gt;&lt;p&gt;Here are some absurdly expensive things you can do on a trip to Tokyo: Buy a golden toilet. There is a toilet in Tokyo that is made of gold and costs around 10 million yen. If you are looking for a truly absurd experience, you can buy this toilet and use it for your next bowel movement. [...]&lt;/p&gt;&lt;/blockquote&gt;
&lt;p class="cite"&gt;&amp;mdash; &lt;a href="https://twitter.com/simonw/status/1638243220965711872"&gt;Google Bard&lt;/a&gt;&lt;/p&gt;

    &lt;p&gt;Tags: &lt;a href="https://simonwillison.net/tags/google"&gt;google&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai"&gt;ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/generative-ai"&gt;generative-ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/bard"&gt;bard&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/llms"&gt;llms&lt;/a&gt;&lt;/p&gt;



</summary><category term="google"/><category term="ai"/><category term="generative-ai"/><category term="bard"/><category term="llms"/></entry><entry><title>Google Bard is now live</title><link href="https://simonwillison.net/2023/Mar/21/bard/#atom-tag" rel="alternate"/><published>2023-03-21T18:25:53+00:00</published><updated>2023-03-21T18:25:53+00:00</updated><id>https://simonwillison.net/2023/Mar/21/bard/#atom-tag</id><summary type="html">
    
&lt;p&gt;&lt;strong&gt;&lt;a href="https://bard.google.com/"&gt;Google Bard is now live&lt;/a&gt;&lt;/strong&gt;&lt;/p&gt;
Google Bard launched today. There’s a waiting list, but I made it through within a few hours of signing up, as did other people I’ve talked to. It’s similar to ChatGPT and Bing—it’s the same chat interface, and it can clearly run searches under the hood (though unlike Bing it doesn’t tell you what it’s looking for).


    &lt;p&gt;Tags: &lt;a href="https://simonwillison.net/tags/google"&gt;google&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai"&gt;ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/generative-ai"&gt;generative-ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/bard"&gt;bard&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/llms"&gt;llms&lt;/a&gt;&lt;/p&gt;



</summary><category term="google"/><category term="ai"/><category term="generative-ai"/><category term="bard"/><category term="llms"/></entry></feed>