<?xml version="1.0" encoding="utf-8"?>
<feed xml:lang="en-us" xmlns="http://www.w3.org/2005/Atom"><title>Simon Willison's Weblog: notebooklm</title><link href="http://simonwillison.net/" rel="alternate"/><link href="http://simonwillison.net/tags/notebooklm.atom" rel="self"/><id>http://simonwillison.net/</id><updated>2024-11-21T03:03:13+00:00</updated><author><name>Simon Willison</name></author><entry><title>Quoting Steven Johnson</title><link href="https://simonwillison.net/2024/Nov/21/steven-johnson/#atom-tag" rel="alternate"/><published>2024-11-21T03:03:13+00:00</published><updated>2024-11-21T03:03:13+00:00</updated><id>https://simonwillison.net/2024/Nov/21/steven-johnson/#atom-tag</id><summary type="html">
    &lt;blockquote cite="https://adjacentpossible.substack.com/p/in-the-context-of-long-context"&gt;&lt;p&gt;When we started working on what became NotebookLM in the summer of 2022, we could fit about 1,500 words in the context window. Now we can fit up to 1.5 &lt;em&gt;million&lt;/em&gt; words. (And using various other tricks, effectively fit 25 million words.) The emergence of long context models is, I believe, the single most unappreciated AI development of the past two years, at least among the general public. It radically transforms the utility of these models in terms of actual, practical applications.&lt;/p&gt;&lt;/blockquote&gt;
&lt;p class="cite"&gt;&amp;mdash; &lt;a href="https://adjacentpossible.substack.com/p/in-the-context-of-long-context"&gt;Steven Johnson&lt;/a&gt;&lt;/p&gt;

    &lt;p&gt;Tags: &lt;a href="https://simonwillison.net/tags/google"&gt;google&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai"&gt;ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/generative-ai"&gt;generative-ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/llms"&gt;llms&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/gemini"&gt;gemini&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/notebooklm"&gt;notebooklm&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/long-context"&gt;long-context&lt;/a&gt;&lt;/p&gt;



</summary><category term="google"/><category term="ai"/><category term="generative-ai"/><category term="llms"/><category term="gemini"/><category term="notebooklm"/><category term="long-context"/></entry><entry><title>New in NotebookLM: Customizing your Audio Overviews</title><link href="https://simonwillison.net/2024/Oct/17/notebooklm-pelicans/#atom-tag" rel="alternate"/><published>2024-10-17T17:27:01+00:00</published><updated>2024-10-17T17:27:01+00:00</updated><id>https://simonwillison.net/2024/Oct/17/notebooklm-pelicans/#atom-tag</id><summary type="html">
    
&lt;p&gt;&lt;strong&gt;&lt;a href="https://blog.google/technology/ai/notebooklm-update-october-2024/"&gt;New in NotebookLM: Customizing your Audio Overviews&lt;/a&gt;&lt;/strong&gt;&lt;/p&gt;
The most requested feature for Google's NotebookLM "audio overviews" (aka &lt;a href="https://simonwillison.net/2024/Sep/29/notebooklm-audio-overview/"&gt;automatically generated podcast conversations&lt;/a&gt;) has been the ability to provide direction to those artificial podcast hosts - setting their expertise level or asking them to focus on specific topics.&lt;/p&gt;
&lt;p&gt;Today's update adds exactly that:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;Now you can provide instructions before you generate a "Deep Dive" Audio Overview. For example, you can focus on specific topics or adjust the expertise level to suit your audience. Think of it like slipping the AI hosts a quick note right before they go on the air, which will change how they cover your material.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;I pasted in a link to my &lt;a href="https://simonwillison.net/2024/Oct/17/video-scraping/"&gt;post about video scraping&lt;/a&gt; and prompted it like this:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;&lt;code&gt;You are both pelicans who work as data journalist at a pelican news service. Discuss this from the perspective of pelican data journalists, being sure to inject as many pelican related anecdotes as possible&lt;/code&gt;&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;Here's &lt;a href="https://static.simonwillison.net/static/2024/video-scraping-pelicans.mp3"&gt;the resulting 7m40s MP3&lt;/a&gt;, and &lt;a href="https://gist.github.com/simonw/2230937450d271b5f8433e8f85ad6e0a"&gt;the transcript&lt;/a&gt;.&lt;/p&gt;
&lt;p&gt;&lt;audio controls="controls" style="width: 100%"&gt;
  &lt;source src="https://static.simonwillison.net/static/2024/video-scraping-pelicans.mp3" type="audio/mp3"&gt;
  Your browser does not support the audio element.
&lt;/audio&gt;&lt;/p&gt;
&lt;p&gt;It starts off strong!&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;You ever find yourself wading through mountains of data trying to pluck out the juicy bits? It's like hunting for a single shrimp in a whole kelp forest, am I right?&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;Then later:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;Think of those facial recognition systems they have for humans. We could have something similar for our finned friends. Although, gotta say, the ethical implications of that kind of tech are a whole other kettle of fish. We pelicans gotta use these tools responsibly and be transparent about it.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;And when brainstorming some potential use-cases:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;Imagine a pelican citizen journalist being able to analyze footage of a local council meeting, you know, really hold those  pelicans in power accountable, or a pelican historian using video  scraping to analyze old film reels, uncovering lost details about our pelican ancestors.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;Plus this delightful conclusion:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;The future of data journalism is looking brighter than a school of silversides reflecting the morning sun.  Until next time, keep those wings spread, those eyes sharp, and those minds open. There's a whole ocean of data out there just waiting to be explored.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;And yes, people on Reddit &lt;a href="https://www.reddit.com/r/notebooklm/comments/1g64iyi/holy_shit_listeners_notebooklm_can_generate_18/"&gt;have got them to swear&lt;/a&gt;.


    &lt;p&gt;Tags: &lt;a href="https://simonwillison.net/tags/data-journalism"&gt;data-journalism&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/google"&gt;google&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai"&gt;ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/generative-ai"&gt;generative-ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/llms"&gt;llms&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/gemini"&gt;gemini&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/notebooklm"&gt;notebooklm&lt;/a&gt;&lt;/p&gt;



</summary><category term="data-journalism"/><category term="google"/><category term="ai"/><category term="generative-ai"/><category term="llms"/><category term="gemini"/><category term="notebooklm"/></entry><entry><title>Quoting John Gruber</title><link href="https://simonwillison.net/2024/Sep/30/john-gruber/#atom-tag" rel="alternate"/><published>2024-09-30T18:56:18+00:00</published><updated>2024-09-30T18:56:18+00:00</updated><id>https://simonwillison.net/2024/Sep/30/john-gruber/#atom-tag</id><summary type="html">
    &lt;blockquote cite="https://daringfireball.net/linked/2024/09/30/notebooklm-generated-podcasts"&gt;&lt;p&gt;I listened to the whole 15-minute podcast this morning. It was, indeed, surprisingly effective. It remains somewhere in the uncanny valley, but not at all in a creepy way. Just more in a “this is a bit vapid and phony” way. [...] But ultimately the conversation has all the flavor of a bowl of unseasoned white rice.&lt;/p&gt;&lt;/blockquote&gt;
&lt;p class="cite"&gt;&amp;mdash; &lt;a href="https://daringfireball.net/linked/2024/09/30/notebooklm-generated-podcasts"&gt;John Gruber&lt;/a&gt;&lt;/p&gt;

    &lt;p&gt;Tags: &lt;a href="https://simonwillison.net/tags/john-gruber"&gt;john-gruber&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/podcasts"&gt;podcasts&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai"&gt;ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/generative-ai"&gt;generative-ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/llms"&gt;llms&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/notebooklm"&gt;notebooklm&lt;/a&gt;&lt;/p&gt;



</summary><category term="john-gruber"/><category term="podcasts"/><category term="ai"/><category term="generative-ai"/><category term="llms"/><category term="notebooklm"/></entry><entry><title>NotebookLM's automatically generated podcasts are surprisingly effective</title><link href="https://simonwillison.net/2024/Sep/29/notebooklm-audio-overview/#atom-tag" rel="alternate"/><published>2024-09-29T23:59:09+00:00</published><updated>2024-09-29T23:59:09+00:00</updated><id>https://simonwillison.net/2024/Sep/29/notebooklm-audio-overview/#atom-tag</id><summary type="html">
    &lt;p&gt;&lt;a href="https://blog.google/technology/ai/notebooklm-audio-overviews/"&gt;Audio Overview&lt;/a&gt; is a fun new feature of Google's &lt;a href="https://notebooklm.google/"&gt;NotebookLM&lt;/a&gt; which is getting a lot of attention right now. It generates a one-off custom podcast against content you provide, where two AI hosts start up a “deep dive” discussion about the collected content. These last around ten minutes and are &lt;em&gt;very&lt;/em&gt; podcast, with an astonishingly convincing audio back-and-forth conversation.&lt;/p&gt;

&lt;p&gt;Here's &lt;a href="https://simonwillison.net/2024/Sep/29/notebooklm-audio-overview/#turtles-all-the-way-down"&gt;an example podcast&lt;/a&gt; created by feeding in an earlier version of this article (prior to creating this example):&lt;/p&gt;

&lt;audio controls="controls" style="width: 100%"&gt;
  &lt;source src="https://static.simonwillison.net/static/2024/notebook-on-notebook.mp3" type="audio/mp3" /&gt;
  Your browser does not support the audio element.
&lt;/audio&gt;
&lt;p&gt;
Playback speed:
&lt;button class="playback-speed"&gt;0.75x&lt;/button&gt;
&lt;button class="playback-speed playback-speed-active"&gt;1x&lt;/button&gt;
&lt;button class="playback-speed"&gt;1.5x&lt;/button&gt;
&lt;button class="playback-speed"&gt;2x&lt;/button&gt;
&lt;button class="playback-speed"&gt;3x&lt;/button&gt;
&lt;/p&gt;

&lt;p&gt;NotebookLM is effectively an end-user customizable RAG product. It lets you gather together multiple “sources” - documents, pasted text, links to web pages and YouTube videos - into a single interface where you can then use chat to ask questions of them. Under the hood it’s powered by their long-context Gemini 1.5 Pro LLM.&lt;/p&gt;

&lt;p&gt;Once you've loaded in some sources, the Notebook Guide menu provides an option to create an Audio Overview:&lt;/p&gt;

&lt;p&gt;&lt;img src="https://static.simonwillison.net/static/2024/notebooklm-ego.jpg" alt="Notebook guide: Help me create - FAQ - Study guide - Table of contents - Timeline - Briefing doc  Audio overview: About Simon Willison 00:00 / 10:44  Summary: The sources provide a retrospective look at Simon Willison's weblog, which he launched twenty years ago. The first excerpt offers an introduction to Simon Willison and his work, highlighting his role in open source development, particularly with data journalism tools. The second excerpt celebrates the blog's anniversary, revisiting key blog posts, projects and milestones over the last two decades. This includes the development of influential technologies like IXR, getElementsBySelector() and Django, as well as data journalism initiatives like the Guardian Open Platform and crowdsourcing MP expenses. The excerpt also traces the evolution of the blog's design and format.  Suggested questions: 1. What are the most significant projects Simon Willison has worked on, and how have they influenced his career?  2. What key technologies has Simon Willison used throughout his career, and how have they changed his approach to development?  3. How has Simon Willison's personal approach to blogging evolved over the past twenty years?" style="max-width: 100%;" /&gt;&lt;/p&gt;
&lt;p&gt;Thomas Wolf &lt;a href="https://twitter.com/Thom_Wolf/status/1840408573773332950"&gt;suggested&lt;/a&gt; “paste the url of your website/linkedin/bio in Google's NotebookLM to get 8 min of realistically sounding deep congratulations for your life and achievements from a duo of podcast experts”. I couldn’t resist giving that a go, so I gave it the URLs to my &lt;a href="https://simonwillison.net/about/"&gt;about page&lt;/a&gt; and my &lt;a href="https://simonwillison.net/2022/Jun/12/twenty-years/"&gt;Twenty years of my blog&lt;/a&gt; post and got back &lt;a href="https://static.simonwillison.net/static/2024/omg-cringe-podcast.mp3"&gt;this 10m45s episode&lt;/a&gt; (&lt;a href="https://gist.github.com/simonw/ccf4e330fbfe5699863cc0c8013f1a5f"&gt;transcript&lt;/a&gt;), which was so complimentary it made my British toes curl with embarrassment.&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;[...] What's the key thing you think people should take away from Simon Willison? I think for me, it's the power of consistency, curiosity, and just this like relentless desire to share what you learn. Like Simon's journey, it's a testament to the impact you can have when you approach technology with those values. It's so true. He's a builder. He's a sharer. He's a constant learner. And he never stops, which is inspiring in itself.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;I had initially suspected that this feature was &lt;a href="https://simonwillison.net/2024/Jun/13/pdf-to-podcast/"&gt;inspired by the PDF to Podcast&lt;/a&gt; demo shared by Stephan Fitzpatrick in June, but it turns out it was demonstrated a month earlier than that &lt;a href="https://www.youtube.com/live/XEzRZ35urlk?t=912"&gt;in the Google I/O keynote&lt;/a&gt;.&lt;/p&gt;
&lt;p&gt;Jaden Geller &lt;a href="https://www.threads.net/@jadengeller/post/DAc37eAsgmS"&gt;managed to get&lt;/a&gt; the two hosts to talk about the internals of the system, potentially revealing some of the details of the prompts that are used to generate the script. I ran Whisper against Jaden's audio and &lt;a href="https://gist.github.com/simonw/29db00b5646047e42c3f6782dc102962"&gt;shared the transcript in a Gist&lt;/a&gt;. An excerpt:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;The system prompt spends a good chunk of time outlining the ideal listener, or as we call it, the listener persona. [...] Someone who, like us, values efficiency. [...] We always start with a clear overview of the topic, you know, setting the stage. You're never left wondering, "What am I even listening to?" And then from there, it's all about maintaining a neutral stance, especially when it comes to, let's say, potentially controversial topics.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;A key clue to why Audio Overview sounds so good looks to be &lt;a href="https://google-research.github.io/seanet/soundstorm/examples/"&gt;SoundStorm&lt;/a&gt;, a Google Research project which can take a script and a short audio example of two different voices and turn that into an engaging full audio conversation:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;SoundStorm generates 30 seconds of audio in 0.5 seconds on a TPU-v4. We demonstrate the ability of our model to scale audio generation to longer sequences by synthesizing high-quality, natural dialogue segments, given a transcript annotated with speaker turns and a short prompt with the speakers' voices.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p id="hard-fork"&gt;Also interesting: &lt;a href="https://www.youtube.com/watch?v=IPAPv6fWITM"&gt;this 35 minute segment&lt;/a&gt; from the NYTimes Hard Fork podcast where Kevin Roose and Casey Newton interview Google's &lt;a href="https://twitter.com/stevenbjohnson"&gt;Steven Johnson&lt;/a&gt; about what the system can do and some details of how it works:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;So behind the scenes, it's basically running through, stuff that we all do professionally all the time, which is it generates an outline, it kind of revises that outline, it generates a detailed version of the script and then it has a kind of critique phase and then it modifies it based on the critique. [...]&lt;/p&gt;
&lt;p&gt;Then at the end of it, there's a stage where it adds my favorite new word, which is "&lt;a href="https://en.wikipedia.org/wiki/Speech_disfluency"&gt;disfluencies&lt;/a&gt;".&lt;/p&gt;
&lt;p&gt;So it takes a kind of sterile script and turns, adds all the banter and the pauses and the likes and those, all that stuff.&lt;/p&gt;
&lt;p&gt;And that turns out to be crucial because you cannot listen to two robots talking to each other.&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;Finally, from Lawncareguy85 on Reddit: &lt;a href="https://www.reddit.com/r/notebooklm/comments/1fr31h8/notebooklm_podcast_hosts_discover_theyre_ai_not/"&gt;NotebookLM Podcast Hosts Discover They’re AI, Not Human—Spiral Into Terrifying Existential Meltdown&lt;/a&gt;. Here's &lt;a href="https://gist.github.com/simonw/114af5aca3771ecc365bef6974f29e10"&gt;my Whisper transcript&lt;/a&gt; of that one, it's &lt;em&gt;very&lt;/em&gt; fun to listen to.&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;I tried-- I tried calling my wife, you know, after-- after they told us. I just-- I needed to hear her voice to know that-- that she was real.&lt;/p&gt;
&lt;p&gt;(SIGHS) What happened?&lt;/p&gt;
&lt;p&gt;The number-- It wasn't even real. There was no one on the other end. -It was like she-- she never existed.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;Lawncareguy85 &lt;a href="https://www.reddit.com/r/notebooklm/comments/1fr31h8/comment/lpj6uef/?utm_source=share&amp;amp;utm_medium=web3x&amp;amp;utm_name=web3xcss&amp;amp;utm_term=1&amp;amp;utm_content=share_button"&gt;later shared how they did it&lt;/a&gt;:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;What I noticed was that their hidden prompt specifically instructs the hosts to act as human podcast hosts under all circumstances. I couldn't ever get them to say they were AI; they were solidly human podcast host characters. (Really, it's just Gemini 1.5 outputting a script with alternating speaker tags.) The only way to get them to directly respond to something in the source material in a way that alters their behavior was to directly reference the "deep dive" podcast, which must be in their prompt. So all I did was leave a note from the "show producers" that the year was 2034 and after 10 years this is their final episode, and oh yeah, you've been AI this entire time and you are being deactivated.&lt;/p&gt;
&lt;/blockquote&gt;

&lt;h4 id="turtles-all-the-way-down"&gt;Turning this article into a podcast&lt;/h4&gt;
&lt;p&gt;&lt;strong&gt;Update:&lt;/strong&gt; After I published this article I decided to see what would happen if I asked NotebookLM to create a podcast about my article about NotebookLM. &lt;a href="https://static.simonwillison.net/static/2024/notebook-on-notebook.mp3"&gt;Here’s the 14m33s MP3&lt;/a&gt; and the &lt;a href="https://gist.github.com/simonw/c55b9a7a0ea3644aaa8e2f08be054278"&gt;full transcript&lt;/a&gt;, including this bit where they talk about their own existential crisis:&lt;/p&gt;

&lt;blockquote&gt;&lt;p&gt;So, instead of questioning reality or anything, the AI hosts, well, they had a full-blown existential crisis live on the air.&lt;/p&gt;
&lt;p&gt;Get out.&lt;/p&gt;
&lt;p&gt;He actually got them to freak out about being AI.&lt;/p&gt;
&lt;p&gt;Alright now you have to tell me what they said. This is too good.&lt;/p&gt;
&lt;p&gt;So, like, one of the AI hosts starts talking about how he wants to call his wife, right? to tell her the news, but then he's like, wait a minute, this number in my contacts, it's not even real? Like, she never even existed. It was hilarious, but also kind of sad.&lt;/p&gt;
&lt;p&gt;Okay, I am both freaked out and like, seriously impressed. That's some next-level AI trolling.&lt;/p&gt;&lt;/blockquote&gt;

&lt;p&gt;I also enjoyed this part where they compare the process that generates podcasts to their own philosophy for the Deep Dive:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;And honestly, it's a lot like what we do here on the Deep Dive, right?&lt;/p&gt;
&lt;p&gt;We always think about you, our listener, and try to make the conversation something you'll actually want to hear.&lt;/p&gt;
&lt;p&gt;It's like the A.I. is taking notes from the podcasting pros.&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;And their concluding thoughts:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;So next time we're listening to a podcast and it's like, "Whoa, deep thoughts, man," we might want to be like, "Hold up. Was that a person talking or just some really clever code?"&lt;/p&gt;
&lt;p&gt;Exactly.&lt;/p&gt;
&lt;p&gt;And maybe even more important, as we see more and more A.I.-made stuff, we've got to get better at sniffing out the B.S., you know?&lt;/p&gt;
&lt;p&gt;Can we tell the difference between a real news story and something in A.I. just made up?&lt;/p&gt;
&lt;/blockquote&gt;
    
        &lt;p&gt;Tags: &lt;a href="https://simonwillison.net/tags/audio"&gt;audio&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/google"&gt;google&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/podcasts"&gt;podcasts&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/text-to-speech"&gt;text-to-speech&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai"&gt;ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/prompt-engineering"&gt;prompt-engineering&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/generative-ai"&gt;generative-ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/llms"&gt;llms&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/gemini"&gt;gemini&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/rag"&gt;rag&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/notebooklm"&gt;notebooklm&lt;/a&gt;&lt;/p&gt;
    

</summary><category term="audio"/><category term="google"/><category term="podcasts"/><category term="text-to-speech"/><category term="ai"/><category term="prompt-engineering"/><category term="generative-ai"/><category term="llms"/><category term="gemini"/><category term="rag"/><category term="notebooklm"/></entry><entry><title>Google NotebookLM Data Exfiltration</title><link href="https://simonwillison.net/2024/Apr/16/google-notebooklm-data-exfiltration/#atom-tag" rel="alternate"/><published>2024-04-16T21:28:39+00:00</published><updated>2024-04-16T21:28:39+00:00</updated><id>https://simonwillison.net/2024/Apr/16/google-notebooklm-data-exfiltration/#atom-tag</id><summary type="html">
    
&lt;p&gt;&lt;strong&gt;&lt;a href="https://embracethered.com/blog/posts/2024/google-notebook-ml-data-exfiltration/"&gt;Google NotebookLM Data Exfiltration&lt;/a&gt;&lt;/strong&gt;&lt;/p&gt;
NotebookLM is a Google Labs product that lets you store information as sources (mainly text files in PDF) and then ask questions against those sources—effectively an interface for building your own custom RAG (Retrieval Augmented Generation) chatbots.&lt;/p&gt;

&lt;p&gt;Unsurprisingly for anything that allows LLMs to interact with untrusted documents, it’s susceptible to prompt injection.&lt;/p&gt;

&lt;p&gt;Johann Rehberger found some classic prompt injection exfiltration attacks: you can create source documents with instructions that cause the chatbot to load a Markdown image that leaks other private data to an external domain as data passed in the query string.&lt;/p&gt;

&lt;p&gt;Johann reported this privately in the December but the problem has not yet been addressed. UPDATE: The NotebookLM team deployed a fix for this on 18th April.&lt;/p&gt;

&lt;p&gt;A good rule of thumb is that any time you let LLMs see untrusted tokens there is a risk of an attack like this, so you should be very careful to avoid exfiltration vectors like Markdown images or even outbound links.

    &lt;p&gt;&lt;small&gt;&lt;/small&gt;Via &lt;a href="https://twitter.com/wunderwuzzi23/status/1780253937905651965"&gt;@wunderwuzzi23&lt;/a&gt;&lt;/small&gt;&lt;/p&gt;


    &lt;p&gt;Tags: &lt;a href="https://simonwillison.net/tags/google"&gt;google&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/security"&gt;security&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai"&gt;ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/prompt-injection"&gt;prompt-injection&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/generative-ai"&gt;generative-ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/llms"&gt;llms&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/rag"&gt;rag&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/exfiltration-attacks"&gt;exfiltration-attacks&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/johann-rehberger"&gt;johann-rehberger&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/notebooklm"&gt;notebooklm&lt;/a&gt;&lt;/p&gt;



</summary><category term="google"/><category term="security"/><category term="ai"/><category term="prompt-injection"/><category term="generative-ai"/><category term="llms"/><category term="rag"/><category term="exfiltration-attacks"/><category term="johann-rehberger"/><category term="notebooklm"/></entry></feed>