<?xml version="1.0" encoding="utf-8"?>
<feed xml:lang="en-us" xmlns="http://www.w3.org/2005/Atom"><title>Simon Willison's Weblog: trust</title><link href="http://simonwillison.net/" rel="alternate"/><link href="http://simonwillison.net/tags/trust.atom" rel="self"/><id>http://simonwillison.net/</id><updated>2024-06-10T20:19:13+00:00</updated><author><name>Simon Willison</name></author><entry><title>Thoughts on the WWDC 2024 keynote on Apple Intelligence</title><link href="https://simonwillison.net/2024/Jun/10/apple-intelligence/#atom-tag" rel="alternate"/><published>2024-06-10T20:19:13+00:00</published><updated>2024-06-10T20:19:13+00:00</updated><id>https://simonwillison.net/2024/Jun/10/apple-intelligence/#atom-tag</id><summary type="html">
    &lt;p&gt;Today's WWDC keynote finally revealed Apple's new set of AI features. The AI section (Apple are calling it Apple Intelligence) started over an hour into the keynote - &lt;a href="https://www.youtube.com/live/RXeOiIDNNek?t=3870s"&gt;this link&lt;/a&gt; jumps straight to that point in the archived YouTube livestream, or you can watch it embedded here:&lt;/p&gt;

&lt;iframe style="max-width: 100%" width="560" height="315" src="https://www.youtube-nocookie.com/embed/RXeOiIDNNek?start=3870" frameborder="0" allow="autoplay; encrypted-media" allowfullscreen="allowfullscreen"&gt; &lt;/iframe&gt;

&lt;p&gt;There's also a detailed Apple newsroom post: &lt;a href="https://www.apple.com/newsroom/2024/06/introducing-apple-intelligence-for-iphone-ipad-and-mac/"&gt;Introducing Apple Intelligence, the personal intelligence system that puts powerful generative models at the core of iPhone, iPad, and Mac&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;There are a lot of interesting things here. Apple have a strong focus on privacy, finally taking advantage of the Neural Engine accelerator chips in the A17 Pro chip on iPhone 15 Pro and higher and the M1/M2/M3 Apple Silicon chips in Macs. They're using these to run on-device models - I've not yet seen any information on which models they are running and how they were trained.&lt;/p&gt;
&lt;h4 id="on-device-models"&gt;On-device models that can outsource to Apple's servers&lt;/h4&gt;
&lt;p&gt;Most notable is their approach to features that don't work with an on-device model. At &lt;a href="https://www.youtube.com/live/RXeOiIDNNek?t=4483s"&gt;1h14m43s&lt;/a&gt;:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;When you make a request, Apple Intelligence analyses whether it can be processed on device. If it needs greater computational capacity, it can draw on Private Cloud Compute, and send only the data that's relevant to your task to be processed on Apple Silicon servers.&lt;/p&gt;
&lt;p&gt;Your data is never stored or made accessible to Apple. It's used exclusively to fulfill your request.&lt;/p&gt;
&lt;p&gt;And just like your iPhone, independent experts can inspect the code that runs on the servers to verify this privacy promise.&lt;/p&gt;
&lt;p&gt;In fact, Private Cloud Compute cryptographically ensures your iPhone, iPad, and Mac will refuse to talk to a server unless its software has been publicly logged for inspection.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;There's some fascinating computer science going on here! I'm looking forward to learning more about this - it sounds like the details will be public by design, since that's key to the promise they are making here.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Update&lt;/strong&gt;: Here are &lt;a href="https://security.apple.com/blog/private-cloud-compute/"&gt;the details&lt;/a&gt;, and they are indeed extremely impressive - more of &lt;a href="https://simonwillison.net/2024/Jun/11/private-cloud-compute/"&gt;my notes here&lt;/a&gt;.&lt;/p&gt;
&lt;h4 id="ethical-ai-images"&gt;An ethical approach to AI generated images?&lt;/h4&gt;
&lt;p&gt;Their approach to generative images is notable in that they're shipping an on-device model in a feature called Image Playground, with a very important limitation: it can only output images in one of three styles: sketch, illustration and animation.&lt;/p&gt;
&lt;p&gt;This feels like a clever way to address some of the ethical objections people have to this specific category of AI tool:&lt;/p&gt;
&lt;ol&gt;
&lt;li&gt;If you can't create photorealistic images, you can't generate deepfakes or offensive photos of people&lt;/li&gt;
&lt;li&gt;By having obvious visual styles you ensure that AI generated images are instantly recognizable as such, without watermarks or similar&lt;/li&gt;
&lt;li&gt;Avoiding the ability to clone specific artist's styles further helps sidestep ethical issues about plagiarism and copyright infringement&lt;/li&gt;
&lt;/ol&gt;
&lt;p&gt;The social implications of this are interesting too. Will people be more likely to share AI-generated images if there are no awkward questions or doubts about how they were created, and will that help it more become socially acceptable to use them?&lt;/p&gt;
&lt;p&gt;I've not seen anything on how these image models were trained. Given their limited styles it seems possible Apple used entirely ethically licensed training data, but I'd like to see more details on this.&lt;/p&gt;
&lt;h4 id="app-intents-prompt-injection"&gt;App Intents and prompt injection&lt;/h4&gt;
&lt;p&gt;Siri will be able to both access data on your device and trigger actions based on your instructions.&lt;/p&gt;
&lt;p&gt;This is the exact feature combination that's &lt;a href="https://simonwillison.net/2023/Apr/14/worst-that-can-happen/#rogue-assistant"&gt;most at risk from prompt injection attacks&lt;/a&gt;: what happens if someone sends you a text message that tricks Siri into forwarding a password reset email to them, and you ask for a summary of that message?&lt;/p&gt;
&lt;p&gt;Security researchers will no doubt jump straight onto this as soon as the beta becomes available. I'm fascinated to learn what Apple have done to mitigate this risk.&lt;/p&gt;
&lt;h4 id="siri-plus-chatgpt"&gt;Integration with ChatGPT&lt;/h4&gt;
&lt;p&gt;Rumors broke last week that Apple had signed a deal with OpenAI to use ChatGPT. That's now been confirmed: here's &lt;a href="https://openai.com/index/openai-and-apple-announce-partnership/"&gt;OpenAI's partnership announcement&lt;/a&gt;:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;Apple is integrating ChatGPT into experiences within iOS, iPadOS, and macOS, allowing users to access ChatGPT’s capabilities—including image and document understanding—without needing to jump between tools.&lt;/p&gt;
&lt;p&gt;Siri can also tap into ChatGPT’s intelligence when helpful. Apple users are asked before any questions are sent to ChatGPT, along with any documents or photos, and Siri then presents the answer directly.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;The keynote talks about that at &lt;a href="https://www.youtube.com/live/RXeOiIDNNek?t=5781s"&gt;1h36m21s&lt;/a&gt;. Those prompts to confirm that the user wanted to share data with ChatGPT are very prominent in the demo!&lt;/p&gt;
&lt;p&gt;&lt;img src="https://static.simonwillison.net/static/2024/siri-chatgpt-loop.gif" alt="Animated screenshot. User says to Siri: I have fresh salmon, lemons, tomatoes. Help me plan a 5-course meal with a dish for each taste bud. Siri shows a dialog Do you want me to use ChatGPT to do that? User clicks Use ChatGPT and gets a generated response." style="max-width: 100%;" /&gt;&lt;/p&gt;
&lt;p&gt;This integration (with GPT-4o) will be free - and Apple don't appear to be charging for their other server-side AI features either. I guess they expect the supporting hardware sales to more than cover the costs of running these models.&lt;/p&gt;
    
        &lt;p&gt;Tags: &lt;a href="https://simonwillison.net/tags/apple"&gt;apple&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ethics"&gt;ethics&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/privacy"&gt;privacy&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/security"&gt;security&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/trust"&gt;trust&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai"&gt;ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/openai"&gt;openai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/prompt-injection"&gt;prompt-injection&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/generative-ai"&gt;generative-ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/chatgpt"&gt;chatgpt&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/llms"&gt;llms&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/apple-intelligence"&gt;apple-intelligence&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai-ethics"&gt;ai-ethics&lt;/a&gt;&lt;/p&gt;
    

</summary><category term="apple"/><category term="ethics"/><category term="privacy"/><category term="security"/><category term="trust"/><category term="ai"/><category term="openai"/><category term="prompt-injection"/><category term="generative-ai"/><category term="chatgpt"/><category term="llms"/><category term="apple-intelligence"/><category term="ai-ethics"/></entry><entry><title>Update on the Recall preview feature for Copilot+ PCs</title><link href="https://simonwillison.net/2024/Jun/7/update-on-the-recall-preview/#atom-tag" rel="alternate"/><published>2024-06-07T17:30:40+00:00</published><updated>2024-06-07T17:30:40+00:00</updated><id>https://simonwillison.net/2024/Jun/7/update-on-the-recall-preview/#atom-tag</id><summary type="html">
    
&lt;p&gt;&lt;strong&gt;&lt;a href="https://blogs.windows.com/windowsexperience/2024/06/07/update-on-the-recall-preview-feature-for-copilot-pcs/"&gt;Update on the Recall preview feature for Copilot+ PCs&lt;/a&gt;&lt;/strong&gt;&lt;/p&gt;
This feels like a very good call to me: in response to &lt;a href="https://simonwillison.net/2024/Jun/1/stealing-everything-youve-ever-typed/"&gt;widespread criticism&lt;/a&gt; Microsoft are making Recall an opt-in feature (during system onboarding), adding encryption to the database and search index beyond just disk encryption and requiring Windows Hello face scanning to access the search feature.

    &lt;p&gt;&lt;small&gt;&lt;/small&gt;Via &lt;a href="https://www.wired.com/story/microsoft-recall-off-default-security-concerns/"&gt;Wired: Microsoft Will Switch Off Recall by Default After Security Backlash&lt;/a&gt;&lt;/small&gt;&lt;/p&gt;


    &lt;p&gt;Tags: &lt;a href="https://simonwillison.net/tags/microsoft"&gt;microsoft&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/privacy"&gt;privacy&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/security"&gt;security&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/trust"&gt;trust&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/windows"&gt;windows&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai"&gt;ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/recall"&gt;recall&lt;/a&gt;&lt;/p&gt;



</summary><category term="microsoft"/><category term="privacy"/><category term="security"/><category term="trust"/><category term="windows"/><category term="ai"/><category term="recall"/></entry><entry><title>Quoting Zac Bowden</title><link href="https://simonwillison.net/2024/Jun/7/zac-bowden/#atom-tag" rel="alternate"/><published>2024-06-07T17:23:54+00:00</published><updated>2024-06-07T17:23:54+00:00</updated><id>https://simonwillison.net/2024/Jun/7/zac-bowden/#atom-tag</id><summary type="html">
    &lt;blockquote cite="https://www.windowscentral.com/software-apps/windows-11/microsoft-has-lost-trust-with-its-users-windows-recall-is-the-last-straw"&gt;&lt;p&gt;In fact, Microsoft goes so far as to promise that it cannot see the data collected by Windows Recall, that it can't train any of its AI models on your data, and that it definitely can't sell that data to advertisers. All of this is true, but that doesn't mean people believe Microsoft when it says these things. In fact, many have jumped to the conclusion that even if it's true today, it won't be true in the future.&lt;/p&gt;&lt;/blockquote&gt;
&lt;p class="cite"&gt;&amp;mdash; &lt;a href="https://www.windowscentral.com/software-apps/windows-11/microsoft-has-lost-trust-with-its-users-windows-recall-is-the-last-straw"&gt;Zac Bowden&lt;/a&gt;&lt;/p&gt;

    &lt;p&gt;Tags: &lt;a href="https://simonwillison.net/tags/microsoft"&gt;microsoft&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/privacy"&gt;privacy&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/trust"&gt;trust&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/windows"&gt;windows&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai"&gt;ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/recall"&gt;recall&lt;/a&gt;&lt;/p&gt;



</summary><category term="microsoft"/><category term="privacy"/><category term="trust"/><category term="windows"/><category term="ai"/><category term="recall"/></entry><entry><title>The AI trust crisis</title><link href="https://simonwillison.net/2023/Dec/14/ai-trust-crisis/#atom-tag" rel="alternate"/><published>2023-12-14T16:14:11+00:00</published><updated>2023-12-14T16:14:11+00:00</updated><id>https://simonwillison.net/2023/Dec/14/ai-trust-crisis/#atom-tag</id><summary type="html">
    &lt;p&gt;Dropbox added some &lt;a href="https://help.dropbox.com/view-edit/dropbox-ai-how-to"&gt;new AI features&lt;/a&gt;. In the past couple of days these have attracted a firestorm of criticism. Benj Edwards rounds it up in  &lt;a href="https://arstechnica.com/information-technology/2023/12/dropbox-spooks-users-by-sending-data-to-openai-for-ai-search-features/"&gt;Dropbox spooks users with new AI features that send data to OpenAI when used&lt;/a&gt;.&lt;/p&gt;
&lt;p&gt;The key issue here is that people are worried that their private files on Dropbox are being passed to OpenAI to use as training data for their models - a claim that is strenuously denied by Dropbox.&lt;/p&gt;
&lt;p&gt;As far as I can tell, Dropbox built some sensible features - summarize on demand, "chat with your data" via Retrieval Augmented Generation - and did a moderately OK job of communicating how they work... but when it comes to data privacy and AI, a "moderately OK job" is a failing grade. Especially if you hold as much of people's private data as Dropbox does!&lt;/p&gt;
&lt;p&gt;Two details in particular seem really important. Dropbox have an &lt;a href="https://www.dropbox.com/ai-principles"&gt;AI principles document&lt;/a&gt; which includes this:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;Customer trust and the privacy of their data are our foundation. We will not use customer data to train AI models without consent.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;They also have a checkbox &lt;a href="https://www.dropbox.com/account/ai"&gt;in their settings&lt;/a&gt; that looks like this:&lt;/p&gt;
&lt;p&gt;&lt;img src="https://static.simonwillison.net/static/2023/dropbox-third-party.png" alt="Third-party AI: Use artificial intelligence (Al) from third-party partners so you can work faster in Dropbox. We only use technology partners we have vetted. Your data is never used to train their internal models, and is deleted from third-party servers within 30 days. Learn more. There is a toggle set to On." style="max-width: 100%;" /&gt;&lt;/p&gt;
&lt;p&gt;&lt;em&gt;&lt;strong&gt;Update:&lt;/strong&gt; Some time between me publishing this article and four hours later, that link stopped working.&lt;/em&gt;&lt;/p&gt;
&lt;p&gt;I took that screenshot on my own account. It's toggled "on" - but I never turned it on myself.&lt;/p&gt;
&lt;p&gt;Does that mean I'm marked as "consenting" to having my data used to train AI models?&lt;/p&gt;
&lt;p&gt;I don't think so: I think this is a combination of confusing wording and the eternal vagueness of what the term "consent" means in a world where everyone agrees to the terms and conditions of everything without reading them.&lt;/p&gt;
&lt;p&gt;But a LOT of people have come to the conclusion that this means their private data - which they pay Dropbox to protect - is now being funneled into the OpenAI training abyss.&lt;/p&gt;
&lt;h4 id="people-dont-believe-openai"&gt;People don't believe OpenAI&lt;/h4&gt;
&lt;p&gt;Here's copy from that Dropbox preference box, talking about their "third-party partners" - in this case OpenAI:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;Your data is never used to train their internal models, and is deleted from third-party servers within 30 days.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;It's increasing clear to me like people simply &lt;strong&gt;don't believe OpenAI&lt;/strong&gt; when they're told that data won't be used for training.&lt;/p&gt;
&lt;p&gt;What's really going on here is something deeper then: AI is facing a crisis of trust.&lt;/p&gt;
&lt;p&gt;I quipped &lt;a href="https://twitter.com/simonw/status/1735086765814542802"&gt;on Twitter&lt;/a&gt;:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;"OpenAI are training on every piece of data they see, even when they say they aren't" is the new "Facebook are showing you ads based on overhearing everything you say through your phone's microphone"&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;Here's what I meant by that.&lt;/p&gt;
&lt;h4 id="facebook-dont-spy-microphone"&gt;Facebook don't spy on you through your microphone&lt;/h4&gt;
&lt;p&gt;Have you heard the one about Facebook spying on you through your phone's microphone and showing you ads based on what you're talking about?&lt;/p&gt;
&lt;p&gt;This theory has been floating around for years. From a technical perspective it should be easy to disprove:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;Mobile phone operating systems don't allow apps to invisibly access the microphone.&lt;/li&gt;
&lt;li&gt;Privacy researchers can audit communications between devices and Facebook to confirm if this is happening.&lt;/li&gt;
&lt;li&gt;Running high quality voice recognition like this at scale is extremely expensive - I had a conversation with a friend who works on server-based machine learning at Apple a few years ago who found the entire idea laughable.&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;The non-technical reasons are even stronger:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;Facebook say they aren't doing this. The risk to their reputation if they are caught in a lie is astronomical.&lt;/li&gt;
&lt;li&gt;As with many conspiracy theories, too many people would have to be "in the loop" and not blow the whistle.&lt;/li&gt;
&lt;li&gt;Facebook don't need to do this: there are much, much cheaper and more effective ways to target ads at you than spying through your microphone. These methods have been working incredibly well for years.&lt;/li&gt;
&lt;li&gt;Facebook gets to show us thousands of ads a year. 99% of those don't correlate in the slightest to anything we have said out loud. If you keep rolling the dice long enough, eventually a coincidence will strike.&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;Here's the thing though: &lt;em&gt;none of these arguments matter&lt;/em&gt;.&lt;/p&gt;
&lt;p&gt;If you've ever experienced Facebook showing you an ad for something that you were talking about out-loud about moments earlier, you've already dismissed everything I just said. You have personally experienced anecdotal evidence which overrides all of my arguments here.&lt;/p&gt;
&lt;p&gt;Here's a Reply All podcast episode from Novemember 2017 that explores this issue: &lt;a href="https://gimletmedia.com/shows/reply-all/z3hlwr"&gt;109 Is Facebook Spying on You?&lt;/a&gt;. Their conclusion: Facebook are not spying through your microphone. But if someone already believes that there is no argument that can possibly convince them otherwise.&lt;/p&gt;
&lt;p&gt;I've experienced this effect myself - over the past few years I've tried talking people out of this, as part of my own personal fascination with how sticky this conspiracy theory is.&lt;/p&gt;
&lt;p&gt;The key issue here is the same as the OpenAI training issue: people &lt;strong&gt;don't believe&lt;/strong&gt; these companies when they say that they aren't doing something.&lt;/p&gt;
&lt;p&gt;One interesting difference here is that in the Facebook example people have personal evidence that makes them believe they understand what's going on.&lt;/p&gt;
&lt;p&gt;With AI we have almost the complete opposite: AI models are weird black boxes, built in secret and with no way of understanding what the training data was or how it influences the model.&lt;/p&gt;
&lt;p&gt;As with so much in AI, people are left with nothing more than "vibes" to go on. And the vibes are bad.&lt;/p&gt;
&lt;h4 id="this-really-matters"&gt;This really matters&lt;/h4&gt;
&lt;p&gt;Trust is really important. Companies lying about what they do with your privacy is a very serious allegation.&lt;/p&gt;
&lt;p&gt;A society where big companies tell blatant lies about how they are handling our data - and get away with it without consequences - is a very unhealthy society.&lt;/p&gt;
&lt;p&gt;A key role of government is to prevent this from happening. If OpenAI are training on data that they said they wouldn't train on, or if Facebook are spying on us through our phone's microphones, they should be hauled in front of regulators and/or sued into the ground.&lt;/p&gt;
&lt;p&gt;If we believe that they are doing this without consequence, and have been getting away with it for years, our intolerance for corporate misbehavior becomes a victim as well. We risk letting companies get away with real misconduct because we incorrectly believed in conspiracy theories.&lt;/p&gt;
&lt;p&gt;Privacy is important, and very easily misunderstood. People both overestimate and underestimate what companies are doing, and what's possible. This isn't helped by the fact that AI technology means the scope of what's possible is changing at a rate that's hard to appreciate even if you're deeply aware of the space.&lt;/p&gt;
&lt;p&gt;If we want to protect our privacy, we need to understand what's going on. More importantly, we need to be able to trust companies to honestly and clearly explain what they are doing with our data.&lt;/p&gt;
&lt;p&gt;On a personal level we risk losing out on useful tools. How many people cancelled their Dropbox accounts in the last 48 hours? How many more turned off that AI toggle, ruling out ever evaluating if those features were useful for them or not?&lt;/p&gt;
&lt;h4 id="what-can-we-do"&gt;What can we do about it?&lt;/h4&gt;
&lt;p&gt;There is something that the big AI labs could be doing to help here: tell us how you are training!&lt;/p&gt;
&lt;p&gt;The fundamental question here is about training data: what are OpenAI using to train their models?&lt;/p&gt;
&lt;p&gt;And the answer is: we have no idea! The entire process could not be more opaque.&lt;/p&gt;
&lt;p&gt;Given that, is it any wonder that when OpenAI say "we don't train on data submitted via our API" people have trouble believing them?&lt;/p&gt;
&lt;p&gt;The situation with ChatGPT itself is even more messy. OpenAI say that they DO use ChatGPT interactions to improve their models - even those from paying customers, with the exception of the "call us" priced &lt;a href="https://openai.com/blog/introducing-chatgpt-enterprise"&gt;ChatGPT Enterprise&lt;/a&gt;.&lt;/p&gt;
&lt;p&gt;If I paste a private document into ChatGPT to ask for a summary, will snippets of that document be leaked to future users after the next model update? Without more details on HOW they are using ChatGPT to improve their models I can't come close to answering that question.&lt;/p&gt;
&lt;p&gt;Clear explanations of how this stuff works could go a long way to improving the trust relationship OpenAI have with their users, and the world at large.&lt;/p&gt;
&lt;p&gt;Maybe take a leaf from large scale platform companies. They publish public post-mortem incident reports on outages, to regain trust with their customers through transparency about exactly what happened and the steps they are taking to prevent it from happening again. Dan Luu has collected a &lt;a href="https://github.com/danluu/post-mortems"&gt;great list of examples&lt;/a&gt;.&lt;/p&gt;
&lt;h4 id="opportunity-local-models"&gt;An opportunity for local models&lt;/h4&gt;
&lt;p&gt;One consistent theme I've seen in conversations about this issue is that people are much more comfortable trusting their data to local models that run on their own devices than models hosted in the cloud.&lt;/p&gt;
&lt;p&gt;The good news is that local models are consistently both increasing in quality and shrinking in size.&lt;/p&gt;
&lt;p&gt;I figured out how to run Mixtral-8x7b-Instruct &lt;a href="https://fedi.simonwillison.net/@simon/111577242044966329"&gt;on my laptop&lt;/a&gt; last night - the first local model I've tried which really does seem to be equivalent in quality to ChatGPT 3.5.&lt;/p&gt;
&lt;p&gt;Microsoft's &lt;a href="https://www.microsoft.com/en-us/research/blog/phi-2-the-surprising-power-of-small-language-models/"&gt;Phi-2&lt;/a&gt; is a fascinating new model in that it's only 2.7 billion parameters (most useful local models start at 7 billion) but claims state-of-the-art performance against some of those larger models. And it looks like they trained it for around $35,000.&lt;/p&gt;
&lt;p&gt;While I'm excited about the potential of local models, I'd hate to see us lose out on the power and convenience of the larger hosted models over privacy concerns which turn out to be incorrect.&lt;/p&gt;
&lt;p&gt;The intersection of AI and privacy is a critical issue. We need to be able to have the highest quality conversations about it, with maximum transparency and understanding of what's actually going on.&lt;/p&gt;
&lt;p&gt;This is hard already, and it's made even harder if we straight up disbelieve anything that companies tell us. Those companies need to earn our trust. How can we help them understand how to do that?&lt;/p&gt;
    
        &lt;p&gt;Tags: &lt;a href="https://simonwillison.net/tags/trust"&gt;trust&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/dropbox"&gt;dropbox&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai"&gt;ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/openai"&gt;openai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/local-llms"&gt;local-llms&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/llms"&gt;llms&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/training-data"&gt;training-data&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/microphone-ads-conspiracy"&gt;microphone-ads-conspiracy&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/digital-literacy"&gt;digital-literacy&lt;/a&gt;&lt;/p&gt;
    

</summary><category term="trust"/><category term="dropbox"/><category term="ai"/><category term="openai"/><category term="local-llms"/><category term="llms"/><category term="training-data"/><category term="microphone-ads-conspiracy"/><category term="digital-literacy"/></entry><entry><title>AI and Trust</title><link href="https://simonwillison.net/2023/Dec/5/ai-and-trust/#atom-tag" rel="alternate"/><published>2023-12-05T21:43:03+00:00</published><updated>2023-12-05T21:43:03+00:00</updated><id>https://simonwillison.net/2023/Dec/5/ai-and-trust/#atom-tag</id><summary type="html">
    
&lt;p&gt;&lt;strong&gt;&lt;a href="https://www.schneier.com/blog/archives/2023/12/ai-and-trust.html"&gt;AI and Trust&lt;/a&gt;&lt;/strong&gt;&lt;/p&gt;
Barnstormer of an essay by Bruce Schneier about AI and trust. It’s worth spending some time with this—it’s hard to extract the highlights since there are so many of them.&lt;/p&gt;

&lt;p&gt;A key idea is that we are predisposed to trust AI chat interfaces because they imitate humans, which means we are highly susceptible to profit-seeking biases baked into them.&lt;/p&gt;

&lt;p&gt;Bruce suggests that what’s needed is public models, backed by government funds: “A public model is a model built by the public for the public. It requires political accountability, not just market accountability.”


    &lt;p&gt;Tags: &lt;a href="https://simonwillison.net/tags/bruce-schneier"&gt;bruce-schneier&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/trust"&gt;trust&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai"&gt;ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/generative-ai"&gt;generative-ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/llms"&gt;llms&lt;/a&gt;&lt;/p&gt;



</summary><category term="bruce-schneier"/><category term="trust"/><category term="ai"/><category term="generative-ai"/><category term="llms"/></entry><entry><title>Can We Trust Search Engines with Generative AI? A Closer Look at Bing’s Accuracy for News Queries</title><link href="https://simonwillison.net/2023/Feb/18/can-we-trust-search-engines-with-generative-ai/#atom-tag" rel="alternate"/><published>2023-02-18T18:09:19+00:00</published><updated>2023-02-18T18:09:19+00:00</updated><id>https://simonwillison.net/2023/Feb/18/can-we-trust-search-engines-with-generative-ai/#atom-tag</id><summary type="html">
    
&lt;p&gt;&lt;strong&gt;&lt;a href="https://medium.com/@ndiakopoulos/can-we-trust-search-engines-with-generative-ai-a-closer-look-at-bings-accuracy-for-news-queries-179467806bcc"&gt;Can We Trust Search Engines with Generative AI? A Closer Look at Bing’s Accuracy for News Queries&lt;/a&gt;&lt;/strong&gt;&lt;/p&gt;
Computational journalism professor Nick Diakopoulos takes a deeper dive into the quality of the summarizations provided by AI-assisted Bing. His findings are troubling: for news queries, which are a great test for AI summarization since they include recent information that may have sparse or conflicting stories, Bing confidently produces answers with important errors: claiming the Ohio train derailment happened on February 9th when it actually happened on February 3rd for example.

    &lt;p&gt;&lt;small&gt;&lt;/small&gt;Via &lt;a href="https://twitter.com/ndiakopoulos/status/1626840648002203649"&gt;@ndiakopoulos&lt;/a&gt;&lt;/small&gt;&lt;/p&gt;


    &lt;p&gt;Tags: &lt;a href="https://simonwillison.net/tags/bing"&gt;bing&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/search"&gt;search&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/trust"&gt;trust&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/generative-ai"&gt;generative-ai&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/llms"&gt;llms&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ai-assisted-search"&gt;ai-assisted-search&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/digital-literacy"&gt;digital-literacy&lt;/a&gt;&lt;/p&gt;



</summary><category term="bing"/><category term="search"/><category term="trust"/><category term="generative-ai"/><category term="llms"/><category term="ai-assisted-search"/><category term="digital-literacy"/></entry><entry><title>Wikipedia trust colouring (with demo)</title><link href="https://simonwillison.net/2007/Sep/1/ucsc/#atom-tag" rel="alternate"/><published>2007-09-01T01:42:59+00:00</published><updated>2007-09-01T01:42:59+00:00</updated><id>https://simonwillison.net/2007/Sep/1/ucsc/#atom-tag</id><summary type="html">
    
&lt;p&gt;&lt;strong&gt;&lt;a href="http://trust.cse.ucsc.edu/"&gt;Wikipedia trust colouring (with demo)&lt;/a&gt;&lt;/strong&gt;&lt;/p&gt;
“The text background of Wikipedia articles is colored according to a value of trust, computed from the reputation of the authors who contributed the text, as well as those who edited the text.”

    &lt;p&gt;&lt;small&gt;&lt;/small&gt;Via &lt;a href="http://blog.k1v1n.com/2007/08/trustworthiness-of-wikipedia-pages.html"&gt;Kevin Gamble&lt;/a&gt;&lt;/small&gt;&lt;/p&gt;


    &lt;p&gt;Tags: &lt;a href="https://simonwillison.net/tags/kevin-gamble"&gt;kevin-gamble&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/trust"&gt;trust&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/ucsc"&gt;ucsc&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/wikipedia"&gt;wikipedia&lt;/a&gt;&lt;/p&gt;



</summary><category term="kevin-gamble"/><category term="trust"/><category term="ucsc"/><category term="wikipedia"/></entry><entry><title>An OpenID is not an account!</title><link href="https://simonwillison.net/2007/Jan/10/account/#atom-tag" rel="alternate"/><published>2007-01-10T10:53:35+00:00</published><updated>2007-01-10T10:53:35+00:00</updated><id>https://simonwillison.net/2007/Jan/10/account/#atom-tag</id><summary type="html">
    &lt;p&gt;I'm excited to see that OpenID has finally started to gain serious traction outside of the Identity community. Understandably, misconceptions about OpenID continue to crop-up. The one I want to address in this entry is the idea that an OpenID can be used as a replacement for a regular user account.&lt;/p&gt;

&lt;p&gt;&lt;em&gt;&lt;strong&gt;Update at 23:55pm:&lt;/strong&gt; I originally tried to illustrate this misconception with a quote &lt;a href="http://www.docuverse.com/blog/donpark/2007/01/09/openid-blues"&gt;from Don Park&lt;/a&gt;; unfortunately I misunderstood the quote and consequently misrepresented his position, for which I apologise unreservedly.&lt;/em&gt;&lt;/p&gt;
 
&lt;p&gt;The &lt;a href="http://web.archive.org/web/20050521002209/www.danga.com/openid/"&gt;old OpenID homepage&lt;/a&gt; (which I miss; the &lt;a href="http://openid.net/"&gt;new one&lt;/a&gt; uses jargon-heavy terms like "a free framework for user-centric digital identity") included this in nice big letters:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;&lt;strong&gt;What about trust?&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;This is not a trust system. Trust requires identity first.&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;OpenID solves the identity problem, not the trust problem. When a user authenticates with OpenID, what they are doing is stating &lt;strong&gt;"I have the ability to prove my ownership of this URL"&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;I used the phrase "have the ability" deliberately. Just because the OpenID authentication was successful doesn't mean that the user is the only person who can authenticate against that OpenID. It would be trivial to create the OpenID equivalent of &lt;a href="http://www.mailinator.com/"&gt;Mailinator&lt;/a&gt;: an identity provider that says "Yes, that's them!" to any identity request. I'm tempted to build it just to emphasize this point! &lt;strong&gt;Update:&lt;/strong&gt; Jayant Gandhi &lt;a href="http://www.jkg.in/openid/"&gt;has built one&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;The key thing here is that you should never trust an OpenID on its own. It could be a real person, or it could be a spammer, psycopath or general undesirable.&lt;/p&gt;

&lt;p&gt;Does this mean OpenID is completely useless? Absolutely not! You just have to think of it in the same way that you think of username and password combinations: as the "key" to an account.&lt;/p&gt;

&lt;p&gt;Most web application signup processes work something like this:&lt;/p&gt;

&lt;ol&gt;
 &lt;li&gt;Bob selects a username&lt;/li&gt;
 &lt;li&gt;Bob enters a password, twice&lt;/li&gt;
 &lt;li&gt;Bob enters his e-mail address&lt;/li&gt;
 &lt;li&gt;Bob clicks a validation link in an e-mail sent to that address&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;Some sites throw a CAPTCHA in there for good measure.&lt;/p&gt;

&lt;p&gt;OpenID replaces at most the first two steps of that registration process. Instead of having a user set up a new password you get them to authenticate with their OpenID &lt;em&gt;at the start of the process&lt;/em&gt;. After that you might still want them to pick a username (especially if you are integrating OpenID in to an existing account system) and you'll almost certainly want them to jump through the e-mail and/or CAPTCHA steps.&lt;/p&gt;

&lt;p&gt;In the future, they can sign in to your site using their OpenID rather than having to dig around for whichever username and password they used.&lt;/p&gt;

&lt;p&gt;A nice thing about the above flow is that it demonstrates how easy it should be to add OpenID support to an existing Web application. If you've already got a user account system that's fine - just give your users a mechanism to associate an OpenID with their existing account so they can log in without using their password. You could even require new users to set up a full account (complete with password that they never intend to use) and then associate it with their OpenID, although doing so eliminates the lower barrier to entry advantage for OpenID users.&lt;/p&gt;

&lt;p&gt;The trust issue is now null and void. An OpenID is just as trustworthy as a regular username and password (which could have been posted to &lt;a href="http://www.bugmenot.com/"&gt;bugmenot&lt;/a&gt; and shared with thousands of people).&lt;/p&gt;

&lt;p&gt;One last time: an OpenID is not an account. Just treat it as an alternative to a traditional username and password and you can't go wrong.&lt;/p&gt;
    
        &lt;p&gt;Tags: &lt;a href="https://simonwillison.net/tags/identity"&gt;identity&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/openid"&gt;openid&lt;/a&gt;, &lt;a href="https://simonwillison.net/tags/trust"&gt;trust&lt;/a&gt;&lt;/p&gt;
    

</summary><category term="identity"/><category term="openid"/><category term="trust"/></entry></feed>