Close Menu
The Washington FeedThe Washington Feed

    Subscribe to Updates

    Get the latest creative news from FooBar about art, design and business.

    What's Hot

    Apple’s new live translation feature for AirPods won’t be available in the EU at launch

    September 12, 2025

    West Seattle Blog… | TRAFFIC ALERT: 42nd/Alaska signal dark

    September 12, 2025

    South African rapist Thabo Bester loses bid to block Netflix film

    September 12, 2025
    Facebook X (Twitter) Instagram
    Facebook X (Twitter) Instagram
    The Washington FeedThe Washington Feed
    Subscribe
    • Home
    • World
    • US
    • seattle
    • Politics
    • Business
    • Tech
    • Contact Us
    The Washington FeedThe Washington Feed
    Home»Tech»Anthropic users face a new choice – opt out or share your data for AI training
    Tech

    Anthropic users face a new choice – opt out or share your data for AI training

    adminBy adminAugust 28, 2025No Comments4 Mins Read
    Facebook Twitter Pinterest LinkedIn Tumblr Email
    Share
    Facebook Twitter LinkedIn Pinterest Email


    Anthropic is making some big changes to how it handles user data, requiring all Claude users to decide by September 28 whether they want their conversations used to train AI models. While the company directed us to its blog post on the policy changes when asked about what prompted the move, we’ve formed some theories of our own.

    But first, what’s changing: previously, Anthropic didn’t use consumer chat data for model training. Now, the company wants to train its AI systems on user conversations and coding sessions, and it said it’s extending data retention to five years for those who don’t opt out.

    That is a massive update. Previously, users of Anthropic’s consumer products were told that their prompts and conversation outputs would be automatically deleted from Anthropic’s back end within 30 days “unless legally or policy‑required to keep them longer” or their input was flagged as violating its policies, in which case a user’s inputs and outputs might be retained for up to two years.

    By consumer, we mean the new policies apply to Claude Free, Pro, and Max users, including those using Claude Code. Business customers using Claude Gov, Claude for Work, Claude for Education, or API access will be unaffected, which is how OpenAI similarly protects enterprise customers from data training policies.

    So why is this happening? In that post about the update, Anthropic frames the changes around user choice, saying that by not opting out, users will “help us improve model safety, making our systems for detecting harmful content more accurate and less likely to flag harmless conversations.” Users will “also help future Claude models improve at skills like coding, analysis, and reasoning, ultimately leading to better models for all users.”

    In short, help us help you. But the full truth is probably a little less selfless.

    Like every other large language model company, Anthropic needs data more than it needs people to have fuzzy feelings about its brand. Training AI models requires vast amounts of high-quality conversational data, and accessing millions of Claude interactions should provide exactly the kind of real-world content that can improve Anthropic’s competitive positioning against rivals like OpenAI and Google.

    Techcrunch event

    San Francisco
    |
    October 27-29, 2025

    Beyond the competitive pressures of AI development, the changes would also seem to reflect broader industry shifts in data policies, as companies like Anthropic and OpenAI face increasing scrutiny over their data retention practices. OpenAI, for instance, is currently fighting a court order that forces the company to retain all consumer ChatGPT conversations indefinitely, including deleted chats, because of a lawsuit filed by The New York Times and other publishers.

    In June, OpenAI COO Brad Lightcap called this “a sweeping and unnecessary demand” that “fundamentally conflicts with the privacy commitments we have made to our users.” The court order affects ChatGPT Free, Plus, Pro, and Team users, though enterprise customers and those with Zero Data Retention agreements are still protected.

    What’s alarming is how much confusion all of these changing usage policies are creating for users, many of whom remain oblivious to them.

    In fairness, everything is moving quickly now, so as the tech changes, privacy policies are bound to change. But many of these changes are fairly sweeping and mentioned only fleetingly amid the companies’ other news. (You wouldn’t think Tuesday’s policy changes for Anthropic users were very big news based on where the company placed this update on its press page.)

    But many users don’t realize the guidelines to which they’ve agreed have changed because the design practically guarantees it. Most ChatGPT users keep clicking on “delete” toggles that aren’t technically deleting anything. Meanwhile, Anthropic’s implementation of its new policy follows a familiar pattern.

    How so? New users will choose their preference during signup, but existing users face a pop-up with “Updates to Consumer Terms and Policies” in large text and a prominent black “Accept” button with a much tinier toggle switch for training permissions below in smaller print – and automatically set to “On.”
    As observed earlier today by The Verge, the design raises concerns that users might quickly click “Accept” without noticing they’re agreeing to data sharing.

    Meanwhile, the stakes for user awareness couldn’t be higher. Privacy experts have long warned that the complexity surrounding AI makes meaningful user consent nearly unattainable. Under the Biden Administration, the Federal Trade Commission even stepped in, warning that AI companies risk enforcement action if they engage in “surreptitiously changing its terms of service or privacy policy, or burying a disclosure behind hyperlinks, in legalese, or in fine print.”

    Whether the commission — now operating with just three of its five commissioners — still has its eye on these practices today is an open question, one we’ve put directly to the FTC.



    Source link

    Share. Facebook Twitter Pinterest LinkedIn Tumblr Email
    admin
    • Website

    Related Posts

    Apple’s new live translation feature for AirPods won’t be available in the EU at launch

    September 12, 2025

    Box CEO Aaron Levie on AI’s ‘era of context’

    September 12, 2025

    Kids in the UK are hacking their own schools for dares and notoriety

    September 12, 2025
    Leave A Reply Cancel Reply

    Demo
    Our Picks
    Stay In Touch
    • Facebook
    • Twitter
    • Pinterest
    • Instagram
    • YouTube
    • Vimeo
    Don't Miss

    Apple’s new live translation feature for AirPods won’t be available in the EU at launch

    Tech September 12, 2025

    One of the headlining features of Apple’s new AirPods Pro 3 was the ability to…

    West Seattle Blog… | TRAFFIC ALERT: 42nd/Alaska signal dark

    September 12, 2025

    South African rapist Thabo Bester loses bid to block Netflix film

    September 12, 2025

    Robert Kraft son Josh Kraft drops out of Boston mayor race against Michelle Wu

    September 12, 2025

    Subscribe to Updates

    Get the latest creative news from SmartMag about art & design.

    About Us

    At TheWashingtonFeed.com, we are committed to delivering accurate, timely, and relevant news from around the world. Whether it’s breaking developments in U.S. politics, major international affairs, or the latest trends in technology, our mission is to keep our readers informed with fact-driven journalism and insightful analysis.

    Email Us: Confordev@gmail.com

    Our Picks

    South African rapist Thabo Bester loses bid to block Netflix film

    September 12, 2025

    Ros Atkins on… Israel’s war in Gaza and proportionality

    September 12, 2025

    Young fans and critics debate his political legacy

    September 12, 2025

    Subscribe to Updates

    Get the latest creative news from FooBar about art, design and business.

    Facebook X (Twitter) Instagram Pinterest
    • Contact Us
    • About Us
    • Privacy Policy
    • Terms and Condition
    © 2025 ThemeSphere. Designed by ThemeSphere.

    Type above and press Enter to search. Press Esc to cancel.