Your AI chatbot history is being collected and used
View in browser ([link removed] )
January 13th, 2026 // Did someone forward you this newsletter? Sign up to receive your own copy here. ([link removed] )
You signed an AI privacy policy. What did you agree to?
Most of us agree to thousands of words of legal text without reading a single sentence.
For example, OpenAI’s privacy policy is 3,000 words; Anthropic’s is 5,000 words. Yet most of us click accept in under five seconds.
Participation in today’s internet effectively requires consenting to terms that, in practice, strip away privacy. That is the price of entry.
But what are we exactly signing? In this newsletter, we examine the privacy policies of leading AI companies, highlight recent changes they are making to reduce user privacy, and discuss what can be done to reclaim privacy in a digital ecosystem built on access to our data.
// TL/DR
When we use AI chatbots, what data is being collected and stored?
We did a close read of the privacy policies of three leading AI companies, OpenAI ([link removed] ) , Anthropic ([link removed] ) , and Perplexity ([link removed] ) (each hyperlink goes to the company’s privacy policy so you can read it yourself, or ironically, ask AI to summarize it).
Data collection
All three AI companies largely collect the same data. In addition to account identifiers (name, email, login credentials), the companies collect all user content—everything that’s inputted by a user into their chatbot (prompts, responses, etc.), as well as what the chatbot outputs in its responses. This includes written content, images, and other files. They also collect technical data (IP addresses, device and browser information, cookies, log files, etc.).
Data use
It should come as no surprise that AI companies are collecting the data we make available to their chatbots. What is surprising is how that data is used, and how the content from seemingly private conversations can find its way back into LLMs. All three AI companies report using the data they collect to maintain their products, improve them over time, ensure security, prevent abuse, and comply with legal obligations.
That may sound innocuous, but it means that AI companies are training their AI models on user data. In 2025, Anthropic announced ([link removed] ) a change to its privacy policy that allowed Claude to train on consumer chats unless users opt out, and it extended retention up to five years for users who allow training. OpenAI already had a similar policy ([link removed] ) in place. AI companies tend to exempt enterprise and corporate accounts from this data use, but individual users should expect that everything they input into an AI chatbot is fed back into the model, unless they opt out (more on how to do that below).
Data disclosure
AI companies retain the right to disclose the data they collect about a user. That’s to be expected, and all three AI companies have similar disclosure clauses in their privacy policies. They can share your data—identifiable to you or de-identified—with third parties (many AI companies rely on a vast network of vendors and other service providers), internal teams, and external law enforcement (when required by law). In the future, those third parties could include advertising partners; similar to the surveillance capitalism model popular in social media, the content we generate in an AI chatbot interaction could one day be used to advertise to us ([link removed] ) .
In past newsletters, we’ve explored the tensions and tradeoffs between online safety and online privacy. Those tensions exist here, too: In an effort to avoid AI-induced psychosis ([link removed] ) or AI chatbots supporting someone to cause harm to themselves or someone else ([link removed] ) , AI companies retain the right to review your AI chat history and refer you to law enforcement.
In 2025, OpenAI issued a statement ([link removed] ) outlining new safety measures, which represent a form of safety-intended surveillance ([link removed] ) : “When we detect users who are planning to harm others, we route their conversations to specialized pipelines where they are reviewed by a small team trained on our usage policies and who are authorized to take action, including banning accounts. If human reviewers determine that a case involves an imminent threat of serious physical harm to others, we may refer it to law enforcement.”
Data retention
A person using an AI chatbot has tacitly agreed that the AI company has few or no limits on the retention of that data. OpenAI’s privacy policy states: “We’ll retain your Personal Data for only as long as we need…” Anthropic’s and Perplexity’s policies say something similar. There is no explicit limit to when data will be deleted or destroyed.
Children’s data
OpenAI and Perplexity’s policies state that their services are not intended for children under the age of 13, and it does not knowingly collect data from children under 13 (it has recently released age prediction ([link removed] ) technology to estimate the age of their users based on chat history). For Anthropic, the age cut-off is 18. If the platforms discover that a minor is using their chatbots, they will disable that account and delete the associated data.
Anthropic allows ([link removed] ) minors to use its models indirectly through third‑party applications (e.g., education apps) if those developers meet specified child‑safety and privacy requirements.
// Trained by default
Anthropic’s privacy policy acknowledges that users have certain rights, especially in jurisdictions like the EU, but it states: “However, please be aware that these rights are limited, and that the process by which we may need to action your requests regarding our training dataset are complex.”
A 2025 study ([link removed] ) by researchers at Stanford analyzed the privacy policies of six leading AI companies (Amazon, Anthropic, Google, Meta, Microsoft, and OpenAI) and raised concerns about the lack of privacy, transparency, and ability for users to opt out of data collection.
-
The study found that all six companies train on their users’ chat data by default, meaning users have opted in to having their data collected and used. “Most users do not change defaults…This is why default collection of chat data for model training by all developers has far-reaching implications for data privacy.”
-
The study also found that an AI company’s privacy policy wasn’t the only source of its data practices. They found that “there were data practices discussed in branch policies that were not disclosed in the main privacy policy…such as how user inputs contribute to model training, whether human reviewers are involved in the process of reviewing inputs, or whether a model is trained on deidentified inputs or on personal data from the internet.”
The study’s lead author, Jennifer King, Privacy and Data Policy Fellow at the Stanford Institute for Human-Centered AI ([link removed] ) , said, “As a society, we need to weigh whether the potential gains in AI capabilities from training on chat data are worth the considerable loss of consumer privacy. And we need to promote innovation in privacy-preserving AI, so that user privacy isn’t an afterthought.”
// How to reclaim privacy
The Stanford study on AI company privacy policies made five recommendations to address the concerns and issues identified in analyzing AI company privacy policies:
- Adopt a comprehensive federal privacy regulation. Without greater consumer privacy regulations in the U.S. that protect consumers, the governance of personal data will remain limited, as AI companies dictate terms (instead of comply with laws). And in the absence of federal regulation, good actors (who don't capture or sell/share as much data) are effectively penalized for adopting higher standards.
- Require opt-in for model training. The study recommended that “policymakers should require developers to ask users to opt-in for model training rather than making it the default…Children should never be opted into data collection by default.”
- Clarify data collection practices. Long, complex privacy policies are not a solution for equipping users with information about how their data will be used, especially when those policies are not comprehensive, and data practices are documented in other documents. The authors of the study recommended clarifying, consolidating, and revamping privacy policies to ensure they capture the totality of data governance.
- Limit personal information from inputs by default. AI companies and policymakers should make it easier to edit, remove, or filter sensitive personal information from chat histories.
- Advance privacy-focused innovation. Some models have begun experimenting with privacy-focused innovations similar to Snapchat’s disappearing messages. OpenAI and Google offer users the ability to create “Temporary Chats” that are not used to train models and are stored for a shorter period (30 days).
// What users can do today
AI companies offer the ability to stop data from being used to train models. Here is what individuals can do today:
-
OpenAI: Navigate to the OpenAI Privacy Portal ([link removed] ) and click on “do not train on my content.” Once done, new conversations will not be used to train AI models. Follow the specific steps outlined here ([link removed] ) by OpenAI.
-
Anthropic: Navigate to Claude’s Privacy Settings ([link removed] ) to toggle “Help improve Claude” to off.
-
Perplexity: In Settings, you can disable data retention or operate in Incognito mode. Follow the steps here ([link removed] ) .
This article from The Washington Post ([link removed] ) offers additional specific instructions for reclaiming privacy and opting out of default data-collection practices. (Accessing this article might require creating an account.)
For individuals seeking to audit the other terms of service and privacy policies they’ve unwittingly consented to online, resources are available from organizations that are members of the Project Liberty Alliance.
-
Consumer Reports’ Permission Slip ([link removed] ) is an app that helps users take back control of their data by having an AI agent discover what data online companies are collecting and automatically file a request to delete it.
-
Block Party ([link removed] ) helps users deep-clean their social media accounts, boosting privacy and increasing control over privacy settings.
Before taking these steps, it’s important to understand their limits. In most cases, opting out applies only to future use of data for model training. Companies typically retain broad rights to store and process information for security, abuse prevention, and legal compliance, and removing data from existing training systems may be complex or even impossible.
// A better tech future
A better future depends on technology that gives people greater control over their digital lives. What would this look like? Data is portable, consent is explicit, access is time-bound and revocable, and value can flow back to the people who generate it. That is the promise of a “people’s internet,” built on systems designed to give each of us a voice, choice, and stake rather than a permanent surrender buried in fine print.
Today, the defaults that underpin AI chatbots and tech platforms are set to collect, store, use, and disclose our data. Individuals are not powerless, but reclaiming privacy largely requires effort and vigilance, shifting the burden onto users. As concern about the data economy grows, there is an opportunity for stronger policies like the Digital Choice Act ([link removed] ) , by-design platform changes, and technologies that make privacy the default rather than the exception.
📰 Other notable headlines
// 🏩 A Guardian investigation ([link removed] ) found that Google’s AI overviews provided inaccurate and false information when queried over blood tests. (Free).
// 📞 An article in The Washington Post ([link removed] ) pointed out that Gen Zers aren’t talking on the phone, and it could cost them (in more than social awkwardness). (Paywall).
// 🤔 Large language models don’t “learn”—they copy. And that could change everything for the tech industry, according to an article in The Atlantic ([link removed] ) . (Paywall).
// 📱 An article in Tech Policy Press ([link removed] ) argued that the latest Grok scandal isn't an anomaly. It follows warnings that were ignored. (Free).
// 🤝 An article in WIRED ([link removed] ) asked, is Craigslist the last real place on the internet? Millennials are still using Craigslist to find jobs, love, and even to cast creative projects—eschewing other AI- and algorithm-dominated online spaces. (Paywall).
// 📸 In the age of AI, we endlessly debate what consciousness looks like. An article in The New Yorker ([link removed] ) asked, can a camera see things more clearly? (Paywall).
// 🏛 Google and Character.AI settled the lawsuit in the case of a 14-year-old in Florida who had killed himself after developing a relationship with an AI chatbot, according to an article in The New York Times ([link removed] ) . (Paywall).
// 😍 Can you optimize love? A group of tech executives, app developers, and Silicon Valley philosophers is seeking to streamline the messy matters of the heart, according to an article in The New York Times ([link removed] ) . (Paywall).
// 👀 An op-ed in The New York Times ([link removed] ?) argued that the multi-trillion-dollar battle for your attention is built on a lie. (Paywall).
Partner news
// Slow Violence, Fast Tech: Reclaiming AI for civil and human rights
January 29 | 1-2 PM ET | Virtual
All Tech Is Human ([link removed] ) is hosting a conversation with Malika Saada Saar, Senior Fellow at Brown University and former Global Head of Human Rights at YouTube, examining how AI’s incremental “slow violence” can erode civil and human rights. Register here ([link removed] ) .
// New podcast: How to let a thousand societies bloom
BlockchainGov ([link removed] ) highlights a new episode of the Network Nations podcast featuring Vitalik Buterin. The discussion explores core tensions shaping the future of digitally aligned communities: culture vs. mission, kinship vs. purpose, and permanence vs. mobility. Tune in here ([link removed] ) .
// TRANSFER Download: AliveNET explores the space between human and machine
January 16 | 5-8 PM ET | Nguyen Wahed Gallery, New York City
TRANSFER ([link removed] ) and Nguyen Wahed Gallery present TRANSFER Download: AliveNET ([link removed] ) , an immersive in-person exhibition (Jan. 16 - March 1) examining how perception, identity, and embodiment are shaped through screen-mediated and algorithmic systems. Register for opening night ([link removed] ) .
What did you think of today's newsletter?
We'd love to hear your feedback and ideas. Reply to this email.
// Project Liberty builds solutions that advance human agency and flourishing in an AI-powered world.
Thank you for reading.
Facebook ([link removed] )
LinkedIn ([link removed] )
Twitter ([link removed] )
Instagram ([link removed] )
Project Liberty footer logo ([link removed] )
10 Hudson Yards, Fl 37,
New York, New York, 10001
Unsubscribe ([link removed] ) Manage Preferences ([link removed] )
© 2025 Project Liberty LLC