Back

AI Models Can De‑anonymize Online Accounts, Study Finds

Study Overview

Scientists from Anthropic and ETH Zurich conducted research demonstrating that modern artificial‑intelligence systems, specifically large language models (LLMs), are capable of identifying the real‑world identities behind supposedly anonymous online accounts. The work, released as a preprint on arXiv, describes an automated pipeline that extracts identity‑related signals from public text and then searches for matching profiles across the internet.

Methodology and Experiments

The AI system examined public posts on platforms such as Hacker News, Reddit, and other forums. It looked for clues including personal interests, demographic hints, writing style, and incidental details that users inadvertently reveal. Three key experiments were performed:

  • Matching Hacker News users to their LinkedIn profiles after stripping obvious identifiers.
  • Linking pseudonymous Reddit accounts across different communities.
  • Splitting a single user’s posting history into two separate profiles to test whether the AI could recognize they belong to the same person.

In these tests the LLM‑based approach achieved up to 68% recall with about 90% precision, a performance level that traditional manual or algorithmic techniques failed to approach.

Implications for Online Privacy

The study highlights a potential erosion of the “practical obscurity” many users rely on when posting anonymously. Journalists, whistleblowers, activists, and ordinary individuals often depend on pseudonymity to discuss sensitive topics without exposing their identities. If AI can automate the deanonymization process quickly and cheaply—estimated at between $1 and $4 per profile—the barrier to large‑scale investigations could drop dramatically.

Limitations and Future Directions

Researchers note that their experiments were conducted in controlled environments using only publicly available data, and the paper has not yet undergone peer review. They also deliberately omitted some technical specifics to reduce the risk of misuse. The authors call for further work to understand both the risks and possible defenses, suggesting improvements in privacy tools, platform safeguards, and AI systems designed to protect sensitive information before it is shared.

Conclusion

As AI becomes more adept at analyzing massive volumes of online content, the balance between powerful discovery tools and the protection of personal privacy will become increasingly critical. This research underscores the urgent need for new strategies to preserve anonymity in the digital age.

Used: News Factory APP - news discovery and automation - ChatGPT for Business

Source: Digital Trends

Also available in: