AI Can Now Unmask Anonymous Users with Shocking Accuracy

Researchers use AI to link Reddit throwaway accounts to real identities for under $6 per person

Alex Barrientos Avatar
Alex Barrientos Avatar

By

Image: PickPik

Key Takeaways

Key Takeaways

  • Researchers achieve 68% recall linking anonymous Reddit posts to real identities
  • AI costs $1.41-$5.64 per target to unmask pseudonymous users online
  • Writing style analysis identifies 35% of users across different accounts

Sarah thought she was careful. Using a throwaway account, she vented about her toxic startup job on r/cscareerquestions, shared movie recommendations on r/horror, and discussed her graduate research without naming her university. Different username, no real name, no photos. Anonymous enough, right?

Wrong. Researchers from ETH Zurich and Anthropic just proved that large language models can unmask pseudonymous users like Sarah with startling accuracy. Their ESRC pipeline—Extract signals, Search candidates, Reason over matches, Calibrate precision—achieved 68% recall at 90% precision when linking Hacker News posts to LinkedIn profiles across 89,000 candidates.

The kicker? It costs between $1.41 and $5.64 per target using standard APIs like Grok 4.1 or GPT-5.2.

How Your Writing Style Betrays You

AI analyzes your unstructured text posts to identify unique patterns that traditional methods missed.

Unlike the Netflix Prize attacks of 2008 that achieved just 0.1% recall at 99% precision, LLMs excel at parsing the messy, unstructured prose you actually write online. The system doesn’t need your structured data or manual effort—it reads between the lines of your casual posts about weekend plans, work frustrations, or favorite TV shows.

In Reddit experiments, researchers identified 35% of users linking their past and current accounts at high precision. Users who discussed more than 10 shared movies across subreddits like r/movies and r/horror saw their recall jump to 48.1% at 90% precision. Your taste in horror films just became a fingerprint.

Million-User Pools, Pocket Change Budgets

The technology scales frighteningly well, processing massive user databases for the price of a fancy coffee.

Real-world testing on Anthropic’s anonymized scientist interviews identified 7-27% of participants from transcripts alone. Details like British English usage, specific research libraries, or biology specializations created unique signatures that LLMs could match across platforms.

“LLM agents can figure out who you are from your anonymous online posts,” according to Simon Lermen, the study’s co-author. The method scales to million-user pools while maintaining roughly 35% recall—and those numbers will only improve as models get smarter.

Standard LLM guardrails crumble under simple prompt tweaks, making this capability accessible to anyone willing to spend a few dollars.

Privacy Is Now Pay-to-Play

Your anonymous posting habits need an immediate reality check before someone else connects the dots.

The implications stretch beyond academic curiosity. Journalists and activists face government surveillance risks, while marketers gain hyper-targeted advertising capabilities that make Cambridge Analytica look quaint. Social engineering attacks just got a massive upgrade.

Proposed mitigations include:

  • Platform API rate limits
  • Scraping detection
  • Bulk export restrictions

Your move: delete old posts, limit personal details, or accept that your “anonymous” opinions might not stay that way much longer.

The age of practical obscurity just ended with a whimper and a credit card charge.

Share this

At Gadget Review, our guides, reviews, and news are driven by thorough human expertise and use our Trust Rating system and the True Score. AI assists in refining our editorial process, ensuring that every article is engaging, clear and succinct. See how we write our content here →