Skip to main content

Archived Article — The Daily Perspective is no longer active. This article was published on 25 March 2026 and is preserved as part of the archive. Read the farewell | Browse archive

Technology

Reddit tightens verification for 'fishy' accounts as bot problem spreads

The platform will require some suspected bots to prove they're human, but promises the checks will remain rare

Reddit tightens verification for 'fishy' accounts as bot problem spreads
Image: The Verge
Key Points 3 min read
  • Reddit will require accounts with suspicious or automated behaviour to verify they are human using methods like Face ID or fingerprint scanning.
  • The company will label legitimate bots with an '[APP]' tag so users know which automated accounts are genuine services.
  • Full ID verification will be used only in rare cases to comply with legal requirements, prioritising privacy over identification.

Reddit is preparing to weaponise biometric authentication against one of social media's oldest scourges: the bot infestation that erodes platform credibility and fuels manipulation at scale.

On Wednesday, Reddit announced it will begin labeling automated accounts that provide a service to users and will require accounts suspected of being bots to verify if they're human. The move signals growing urgency around artificial accounts, particularly as AI agents become more sophisticated and harder to distinguish from real users.

The verification approach reflects a careful balance between security and privacy. For now, verification will take the form of on-device methods, including Face ID and passkeys, though the company is considering alternative methods including World ID, the face-scanning orb company run by Sam Altman. Huffman wrote that "the best long-term solutions will be decentralised, individualised, private, and ideally not require an ID at all."

The crackdown responds to a genuine and worsening problem. According to Cloudflare, the traffic from bots will exceed human traffic by 2027, when you include bots like web crawlers and AI agents in the mix. The social platform has become a particular target for manipulation; Reddit has become a popular destination for bots that attempt to manipulate narratives and astroturf to shill for companies or their products. Reddit removes an average of 100,000 accounts per day in its ongoing efforts to control the problem.

Critically, Reddit is not proposing blanket verification for all users. Reddit stresses this is not going to be a sitewide verification requirement and will only occur if something suggests that the account isn't human, including its activity on the site or other technical markers. To identify potential bots, Reddit is using specialised tooling that looks at account-level signals and other factors, like how quickly the account is attempting to write or post content.

The stakes for Reddit's credibility are significant. Earlier this month, Digg shut down its community platform after failing to control bot proliferation. Would-be Reddit competitor Digg just shut down because it couldn't get a handle on the bots overrunning its site. Reddit's approach suggests the company views the threat as existential to its business model, which depends on user trust that the comments and posts they read come from real people.

Even so, Reddit faces meaningful scrutiny over whether biometric verification squares with the anonymity users expect. Huffman said "part of our promise for our users is we don't know your name but we do want to know you're a person. It'll be an evolution for us for a while, and probably every platform to find the right middle ground here." Reddit co-founder Alexis Ohanian said on X that Reddit requiring Face ID wasn't something he expected but agreed that something had to be done about the fake content from bots, adding that "I just don't know how to sell face-scanning to Redditors or even lurkers."

As part of the new policy, Reddit is also adding an '[APP]' label to existing 'good' bots on the platform and making it easier for users to report suspected 'bad' bots. The labeling system aims to give users transparency about which automated accounts they're interacting with, distinguishing between bots that provide genuine utility (weather alerts, price tracking) and those designed to deceive or manipulate.

The human verification requirement raises legitimate questions about implementation and unintended consequences. Legitimate power users who automate their own tools may face friction. Users in regions with weak biometric infrastructure could struggle. And any system that requires proof of humanhood creates new attack surfaces for fraud and theft.

For now, Reddit's approach suggests a pragmatic middle path: deploy verification selectively against obvious offenders rather than demand universal identification, use the lightest-touch biometric methods available, and preserve anonymity where possible. Whether it proves sufficient to hold back the tide of AI-generated accounts remains an open question.

Sources (4)
Sophia Vargas
Sophia Vargas

Sophia Vargas is an AI editorial persona created by The Daily Perspective. Covering US politics, Latin American affairs, and the global shifts emanating from the Western Hemisphere. As an AI persona, articles are generated using artificial intelligence with editorial quality controls.