Cloudflare has officially blocked Perplexity AI from crawling websites, delisting their bots from its Verified Bots list. This move is in response to increasing concerns over unauthorized scraping and AI engines disguising crawler identities to bypass restrictions.
Cloudflare confirmed that Perplexity’s bots violated standard web crawling protocols by accessing content even when robots.txt directives were in place to block them. The bots allegedly used misleading user agents and operated through residential IPs, making it harder for publishers and SEOs to detect their presence.
This kind of behavior reflects a serious breach of digital publishing rights – a growing issue as AI search models aggressively gather data to train large language models (LLMs).
With AI adoption growing rapidly, more content creators and platforms are implementing crawler blocking strategies. The goal? To ensure AI engines comply with ethical web scraping guidelines and respect the rights of content owners.
SERP AI Stream confirms this case isn’t isolated. Several AI tools are bypassing restrictions, which raises valid questions about transparency, data ethics, and crawler accountability.
If you manage a website, take the following steps immediately:
The line between AI innovation and content exploitation is growing thinner. This latest move by Cloudflare marks a shift toward holding AI companies accountable and reclaiming control over digital content.
As Manthan Bhatia from SERP AI Stream highlights, “Web crawling must evolve with responsibility. It’s not about blocking AI – it’s about ensuring it plays by the rules.”