Cloudflare Crawl Endpoint
Posted by jeffpalmer 3 hours ago
Comments
Comment by greatgib 1 hour ago
And once that is setup, and you have you walled garden, then you can present your own api to scrape website. All well done to be used by your llm. But as you know, we are the gate keeper so that the Mafia boss decide what will be the "intermediate" that is proper for itself to let you do what you were doing without intermediary before.
Comment by shadowfiend 1 hour ago
Comment by x0x0 27 minutes ago
It's hard to see how this isn't extorting folks by offering a working solution that, oh, cloudflare doesn't block. As long as you pay Cloudflare.
Perhaps I'm overly cynical, but I'd be quite surprised if cloudflare subjected their own headless browsing to the same rules the rest of the internet gets.
Comment by gruez 16 minutes ago
The docs are pretty equivocal though:
>If you use Cloudflare products that control or restrict bot traffic such as Bot Management, Web Application Firewall (WAF), or Turnstile, the same rules will apply to the Browser Rendering crawler.
It's not just robots.txt. Most (all?) restrictions that apply to outside bots apply to cloudflare's bot as well, at least that's what they're claiming. If they're being this explicit about it, I'm willing to give them the benefit of the doubt until there's evidence to the contrary, rather than being a cynic and assuming the worst.
Comment by arjunchint 2 minutes ago
Comment by jasongill 2 hours ago
Obviously there's good reasons NOT to, but I am surprised they haven't started offering it (as an "on-by-default" option, naturally) yet.
Comment by michaelmior 1 hour ago
It's entirely possible that they're doing this under the hood for cases where they can clearly identify the content they have cached is public.
Comment by binarymax 1 hour ago
Comment by janalsncm 59 minutes ago
Comment by selcuka 1 hour ago
> Cloudflare's network now supports real-time content conversion at the source, for enabled zones using content negotiation headers. Now when AI systems request pages from any website that uses Cloudflare and has Markdown for Agents enabled, they can express the preference for text/markdown in the request. Our network will automatically and efficiently convert the HTML to markdown, when possible, on the fly.
Comment by cmsparks 1 hour ago
Comment by csomar 1 hour ago
Comment by ljm 1 hour ago
And they can pull it off because of their reach over the internet with the free DNS.
Comment by shadowfiend 1 hour ago
Comment by iso-logi 1 hour ago
The fact that 30%+ of the web relies on their caching services, routablility services and DDoS protection services is the main pull.
Their DNS is only really for data collection and to front as "good will"
Comment by subscribed 28 minutes ago
Like there's a difference between dozens of drunk teenagers thrashing the city streets in the illegal street race vs a taxi driver.
Comment by theamk 1 hour ago
> The /crawl endpoint respects the directives of robots.txt files, including crawl-delay. All URLs that /crawl is directed not to crawl are listed in the response with "status": "disallowed".
You don't need any scraping countermeasures for crawlers like those.
Comment by Macha 17 minutes ago
Comment by gruez 13 minutes ago
Given that malicious bots are allegedly spoofing real user agents, "another user agent you have to add to your list" seems like the least of your problems.
Comment by its-kostya 1 hour ago
Comment by giancarlostoro 59 minutes ago
Comment by rrr_oh_man 1 hour ago
Comment by Retr0id 1 hour ago
Comment by david_iqlabs 23 minutes ago
Comment by everfrustrated 1 hour ago
Comment by shadowfiend 1 hour ago
Comment by arjie 24 minutes ago
Comment by echoangle 9 minutes ago
Comment by devnotes77 1 hour ago
Workers-originated requests include a CF-Worker header identifying the workers subdomain, which distinguishes them from regular CDN proxying. You can match on this in a WAF rule or origin middleware.
The trickier issue: rendered requests originate from Cloudflare ASN 13335 with a low bot score, so if you rely on CF bot scores for content protection, requests through their own crawl product will bypass that check. The practical defense is application-layer rate limiting and behavioral analysis rather than network-level scores -- which is better practice regardless.
The structural conflict is real but similar to search engines offering webmaster tools while running the index. The incentives are misaligned, but the individual products have independent utility. The harder question is whether the combination makes it meaningfully harder to build effective bot protection on top of their platform.
Comment by radium3d 1 hour ago
``` write a custom crawler that will crawl every page on a site (internal links to the original domain only, scroll down to mimic a human, and save the output as a WebP screenshot, HTML, Markdown, and structured JSON. Make it designed to run locally in a terminal on a linux machine using headless Google Chrome and take advantage of multiple cores to run multiple pages simultaneously while keeping in mind that it might have to throttle if the server gets hit too fast from the same IP. ```
Might use available open source software such as python, playwright, beautifulsoup4, pillow, aiofiles, trafilatura
Comment by Normal_gaussian 31 minutes ago
You'll still be hand-rolling it if you want to disrespect crawling requirements though.
Comment by supermdguy 17 minutes ago
Comment by Normal_gaussian 21 minutes ago
From the behaviour of our peers, this seems to be the real headline news.
Comment by jppope 1 hour ago
Comment by pupppet 1 hour ago
Comment by skybrian 32 minutes ago
Comment by patchnull 1 hour ago
Comment by binarymax 1 hour ago
Comment by triwats 2 hours ago
Comment by devnotes77 1 hour ago
First, the Cloudflare Crawl endpoint does not require the target site to use Cloudflare. It spins up a headless Chrome instance (via the Browser Rendering API) that fetches and renders any publicly accessible URL. You could crawl a site hosted on Hetzner or a bare VPS with the same call.
Second on pricing: Browser Rendering is only available on the Workers Paid plan ($5/month). It is not part of the free tier. Usage is billed per invocation beyond the included quota - the exact limits are in the Cloudflare docs under Browser Rendering pricing, but for archival use cases with moderate crawl rates you are very unlikely to run into meaningful costs.
The practical gotcha for forum archival is pagination and authentication-gated content. If the forum requires a login to see older posts, a headless browser session with saved cookies would help, but that is more complex to orchestrate than a single-shot fetch.
Comment by babelfish 1 hour ago
Comment by 8cvor6j844qw_d6 2 hours ago
I'll need to test it out, especially with the labyrinth.
Comment by jsheard 1 hour ago
Further down they also mention that the requests come from CFs ASN and are branded with identifying headers, so third party filters could easily block them too if they're so inclined. Seems reasonable enough.
Comment by xhcuvuvyc 2 hours ago
Comment by mdasen 1 hour ago
We're creating an internet that is becoming self-reinforcing for those who already have power and harder for anyone else. As crawling becomes difficult and expensive, only those with previously collected datasets get to play. I certainly understand individual sites wanting to limit access, but it seems unlikely that they're limiting access to the big players - and maybe even helping them since others won't be able to compete as well.
Comment by adi_kurian 1 hour ago
Comment by canpan 1 hour ago
I'm split between: Yes! At last something to get CF protected sites! And: Uh! Now the internet is successfully centralized.
Comment by memothon 2 hours ago
Comment by rvz 1 hour ago
Comment by Imustaskforhelp 1 hour ago
I had the idea after buying https://mirror.forum recently (which I talked in discord and archiveteam irc servers) that I wanted to preserve/mirror forums (especially tech) related [Think TinyCoreLinux] since Archive.org is really really great but I would prefer some other efforts as well within this space.
I didn't want to scrape/crawl it myself because I felt like it would feel like yet another scraping effort for AI and strain resources of developers.
And even when you want to crawl, the issue is that you can't crawl cloudflare and sometimes for good measure.
So in my understanding, can I use Cloudflare Crawl to essentially crawl the whole website of a forum and does this only work for forums which use cloudflare ?
Also what is the pricing of this? Is it just a standard cloudflare worker so would I get free 100k requests and 1 Million per the few cents (IIRC) offer for crawling. Considering that Cloudflare is very scalable, It might even make sense more than buying a group of cheap VPS's
Also another point but I was previously thinking that the best way was probably if maintainers of these forums could give me a backup archive of the forum in a periodic manner as my heart believes it to be most cleanest way and discussing it on Linux discord servers and archivers within that community and in general, I couldn't find anyone who maintains such tech forums who can subscribe to the idea of sharing the forum's public data as a quick backup for preservation purposes. So if anyone knows or maintains any forums myself. Feel free to message here in this thread about that too.
Comment by ipaddr 1 hour ago
You feel better paying someone to do the same thimg?
Comment by Imustaskforhelp 1 hour ago
Also, I am genuinely open to feedback (Like a lot) so just let me know if you know of any other alternative too for the particular thing that I wish to create and I would love to have a discussion about that too! I genuinely wish that there can be other ways and part of the reason why I wrote that comment was wishing that someone who manages forums or knows people who do can comment back and we can have a discussion/something-meaningful!
I am also happy with you also suggesting me any good use cases of the domain in general if there can be made anything useful with it. In fact, I am happy with transferring this domain to you if this is something which is useful to ya or anyone here (Just donate some money preferably 50-100$ to any great charity in date after this comment is made and mail me details and I am absolutely willing to transfer the domain, or if you work in any charity currently and if it could help the charity in any meaningful manner!)
I had actually asked archive team if I could donate the domain to them if it would help archive.org in any meaningful way and they essentially politely declined.
I just bought this domain because someone on HN said mirror.org when they wanted to show someone else mirror and saw the price of the .org domain being so high (150k$ or similar)and I have habit of finding random nice TLD and I found mirror.forum so I bought it
And I was just thinking of hmm what can be a decent idea now that I have bought it and had thought of that. Obviously I have my flaws (many actually) but I genuinely don't wish any harm to anybody especially those people who are passionate about running independent forums in this centralized-web. I'd rather have this domain be expired if its activation meant harm to anybody.
looking forward to discussion with ya.