postalcoder
9 hours ago
There may actually be some utility here. LLM agents refuse to traverse the links. Tested with gemini-3-pro, gpt-5.2, and opus 4.5.
edit: gpt-oss 20B & 120B both eagerly visit it.
devsda
8 hours ago
I wish this came a day earlier.
There is a current "show your personal site" post on top of HN [1] with 1500+ comments. I wonder how many of those sites are or will be hammered by AI bots in the next few days to steal/scrape content.
If this can be used as a temporary guard against AI bots, that would have been a good opportunity to test it out.
aflukasz
an hour ago
AI bots (or clients claiming to be one) appear quite fast on new sites, at least that's what I saw recently in few places. They probably monitor Certificate Transparency logs - you won't hide by avoiding linking. Unless you are ok with staying in the shadow of naked http.
xlii
7 hours ago
I posted my site on the thread.
My site is hosted on Cloudflare and I trust its protection way more than flavor of the month method. This probably won't be patched anytime soon but I'd rather have some people click my link and not just avoid it along with AI because it looks fishy :)
treebeard901
6 hours ago
I've been considering how feasible it would be to build a modern form of the denial of service low orbit ion cannon by having various LLMs hammer sites until they break. I'm sure anything important already has Cloudflare style DDOS mitigation so maybe it's not as effective. Still, I think it's only a matter of time before someone figures it out.
There have been several amplification attacks using various protocols for DDOS too...
devsda
3 hours ago
Yeah I meant using it as an experiment to test with two different links(or domains) and not as a solution to evade bot traffic.
Still, I think it would be interesting to know if anybody noticed a visible spike in bot traffic(especially AI) after sharing their site info in that thread.
testfrequency
7 hours ago
Glad I’m not the only one who felt icky seeing that post.
I agree my tinfoil hat signal told me this was the perfect way to ask people for bespoke, hand crafted content - which of course AI will love to slurp up to keep feeding the bear.
kzalesak
3 hours ago
I think that something specifically intended for this, like Anubis, is a much better option.
subscribed
2 hours ago
Anubis flatly refuses me access to several websites when I'm accessing them with a normal Chromium with enabled JS and whatnot, from a mainstream, typical OS, just with aggressive anti-tracking settings.
Not sure if that's the intended use case. At least Cloudflare politely masks for CAPTCHA.
john01dav
an hour ago
I thought that Anubis solely is proof of work, so I'm very curious as to what's going on here.
fc417fc802
40 minutes ago
What do you mean "refuses"? The worst it should do is serve up a high difficulty proof of work. Unless it gained new capabilities recently?
Are you sure the block isn't due to the authors of those websites using some other tool in addition?
jnrk
7 hours ago
Of course, the downside is that people might not even see your site at all because they’re afraid to click on that suspicious link.
postalcoder
7 hours ago
Site should add a reverse lookup. Provide the poison and antidote.
gala8y
4 hours ago
Bitly does that, just add '+' to Bitly URL (probably other shorteners, too).
briandear
3 hours ago
How is AI viewing content any different from Google? I don’t even use Google anymore because it’s so filled with SEO trash as to be useless for many things.
Zambyte
an hour ago
Try hosting a cgit server on a 1u server in your bedroom and you'll see why.
PUSH_AX
6 hours ago
LLM led scraping might not as it requires an LLM to make a choice to kick it off, but crawling for the purpose of training data is unlikely to be affected.
Barathkanna
7 hours ago
Sounds like a useful signal for people building custom agents or models. Being able to control whether automated systems follow a link via metadata is an interesting lever, especially given how inconsistent current model heuristics are.