* bug#52338: Crawler bots are downloading substitutes
@ 2021-12-06 21:20 Leo Famulari
2021-12-06 22:18 ` bug#52338: [maintenance] hydra: berlin: Create robots.txt Leo Famulari
2021-12-10 21:21 ` Mark H Weaver
0 siblings, 2 replies; 10+ messages in thread
From: Leo Famulari @ 2021-12-06 21:20 UTC (permalink / raw)
To: 52338
I noticed that some bots are downloading substitutes from
ci.guix.gnu.org.
We should add a robots.txt file to reduce this waste.
Specifically, I see bots from Bing and Semrush:
https://www.bing.com/bingbot.htm
https://www.semrush.com/bot.html
^ permalink raw reply [flat|nested] 10+ messages in thread
* bug#52338: [maintenance] hydra: berlin: Create robots.txt.
2021-12-06 21:20 bug#52338: Crawler bots are downloading substitutes Leo Famulari
@ 2021-12-06 22:18 ` Leo Famulari
2021-12-09 13:27 ` bug#52338: Crawler bots are downloading substitutes Mathieu Othacehe
2021-12-10 21:21 ` Mark H Weaver
1 sibling, 1 reply; 10+ messages in thread
From: Leo Famulari @ 2021-12-06 22:18 UTC (permalink / raw)
To: 52338
I tested that `guix system build` does succeed with this change, but I
would like a review on whether the resulting Nginx configuration is
correct, and if this is the correct path to disallow. It generates an
Nginx location block like this:
------
location /robots.txt {
add_header Content-Type text/plain;
return 200 "User-agent: *
Disallow: /nar
";
}
------
* hydra/nginx/berlin.scm (berlin-locations): Add a robots.txt Nginx location.
---
hydra/nginx/berlin.scm | 9 ++++++++-
1 file changed, 8 insertions(+), 1 deletion(-)
diff --git a/hydra/nginx/berlin.scm b/hydra/nginx/berlin.scm
index 1f4b0be..3bb2129 100644
--- a/hydra/nginx/berlin.scm
+++ b/hydra/nginx/berlin.scm
@@ -174,7 +174,14 @@ PUBLISH-URL."
(nginx-location-configuration
(uri "/berlin.guixsd.org-export.pub")
(body
- (list "root /var/www/guix;"))))))
+ (list "root /var/www/guix;")))
+
+ (nginx-location-configuration
+ (uri "/robots.txt")
+ (body
+ (list
+ "add_header Content-Type text/plain;"
+ "return 200 \"User-agent: *\nDisallow: /nar/\n\";"))))))
(define guix.gnu.org-redirect-locations
(list
--
2.34.0
^ permalink raw reply related [flat|nested] 10+ messages in thread
* bug#52338: Crawler bots are downloading substitutes
2021-12-06 22:18 ` bug#52338: [maintenance] hydra: berlin: Create robots.txt Leo Famulari
@ 2021-12-09 13:27 ` Mathieu Othacehe
2021-12-09 15:42 ` Tobias Geerinckx-Rice via Bug reports for GNU Guix
0 siblings, 1 reply; 10+ messages in thread
From: Mathieu Othacehe @ 2021-12-09 13:27 UTC (permalink / raw)
To: Leo Famulari; +Cc: 52338
Hello Leo,
> + (nginx-location-configuration
> + (uri "/robots.txt")
> + (body
> + (list
> + "add_header Content-Type text/plain;"
> + "return 200 \"User-agent: *\nDisallow: /nar/\n\";"))))))
Nice, the bots are also accessing the Cuirass web interface, do you
think it would be possible to extend this snippet to prevent it?
Thanks,
Mathieu
^ permalink raw reply [flat|nested] 10+ messages in thread
* bug#52338: Crawler bots are downloading substitutes
2021-12-09 13:27 ` bug#52338: Crawler bots are downloading substitutes Mathieu Othacehe
@ 2021-12-09 15:42 ` Tobias Geerinckx-Rice via Bug reports for GNU Guix
2021-12-10 16:22 ` Leo Famulari
0 siblings, 1 reply; 10+ messages in thread
From: Tobias Geerinckx-Rice via Bug reports for GNU Guix @ 2021-12-09 15:42 UTC (permalink / raw)
To: Mathieu Othacehe; +Cc: 52338, leo
[-- Attachment #1.1: Type: text/plain, Size: 1469 bytes --]
Mathieu Othacehe 写道:
> Hello Leo,
>
>> + (nginx-location-configuration
>> + (uri "/robots.txt")
It's a micro-optimisation, but it can't hurt to generate ‘location
= /robots.txt’ instead of ‘location /robots.txt’ here.
>> + (body
>> + (list
>> + "add_header Content-Type text/plain;"
>> + "return 200 \"User-agent: *\nDisallow:
>> /nar/\n\";"))))))
Use \r\n instead of \n, even if \n happens to work.
There are many ‘buggy’ crawlers out there. It's in their own
interest to be fussy whilst claiming to respect robots.txt. The
less you deviate from the most basic norm imaginable, the better.
I tested whether embedding raw \r\n bytes in nginx.conf strings
like this works, and it seems to, even though a human would
probably not do so.
> Nice, the bots are also accessing the Cuirass web interface, do
> you
> think it would be possible to extend this snippet to prevent it?
You can replace ‘/nar/’ with ‘/’ to disallow everything:
Disallow: /
If we want crawlers to index only the front page (so people can
search for ‘Guix CI’, I guess), that's possible:
Disallow: /
Allow: /$
Don't confuse ‘$’ with ‘supports regexps’. Buggy bots might fall
back to ‘Disallow: /’.
This is where it gets ugly: nginx doesn't support escaping ‘$’ in
strings. At all. It's insane.
[-- Attachment #1.2: Type: text/plain, Size: 201 bytes --]
geo $dollar { default "$"; } #
stackoverflow.com/questions/57466554
server {
location = /robots.txt {
return 200
"User-agent: *\r\nDisallow: /\r\nAllow: /$dollar\r\n";
}
}
[-- Attachment #1.3: Type: text/plain, Size: 99 bytes --]
*Obviously.*
An alternative to that is to serve a real on-disc robots.txt.
Kind regards,
T G-R
[-- Attachment #2: signature.asc --]
[-- Type: application/pgp-signature, Size: 247 bytes --]
^ permalink raw reply [flat|nested] 10+ messages in thread
* bug#52338: Crawler bots are downloading substitutes
2021-12-09 15:42 ` Tobias Geerinckx-Rice via Bug reports for GNU Guix
@ 2021-12-10 16:22 ` Leo Famulari
2021-12-10 16:47 ` Tobias Geerinckx-Rice via Bug reports for GNU Guix
0 siblings, 1 reply; 10+ messages in thread
From: Leo Famulari @ 2021-12-10 16:22 UTC (permalink / raw)
To: Tobias Geerinckx-Rice; +Cc: othacehe, 52338
[-- Attachment #1: Type: text/plain, Size: 286 bytes --]
On Thu, Dec 09, 2021 at 04:42:24PM +0100, Tobias Geerinckx-Rice wrote:
[...]
> An alternative to that is to serve a real on-disc robots.txt.
Alright, I leave it up to you. I just want to prevent bots from
downloading substitutes. I don't really have opinions about any of the
details.
[-- Attachment #2: signature.asc --]
[-- Type: application/pgp-signature, Size: 833 bytes --]
^ permalink raw reply [flat|nested] 10+ messages in thread
* bug#52338: Crawler bots are downloading substitutes
2021-12-10 16:22 ` Leo Famulari
@ 2021-12-10 16:47 ` Tobias Geerinckx-Rice via Bug reports for GNU Guix
2021-12-11 9:46 ` Mathieu Othacehe
0 siblings, 1 reply; 10+ messages in thread
From: Tobias Geerinckx-Rice via Bug reports for GNU Guix @ 2021-12-10 16:47 UTC (permalink / raw)
To: Leo Famulari; +Cc: othacehe, 52338
[-- Attachment #1: Type: text/plain, Size: 95 bytes --]
Leo Famulari 写道:
> Alright, I leave it up to you.
Dammit.
Kind regards,
T G-R
[-- Attachment #2: signature.asc --]
[-- Type: application/pgp-signature, Size: 247 bytes --]
^ permalink raw reply [flat|nested] 10+ messages in thread
* bug#52338: Crawler bots are downloading substitutes
2021-12-06 21:20 bug#52338: Crawler bots are downloading substitutes Leo Famulari
2021-12-06 22:18 ` bug#52338: [maintenance] hydra: berlin: Create robots.txt Leo Famulari
@ 2021-12-10 21:21 ` Mark H Weaver
2021-12-10 22:52 ` Tobias Geerinckx-Rice via Bug reports for GNU Guix
1 sibling, 1 reply; 10+ messages in thread
From: Mark H Weaver @ 2021-12-10 21:21 UTC (permalink / raw)
To: Leo Famulari, 52338
Hi Leo,
Leo Famulari <leo@famulari.name> writes:
> I noticed that some bots are downloading substitutes from
> ci.guix.gnu.org.
>
> We should add a robots.txt file to reduce this waste.
>
> Specifically, I see bots from Bing and Semrush:
>
> https://www.bing.com/bingbot.htm
> https://www.semrush.com/bot.html
For what it's worth: during the years that I administered Hydra, I found
that many bots disregarded the robots.txt file that was in place there.
In practice, I found that I needed to periodically scan the access logs
for bots and forcefully block their requests in order to keep Hydra from
becoming overloaded with expensive queries from bots.
Regards,
Mark
^ permalink raw reply [flat|nested] 10+ messages in thread
* bug#52338: Crawler bots are downloading substitutes
2021-12-10 21:21 ` Mark H Weaver
@ 2021-12-10 22:52 ` Tobias Geerinckx-Rice via Bug reports for GNU Guix
0 siblings, 0 replies; 10+ messages in thread
From: Tobias Geerinckx-Rice via Bug reports for GNU Guix @ 2021-12-10 22:52 UTC (permalink / raw)
To: Mark H Weaver; +Cc: 52338, leo
[-- Attachment #1: Type: text/plain, Size: 1073 bytes --]
All,
Mark H Weaver 写道:
> For what it's worth: during the years that I administered Hydra,
> I found
> that many bots disregarded the robots.txt file that was in place
> there.
> In practice, I found that I needed to periodically scan the
> access logs
> for bots and forcefully block their requests in order to keep
> Hydra from
> becoming overloaded with expensive queries from bots.
Very good point.
IME (which is a few years old at this point) at least the
highlighted BingBot & SemrushThing always respected my robots.txt,
but it's definitely a concern. I'll leave this bug open to remind
us of that in a few weeks or so…
If it does become a problem, we (I) might add some basic
User-Agent sniffing to either slow down or outright block
non-Guile downloaders. Whitelisting any legitimate ones, of
course. I think that's less hassle than dealing with dynamic IP
blocks whilst being equally effective here.
Thanks (again) for taking care of Hydra, Mark, and thank you Leo
for keeping an eye on Cuirass :-)
T G-R
[-- Attachment #2: signature.asc --]
[-- Type: application/pgp-signature, Size: 247 bytes --]
^ permalink raw reply [flat|nested] 10+ messages in thread
* bug#52338: Crawler bots are downloading substitutes
2021-12-10 16:47 ` Tobias Geerinckx-Rice via Bug reports for GNU Guix
@ 2021-12-11 9:46 ` Mathieu Othacehe
2021-12-19 16:53 ` Mathieu Othacehe
0 siblings, 1 reply; 10+ messages in thread
From: Mathieu Othacehe @ 2021-12-11 9:46 UTC (permalink / raw)
To: Tobias Geerinckx-Rice; +Cc: 52338
Hey,
The Cuirass web interface logs were quite silent this morning and I
suspected an issue somewhere. I then realized that you did update the
Nginx conf and the bots were no longer knocking at our door, which is
great!
Thanks to both of you,
Mathieu
^ permalink raw reply [flat|nested] 10+ messages in thread
* bug#52338: Crawler bots are downloading substitutes
2021-12-11 9:46 ` Mathieu Othacehe
@ 2021-12-19 16:53 ` Mathieu Othacehe
0 siblings, 0 replies; 10+ messages in thread
From: Mathieu Othacehe @ 2021-12-19 16:53 UTC (permalink / raw)
To: 52338-done
> Thanks to both of you,
And closing!
Mathieu
^ permalink raw reply [flat|nested] 10+ messages in thread
end of thread, other threads:[~2021-12-19 16:55 UTC | newest]
Thread overview: 10+ messages (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
2021-12-06 21:20 bug#52338: Crawler bots are downloading substitutes Leo Famulari
2021-12-06 22:18 ` bug#52338: [maintenance] hydra: berlin: Create robots.txt Leo Famulari
2021-12-09 13:27 ` bug#52338: Crawler bots are downloading substitutes Mathieu Othacehe
2021-12-09 15:42 ` Tobias Geerinckx-Rice via Bug reports for GNU Guix
2021-12-10 16:22 ` Leo Famulari
2021-12-10 16:47 ` Tobias Geerinckx-Rice via Bug reports for GNU Guix
2021-12-11 9:46 ` Mathieu Othacehe
2021-12-19 16:53 ` Mathieu Othacehe
2021-12-10 21:21 ` Mark H Weaver
2021-12-10 22:52 ` Tobias Geerinckx-Rice via Bug reports for GNU Guix
Code repositories for project(s) associated with this external index
https://git.savannah.gnu.org/cgit/guix.git
This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.