Port content over

This commit is contained in:
2020-11-24 22:46:02 +00:00
commit 6ca3367104
24 changed files with 1589 additions and 0 deletions

32
src/robots.txt Normal file
View File

@@ -0,0 +1,32 @@
# Being archived in a long-term store is harmful to my privacy. Never
# know when I might need to change something in a hurry
User-Agent: ia_archiver
Disallow: /
User-Agent: archiver
Disallow: /
# Search engines tend to update their indexes fairly quickly, so no
# objections to being indexed by them in general. That said, I want to
# do my own (tiny) part in making Google useless
# not contribute to
User-Agent: indexer
Disallow:
User-agent: Googlebot
Disallow: /
User-Agent: gus
Disallow:
# Research *should* only report anonymised aggregates, I can live with
# that
User-Agent: researcher
Disallow:
# I remain confused by the incluson of proxies in robots.txt, but am
# happy for them to access the site as long as they themselves forbid
# being indexed or archived. I can add exceptions if I find any that
# don't do that
User-Agent: webproxy
Disallow: