maintain own robots.txt instead of using upstraem and disallow some annoying seo marketing crap bots from logs #34
|
@ -57,6 +57,7 @@
|
||||||
mode: 0755
|
mode: 0755
|
||||||
tags:
|
tags:
|
||||||
- authelia-nginx
|
- authelia-nginx
|
||||||
|
- authelia-robots
|
||||||
notify: reload nginx
|
notify: reload nginx
|
||||||
|
|
||||||
- name: template nginx vhost
|
- name: template nginx vhost
|
||||||
|
|
|
@ -61,6 +61,7 @@
|
||||||
- sitemap.xml
|
- sitemap.xml
|
||||||
tags:
|
tags:
|
||||||
- robots
|
- robots
|
||||||
|
- gitea-robots
|
||||||
|
|
||||||
- name: copy gitea templates
|
- name: copy gitea templates
|
||||||
copy:
|
copy:
|
||||||
|
|
|
@ -1,6 +1,37 @@
|
||||||
User-agent: baidu
|
User-agent: baidu
|
||||||
crawl-delay: 1
|
crawl-delay: 1
|
||||||
|
|
||||||
|
{# SEO/SEM #}
|
||||||
|
User-agent: SemrushBot
|
||||||
|
Disallow: /
|
||||||
|
|
||||||
|
{# SEO/SEM #}
|
||||||
|
User-agent: AhrefsBot
|
||||||
|
Disallow: /
|
||||||
|
|
||||||
|
{# SEO/SEM #}
|
||||||
|
User-agent: DataForSeoBot
|
||||||
|
Disallow: /
|
||||||
|
|
||||||
|
{# https://www.ionos.de/terms-gtc/faq-crawler/ #}
|
||||||
|
{# disallowed on principle for not having that page in english #}
|
||||||
|
User-agent: IonCrawl
|
||||||
|
Disallow: /
|
||||||
|
|
||||||
|
{# SEO/SEM #}
|
||||||
|
User-agent: barkrowler
|
||||||
|
Disallow: /
|
||||||
|
|
||||||
|
{# SEO/SEM #}
|
||||||
|
User-agent: seoscanners.net
|
||||||
|
Disallow: /
|
||||||
|
|
||||||
|
{# SEO/SEM #}
|
||||||
|
User-agent: MegaIndex.ru
|
||||||
|
Disallow: /
|
||||||
|
User-agent: megaindex.com
|
||||||
|
Disallow: /
|
||||||
|
|
||||||
User-agent: *
|
User-agent: *
|
||||||
|
|
||||||
Disallow: /*/pulse
|
Disallow: /*/pulse
|
||||||
|
|
|
@ -209,6 +209,17 @@
|
||||||
daemon_reload: true
|
daemon_reload: true
|
||||||
name: jellyfin_auth
|
name: jellyfin_auth
|
||||||
|
|
||||||
|
- name: template robots.txt
|
||||||
|
template:
|
||||||
|
src: "robots.txt.j2"
|
||||||
|
dest: "{{ systemuserlist.jellyfin.home }}/robots.txt"
|
||||||
|
owner: jellyfin
|
||||||
|
group: jellyfin
|
||||||
|
mode: 0755
|
||||||
|
tags:
|
||||||
|
- jellyfin-nginx
|
||||||
|
- jellyfin-robots
|
||||||
|
notify: reload nginx
|
||||||
|
|
||||||
- name: template jellyfin nginx config
|
- name: template jellyfin nginx config
|
||||||
template:
|
template:
|
||||||
|
|
|
@ -69,6 +69,14 @@ server {
|
||||||
return 302 https://$host/web/;
|
return 302 https://$host/web/;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
# jellyfin-web already has a robots.txt file that disallows everything, but we still want to maintain our own.
|
||||||
|
# jellyfin (not -web) will issue a 302 redirect from {{ jellyfin_url }}/robots.txt to {{ jellyfin_url }}/web/robots.txt
|
||||||
|
# where the file is then served from.
|
||||||
|
#
|
||||||
|
# https://github.com/jellyfin/jellyfin-web/blob/master/src/robots.txt
|
||||||
|
location = /robots.txt {
|
||||||
|
alias {{ systemuserlist.jellyfin.home }}/robots.txt;
|
||||||
|
}
|
||||||
location / {
|
location / {
|
||||||
#include /etc/nginx/jellyfin/jellyfin_auth.conf;
|
#include /etc/nginx/jellyfin/jellyfin_auth.conf;
|
||||||
|
|
||||||
|
|
|
@ -0,0 +1,2 @@
|
||||||
|
User-agent: *
|
||||||
|
Disallow: /
|
Loading…
Reference in New Issue