From 4c9c091f7b5240b8c665cc35d634bd45815ab66b Mon Sep 17 00:00:00 2001 From: Alex Xu Date: Wed, 2 Oct 2013 06:37:33 -0400 Subject: fix minor bug, add robots.txt --- www/js/menu.js | 2 +- www/robots.txt | 120 +++++++++++++++++++++++++++++++++++++++++++++++++++++++++ 2 files changed, 121 insertions(+), 1 deletion(-) create mode 100644 www/robots.txt diff --git a/www/js/menu.js b/www/js/menu.js index f2fbacd..2715e26 100644 --- a/www/js/menu.js +++ b/www/js/menu.js @@ -28,7 +28,7 @@ dialogs: document.getElementById("dialogs"), dialog: { options: document.getElementById("options"), - retn: document.getElementById("retn") + retn: document.getElementById("return") }, mainMenu: document.getElementById("main-menu"), main: { diff --git a/www/robots.txt b/www/robots.txt new file mode 100644 index 0000000..5a4e2f8 --- /dev/null +++ b/www/robots.txt @@ -0,0 +1,120 @@ +# +# robots.txt - shamelessly stolen from http://www.wikipedia.org/ +# +# Please note: There are a lot of pages on this site, and there are +# some misbehaved spiders out there that go _way_ too fast. If you're +# irresponsible, your access to the site may be blocked. +# + +# advertising-related bots: +User-agent: Mediapartners-Google* +Disallow: / + +# Wikipedia work bots: +User-agent: IsraBot +Disallow: + +User-agent: Orthogaffe +Disallow: + +# Crawlers that are kind enough to obey, but which we'd rather not have +# unless they're feeding search engines. +User-agent: UbiCrawler +Disallow: / + +User-agent: DOC +Disallow: / + +User-agent: Zao +Disallow: / + +# Some bots are known to be trouble, particularly those designed to copy +# entire sites. Please obey robots.txt. +User-agent: sitecheck.internetseer.com +Disallow: / + +User-agent: Zealbot +Disallow: / + +User-agent: MSIECrawler +Disallow: / + +User-agent: SiteSnagger +Disallow: / + +User-agent: WebStripper +Disallow: / + +User-agent: WebCopier +Disallow: / + +User-agent: Fetch +Disallow: / + +User-agent: Offline Explorer +Disallow: / + +User-agent: Teleport +Disallow: / + +User-agent: TeleportPro +Disallow: / + +User-agent: WebZIP +Disallow: / + +User-agent: linko +Disallow: / + +User-agent: HTTrack +Disallow: / + +User-agent: Microsoft.URL.Control +Disallow: / + +User-agent: Xenu +Disallow: / + +User-agent: larbin +Disallow: / + +User-agent: libwww +Disallow: / + +User-agent: ZyBORG +Disallow: / + +User-agent: Download Ninja +Disallow: / + +# +# Sorry, wget in its recursive mode is a frequent problem. +# Please read the man page and use it properly; there is a +# --wait option you can use to set the delay between hits, +# for instance. +# +User-agent: wget +Disallow: / + +# +# The 'grub' distributed client has been *very* poorly behaved. +# +User-agent: grub-client +Disallow: / + +# +# Doesn't follow robots.txt anyway, but... +# +User-agent: k2spider +Disallow: / + +# +# Hits many times per second, not acceptable +# http://www.nameprotect.com/botinfo.html +User-agent: NPBot +Disallow: / + +# A capture bot, downloads gazillions of pages with no public benefit +# http://www.webreaper.net/ +User-agent: WebReaper +Disallow: / -- cgit v1.2.3-70-g09d2