diff options
author | Alex Xu <alex_y_xu@yahoo.ca> | 2013-10-02 06:37:33 -0400 |
---|---|---|
committer | Alex Xu <alex_y_xu@yahoo.ca> | 2013-10-02 06:37:58 -0400 |
commit | 4c9c091f7b5240b8c665cc35d634bd45815ab66b (patch) | |
tree | 15c0cd73a5f42361126053e45e54051875583630 /www | |
parent | e888c09c869d5b5059eaea1d6bc16d3633ff0782 (diff) | |
download | html5ks-4c9c091f7b5240b8c665cc35d634bd45815ab66b.tar.xz html5ks-4c9c091f7b5240b8c665cc35d634bd45815ab66b.zip |
fix minor bug, add robots.txt
Diffstat (limited to 'www')
-rw-r--r-- | www/js/menu.js | 2 | ||||
-rw-r--r-- | www/robots.txt | 120 |
2 files changed, 121 insertions, 1 deletions
diff --git a/www/js/menu.js b/www/js/menu.js index f2fbacd..2715e26 100644 --- a/www/js/menu.js +++ b/www/js/menu.js @@ -28,7 +28,7 @@ dialogs: document.getElementById("dialogs"), dialog: { options: document.getElementById("options"), - retn: document.getElementById("retn") + retn: document.getElementById("return") }, mainMenu: document.getElementById("main-menu"), main: { diff --git a/www/robots.txt b/www/robots.txt new file mode 100644 index 0000000..5a4e2f8 --- /dev/null +++ b/www/robots.txt @@ -0,0 +1,120 @@ +# +# robots.txt - shamelessly stolen from http://www.wikipedia.org/ +# +# Please note: There are a lot of pages on this site, and there are +# some misbehaved spiders out there that go _way_ too fast. If you're +# irresponsible, your access to the site may be blocked. +# + +# advertising-related bots: +User-agent: Mediapartners-Google* +Disallow: / + +# Wikipedia work bots: +User-agent: IsraBot +Disallow: + +User-agent: Orthogaffe +Disallow: + +# Crawlers that are kind enough to obey, but which we'd rather not have +# unless they're feeding search engines. +User-agent: UbiCrawler +Disallow: / + +User-agent: DOC +Disallow: / + +User-agent: Zao +Disallow: / + +# Some bots are known to be trouble, particularly those designed to copy +# entire sites. Please obey robots.txt. +User-agent: sitecheck.internetseer.com +Disallow: / + +User-agent: Zealbot +Disallow: / + +User-agent: MSIECrawler +Disallow: / + +User-agent: SiteSnagger +Disallow: / + +User-agent: WebStripper +Disallow: / + +User-agent: WebCopier +Disallow: / + +User-agent: Fetch +Disallow: / + +User-agent: Offline Explorer +Disallow: / + +User-agent: Teleport +Disallow: / + +User-agent: TeleportPro +Disallow: / + +User-agent: WebZIP +Disallow: / + +User-agent: linko +Disallow: / + +User-agent: HTTrack +Disallow: / + +User-agent: Microsoft.URL.Control +Disallow: / + +User-agent: Xenu +Disallow: / + +User-agent: larbin +Disallow: / + +User-agent: libwww +Disallow: / + +User-agent: ZyBORG +Disallow: / + +User-agent: Download Ninja +Disallow: / + +# +# Sorry, wget in its recursive mode is a frequent problem. +# Please read the man page and use it properly; there is a +# --wait option you can use to set the delay between hits, +# for instance. +# +User-agent: wget +Disallow: / + +# +# The 'grub' distributed client has been *very* poorly behaved. +# +User-agent: grub-client +Disallow: / + +# +# Doesn't follow robots.txt anyway, but... +# +User-agent: k2spider +Disallow: / + +# +# Hits many times per second, not acceptable +# http://www.nameprotect.com/botinfo.html +User-agent: NPBot +Disallow: / + +# A capture bot, downloads gazillions of pages with no public benefit +# http://www.webreaper.net/ +User-agent: WebReaper +Disallow: / |