summaryrefslogtreecommitdiff
path: root/www
diff options
context:
space:
mode:
authorAlex Xu <alex_y_xu@yahoo.ca>2013-10-02 06:37:33 -0400
committerAlex Xu <alex_y_xu@yahoo.ca>2013-10-02 06:37:58 -0400
commit4c9c091f7b5240b8c665cc35d634bd45815ab66b (patch)
tree15c0cd73a5f42361126053e45e54051875583630 /www
parente888c09c869d5b5059eaea1d6bc16d3633ff0782 (diff)
downloadhtml5ks-4c9c091f7b5240b8c665cc35d634bd45815ab66b.tar.xz
html5ks-4c9c091f7b5240b8c665cc35d634bd45815ab66b.zip
fix minor bug, add robots.txt
Diffstat (limited to 'www')
-rw-r--r--www/js/menu.js2
-rw-r--r--www/robots.txt120
2 files changed, 121 insertions, 1 deletions
diff --git a/www/js/menu.js b/www/js/menu.js
index f2fbacd..2715e26 100644
--- a/www/js/menu.js
+++ b/www/js/menu.js
@@ -28,7 +28,7 @@
dialogs: document.getElementById("dialogs"),
dialog: {
options: document.getElementById("options"),
- retn: document.getElementById("retn")
+ retn: document.getElementById("return")
},
mainMenu: document.getElementById("main-menu"),
main: {
diff --git a/www/robots.txt b/www/robots.txt
new file mode 100644
index 0000000..5a4e2f8
--- /dev/null
+++ b/www/robots.txt
@@ -0,0 +1,120 @@
+#
+# robots.txt - shamelessly stolen from http://www.wikipedia.org/
+#
+# Please note: There are a lot of pages on this site, and there are
+# some misbehaved spiders out there that go _way_ too fast. If you're
+# irresponsible, your access to the site may be blocked.
+#
+
+# advertising-related bots:
+User-agent: Mediapartners-Google*
+Disallow: /
+
+# Wikipedia work bots:
+User-agent: IsraBot
+Disallow:
+
+User-agent: Orthogaffe
+Disallow:
+
+# Crawlers that are kind enough to obey, but which we'd rather not have
+# unless they're feeding search engines.
+User-agent: UbiCrawler
+Disallow: /
+
+User-agent: DOC
+Disallow: /
+
+User-agent: Zao
+Disallow: /
+
+# Some bots are known to be trouble, particularly those designed to copy
+# entire sites. Please obey robots.txt.
+User-agent: sitecheck.internetseer.com
+Disallow: /
+
+User-agent: Zealbot
+Disallow: /
+
+User-agent: MSIECrawler
+Disallow: /
+
+User-agent: SiteSnagger
+Disallow: /
+
+User-agent: WebStripper
+Disallow: /
+
+User-agent: WebCopier
+Disallow: /
+
+User-agent: Fetch
+Disallow: /
+
+User-agent: Offline Explorer
+Disallow: /
+
+User-agent: Teleport
+Disallow: /
+
+User-agent: TeleportPro
+Disallow: /
+
+User-agent: WebZIP
+Disallow: /
+
+User-agent: linko
+Disallow: /
+
+User-agent: HTTrack
+Disallow: /
+
+User-agent: Microsoft.URL.Control
+Disallow: /
+
+User-agent: Xenu
+Disallow: /
+
+User-agent: larbin
+Disallow: /
+
+User-agent: libwww
+Disallow: /
+
+User-agent: ZyBORG
+Disallow: /
+
+User-agent: Download Ninja
+Disallow: /
+
+#
+# Sorry, wget in its recursive mode is a frequent problem.
+# Please read the man page and use it properly; there is a
+# --wait option you can use to set the delay between hits,
+# for instance.
+#
+User-agent: wget
+Disallow: /
+
+#
+# The 'grub' distributed client has been *very* poorly behaved.
+#
+User-agent: grub-client
+Disallow: /
+
+#
+# Doesn't follow robots.txt anyway, but...
+#
+User-agent: k2spider
+Disallow: /
+
+#
+# Hits many times per second, not acceptable
+# http://www.nameprotect.com/botinfo.html
+User-agent: NPBot
+Disallow: /
+
+# A capture bot, downloads gazillions of pages with no public benefit
+# http://www.webreaper.net/
+User-agent: WebReaper
+Disallow: /