-
-
Save hans2103/5729749 to your computer and use it in GitHub Desktop.
# $Id: robots.txt,v magento-specific 2010/28/01 18:24:19 goba Exp $ | |
# Edited: 2014/10/30 by Byte.nl | |
# | |
# robots.txt | |
# | |
# This file is to prevent the crawling and indexing of certain parts | |
# of your site by web crawlers and spiders run by sites like Yahoo! | |
# and Google. By telling these “robots” where not to go on your site, | |
# you save bandwidth and server resources. | |
# | |
# This file will be ignored unless it is at the root of your host: | |
# Used: http://example.com/robots.txt | |
# Ignored: http://example.com/site/robots.txt | |
# | |
# For more informationsk abocut the robots.txt standard, see: | |
# http://www.robotstxt.org/wc/robots.html | |
# | |
# For syntax checking, see: | |
# http://www.sxw.org.uk/computing/robots/check.html | |
# | |
# Prevent blocking URL parameters with robots.txt | |
# Use Google Webmaster Tools > Crawl > Url parameters instead | |
# Website Sitemap | |
Sitemap: http://www.example.com/sitemap.xml | |
# Crawlers Setup | |
User-agent: * | |
Crawl-delay: 10 | |
# Allowable Index | |
# Mind that Allow is not an official standard | |
Allow: /*?p= | |
Allow: /index.php/blog/ | |
Allow: /catalog/seo_sitemap/category/ | |
# Allow: /catalogsearch/result/ | |
Allow: /media/catalog/ | |
# Directories | |
Disallow: /404/ | |
Disallow: /app/ | |
Disallow: /cgi-bin/ | |
Disallow: /downloader/ | |
Disallow: /errors/ | |
Disallow: /includes/ | |
Disallow: /js/ | |
Disallow: /lib/ | |
Disallow: /magento/ | |
# Disallow: /media/ | |
Disallow: /media/captcha/ | |
# Disallow: /media/catalog/ | |
#Disallow: /media/css/ | |
#Disallow: /media/css_secure/ | |
Disallow: /media/customer/ | |
Disallow: /media/dhl/ | |
Disallow: /media/downloadable/ | |
Disallow: /media/import/ | |
#Disallow: /media/js/ | |
Disallow: /media/pdf/ | |
Disallow: /media/sales/ | |
Disallow: /media/tmp/ | |
Disallow: /media/wysiwyg/ | |
Disallow: /media/xmlconnect/ | |
Disallow: /pkginfo/ | |
Disallow: /report/ | |
Disallow: /scripts/ | |
Disallow: /shell/ | |
#Disallow: /skin/ | |
Disallow: /stats/ | |
Disallow: /var/ | |
# Paths (clean URLs) | |
Disallow: /index.php/ | |
Disallow: /catalog/product_compare/ | |
Disallow: /catalog/category/view/ | |
Disallow: /catalog/product/view/ | |
Disallow: /catalog/product/gallery/ | |
Disallow: /catalogsearch/ | |
Disallow: /checkout/ | |
Disallow: /control/ | |
Disallow: /contacts/ | |
Disallow: /customer/ | |
Disallow: /customize/ | |
Disallow: /newsletter/ | |
Disallow: /poll/ | |
Disallow: /review/ | |
Disallow: /sendfriend/ | |
Disallow: /tag/ | |
Disallow: /wishlist/ | |
# Files | |
Disallow: /cron.php | |
Disallow: /cron.sh | |
Disallow: /error_log | |
Disallow: /install.php | |
Disallow: /LICENSE.html | |
Disallow: /LICENSE.txt | |
Disallow: /LICENSE_AFL.txt | |
Disallow: /STATUS.txt | |
Disallow: /get.php # Magento 1.5+ | |
# Paths (no clean URLs) | |
#Disallow: /*.js$ | |
#Disallow: /*.css$ | |
Disallow: /*.php$ | |
Disallow: /*?SID= | |
Disallow: /rss* | |
Disallow: /*PHPSESSID |
Matt Cutts recommends not blocking .css and .js files for better crawling
https://www.youtube.com/watch?v=qglAm8QiX5k#t=1000
Google has come a very long way as far as being able to detect what is correct and what is not on a website. In my opinion this is based on somebody who focuses on search every day not development so we may have differing views. Googlebot should be able to crawl everything unless you are intentionally trying to keep it at bay e.g. building a new site or in the case of any sort of affiliate links or paid advertisements. You must block those. Outside of that I do understand reasons for blocking bots but Google has to become smart enough that you should not need to disallow crawling of e.g. .PHP obviously Google can only crawl HTML, JS, CSS and now even Ajax however because as you guys know PHP is a server-side language it cannot be crawled.
is another thing to think about is simply using regex to remove the .html .php whatever's that we do not want the bot to crawl at the end of the URL. Rather than trying to block it via robots.txt which is in a very inefficient method of crawling the website. Because every time you plug something via robots.txt and you link to anything that links to that file that is blocked Google will crawl it anyway.
If you build a development server you should use a password system to block the bot.
Hi! Thank you for this Magento robots.txt version. I found more examples here http://sam-ecommerce.com/blog/view/magento-robots.txt-file-examples. Could you help me to choose the most appropriate one?
Don't "disallow" /checkout as Google find's checkout pages very interesting nowadays!
added comment to robots.txt
Prevent blocking URL parameters with robots.txt
Use Google Webmaster Tools > Crawl > Url parameters instead