Last active
December 7, 2016 23:40
-
-
Save hans2103/5729749 to your computer and use it in GitHub Desktop.
Magento robots.txt as published on our Byte blog Magento Robots.txt [http://www.byte.nl/blog/magento-robots-txt/]
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
# $Id: robots.txt,v magento-specific 2010/28/01 18:24:19 goba Exp $ | |
# Edited: 2014/10/30 by Byte.nl | |
# | |
# robots.txt | |
# | |
# This file is to prevent the crawling and indexing of certain parts | |
# of your site by web crawlers and spiders run by sites like Yahoo! | |
# and Google. By telling these “robots” where not to go on your site, | |
# you save bandwidth and server resources. | |
# | |
# This file will be ignored unless it is at the root of your host: | |
# Used: http://example.com/robots.txt | |
# Ignored: http://example.com/site/robots.txt | |
# | |
# For more informationsk abocut the robots.txt standard, see: | |
# http://www.robotstxt.org/wc/robots.html | |
# | |
# For syntax checking, see: | |
# http://www.sxw.org.uk/computing/robots/check.html | |
# | |
# Prevent blocking URL parameters with robots.txt | |
# Use Google Webmaster Tools > Crawl > Url parameters instead | |
# Website Sitemap | |
Sitemap: http://www.example.com/sitemap.xml | |
# Crawlers Setup | |
User-agent: * | |
Crawl-delay: 10 | |
# Allowable Index | |
# Mind that Allow is not an official standard | |
Allow: /*?p= | |
Allow: /index.php/blog/ | |
Allow: /catalog/seo_sitemap/category/ | |
# Allow: /catalogsearch/result/ | |
Allow: /media/catalog/ | |
# Directories | |
Disallow: /404/ | |
Disallow: /app/ | |
Disallow: /cgi-bin/ | |
Disallow: /downloader/ | |
Disallow: /errors/ | |
Disallow: /includes/ | |
Disallow: /js/ | |
Disallow: /lib/ | |
Disallow: /magento/ | |
# Disallow: /media/ | |
Disallow: /media/captcha/ | |
# Disallow: /media/catalog/ | |
#Disallow: /media/css/ | |
#Disallow: /media/css_secure/ | |
Disallow: /media/customer/ | |
Disallow: /media/dhl/ | |
Disallow: /media/downloadable/ | |
Disallow: /media/import/ | |
#Disallow: /media/js/ | |
Disallow: /media/pdf/ | |
Disallow: /media/sales/ | |
Disallow: /media/tmp/ | |
Disallow: /media/wysiwyg/ | |
Disallow: /media/xmlconnect/ | |
Disallow: /pkginfo/ | |
Disallow: /report/ | |
Disallow: /scripts/ | |
Disallow: /shell/ | |
#Disallow: /skin/ | |
Disallow: /stats/ | |
Disallow: /var/ | |
# Paths (clean URLs) | |
Disallow: /index.php/ | |
Disallow: /catalog/product_compare/ | |
Disallow: /catalog/category/view/ | |
Disallow: /catalog/product/view/ | |
Disallow: /catalog/product/gallery/ | |
Disallow: /catalogsearch/ | |
Disallow: /checkout/ | |
Disallow: /control/ | |
Disallow: /contacts/ | |
Disallow: /customer/ | |
Disallow: /customize/ | |
Disallow: /newsletter/ | |
Disallow: /poll/ | |
Disallow: /review/ | |
Disallow: /sendfriend/ | |
Disallow: /tag/ | |
Disallow: /wishlist/ | |
# Files | |
Disallow: /cron.php | |
Disallow: /cron.sh | |
Disallow: /error_log | |
Disallow: /install.php | |
Disallow: /LICENSE.html | |
Disallow: /LICENSE.txt | |
Disallow: /LICENSE_AFL.txt | |
Disallow: /STATUS.txt | |
Disallow: /get.php # Magento 1.5+ | |
# Paths (no clean URLs) | |
#Disallow: /*.js$ | |
#Disallow: /*.css$ | |
Disallow: /*.php$ | |
Disallow: /*?SID= | |
Disallow: /rss* | |
Disallow: /*PHPSESSID |
Hi! Thank you for this Magento robots.txt version. I found more examples here http://sam-ecommerce.com/blog/view/magento-robots.txt-file-examples. Could you help me to choose the most appropriate one?
Don't "disallow" /checkout as Google find's checkout pages very interesting nowadays!
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Google has come a very long way as far as being able to detect what is correct and what is not on a website. In my opinion this is based on somebody who focuses on search every day not development so we may have differing views. Googlebot should be able to crawl everything unless you are intentionally trying to keep it at bay e.g. building a new site or in the case of any sort of affiliate links or paid advertisements. You must block those. Outside of that I do understand reasons for blocking bots but Google has to become smart enough that you should not need to disallow crawling of e.g. .PHP obviously Google can only crawl HTML, JS, CSS and now even Ajax however because as you guys know PHP is a server-side language it cannot be crawled.
is another thing to think about is simply using regex to remove the .html .php whatever's that we do not want the bot to crawl at the end of the URL. Rather than trying to block it via robots.txt which is in a very inefficient method of crawling the website. Because every time you plug something via robots.txt and you link to anything that links to that file that is blocked Google will crawl it anyway.
If you build a development server you should use a password system to block the bot.