Skip to content

Instantly share code, notes, and snippets.

@hans2103
Last active December 7, 2016 23:40
Show Gist options
  • Star 7 You must be signed in to star a gist
  • Fork 0 You must be signed in to fork a gist
  • Save hans2103/5729749 to your computer and use it in GitHub Desktop.
Save hans2103/5729749 to your computer and use it in GitHub Desktop.
Magento robots.txt as published on our Byte blog Magento Robots.txt [http://www.byte.nl/blog/magento-robots-txt/]
# $Id: robots.txt,v magento-specific 2010/28/01 18:24:19 goba Exp $
# Edited: 2014/10/30 by Byte.nl
#
# robots.txt
#
# This file is to prevent the crawling and indexing of certain parts
# of your site by web crawlers and spiders run by sites like Yahoo!
# and Google. By telling these “robots” where not to go on your site,
# you save bandwidth and server resources.
#
# This file will be ignored unless it is at the root of your host:
# Used: http://example.com/robots.txt
# Ignored: http://example.com/site/robots.txt
#
# For more informationsk abocut the robots.txt standard, see:
# http://www.robotstxt.org/wc/robots.html
#
# For syntax checking, see:
# http://www.sxw.org.uk/computing/robots/check.html
#
# Prevent blocking URL parameters with robots.txt
# Use Google Webmaster Tools > Crawl > Url parameters instead
# Website Sitemap
Sitemap: http://www.example.com/sitemap.xml
# Crawlers Setup
User-agent: *
Crawl-delay: 10
# Allowable Index
# Mind that Allow is not an official standard
Allow: /*?p=
Allow: /index.php/blog/
Allow: /catalog/seo_sitemap/category/
# Allow: /catalogsearch/result/
Allow: /media/catalog/
# Directories
Disallow: /404/
Disallow: /app/
Disallow: /cgi-bin/
Disallow: /downloader/
Disallow: /errors/
Disallow: /includes/
Disallow: /js/
Disallow: /lib/
Disallow: /magento/
# Disallow: /media/
Disallow: /media/captcha/
# Disallow: /media/catalog/
#Disallow: /media/css/
#Disallow: /media/css_secure/
Disallow: /media/customer/
Disallow: /media/dhl/
Disallow: /media/downloadable/
Disallow: /media/import/
#Disallow: /media/js/
Disallow: /media/pdf/
Disallow: /media/sales/
Disallow: /media/tmp/
Disallow: /media/wysiwyg/
Disallow: /media/xmlconnect/
Disallow: /pkginfo/
Disallow: /report/
Disallow: /scripts/
Disallow: /shell/
#Disallow: /skin/
Disallow: /stats/
Disallow: /var/
# Paths (clean URLs)
Disallow: /index.php/
Disallow: /catalog/product_compare/
Disallow: /catalog/category/view/
Disallow: /catalog/product/view/
Disallow: /catalog/product/gallery/
Disallow: /catalogsearch/
Disallow: /checkout/
Disallow: /control/
Disallow: /contacts/
Disallow: /customer/
Disallow: /customize/
Disallow: /newsletter/
Disallow: /poll/
Disallow: /review/
Disallow: /sendfriend/
Disallow: /tag/
Disallow: /wishlist/
# Files
Disallow: /cron.php
Disallow: /cron.sh
Disallow: /error_log
Disallow: /install.php
Disallow: /LICENSE.html
Disallow: /LICENSE.txt
Disallow: /LICENSE_AFL.txt
Disallow: /STATUS.txt
Disallow: /get.php # Magento 1.5+
# Paths (no clean URLs)
#Disallow: /*.js$
#Disallow: /*.css$
Disallow: /*.php$
Disallow: /*?SID=
Disallow: /rss*
Disallow: /*PHPSESSID
@hans2103
Copy link
Author

added comment to robots.txt

Prevent blocking URL parameters with robots.txt

Use Google Webmaster Tools > Crawl > Url parameters instead

@hans2103
Copy link
Author

Matt Cutts recommends not blocking .css and .js files for better crawling
https://www.youtube.com/watch?v=qglAm8QiX5k#t=1000

https://yoast.com/google-panda-robots-css-js/

@blueprintmrk
Copy link

Google has come a very long way as far as being able to detect what is correct and what is not on a website. In my opinion this is based on somebody who focuses on search every day not development so we may have differing views. Googlebot should be able to crawl everything unless you are intentionally trying to keep it at bay e.g. building a new site or in the case of any sort of affiliate links or paid advertisements. You must block those. Outside of that I do understand reasons for blocking bots but Google has to become smart enough that you should not need to disallow crawling of e.g. .PHP obviously Google can only crawl HTML, JS, CSS and now even Ajax however because as you guys know PHP is a server-side language it cannot be crawled.

is another thing to think about is simply using regex to remove the .html .php whatever's that we do not want the bot to crawl at the end of the URL. Rather than trying to block it via robots.txt which is in a very inefficient method of crawling the website. Because every time you plug something via robots.txt and you link to anything that links to that file that is blocked Google will crawl it anyway.

If you build a development server you should use a password system to block the bot.

@KateGum
Copy link

KateGum commented Aug 18, 2016

Hi! Thank you for this Magento robots.txt version. I found more examples here http://sam-ecommerce.com/blog/view/magento-robots.txt-file-examples. Could you help me to choose the most appropriate one?

@riconeitzel
Copy link

Don't "disallow" /checkout as Google find's checkout pages very interesting nowadays!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment