This tutorial will teach you how to wack out a robots.txt file that you can use to both show all search engine spiders where you put your sitemap.xml file (thanx Duncan ) & stop spiders from indexing those directories that you donât want them to.
1) Begin by making a new Design. Call it something like âBlank Designâ (this can be useful else where too).
Now add the following code to the parse file and save:
[b]2)[/b] Then create and standard page at the highest level (top) of your site (naming it something sensible, it doesnât have to be robots.txt but be sure to change the web path to ârobots.txtâ). Spiders need to see the robots.txt file in the root of the site.
[b]3)[/b] Edit the âPage Contentsâ screen and make sure that it is presenting [b]RAW HTML[/b]. Then add (as much as you want of) the following code (here Iâm disallowing Google Images, Yahoo Media & psbot from indexing our site):
Sitemap: http://www.yourdomain.com/sitemap.xmlUser-agent: Googlebot-Image Disallow: / User-agent: Yahoo-MMCrawler Disallow: / User-agent: Yahoo-MMAudVid Disallow: / User-agent: psbot Disallow: / User-agent: * Disallow: /_admin</pre><br />
See: http://www.google.com/support/webmasters/bin/answer.py?answer=40360 & http://www.google.com/support/webmasters/bin/answer.py?answer=40360 for more info
[b]4)[/b] Make it all go live & then sit back and relax.