hankknight
asked on
robots.txt: exclude all pages that have ? in url
How can I prevent robots from following dynamic pages liks this?
http://www.example.com/page?item=value
http://www.example.com/page?item=value
ASKER
No, my question is how to limit crawling of ALL pages that have ? in it.
http://www.example.com/page?item=value
http://www.example.com/?foo=bar
http://www.example.com/cgi-bin/run.pl?item=value
http://www.example.com/script.php?sid=456345635654656
http://www.example.com/page?item=value
http://www.example.com/?foo=bar
http://www.example.com/cgi-bin/run.pl?item=value
http://www.example.com/script.php?sid=456345635654656
ASKER CERTIFIED SOLUTION
membership
This solution is only available to members.
To access this solution, you must be a member of Experts Exchange.
Disallow: /page
or to disallow all pages (not recommended)
User-agent: *
Disallow: /