My URL structure is set up in two parallels (both lead to the same place ):
- www.mydomain.com/subname
- www.mydomain.com/123
The trouble is is that, the spiders are crawling into things like:
- www.mydomain.com/subname/default_media_function
- www.mydomain.com/subname/map_function
Note that the name "subname" represents thousands of different pages on my site that all have that same function.
And they are erroring out because those links are strictly for JSON or AJAX purposes and not actual links. I would like to block them from accessing those pages, but how would I do that if the URL contains a variable?
Would this work in robots.txt?
Disallow: /map_function