![]() |
robots.txt for 3.8.2 - Any Ideas?
Hello and Greetings,
I have just noticed that Google Webmaster Tools is complaining about a LOT of URLs being restricted by my robots.txt file. Is anyone else having this problem? If not, can I get an example of a robots.txt written for 3.8.2? I tweaked mine thinking that I was fixing a duplicate content problem but I apparently crossed the line on it :D Any ideas, suggestions greatly appreciated. TIA |
What is it currently?
|
Sorry, should have thought to post it:
Quote:
|
Two points in addition to above robots.txt:-
1. We should not give out our admincp directory in robots.txt as it makes the location displayable to the world. What is the use of renaming admincp feature then? 2. Also its good to give a referance to our sitemap at the end of robots.txt Sitemap: http://site.com/sitemap_index.xml.gz |
Thanks for the comments, had not thought about the sitemap reference in there. thanks for that. I double password protect my admincp folder though I could easily take it out of the list altogether as the bots can not access it anyway so thanks for that comment as well.
|
Anyone see any problem with the content of this robots.txt or have any idea how to fix the "google" complaining about the restrictions? TIA
|
Quote:
I pretty much followed the advice in this article, and have had not complaints from google: http://www.theadminzone.com/forums/s...ad.php?t=19872 |
As a matter of fact I did use those guidelines to construct my robots.txt (and the follow on suggestions in that thread). I forgot about that, thanks for reminding me about it.
|
I do have a follow on question on the robots.txt file that I am currently using. I have the vbulletin blog software installed on this site as well as wordpress. I have not disallows in this robots.txt for any blog files. I would not have thought anything about it except that I just looked at my sitemap and see a huge number of URLs for blog stuff that doesn't really exist like archives from 1983?
Anyone have a suggestion about a sensible robots.txt entry for both the vbulletin blog and wordpress? TIA for any ideas. |
Sounds like a sitemap issue not a robot.txt issue. I know that's not an answer per say, but I'd be looking at why your sitemap contains links that don't exist, instead.
|
I can exclude it from the sitemap for sure but I thought it was pretty strange to be referencing archives from 1970 through current. The links looks like:
hxxp://www.mysite.com/blog.php?do=list&m=12&y=1970 hxxp://www.mysite.com/blog.php?u=1&m=12&y=1970 hxxp://www.mysite.com/blog.php?do=list&m=11&y=1970 hxxp://www.mysite.com/blog.php?u=1&m=11&y=1970 hxxp://www.mysite.com/blog.php?do=list&m=10&y=1970 hxxp://www.mysite.com/blog.php?u=1&m=10&y=1970 hxxp://www.mysite.com/blog.php?do=list&m=9&y=1970 hxxp://www.mysite.com/blog.php?u=1&m=9&y=1970 hxxp://www.mysite.com/blog.php?do=list&m=8&y=1970 hxxp://www.mysite.com/blog.php?u=1&m=8&y=1970 hxxp://www.mysite.com/blog.php?do=list&m=7&y=1970 hxxp://www.mysite.com/blog.php?u=1&m=7&y=1970 hxxp://www.mysite.com/blog.php?do=list&m=6&y=1970 hxxp://www.mysite.com/blog.php?u=1&m=6&y=1970 hxxp://www.mysite.com/blog.php?do=list&m=5&y=1970 hxxp://www.mysite.com/blog.php?u=1&m=5&y=1970 hxxp://www.mysite.com/blog.php?do=list&m=4&y=1970 hxxp://www.mysite.com/blog.php?u=1&m=4&y=1970 hxxp://www.mysite.com/blog.php?do=list&m=3&y=1970 hxxp://www.mysite.com/blog.php?u=1&m=3&y=1970 hxxp://www.mysite.com/blog.php?do=list&m=2&y=1970 hxxp://www.mysite.com/blog.php?u=1&m=2&y=1970 hxxp://www.mysite.com/blog.php?do=list&m=1&y=1970 hxxp://www.mysite.com/blog.php?u=1&m=1&y=1970 The sitemap software is finding this for every year to current? |
Those are sorting URL's that google finds and goes crazy on. Block off blog.php and call it a day.
It will also do this for the calendar, usually on newer vBulletin sites for some reason if you use the site:http://sitename.com you'll see 500 pages of calender sorting URL's. Its fine to have the warning messages in GWT that tell you access is restricted to those URL's, they provide no value what so ever to your sites rankings, and should be blocked off as with most sorting URL's since it will just be duplicated content from else where. |
Thanks very much for the reply. I appreciate it. It fits with what I suspected on the duplicate content issues I am trying to solve. I will add blog.php to my robots.txt file and exclude it from my sitemap.
thanks again. --------------- Added [DATE]1240771303[/DATE] at [TIME]1240771303[/TIME] --------------- Thanks very much for the reply. I appreciate it. It fits with what I suspected on the duplicate content issues I am trying to solve. I will add blog.php to my robots.txt file and exclude it from my sitemap. thanks again. --------------- Added [DATE]1240797229[/DATE] at [TIME]1240797229[/TIME] --------------- Quote:
TIA for a reply. |
All times are GMT. The time now is 05:31 PM. |
Powered by vBulletin® Version 3.8.12 by vBS
Copyright ©2000 - 2025, vBulletin Solutions Inc.
X vBulletin 3.8.12 by vBS Debug Information | |
---|---|
|
|
![]() |
|
Template Usage:
Phrase Groups Available:
|
Included Files:
Hooks Called:
|