![]() |
Yet another ROBOTS.txt file question (sorry)
Can someone give me some advice on this please.
I have a public section and a private section of my forum. I want to allow bots from search engines to crawl and index my public forum in order to get some traffic through it. But I also want to absolutely deny any bots at all from access and crawling anything else. Can I specify a robots.txt that says allow 1 forum category and disallow everywhere else? I have searched but can't find any information on these specifics. Obviously I found lots of robots.txt stuff, but nothing that can be of specific help. Thanks for your help. |
Nope, that's not what robots.txt is for anyway.
|
ok thanks, so how can I achieve what I described?
any help on that? Thanks |
ACP->Forums & Moderators->Forum Permissions
Change the 'Unregistered/Not Logged In' permissions for forums you don't want viewed to all 'No'. With that setting, the forums won't even exist to robots or unregistered users. |
oh as simple as that, ok didn't realise, im still new to most of this. So what is the point of the robots.txt other than to deny all bots? THANKS
|
Robots.txt is good for asking robots not to access things like register.php, search.php, etc.
Note I said 'asking'. Not all robots obey robots.txt. |
for what reason would you do that? BTW thanks again for your patience and help
|
Many pages serve no purpose to a robot and are just a waste of bandwidth when they crawl those pages.
Register.php, login.php, search.php, subscription.php, profile.php are just a few of those types of pages. |
There are threads around with lists of the pages you might want to include in your robots.txt block.
Note that the blocking suggested above blocks not just robots but also unregistered visitors. If that is your intention, fine. If you want to allow everyone to see a particular forum and just don't want that forum's threads appearing in search engines then I don't believe you can do that with vB. |
Quote:
|
You said what, that it can't be done with vB? ;)
In fact, I made that post hoping someone would tell me I was wrong and point to a way one could allow visitors into a forum but prevent that forum's threads from showing up in the SEs. |
If the forum is set to private spiders will not be able to access it anyway
|
Setting it to private blocks unregistered visitors.
I know the logic may sound screwed up, but there is sense in such a setup. For example, where you want to allow links in the sub-forum but want to dissuade people posting links in there for the SEO benefit. |
Quote:
Quote:
|
ok all good info thanks for your time everyone.
I actually want a small section of my forum to be PUBLIC and I want search engines to index it. I just dont want to have 50 bots killing my bandwidth. any answer to that? Thanks a million for your help. |
bump...
1. Create a robots.txt to eliminate pages like profiles, search, etc from spiders (This will also help eliminate profiles and other pages from coming up in searches by potential traffic.) 2. Create permissions per forum for unregistered visitors like 'nhawk' described and be careful as visitors will not be able to view these forums either depending on configurations. 3. If you really wish to control which robots/spiders are accessing your forums and utilizing bandwidth, consider this add on. I do however recommend you read all 350 posts in order to guarantee your success. There is a lot to read on user agents even beyond the resources available within the tread in order to properly use this program and ensure you are not shutting out spiders that could help your traffic and cause. Ban Spiders by User Agent by Simon Lloyd |
here is the start for you, then get robots you don't like covered by
User-agent: Name of Robot Disallow: / or ban them ^ just like said above robot.txt Code:
User-agent: * |
thanks a lot for this zascok.
if I dont want to ban any spiders can I just copy and paste this into a txt file and leave it at that? |
yup and nope you can't leave it at that you gotta up it into the root of your forum :)
|
Indeed. The file should be located at www.yourdomain.com/robot.txt . Right in your root assuming that you have that access. That is where Google, as well as other well behaved bots will look for it. It gets more complicated if you are on a shared server through a provider that gives you a default address such as www.sharedservercompany.com/yourusername or something similar. In that situation, you would require you host to assist if they are able.
--------------- Added [DATE]1344202148[/DATE] at [TIME]1344202148[/TIME] --------------- Curiously, Within your code you do not have a "forums/" prefix on your items. That would be required if your forum was located within a forums folder in the root, correct? |
Quote:
all the same with /forums on front of each line for the forum itself, the rest is up to what you have in the root. I just don't have anything else but forum :) so it's right in the top. Code:
Disallow: /forums/*.js |
To clarify:
Code:
Disallow: /forums/*.js Code:
Disallow: /*.js Thanks in advance zascok! --------------- Added [DATE]1344206047[/DATE] at [TIME]1344206047[/TIME] --------------- Quote:
|
Quote:
I meant if I don't want to ban any bots or spiders from my site, I just want to limit them to the usual stuff, I can just leave your example file as it is, unedited, and upload that to forum root right? Thanks |
That is correct. Put robots.txt into your root with the contents that zascok gracefully provided. :)
|
All times are GMT. The time now is 08:47 PM. |
Powered by vBulletin® Version 3.8.12 by vBS
Copyright ©2000 - 2025, vBulletin Solutions Inc.
X vBulletin 3.8.12 by vBS Debug Information | |
---|---|
|
|
![]() |
|
Template Usage:
Phrase Groups Available:
|
Included Files:
Hooks Called:
|