![]() |
perfect robots.txt for vb 3.7?
Does anyone care to share the perfect robots.txt for vb 3.7
here is what i have Code:
User-agent: * |
You should have that in the domain root.
|
don't mind me sounding like a newbie here but i added the content to a robots.txt file, upped it to my server html root folder like 8 hrs ago..now when i check my who's online list, i still see google bots murdering my calender.php, login.php...etc...
anything i did wrong? |
Google only checks robots.txt every day or two, I think. Give it a little time.
|
Quote:
|
once every 24hrs :)
|
yes apparently it worked like 5 minutes after i posted... only the important files like forumdisplay, showpost, showthread etc we being scanned. thanks for this mate. cheers. :)
|
Make sure you add in
User-agent: Fasterfox Disallow: / I managed to catch a thread on my board where people were discussing have Fasterfox set as turbo charged...my loads went through the roof !! |
Can we combine these 2?
Quote:
Quote:
|
Quote:
This is how mine looks (part of it anyway) Code:
User-agent: Fasterfox |
Quote:
yes you can |
Thank you, sir. ;)
|
Was there ever anything to stop Google Web Accelerator ?
|
yeah i would like to know that also.
|
is it a bit of a security issue telling everyone where your admincp is?
|
I don't see how as it is the default directory name for vb.
|
Quote:
Considering all the brute force tools available to script kiddies I say out of sight, out of mind. |
All they would have to do was know any names of the default files in the admincp directory and they could find the directory that way. So even that isn't failsafe.
|
Quote:
how? http://www.domain.com/forum/yabbadabbado/index.php |
If they can get on the server, they can find the directory. There is no foolproof way to stop someone who wants to get to it bad enough. All you can do is slow them down. But good luck with your theory. ;)
|
Quote:
Code:
// ****** PATH TO ADMIN & MODERATOR CONTROL PANELS ****** I only allow people http access to my server. Apart from guessing I'm a big fan of the flintstones it does provide extra hoops that any hacker will have to jump through. Not theoretical hoops either : They have to guess another defacto password. If someone has access to my server through ssh or whatever then it's already too late. From reading on here 90% of board hacks are through a browser simply putting admincp after the board url and guessing. A large percentage of these 'hacks' are inside jobs by disgruntled mods who presumably know their way around a standard installation. Why exclude something by name when googlebot only follows links and the admin link will never appear for guests? Are you putting your admin cp in a sitemap too?! (j/k re sitemap!) Either way it's not a major issue but I just don't see the point in listing it. |
Quote:
I've never heard of this before, I'll add it if it helps, but I really hate messing with my robots.txt file, I don't want to stop the bots from me site :D thanks -Brandon |
Quote:
There's a few other download accelerators too but some allow surfers spoof their useragent which is unfortunate. |
Does the fasterfox addon in your robots.txt completely block anyone using fasterfox or just blocks the functionality of the add-on on the site?
|
@Phooey
I'd assume it just blocks the functionality of the add-on I run multiple websites on my server...only one is semi-large and having issues with spiders hogging all the resources (crashing my database multiple times a day due to having too many connections). Should I upload that file into the root folder of that one site or should I modify it and upload it into my server root folder where my xxxx.com, yyyyy.com, zzzzz.com, etc folders are located for my individual websites? I'm using 3.6.9 (soon to upgrade to 3.6.10--not quite ready for the jump to 3.7 but eventually will do so), so I'm gonna have to modify this a little to suit my needs prolly. |
Quote:
Here's what I use : Code:
User-agent: * |
Heres more information on Robots.txt for vBulletin including some other files and folders you should block out.
After you upload it, it could take up to 24hrs for spiders to find and start to follow the rules of robots.txt most only download it once a day. |
Too bad that link doesn't work.
|
It does work....
|
All times are GMT. The time now is 04:57 PM. |
Powered by vBulletin® Version 3.8.12 by vBS
Copyright ©2000 - 2025, vBulletin Solutions Inc.
X vBulletin 3.8.12 by vBS Debug Information | |
---|---|
|
|
![]() |
|
Template Usage:
Phrase Groups Available:
|
Included Files:
Hooks Called:
|