![]() |
Resolve IPs, cool thanks man! :)
BTW, congrats on the visitors, i hope that everysingle one of them joins (same for my site :bandit: ) :):):) |
Damn Google just wont visit me :(
|
I got like 6 bots on right now. googlebots. =|
|
Quote:
There are currently 0 members and 49 guests on the boards.[/high] That helped you break the record of visitors also, heh. Still having Google now onto your board... |
AUTO-FORWARDING
I'll put this out there for those of you who want to send users to the real forum page and spiders to the archive. I hope you don't get banned for link cloaking... see my last post. PHP Code:
|
]
Quote:
|
You can't do it, due to the way it is now set the database in vBulletin, at least not to my knowledge.
Code:
$loggedins = $DB_site->query_first(" |
KuraFire has released a hack to modify the whois online :) after my suggestion and request.
Online 15 users Staff: 3 (user1,user2,user3) Members: 7 (user4,user5,user6,user7,user8,user9,user10) Guests: 5 (user11,user12,user13,user14,user15) Maybe guests can now be split into: Guests: 2 (user11,user12) Search Engines: 3 (user13,user14,user15) |
You cannot Floris, that's why I posted the query.
|
When I typed my post, your post wasn't there yet :)
teck: Today 05:31 PM xip: Today 05:32 PM Idea: can't we rewrite it to have a seperate usergroup id for bots from search engine? like the ban script- funtion, but instead of banning, showing its a search engine bot: google. :) |
]:(
|
The way it works now vBulletin is this:
Every time an user (guest or member) enters the site, a unique session is created, that is automatically deleted after 900 seconds, if not in use anymore. The highlighted part, userid=0, reflects only the guests, since they have no userid's. So the query counts the the sessions opened by those users, not their user agent or any other ident method. Unfortunately, there is no way around this... is not as simple as it is with members. |
something slightly similar (see attachment)
search engines are listed in italics wooolF[RM] yes it is possible although it would add 1 extra query to index.php - covering every search engine isnt realistic (new ones everyday etc) but doing the major ones is fairly easy. i'll get floris to test some bits tonight if he's around on irc |
Quote:
deny from 12.148.209. those deny's block more that just the monkeys at crawler918.com my limit section in htaccess looks like: Code:
<limit GET POST> also in my robots.txt i've got: # allow everyone else User-agent: * Disallow: # block turnitin.com User-agent: TurnitinBot Disallow: / www.turnitin.com might be a good cause for teachers - but they charge for accessing the data they've collected - so i'd rather not have them using my bandwidth/server load for free. So they can stay off my site until they decide to give a little back. |
Hmm, I did a WHOIS on their company (crawler918) and it came up with those 2 IPs...
Just curious, what results you got (other names)? |
][ 19:31:44 ] _? ? /dns [ www.crawler918.com ] ...
[ 19:31:45 ] _? ? Failed to resolve : [ no such user ] |
]
Quote:
Also 1 extra query is not that much... maybe just adding it to who's online... Thanx for the effort! :D |
Wow, that was easy, only 5 lines of code or something :)
|
]as said earlier, looks sexy :) if you could also release it instead of teasing me :p ;)
|
If inp allows me to make a release, sure :)
|
release? it's just an addon to existed hack... uhm...
|
"I will release it"
"I will addon it" I will go with 'release'. |
sorry... /me hides in the nearest bush and cries silently...
|
w00t, this is to l33t teck :)
FINALLY !!!!!!!!!!!!!!!! 43 Google BOTS crawling since yesterday and STILL AT IT !!!!!! Thanks TECK :) HOTM for this HACK !!!!!!! |
im so lucky...im mostly experiencing the turnitin.com crawl.
added them to robots.txt, but have to wait till their cached version expires. |
you can add a broad deny for them for 48 hrs (time it takes for their cache of robots.txt to expire)
## turnitin.com deny from 64.140.49 remember to remove it thou, it blocks a few more than turniton.com but they dont own their own ip block... >as said earlier, looks sexy if you could also release it instead of teasing me just cleaning stuff up atm need to write instructions as well :/ |
]oki, I'll just hang on, thanx for the job u guys do :)
|
Quote:
About the index.php file, you said is possible to be done also, can you post the code? I would like to see please, so I can learn from you a tip. Is not possible into my eyes... |
Quote:
Please don't panic if the links are dropped in a week or 2, is normal... they are moved from the "fast" crawl to the deep one. |
If you want to display nice names for your crawlers, instead of "Guest", see attached file (20 seconds install).
All you have to do is to add your crawler name and IP part. NOTE: Pay attention to the commas, when you add each crawler. Notice that the last one doesn't have a comma at the end. |
]
Quote:
|
You guys should have just waited.
|
]
Quote:
|
Example of the script in action for crawler name instead of Guest...
Hmmm, 27 Google crawlers not chewing the web site... |
Guys, if you get new crawler IP's, please post them here so everyone can add them...
Thanks. |
This is why our script is better, it doesn't care about the IP
Here are some screenshots for inph to link to. |
He will soon release his addon, which will adjust the nosessioshash part and makes guest turn into the bot on online.php and shows how many bots are online on index.php at whois online section (following me?) hehe
|
Quote:
You should release it so everyone can use it. |
Quote:
Just sit down and wait :banana: |
Well, you posted screenshots, so I presumed is done.
Then you should wait before you post anything... :p And I don't like to sit down. :banana: |
All times are GMT. The time now is 10:35 PM. |
Powered by vBulletin® Version 3.8.12 by vBS
Copyright ©2000 - 2025, vBulletin Solutions Inc.
X vBulletin 3.8.12 by vBS Debug Information | |
---|---|
|
|
![]() |
|
Template Usage:
Phrase Groups Available:
|
Included Files:
Hooks Called:
|