Posts: 1,868
Threads: 134
Joined: Sep 2014
stryder
Jan 15, 2026 05:41 PM
(This post was last modified: Jan 16, 2026 03:08 AM by Neuro.)
Update:
I've made a slight change as just blocking entire ranges of IPs globally was extremely drastic. I do intend to try and whittle down as to which IP ranges need to be penalised for scraping, however for the time being I've rigged a change to match the method I currently have for dealing with older bots that were too active.
It's basically a flooding control that will only activate during site active periods. (Those periods are based upon observations of global activity as well as user activity.)
So it means the scrapers can sort of be active, but it will be hit or miss as to whether they window for access is available. It means humans caught behind the IP ranges can still access the site, although it will be intermittent (they'd have to message me or post to the forum to let me know if they are affected)
(I have to admit I prefer the "rawdog" method of site operations where I don't use third-party API's, so it limits what I can do as I'm not using Cloudflare, GeoIP etc which would of easily tackled the issue.)
Anyway we'll see if this curbs the site from hitting the resource/processor ceiling.
Further Update:
Looks like just a time period throttle isn't enough on its own. So I'm trying some small changes that I hope doesn't upset the membership.
You'll notice that when not logged in, you'll not see as many forums or the whosonline information. This has been done to reduce how much scrapers and unauthorised spiders have to access. Normal spiders that have been added to the list however will still be able to access the forum, so posts/threads can still be found for search engines.
I added a nonpost subforum that is only viewable (or should only be viewable) while not logged in. It's aimed at the bot/spider/scrapers and their owners. At least an attempt to point out that it's not okay to scrape abusively and that something will be done when they do. Not that it will make a blind bit of difference, but at least no one can say "We didn't know that!"
Anyway I'll continue to look to see what effective ways I have to cut the scraper abuse, as I really don't want to keep the IPs on the ban list considering just how many I've assigned to it currently.