Post contents (and a mirror):

BotDefense is wrapping up operations

TL;DR below.

When we announced the BotDefense project in 2019, we had no idea how large the project would become. Our initial list of bots was just 879 accounts. Most of them were annoying rather than outright malicious.

Since then, we’ve witnessed the rise of malicious bots being used to farm karma for the purpose of spamming and scamming users across Reddit and we’ve done our best to help communities stem the tide. We spent countless hours finding and reviewing accounts, writing code to automate detections, and reviewing appeals (mostly from outright criminals and karma farmers definitely running bots, but we typically unban about 4 accounts per month, and unlike similar bots an unban means that we unban the account everywhere we banned it).

Along the way, we’ve struggled with the scope of the problem, rewritting our back-end code multiple times and figuring out how to scale to the 3,650 subreddits that BotDefense now moderates. We came up with new algorithms to identify content theft, reduce the number of times we accidentally ban an innocent account, and more. In January of 2023, we added an incredible 10,070 bots to our ban list which now stands at an incredible 144,926 accounts.

Like many anti-abuse projects on Reddit, we’ve done all of this for free while putting up with Reddit’s penchant for springing detrimental changes on developers and moderators (e.g., adding API limits without advance notice and blocking Pushshift) and figuring out workarounds for numerous scalability issues that Reddit never seems to fix. Without Pushshift, the number of malicious bots we were able to ban dropped to 5,517 in May.

Now, Reddit has changed the Reddit API terms to destroy third-party apps and harm communities. A group of developers and moderators tried to convince Reddit to not continue down this path and communities protested like never before, but that was all in vain. Reddit is so brazenly hostile to moderators and developers that the CEO of Reddit has referred to us as “landed gentry”.

With these changes and in this environment, we no longer believe we can effectively perform our mission. The community of users and moderators submitting accounts to us depend on Pushshift, the API, and third-party apps. And we would be deluding ourselves if we believed any assurances from Reddit given the track record of broken promises. Investing further resources into Reddit as a platform presents significant risks, and it’s safer to allocate one’s time, energy, and passions elsewhere.

Therefore, we have already disabled submissions of new accounts and our back-end analytics, and we will be disabling future actions on malicious and annoying bots. We will continue to review appeals and process unbans for a minimum of 90 days, or until Reddit breaks the code running BotDefense.

We’d rather be figuring out how to combat the influx of ChatGPT bots flooding Reddit, temu bots flooding subreddits with fake comments, and every other malicious bot out there, of course.

At this time, we advise keeping BotDefense as a moderator through October 3rd so any future unbans can be processed. We will provide updates if the situation changes or if we have any other news to share.

Finally, I want to thank all of the users and moderators who have contributed accounts, my co-moderators who have helped review countless accounts, and to all of the communities that have trusted us with helping moderate their subreddits.

Regards.

— dequeued

TL;DR With the API changes now in place, we no longer believe we can effectively perform our mission so we are sunsetting BotDefense. We recommend keeping BotDefense on as a moderator through October 3rd so any unbans can be processed.

  • @NorthCountryHermit@lemm.ee
    link
    fedilink
    English
    29 months ago

    I reported a bunch of spam-link bots, porn-bots posting CSAM-adjacent stuff, and a slew of stuff that was very obviously and blatantly in violation of Reddit ToS.

    All the links/content remained and my account got suspended for “abusing the report function”. Considering the content I reported, it’s a safe guess that Reddit admin and moderation teams benefit from having such material on the site. Not sure how, or why. But their actions seem to indicate something.

    Then you look at the controversies Reddit has had… that one dude’s suicide, Ghislaine Maxwell holding a powermod position on front-page subs, the powermod controversy wherein a dozen or so mods had thousands of subreddits under their control.

    There’s also been a huge attitude shift. While most of it seems the same, there’s a very strong anti-American presence now that seems to make sure they fit that sentiment into every thread whether it’s justified, related, or completely out of place.

    The Reddit algorithms or whatever also seem to love pushing certain topics to the point of stripping it of any/all meaning and turning into propaganda.