Instance Admins: Check Your Instance for Vote Manipulation Accounts [PSA]
Over the past 5-6 months, I've been noticing a lot of new accounts spinning up that look like this format:
https://instance.xyz/u/gmbpjtmt
https://instance.xyz/u/tjrwwiif
https://instance.xyz/u/xzowaikv
What are they doing?
They're boosting and/or downvoting mostly, if not exclusively, US news and politics posts/comments to fit their agenda.
Edit: Could also be manipulating other regional news/politics, but my instance is regional and doesn't subscribe to those which limits my visibility into the overall manipulation patterns.
What do these have in common?
Most are on instances that have signups without applications (I'm guessing the few that are on instances with applications may be from before those were enabled since those are several months old, but just a guess; they could have easily just applied and been approved.)
Most are random 8-character usernames (occasionally 7 or 9 characters)
Most have a common set of users they're upvoting and/or downvoting consistently
No posts/comments
No avatar or bio (that's pretty common in general, but combine it with the other common attributes)
Update: Have had several anonymous reports (thanks!) that these users are registering with an @sharklasers.com email address which is a throwaway email service.
What can you, as an instance admin, do?
Keep an eye on new registrations to your instance. If you see any that fit this pattern, pick a few (and a few off this list) and see if they're voting along the same lines. You can also look in the login_token table to see if there is IP address overlap with other users on your instance and/or any other of these kinds of accounts.
You can also check the local_user table to see if the email addresses are from the same provider (not a guaranteed way to match them, but it can be a clue) or if they're they same email address using plus-addressing (e.g. [email protected], [email protected], etc).
Why are they doing this?
Your guess is as good as mine, but US elections are in a few months, and I highly suspect some kind of interference campaign based on the volume of these that are being spun up and the content that's being manipulated. That, or someone, possibly even a ghost or an alien life form, really wants the impression of public opinion being on their side. Just because I don't know exactly why doesn't mean that something fishy isn't happening that other admins should be aware of.
Who are the known culprits?
These are ones fitting that pattern which have been identified. There are certainly more, but these have been positively identified. Some were omitted since they were more garden-variety "to win an argument" style manipulation.
These all seem to be part of a campaign. This list is by no means comprehensive, and if there are any false positives, I do apologize. I've tried to separate out the "garden variety" type from the ones suspected of being part of a campaign, but may have missed some.
Edit: If you see anyone from your instance on here, please please please verify before taking any action. I'm only able to cross-check these against the content my instance is aware of.
What surprises me is that these seem to be all on other instances - including a few big ones like just.works - rather than someone spinning up their own instance to create unlimited accounts to downvote/spam/etc.
Not really: if you're astroturfing, you don't do all your astroturfing from a single source because that makes it so obvious even a blind person could see it and sort it out.
You do it from all over the places, mixed in with as much real user traffic as you can, and then do it steadily and without being hugely bursty from a single location.
Humans are very good at pattern matching and recognition (which is why we've not all been eaten by tigers and leopards) and will absolutely spot the single source, or extremely high volume from a single source, or even just the looks-weird-should-investigate-more pattern you'd get from, for example, exactly what happened to cause this post.
TLDR: they're doing this because they're trying to evade humans and ML models by spreading the load around, making it not a single source, and also trying to mix it in with places that would also likely have substantial real human traffic because uh, that's what you do if you're hoping to not be caught.
After digging into it, we banned the two sh.itjust.works accounts mentioned in this post. A quick search of the database did not reveal any similar accounts, though that doesn't mean they aren't there.
My bachelor's thesis was about comment amplifying/deamplifying on reddit using Graph Neural Networks (PyTorch-Geometric).
Essentially: there used to be commenters who would constantly agree / disagree with a particular sentiment, and these would be used to amplify / deamplify opinions, respectively. Using a set of metrics [1], I fed it into a Graph Neural Network (GNN) and it produced reasonably well results back in the day. Since Pytorch-Geomteric has been out, there's been numerous advancements to GNN research as a whole, and I suspect it would be significantly more developed now.
Since upvotes are known to the instance administrator (for brevity, not getting into the fediverse aspect of this), and since their email addresses are known too, I believe that these two pieces of information can be accounted for in order to detect patterns. This would lead to much better results.
In the beginning, such a solution needs to look for patterns first and these patterns need to be flagged as true (bots) or false (users) by the instance administrator - maybe 200 manual flaggings. Afterwards, the GNN could possibly decide to act based on confidence of previous pattern matching.
This may be an interesting bachelor's / master's thesis (or a side project in general) for anyone looking for one. Of course, there's a lot of nuances I've missed. Plus, I haven't kept up with GNNs in a very long time, so that should be accounted for too.
Edit: perhaps IP addresses could be used too? That's one way reddit would detect vote manipulation.
[1] account age, comment time, comment time difference with parent comment, sentiment agreement/disgareement with parent commenters, number of child comments after an hour, post karma, comment karma, number of comments, number of subreddits participated in, number of posts, and more I can't remember.
That would definitely work for rooting out ones local to an instance, but not cross-instance. For example, none of these were local to my instance, so I don't have email or IP data for those and had to identify them based on activity patterns.
I worked with another instance admin who did have one of these on their instance, and they confirmed IP and email provider overlap of those accounts as well as a local alt of an active user on another instance. Unfortunately, there is no way to prove that the alt on that instance actually belongs to the "main" alt on another instance. Due to privacy policy conflicts, they couldn't share the actual IP/email values but could confirm that there was overlap among the suspect accounts.
Admins could share IP and email info and compare, but each instance has its own privacy policy which may or may not allow for that (even for moderation purposes). I'm throwing some ideas around with other admins to find a way to share that info that doesn't violate the privacy of any instances' users. My current thought was to share a hash of the IP address, IP subnet, email address, and email provider. That way those hashes could be compared without revealing the actual values. The only hiccup with that is that it would be incredibly easy to generate a rainbow table of all IPv4 addresses to de-anonymize the IP hashes, so I'm back to square one lol.
Yes, this would essentially be a detecting mechanism for local instances. However, a network trained on all available federated data could still yield favorable results. You may just end up not needing IP Addresses and emails. Just upvotes / downvotes across a set of existing comments would even help.
The important point is figuring out all possible data you can extract and feed it to a "ML" black box. The black box can deal with things by itself.
I can only verify them based on the posts/comment votes my instance is aware of. That said, I do have sufficient data and enough overlap to establish a connection/pattern.
I just had a look at https://lemy.lol/, and they have email verification enabled, so it's not just people finding instances without email check to spam account on there.
I think what we need is an automated solution which flags groups of accounts for suspect vote manipulation.
We appreciate the work you put into this, and I imagine it took some time to put together. That will only get harder to do if someone / some entity puts money into it.
Yeah, this definitely seems more like script kiddie than adversarial nation-state. We're not big enough here, yet anyway, that I think we'd be attracting that kind of attention and effort. However, it is a good practice run for identifying this kind of thing.
It's easy on Reddit because they have their own username generator when you sign up, but the usernames being used here are very telling. Random letters is literally the absolute bare minimum effort for randomly generating usernames. A competent software engineer could make something substantially better in an afternoon and I feel like an adversarial nation-state would be using something like a small language model trained solely on large lists of scraped usernames.
On the other hand, any automated solution will be possible to work around. Such a system would be open source like the rest of Lemmy and you'd know exactly the criteria you need to live up to to avoid getting hit by the filter.
What if the tool was more of a toolbox, where each instance could configure it the way that they want (ex. Thresholds before something is flagged, etc.) Similar to how automod works, where the options are well known but it's hard to tell what any particular space is running behind the scenes.
At the very least, tools like this can make it harder for silent vote manipulation even if it doesn't stop it entirely
As an end user, ie. not someone who either hosts an instance or has extra permissions, can we in anyway see who voted on a post or comment?
I'm asking because over the time I've been here, I've noticed that many, but not all, posts or comments attract a solitary down vote.
I see this type of thing all over the place. Sometimes it's two down votes, indicating that it happens more than once.
I note that human behaviour might explain this to some extent, but the voting happens almost immediately, in the face of either no response, or positive interactions.
As a regular user, I don't think there's much you can do, unfortunately (though thank you for your willingness to help!). Sometimes you can look at a post/comment from Kbin to see the votes, but I think Mbin only shows the upvotes. Most former kbin instances, I believe, switched to mbin when development on kbin stalled.
The solitary downvotes are annoying for sure. "Some people, sigh" is just my response to that. I just ignore those.
Re: Downvote bots. I can't say they're necessarily bots, but my instance has scripts that flag accounts that exclusively give out downvotes and then bans them. That's about the best I can do, at present, to counter those for my users.
Re: Downvote bots. I can't say they're necessarily bots, but my instance has scripts that flag accounts that exclusively give out downvotes and then bans them. That's about the best I can do, at present, to counter those for my users.
It is usually not a good idea to specify what your exact metrics are for a ban. A bad actor could see that and then get around it by randomly upvoting something every now and then.
I was replying to a comment in a big news community about 5 months ago. It took me probably 2 minutes, at most, to compose my reply. By the time I submitted the comment (which triggered the vote counts to update in the app), the comment I was replying to had received ~17 downvotes. This wasn't a controversial comment or post, mind you.
17 votes in under 2 minutes on a comment is a bit unusual, so I pulled up the vote viewer to see who all had downvoted it so quickly. Most of them were these random 8 character usernames like are shown in the post.
From there, I went to the DB to look at the timestamps on those votes, and they were all rapid-fire, back to back. (e.g. someone put the comment AP ID into a script and sent their bot swarm after it)
So that's when I realized something fishy was happening and dug deeper. Looking at what was upvoted from those, however, revealed more than what they were downvoting. Have been keeping an eye out for those type of accounts since. They stopped registering for a while, but then they started coming up again within the last week or two.
I wonder if private voting will make it too difficult to discover
Depends how it's implemented. If the random usernames that are supplied from the private votes are random for each vote, that would make it nearly impossible to catch (and would also clutter the person table on instances with junk, one-off entries). If the private voting accounts are static and always show up with the same identifier, I don't think it would make it much more difficult than it is now with these random user accounts being used. The kicker would be that only the private version of the account would be actionable.
The only platform with private voting I know of right now is Piefed, and I'm not sure if the private voting usernames are random each time or static (I think they're static and just not associated with your main profile). All that said, I'm not super clear on how private voting is implemented.
But this is SOO tedious. The annoying bit is it could just be one person who set it up over a weekend, has a script that they plug into when wanting to be a troll, and now all admins/mods have to do more work.
You're fighting the good fight! So annoying that folks are doing it on freaking lemmy.
I wonder if there's a way for admins to troll back. Like instead of banning the accounts, send them into a captcha loop with unsolvable or progressively harder captchas (or ones designed to poison captcha solving bots' training).
Is there any existing opensource tool for manipulation detection for lemmy? If not we should create one to reduce the manual workload for instance admins
If there were, upbotters would use it to verify that new bottling methods weren't detectable. There's a reason why reddit has so much obfuscation around voting and bans.
I mean if a new account or an account with no content on it starts downvoting a lot of things or upvoting a lot of things that's generally a red flag that it's a vote manipulation account. It's not always but it's usually pretty obvious when it actually is. A person who spends their entire time downvoting everything they see, or downvoting things randomly is likely one of those bots.
Could they come up with ways around it? Sure by participating and looking like real users with post and comment history. Though that requires effort and would slow them down majorly, so it's something that they're very unlikely to do.
Good point, but is it then possible to come up with detection algorithms that makes it hard for upbotters even if they know the algorithm? I think that would be more ideal than security through obfuscation but not sure how feasible that is
What stops the botters from setting up their own instances to create unlimited users for manipulating votes?
Nothing, really. Though bad instances like that would be quickly defederated from most. But yeah, admins would have to keep an eye on things to determine that and take action.
They usually get found out pretty easily and then defederated by everyone. There's a service called fediseer which allows instance admins to flag instances as harmful, which other admins can use to determine if they should block an instance.
In order for that to really work they would have to rotate between a lot of domain names either by changing their own instance's domain or using a proxy. Either way they'd run out of domains rather quickly.
It's way easier for them to just get accounts on the big servers and hide there as if they were normal lurking users.
Im fully against public down votes becaue I already see people calling out other users by their name in threads they're not even part of. Theres no world where that behavior gets better when you give them more tools to witch hunt. Lemmy is as much an insular echo chamber as any social media and there are plenty of users dedicated to keeping it that way.
I think retaliatory downvotes happen either way if you're in an argument. Same with report abuse, which, if it happens to a high degree, would be the moderator's responsibility to ban the perpetrator (reports here are not anonymous like they were on Reddit).
Also, if there's someone with an abusive mind, they can easily use another instance that shows Activity to identify downvoters. The vote is public either way for federation purposes, they're just hidden from certain instances - at least on the user level, but they're still there technically.
Ethically, I can't (and won't). I'm only comfortable and confident enough to share the list of sockpuppet accounts I've confirmed and provide the information necessary to detect them. I did list the topics I'm aware of (US news and politics), but I'm only able to see activity based on what my instance knows about. So they may be manipulating other communities, but if my instance doesn't subscribe to them (or they're by posters that have been banned), I have no way of seeing it.
That's actually why I posted this. My visibility is limited, so once I identified the pattern, I'm passing that along to other admins for awareness.
@[email protected] I have cleaned these and some other bot accounts from my instance. I was ok to open registrations to this point because we were able to get reports for almost every activity and we could easily manage them. But unfortunately Lemmy does not have a regulatory mechanism for votes, so I'll keep it manual approval until then.
Also it looks like they're manually creating accounts since we had captcha + email approval in our instance from the beginning. So this means that even with manual approvals, a botnet can be created – just in a delayed manner.
Yep, seems manual or at least only partially automated based on feedback from other admins.
Also yeah, unfortunately, Lemmy doesn't have the ability to report users to their home admins, just content they post. Not sure if that's a moderation feature that's in the pipeline or not (haven't checked for a bit).
I have a manual process for admitting people, do I need to do anything if I know exactly who is on my instance, or do I need to do anything to protect my instance from other bad acting instances (beyond defederating, which I do when I notice a lot of spam). Any queries you recommend?
I have a manual process for admitting people, do I need to do anything if I know exactly who is on my instance,
With that in place, I wouldn't think so. I'm in the same boat with a small instance that has always used applications. The problematic accounts I've noticed are all using these random, 8-character names and seem to be setting up shop across open instances w/o applications. So chances are, if you're manually admitting people, you'd have noticed these already and likely not approved them.
do I need to do anything to protect my instance from other bad acting instances
Unfortunately, defederating only protects your instance's users from being impacted by the manipulations. Beyond that, it's less a bad instance rather than them being taken advantage of (kind of like our persistent troll who instance hops every few days).
For now, I've just banned the vote manipulation accounts and moved on (this PSA notwithstanding lol) I wouldn't consider these a "defederation worthy" offense. When I do defed, it's for bigger reasons or just temporary due to spam (sometimes admins can't deal with it right away but it's causing a huge problem now and I need to do something in the short term).
Queries, I do have some, but they're ugly AF. lol. I should prob look into starting a Matrix room or admin community where we can share and improve each others' utility scripts.
Queries, I do have some, but they’re ugly AF. lol. I should prob look into starting a Matrix room or admin community where we can share and improve each others’ utility scripts.
That's pretty much the official Lemmy space's Moderation tools room, right?
Another data point in favor of supporters of Dead Internet Theory .
Also, this is one more example of why it would be better if instances charged a little bit from everyone: spammers will rather run things from their own machines (or some illegal botnet) than paying something with a credit card.
you’d just get a bunch of chargebacks from stolen credit cards lol.
Criminals use stolen credit cards for high value items that can be sold quickly. If criminals really wanted to do mass manipulation via AP servers, it will be easier/faster/cheaper for them to spin up their own servers than signing up for paid accounts.
The one counter-argument that I would accept though: what if bad actors running psyops become commercial providers to attract legit customers and mix it with their agents?
That's one thing that nobody really ever talks about when it comes to discussing payment verification. The fact that the people who are willing to commit scams and fraud are also willing to steal credit or debit cards.
Yeah nah man. I’m poor as fuck. Like usually have cents left if my bank account poor, and don’t always have meals poor. I ain’t paying a penny, even though I love lemmy.
Everytime someone uses phrases like "Dead internet theory", I assume they're some crab living under a rock with limited real world experiences. And it fits every time.
Just beware the bias here. Only pointing this out for people working against their agenda doesn't mean there isn't the same going on for the other side...just that they aren't going to point it out as it doesn't help their agenda.
I have been calling this shit since DNc did that switcheroo trick. For which I was needlessly down voted. I just blocked news and politics subs.
Good to see some hard evidence that we got DNC komissars around.
People can support kamala all they want but when somebody is spending money to set narratives, even supporters should raise an eye brow. What is this really about?
Their entire agenda hangs on people not jumping ship because everybody just loves Democrats so much and the alternatives are not popular enough.
Nobody is even arguing that they like Democrats anymore, it's just yapping about how Biden (and now Kamala) is the only viable option to win. If it turns out alternatives are actually popular enough that enough people are willing to jump ship the entire narrative sinks.
Lemmy should have the option to defederate from instances depending on automated criteria. Sign ups without admin checks are a great attribute to use for defederation, because it leads to such abuse. I've finally blocked most communities and instances that have news about US politics and have a clean feed, but for newcomers, that shit is everywhere.
It's not a native feature, but some instances have a script or plugin (not super familiar with it beyond a general awareness of its existence) that can tie their federation allow/block lists with Fediseer. So, like, if an instance gets censured by a bunch of other instances you're on good terms with, it can automatically pick that up and add it to your block list.
I don't hate the idea of that, and I have seen it protect a few instances from several spam waves, but I haven't implemented it myself.
Users could also be doing and reporting the checking up - if votes were transparent - and they would be able to do it on far wider scale. Oh those leopards, eating your faces, vote obfuscation proponents.
Lemmy should do something like make captcha and email verification the default in the next version, and reject federation from anyone with a lower version. If we accept federation from any instance where this was never turned on, banning accounts one by one is worse than Sisyphean. They'll just keep finding more vulnerable instances that are already trusted and abuse them to spam the rest of the fediverse.
If admins want to manually turn it off, then they should be prepared to manage that.
It's painfully obvious lemmy is overrun with astroturf. Kamala spam has been oppressive and it's just cringe most of the time. I refuse to believe that most of the real users here are that cringe. Also, I support Kamala.
I'm really not sure. 47k monthly active users, between 30% to 50% of them not American, and those who are are already going to vote Democrat, is it really worth the hassle?
Its funny because there's a user assuming the astro turf is going the other direction and they're getting mostly up voted. Too many people are thinking with their feelings instead of their brain.
The blue wave doesn't care about wisdom or agency any more than MAGA. The masses mistake revolutionary and Russian agent in false dichotomy. And, the .world mods are more than complicit.
The majority here will hate you for truth. There are better venues for it.