this post was submitted on 31 Jul 2023
117 points (85.0% liked)

Meta (lemm.ee)

3473 readers
17 users here now

lemm.ee Meta

This is a community for discussion about this particular Lemmy instance.

News and updates about lemm.ee will be posted here, so if that's something that interests you, make sure to subscribe!


Rules:


If you're a Discord user, you can also join our Discord server: https://discord.gg/XM9nZwUn9K

Discord is only a back-up channel, [email protected] will always be the main place for lemm.ee communications.


If you need help with anything, please post in !support instead.

founded 1 year ago
MODERATORS
 

There's a new one suddenly popping up in my feed but obviously the reports are being "resolved" by the mods of that community. They suggested to me that I block their community but I will not because that is how you get a cesspit of an instance. How do we report disinformation communities straight to the admins?

Edit: the admins did remove the community in question so I'm going to take that as the official stance on disinformation communities and also assume that any type of community (right wing, left wing, or other) that are intentionally spreading disinformation will be removed. That makes me feel much better about the situation since this type of thing is pretty much guaranteed to pop up again.

you are viewing a single comment's thread
view the rest of the comments
[–] [email protected] 5 points 1 year ago (1 children)

I'm also not advocating for killing trolls that bother me... so take care not to belabor a quick metaphor.

The vast majority of disinformation comes in a few key topics related to current hot button political issues and is generally pushed by recognizable sources. It's not unreasonable to expect admins to check into user reports of disinformation and organized trolling against known sources. I'm not an admin so I'm not going to write up the specific criteria right here and now.

Choosing not to do so is also a conscious choice to host such content.

[–] [email protected] -1 points 1 year ago (1 children)

Hey, it’s okay to break down a metaphor if I don’t think it’s applicable to the conversation.

Yes, totally I agree with you, I think admins should review reported content and do some investigation if needed.

I guess I have a problem with removing users and communities based on someone’s opinion of the content itself. Vote manipulation, brigading, creating multiple accounts to push agenda, repeated automated posting, and even organized trolling like you mentioned are not direct opinions on the content posted. They are clearly defined and relatively easy to identify. “Disinformation,” “recognizable sources,” and “hot button political issues” are direct opinions about the content or subject of a post or community. They are not clearly defined and differ greatly from person to person.

I asked you to suggest a definition or criteria of disinformation to move us from the “what” to the “how.” Thinking about how this might be regulated practically might help you understand why I think it’s problematic to remove users and communities based solely on someone’s opinion of their content.

[–] [email protected] 4 points 1 year ago* (last edited 1 year ago)

Believe me I do understand why it could be considered problematic. My disagreement stems from the idea that it's better to have no policy rather than an imperfect policy or one that relies on some discretion.

My point in highlighting that disinformation centers around a few hot button issues is to reinforce that we're not talking about some nebulous or opinion-driven debate; rather there are a few key disinformation strategies that take advantage of the "bullshit asymmetry" to poison real discussion. They are easily identified because they are well documented and reported on.

I'm simply unconvinced by arguments that it's too hard to identify and nip such malicious communities in the bud. Even less so by arguments that doing so is somehow a slippery slope.