UPDATE: proof is at https://piefed.social/c/fediverse/p/2035409/proof-of-ai-assisted-political-profiling-by-unruffled-lemmy-dbzer0-com. The main instance is lemmy.dbzer0.com but anarchist.nexus and quokka.au share admin/mod teams so those two are suspect also. I recently discovered that some popular federated instances have been using LLM-assisted moderation tooling that evaluates whether someone has said something bannable. They do this by running a script/app that sends the […]
as seen here and here, some instances are feeding posts wholesale to prompts, for what seem like extremely unsound reasons to me
I’d rather make this a learning moment than a purging moment.
absolutely! please show me you’ve learned from this moment by taking a stance against the use of LLMs anywhere in the moderation process and making it part of your instance’s policies going forward.
perform away! please keep in mind that Lemmy and ActivityPub are communities, and the only proof of intent we have is through action. a change in policy might feel performative, but it gives me something I can boost as a positive example and a sign that things will be better going forward.
We’re probably long overdue with an official meta post in our instance about this whole shitshow that we can link on the next time this disinfo is pushed.
absolutely! please show me you’ve learned from this moment by taking a stance against the use of LLMs anywhere in the moderation process and making it part of your instance’s policies going forward.
I think that would just be performative at this point, but I’ll discuss with the team.
perform away! please keep in mind that Lemmy and ActivityPub are communities, and the only proof of intent we have is through action. a change in policy might feel performative, but it gives me something I can boost as a positive example and a sign that things will be better going forward.
We’re probably long overdue with an official meta post in our instance about this whole shitshow that we can link on the next time this disinfo is pushed.