Hi #fediverse! I'm working on Hackers' Pub, a small #ActivityPub-powered social platform for developers and tech folks.
We're currently drafting a content #moderation (#flag/#report) system and would really appreciate any feedback from those who have experience with federated moderationβwe're still learning.
Some ideas we're exploring:
Protecting reporter anonymity while giving reported users enough context to understand and improve
Graduated responses (warning β content removal β suspension) rather than jumping to bans
Using LLM to help match reports to code of conduct provisions
Supporting ActivityPub Flag activity for cross-instance reports
Our guiding principle is that moderation should be about growth, not punishment. Expulsion is the last resort.
Here's the full draft if you're curious: https://github.com/hackers-pub/hackerspub/issues/192.
If you've dealt with moderation in federated contexts, what challenges did you run into? What worked well? We'd love to hear your thoughts.