Skip Navigation

You're viewing a single thread.

22 comments
  • Is the fediverse doing anything better?

    • Fediverse is a bunch of independent websites potentially connected by compatible software, not one entity, so there’s not really a basis for comparison. You could ask about individual instances. But also it’s about “failing to cooperate with a probe into anti-child abuse practices”, not hosting or failing to moderate material. Australian law says they can asks sites about their policies and they have to at least respond.

      • The article has their response. Given their warning to google as well, apparently the responses also have to be good enough for them.

        • They said

          X's noncompliance was more serious, the regulator said, including failure to answer questions about how long it took to respond to reports of child abuse, steps it took to detect child abuse in livestreams and its numbers of content moderation, safety and public policy staff.

          So yes, all the questions need to be at least addressed and probably saying “we don’t do that because Elron doesn’t care about it” wouldn’t suffice either.

          • Cool, see my first comment again

            • You mean, is the Fediverse doing any better? Why would I need to read that again?

              • because we've gone in a circle of me asking if the site we are on right now is doing anything better with regards to this problematic material, since folks seem to care about Twitters failure to address it themselves. You respond that it's not about their lack of addressing the material, but they're lack of a response to the regulatory inquiry. I point out that they did respond, and your response is that oh they actually need to have a good answer of how they are addressing the material. Which is the same premise as the article and what my first comment was about. It's hypocrisy, because the standard isnt being applied to the fediverse, no one is up in arms about our lack of automatic detection of problematic material or surveillance of private messaging. Because we care about privacy when we're not being blinded by well intentioned Musk hate.

                • I posted from the article that they didn't respond to several questions:

                  X's noncompliance was more serious, the regulator said, including failure to answer questions about how long it took to respond to reports of child abuse, steps it took to detect child abuse in livestreams and its numbers of content moderation, safety and public policy staff.

                  I speculated that probably they also need adequate responses, but that's not what the article or the fine is about.

                  If one of the individual sites in the Fediverse was asked by Australian regulators, I bet they'd respond fully. It's not quite the same situation as Twitter, either - none of these sites are large enough to require many staff members, and don't have their own live streaming platform.

    • Since there is no hierarchical top general moderator/admin and every instance is under supervision by the respective owners of these instances, responsibility of safety is technically forwarded to individual instance admins as far as their instance goes. Or that's what I make of it at least, anyone feel free to correct me if I'm wrong. Also, the above conclusion does not include any possible random future law made up to state differently (decision-making entities have weird unpredictable logics... 😅)

      As far as for Mastodon itself, it could use some upgrades in its user management and reporting features, though (an option to automate instant reactions (like tempban until reviewed) on certain categories of reports (like child abuse and extreme/shocking violence) to prevent anyone reported for those kinds of things actively being able to continue until an admin sees and processes the report and reports are definitely not visible enough yet).

      • And things like automatic detection and direct message surveillance like these regulators are asking for?

        • Well, if those become necessary I'll just have to add Mastodon, along with anything known too well, in the bin for government-ruined software and start using hidden services... I will never willfully comply to spyware, not even (read: especially not) government-approved ones.

          I have no idea if Mastodon has any plans adding those to the instance software though... Probably will if they get lawfully obligated I suppose, but I still sincerely hope not (as I still also sincerely hope this proposal gets dismissed for the obvious contradicting privacy laws it breaks and the vulnerabilities of backdoors).

    • "The Fediverse" is about 13,000 separate services that are each individually responsible for illegal content on their systems. Some probably aren't doing a good enough job, but most of them are and they've mostly defederated the ones that fail to do so.

      And why wouldn't they? Many hands make light work and the fediverse has tens of thousands of moderators to deal with far fewer posts that the X network. Twitter had a decent moderation team once, but Musk has gutted the team.

You've viewed 22 comments.