The policy changes come after an NBC News investigation last month into child safety on the platform.

  • okawari@kbin.social
    link
    fedilink
    arrow-up
    25
    ·
    edit-2
    1 year ago

    Sadly, I’m sure any social platform where one can make their own private community (actually private or perceived to be private) will have more of these than most of us think. Its just that we don’t see them.

    I’m also not surprised that services like discord is seemingly relaxed at moderating them, as its a problem that is invisible to most users. Moderating is expensive, and unless it hurts public opinion, seemingly its not worth it for them

    • Itty53@kbin.social
      link
      fedilink
      arrow-up
      3
      ·
      edit-2
      1 year ago

      This is the unfortunate and altogether wrong (morally, it’s not incorrect) truth of the corporate structure. If you can’t point to an ROI they don’t want to spend money on it. The tech world has gone to great collective lengths to push the value of prevention and best practices even if they take a bit of time and money, and I really wish they’d take up that same passion for this kind of shit.

      I also really wish the public would treat these companies as the conspirators they are too, but I know they won’t. They never have.

      The gross truth is there’s a not-inconsequential number of pedophiles in the tech world compared to other industries. It’s the kind of world where you’re not likely to stumble into it online unless you’re well versed in the technology we all collectively call “the internet”. Speaking personally I’ve encountered it myself multiple times, technologically literate people who think they can evade or even worse, relate to others in that industry. Might be an unpopular opinion but I’m in that industry too, I’m not talking about just your industry, reader.