cross-posted from: https://lemmy.world/post/20664372

X (formerly Twitter) claims that non-consensual nudity is not tolerated on its platform. But a recent study shows that X is more likely to quickly remove this harmful content—sometimes known as revenge porn or non-consensual intimate imagery (NCII)—if victims flag content through a Digital Millennium Copyright Act (DMCA) takedown rather than using X’s mechanism for reporting NCII.

  • Peruvian_Skies@sh.itjust.works
    link
    fedilink
    English
    arrow-up
    70
    ·
    edit-2
    2 months ago

    In other words, the consent of a corporation is more important than the consent of a human being… for the public distribution of that human being’s likeness in an intimate context. Holy dystopia, conservatives are fucked in the head.

    • shalafi@lemmy.world
      link
      fedilink
      English
      arrow-up
      3
      ·
      2 months ago

      The likelihood of being sued is the issue. The DMCA has tons of case law behind it making it a powerful tool for the media companies. Not so much for NCII. Also, individuals usually don’t have legal teams.

      So as with much fuckery in the modern world, we can use the rules as well as they can. If someone’s image is shared as NCII, they only need flag it as a DMCA violation. Problem being, who is going to know that trick?