• Grippler
    link
    fedilink
    arrow-up
    8
    ·
    5 months ago

    The images that are flagged by such scanning, local or server side, will have to be manually verified to avoid false persecution. Someone will have to look at the private images you’ve sent that might get flagged.

    These systems have huge margins of error and are incredibly inaccurate, so there will be a significant task in manually verifying everything. And do you trust some government random employee (or just the departments general IT practices or ability to not be hacked) with not leaking your nudes or personal images? I sure as hell don’t.

    And even if this is handled perfectly and all government employees are super super honorable standup citizens that never do anything slightly wrong ever…There are still malicious governments that persecute minorities, I doubt they will handle these backdoors in digital privacy very well.

    • mihor@lemmy.ml
      link
      fedilink
      arrow-up
      8
      ·
      5 months ago

      So if I send a photo of our kids playing naked in a baby pool to my wife through signal, some slimy-ass eurocrat in some IT center will be able to ‘manually verify’ the photo of my naked kids?? Are you mentally sound??

      I hate pedos just as much as every other sane parent, perhaps even more so (I’d love to wear “Why, Garry, why?!” t-shirt all day every day). But to hell with this stupid idea that some slimy scumbags will be able to browse my own photos of my own kids. Hell, even any random photo I take, it’s my business and nobody elses! Go catch pedos the proper way instead, work a little, we don’t need Gestapo or Stasi to hover over everything we do or photograph.

    • eveninghere@beehaw.org
      link
      fedilink
      arrow-up
      1
      arrow-down
      1
      ·
      5 months ago

      They say they the images are merely matched to pre-determined images found on the web. You’re talking about a different scenario where AI detects inappropriate contents in an image.

      • Grippler
        link
        fedilink
        arrow-up
        2
        ·
        edit-2
        5 months ago

        It will detect known images and potential new images…how do you think it will the potential new and unknown images?

          • Grippler
            link
            fedilink
            arrow-up
            2
            ·
            edit-2
            5 months ago

            Literally the article linked in the OP…

            Article 10a, which contains the upload moderation plan, states that these technologies would be expected “to detect, prior to transmission, the dissemination of known child sexual abuse material or of new child sexual abuse material.”

            • eveninghere@beehaw.org
              link
              fedilink
              arrow-up
              1
              ·
              5 months ago

              My bad. But that phrasing is super stupid, honestly. What company would want to promise to detect new child sex abuse material? Impossible to avoid false negatives.