In the days after the US Department of Justice (DOJ) published 3.5 million pages of documents related to the late sex offender Jeffrey Epstein, multiple users on X have asked Grok to “unblur” or remove the black boxes covering the faces of children and women in images that were meant to protect their privacy.

  • ToTheGraveMyLove@sh.itjust.works
    link
    fedilink
    English
    arrow-up
    68
    ·
    edit-2
    10 hours ago

    Are these people fucking stupid? AI can’t remove something hardcoded to the image. The only way for it to “remove” it is by placing a different image over it, but since it has no idea what’s underneath, it would literally just be making up a new image that has nothing to do with the content of the original. Jfc, people are morons. I’m disappointed the article doesn’t explicitly state that either.

    • mfed1122@discuss.tchncs.de
      link
      fedilink
      English
      arrow-up
      1
      ·
      2 hours ago

      They think that the AI is smart enough to deduce from the pixels around it what the original face must have looked like, even though there’s actually no reason why there should be a strict causal relationship between those things.

    • usualsuspect191@lemmy.ca
      link
      fedilink
      English
      arrow-up
      35
      ·
      edit-2
      8 hours ago

      The black boxes would be impossible, but there are some types of blur that keep enough of the original data they can be undone. There was a pedofile that used a swirl to cover his face in pictures and investigators were able to unswirl the images and identify him.

      With how the rest of it has gone it wouldn’t surprise me if someone was incompetent enough to use a reversible one, although I have doubts Grok would do it properly.

      Edit: this technique only works for video, but maybe if there are several pictures of the same person all blurred it could be used there too?

      https://youtu.be/acKYYwcxpGk

      • floquant@lemmy.dbzer0.com
        link
        fedilink
        English
        arrow-up
        3
        ·
        3 hours ago

        Yeah, but this type of machine learning and diffusion models used in image genAI are almost completely disjoint

        • usualsuspect191@lemmy.ca
          link
          fedilink
          English
          arrow-up
          1
          ·
          2 hours ago

          Agree with you there. Just pointing out that in theory and with the right technique, some blurring methods can be undone. Grok most certainly is the wrong tool for the job.

      • BarneyPiccolo@lemmy.today
        link
        fedilink
        English
        arrow-up
        14
        ·
        9 hours ago

        Several years ago, authorities were searching the world for a guy who had been going around the world, molesting children, photographing them, and distributing them on the Internet. He was often in the photos, but he had chosen to use some sort of swirl blur on his face to hide it. The authorities just “unswirled” it, and there was his face, in all those photos of abused children.

        They caught him soon after.

      • Barracuda@lemmy.zip
        link
        fedilink
        English
        arrow-up
        11
        ·
        9 hours ago

        A swirl is a distortion that is non-destructive. Am anonymity blur averages out pixels over a wide area in a repetitive manner, which destroys information. Would it be possible to reverse? Maybe a little bit. Maybe one pixel out of every %, but there wouldn’t be any way to prove the accuracy of that pixel and there would be massive gaps in information.

        • altkey (he\him)@lemmy.dbzer0.com
          link
          fedilink
          English
          arrow-up
          8
          ·
          7 hours ago

          Swirl is destfuctive like almost everything in raster graphics with recompressing, but unswirling it back makes a good approximation in somehow reduced quality. If the program or a code of effect is known, e.g. they did it in Photoshop, you just drag a slider to the opposite side. Coming to think of it, it could be a nice puzzle in an adventure game or one another kind of captcha.

          • Barracuda@lemmy.zip
            link
            fedilink
            English
            arrow-up
            1
            ·
            1 hour ago

            You’re right. I meant more by “non-destructive” that it is, depending on factors like intensity and known algorithm, reversible.

      • priapus@piefed.social
        link
        fedilink
        English
        arrow-up
        1
        ·
        edit-2
        6 hours ago

        This is true that some blurs could be undone, but the ones used in the files are definitely destructive and cannot be undone. Grok and any other image generation tool is also definitely not capable of doing it. It requires knowledge of how it was blurred so you can use the same algorithm to undo it, models simply guess what it should look like.

    • Pyr@lemmy.ca
      link
      fedilink
      English
      arrow-up
      9
      arrow-down
      2
      ·
      edit-2
      10 hours ago

      There was someone who reported that due to the incompetence of whitehouse staffers, some of the Epstein files had simply been “redacted” in ms word by highlighting the text black, so people were actually able to remove the redactions by turning the pdf back into word and removing the black highlighting to reveal the text.

      Who knows if some of the photos might be the same issue.

    • pkjqpg1h@lemmy.zip
      link
      fedilink
      English
      arrow-up
      2
      ·
      9 hours ago

      Actually, there is a short video on that page that explains this with examples