• P1r4nha@feddit.de
    link
    fedilink
    English
    arrow-up
    9
    ·
    1 year ago

    True, but only when you expect it. I’ve seen real pictures with weird lighting before and if I didn’t know they were real, I would’ve thought of Photoshop. With some experience you know what to look for, but there have already been plenty of studies showing that AI persons can not properly be distinguished from real people in pictures.

    • Otter@lemmy.ca
      link
      fedilink
      English
      arrow-up
      2
      ·
      1 year ago

      Definitely true

      I wonder if we’re going to end up with a new field of forensic medicine determining if media is real based on subtle anatomy/biomechanics details. Even if the person is real, a particular photo or video might not be

      • P1r4nha@feddit.de
        link
        fedilink
        English
        arrow-up
        6
        ·
        1 year ago

        Also at what point is an image altered? Every camera has an ISP that improves image quality, so that shouldn’t be counted as an alteration, even if it can impact the image quite significantly. Then there’s manual tools like photoshop, which can do things like the ISP, but also a lot more and then you got some “AI” tools that blur the lines even more. At least CNNs are just filter banks with learned kernels (to be very reductive). The cut is a bit clearer with diffusion image generation and similar tech, as that stuff is just clearly fake, but what about img-to-img diffusion?