If AI and deep fakes can listen to a video or audio of a person and then are able to successfully reproduce such person, what does this entail for trials?

It used to be that recording audio or video would give strong information which often would weigh more than witnesses, but soon enough perfect forgery could enter the courtroom just as it’s doing in social media (where you’re not sworn to tell the truth, though the consequences are real)

I know fake information is a problem everywhere, but I started wondering what will happen when it creeps in testimonies.

How will we defend ourselves, while still using real videos or audios as proof? Or are we just doomed?

  • ConstipatedWatson@lemmy.worldOP
    link
    fedilink
    arrow-up
    2
    ·
    26 days ago

    True, sooner or later there might be ways to make sure that a picture or video are digitally signed and probably it would be very hard to crack, but theoretically a fake video might still pass for real (though it would require a lot of resources to make that happen)

    • SirEDCaLot@lemmy.today
      link
      fedilink
      arrow-up
      1
      ·
      25 days ago

      More likely, most of the sources that produce photos and videos would not be using the digital signatures. Professional cameras for journalists probably would have the signature chip. Cheapo Chinese surveillance cameras? Unlikely.