If AI and deep fakes can listen to a video or audio of a person and then are able to successfully reproduce such person, what does this entail for trials?
It used to be that recording audio or video would give strong information which often would weigh more than witnesses, but soon enough perfect forgery could enter the courtroom just as it’s doing in social media (where you’re not sworn to tell the truth, though the consequences are real)
I know fake information is a problem everywhere, but I started wondering what will happen when it creeps in testimonies.
How will we defend ourselves, while still using real videos or audios as proof? Or are we just doomed?
I think other answers here are more essential - chain of custody, corroborating evidence, etc.
That said, Leica has released a camera that digitally signs its images, and other manufacturers are working on similar things. That will allow people to verify whether the image is original or has been edited. From what I understand Leica has some scheme where you can sign images when you update them too, so there’s a whole chain of documentation. Here’s a brief article
It’s an interesting experiment, but why would we trust everything that Leica supposedly verified? The same shit with digital signatures and blockchain stuff. We are at the gates of the world where we have zero trust by default and would only intentionally outsource verification to third parties we trust, because penalties for mistakes are growing each day.
I don’t think we should inherently. I’ve thought about the idea of digitally signed photos and it seems sound unless someone is quite clever with electronics. I’m guessing there’s some embedded key on the camera that is hard but maybe not impossible to access. If people can hack Teslas for “full autopilot” or run Doom on an ATM machine I’m not confident that this kind of encryption will never be cracked. However, I would hope an expert witness would also examine the camera that supposedly took the picture. I would think it to be impossible for someone to acquire the key without a 3rd party detecting the intrusion.
Today we have EXIFs and it’s better to wipe them all of these for privacy reasons. Because every picture you take otherwise contains a lot of your data like geoloc, model, exposuer, etc. That’s the angle they are yet to tackle - because most of these things are also leave us vulnerable.
Well as I said, I think there’s a collection of things we already use for judging what’s true, this would just be one more tool.
A cryptographic signature (in the original sense, not just the Bitcoin sense) means that only someone who possesses a certain digital key is able to sign something. In the case of a digitally signed photo, it verifies “hey I, key holder, am signing this file”. And if the file is edited, the signed document won’t match the tampered version.
Is it possible someone could hack and steal such a key? Yes. We see this with certificates for websites, where some bad actor is able to impersonate a trusted website. (And of course when NFT holders get their apes stolen)
But if something like that happened it’s a cause for investigation, and it leaves a trail which authorities could look into. Not perfect, but right now there’s not even a starting point for “did this image come from somewhere real?”
A camera that authenticates the timestamp and contents of an image is great. But it’s still limited. If I take that camera, mount it on a tripod, and take a perfect photograph of a poster of Van Gogh’s Starry Night, the resulting image will be yet another one of millions of similar copies, only with a digital signature proving that it was a newly created image today, in 2024.
Authenticating what the camera sensor sees is only part of the problem, when the camera can be shown fake stuff, too. Special effects have been around for decades, and practical effects are even older.
Fake evidence, e.g. forged documents, are not not new things. They take things like origin, chain of custody etc into account.
Sure, but if you meet up with someone and they later have an audio recording that is completely fabricated from the real audio, there’s nothing for chain of anything. Audio used to be damning evidence and was fairly easily discoverable if it was hacked together to try to sound different. If that goes away, then it just becomes useless as evidence.
You can’t just use an audio file by itself. It has to come from somewhere.
The courts already have a system in place that if someone seeks to introduce a screenshot of a text message, or a printout of a webpage, or a VHS tape with video, or just a plain audio file, needs to be able to introduce that as evidence, with someone who testifies that it is real and that it is accurate, with an opportunity for others to question and even investigate where it came from and how it was made/stored/copied.
If I just show up to a car accident case with an audio recording that I claim is the other driver admitting that he forgot to look before turning, that audio is gonna do basically nothing unless and until I show that I had a reason to be making that recording while talking to him, why I didn’t give it to the police who wrote the accident report that day, etc. And even then, the other driver can say “that’s not me and I don’t know what you think that recording is” and we’re still back to a credibility problem.
We didn’t need AI to do impressions of people. This has always been a problem, or a non-problem, in evidence.
It becomes useless as evidence unless you can establish authenticity. It just makes audio recordings more in a class with text documents; perfectly fakeable, but admissible with the right supporting information. So I agree it’s a change, but it’s not the end of audio evidence, and it’s a change in a direction which courts already have experience.
I’m not a tech person, so I’ll take the lowest hanging fruit. The obvious answer is to write a program that can detect AI. Then there will be a competition between AI fakes and AI detection. This is similar to what we have in sports. There are forbidden enhancement procedures (e.g., steroids, blood doping, etc.) that have to keep improving in subtlety so not to be detected by Anti cheating measures.
That’s essentially how Generative adversarial networks work, and the effect is that the generative program gets better at making its fakes be undetectable
Fake does not change what actually happened. Just look for facts in the real world that support the theory. Remember, photoshop existed before AI. We have DNA checks today.
Fake does not change what actually happened
Seriously? You want to charge people just because the fake photo/video could have incriminate them? Welcome to the minority report…
As someone who works in the field of criminal law (in Europe, and I would be shocked if it wasn’t the same in the US) - I’m not actually very worried about this. By that I don’t mean to say it’s not a problem, though.
The risk of evidence being tampered with or outright falsified is something that already exists, and we know how to deal with it. What AI will do is lower the barrier for technical knowledge needed to do it, making the practice more common.
While it’s pretty easy for most AI images to be spotted by anyone with some familiarity with them, they’re only going to get better and I don’t imagine it will take very long before they’re so good the average person can’t tell.
In my opinion this will be dealt with via two mechanisms:
-
Automated analysis of all digital evidence for signatures of AI as a standard practice. Whoever can be the first person to land contracts with police departments to provide bespoke software for quick forensic AI detection is going to make a lot of money.
-
A growth in demand for digital forensics experts who can provide evidence on whether something is AI generated. I wouldn’t expect them to be consulted on all cases with digital evidence, but for it to become standard practice where the defence raises a challenge about a specific piece of evidence during trial.
Other than that, I don’t think the current state of affairs when it comes to doctored evidence will particularly change. As I say, it’s not a new phenomenon, so countries already have the legal and procedural framework in place to deal with it. It just needs to be adjusted where needed to accommodate AI.
What concerns me much more than the issue you raise is the emergence of activities which are uniquely AI dependent and need legislating for. For example, how does AI generated porn of real people fit into existing legislation on sex offences? Should it be an offence? Should it be treated differently to drawing porn of someone by hand? Would this include manually created digital images without the use of AI? If it’s not decided to be illegal generally, what about when it depicts a child? Is it the generation of the image that should be regulated, or the distribution? That’s just one example. What about AI enabled fraud? That’s a whole can of worms in itself, legally speaking. These are questions that in my opinion are beyond the remit of the courts and will require direction from central governments and fresh, tailor made legislation to deal with.
-
Maybe each camera could have a unique private key that it could use to watermark keyframes with a hash of the frames themselves.
Usually I see non-technical people throw ideas like this and they’re stupid, but I’ve been thinking about this for a few minutes and it’s actually kinda smart
I think that’s exactly how it’s going to work - you can’t force all ‘fake’ sources to have signatures- it’s too easy to make one without one for malicious reasons. Instead you have to create trusted sources of real images. Much easier and more secure
How would you prove that the camera itself is real, is the only device with access to the private key and isn’t falsifying it’s video feed?
The sort of case I was thinking of is if different parties present different versions of an image or video and you want to establish which version is altered and which is original.
You still have the same problem though. You can produce a camera in court and reject one of the images, but you still need to prove that the camera wasn’t tampered with and it was the one at the scene of the crime.
Leica has one camera that does this, and others are working on them. Just posted this link in another comment
The camera can sign things however it wishes, but that doesn’t automatically make the camera trustworthy.
In the same sense, I can sign any number of documents claiming to have seen a crime take place but that doesn’t make it sufficient evidence.
It’s a scary question, made a lot less scary by whoever it was that said “you know, I guess we’ve had text deepfakes a long time”
Eventually people just know it could be fake, so they look for other ways of verifying. The inevitability and the scale of it mean that, at the very least, we’ll have all our brainpower on it eventually.
It’s the meantime where shit could get wild.
One step closer to requiring smart phones to track an individual for their alibi
Disclaimer: I’m not an expert, just an interested amateur wanting to chat and drawing from comparisons from past leaps in tech and other conversations/videos.
For a time expert analysis will probably work. For instance, the “click here to prove your not a robot” boxes can definitely be clicked by robots, but for now the robot moves in detectably different way. My guess is that, for at least a while, AI content will be different from actual video in ways like code. There will probably be an arms race of sorts between AI and methods to detect AI.
Other forms of evidence like DNA, eyewitness accounts, cell phone tracking etc. will likely help mitigate deceitful AI somewhat. My guess is that soon video/audio will no longer be considered as ironclad as it was even a few years ago.
There are discussions about making AI tools have a digital “watermark” than can be used to identify AI-generated content. Of course this won’t help with black market-type programs, but it will keep most people out of the “deep fake for trials” game.
When it comes to misinformation on social media though, well…it’s probably going to get crazy. The last decade or so has been a race at an unprecedented scale to try and keep up with BS “proof”, psuedoscience, etc. Sadly those on the side of truth haven’t always won. The only answer I have for that is making sure people are educated about how to deal with misinformation and deepfakes - eg. awareness they exist, identifying reputable sources and expert consensus, and so on.
For the longest time now, from before AI, before NFT was a thing i had an idea to incorporating blockchain tech into real life media footage to combat the rise of misinformation.
The metadata, original author would be stored on this chain the moment footage is recorded. The biggest challenge is that this means the devices themselves need to be connected.
Adoption would be slow but i imagined news and official channels make use of this tech first. Eventually all footage outside of this will be seen as not trustworthy
Then NFTBros came along and people have shit on this idea ever since. Some days i feel that was a conspiracy to ruim out perception of potential but more likely humans where just greedy.
I still believe this could work. Detailed example below:
The system works with a fair amount of transparency, verifiable digital signatures for recording devices and their owners. Professional cameras and organizations would have publicly known IDs, while individuals could choose to remain pseudonymous authors but would need to build credibility over time.
Let’s say BBC records an interview. When viewers watch this content on any platform, they can access blockchain verification through an embedded interface (perhaps a small icon in the corner). This shows the complete chain of custody from recording to broadcast.
The system verifies content through computational comparisons. When a raw interview is edited into a final piece:
- Each original clip has a unique blockchain signature
- The final edited version’s signature can be compared against source material
- Automated analysis shows what percentage of original footage matches
- Modifications like color correction or audio adjustments are detected through signature differences
- Additional elements like station logos or intro sequences have their own verified identifiers
Because it’s insanely idiotic. Signing videos is one thing.
Hooking it into blockchain bullshit is entirely deranged. It adds a bunch of complexity to provide literally zero benefit in any possible context.
I am not sure what you think blockchain actually is but in essence its a decentralized ledger of signatures.
Not coins, no sellable goods. Just that. Computers connected in a network to verify the correctness of a cloud ledger.
So if you say signing footage is one thing how do you propose a laymen can verify that signature without centralized databank.
I understand some people may not mind centralized authority but i prefer against it.
I am willing to hear peoples thoughts on this. I am not pro or against blockchain or any form of technology. With the information i have this just seems like a reasonable and practical solution.
I am well aware of what it is. It serves no purpose and provides no benefit.
Ignoring the fact that hardware signing doesn’t validate inputs as “real”, because it’s entirely possible to replicate the actual signals entering the camera, and the fact that the entire premise by definition would be a terrible power grab by big hardware/software tools, the very obvious way to implement such an approach would be the exact same system as certificate authorities. You have to have actual root certificate signers.
Blockchain is horseshit and serves no purpose.
That hardware inputs can be faked is part of my reasoning here because there would be transparency of the source of footage.
If a reputable journalists fake their own footage and it would be found out their credibility would be gone.
If they often rely on borrowing footage and don’t fact check it. Credibility will degrade as well.
Journalist media that does their work and only uses credible sources will thrive.
My solution isn’t about who or how signature gets created but how ordinary people can check for themselves where a clip within footage originates from.
I am fine with inventing a new system that does this and call it something else than blockchain. But my understanding is that it does pretty much provide this functionality in a robust manner.
Also typing these comments on the go caused me to lose something dear to me on public transport. I am very sad now and probably wont engage further.
Again, you have to completely ignore that the core premise is evil intended to give big players even stronger monopoly control. It’s anti-free in every sense, and as an added bonus, would very certainly make possession of specific hardware sufficient to be executed in some countries, because everything it has ever captured would be tracked to it.
But if you do that, there is already a system that does exactly what you’re asking. You don’t need to invent anything. It’s certificate authorities.
Sure i’ll have a look at decentralized certificate authorities options.
Very possibles to adapt my idea to whatever technology provides those function honestly.
The only actual connection i have with blockchain is that reading about it when it was new directly inspired in me a possible way to combat fake news.
Science has proven that the entire model of human memory as factual testimony is a fallacy. That came out long before AI in the public space. I don’t think anyone has addressed that revelation. I doubt anyone will address this one. Hell, there are still people sketching the courtroom like cameras don’t exist. A president can stage a failed coup and a SC judge can fly the traitor’s flag and there are no consequences for either.
So what will be done, absolutely nothing, unless some billionaires stage a proxy war over it.
We’re not. Its going to upend our already laughably busted “justice” system to new unknown heights of cartoonish malfeasance.
Eventually, we will just have to accept that photographic proof is no longer proof.
There are ways that you could guarantee an image is valid. You would need a hardware security module inside the camera, which signs a hash of the picture with its own built-in security key that can’t be extracted and a serial number that it generates. That can prove that an image came from a particular camera, and if you change even one pixel of that image the signature won’t match anymore. I don’t see this happening anytime soon. Not mainstream at least. There are one or two camera manufacturers that offer this as a feature, but it’s not on things like surveillance cameras or cell phones nor will it be anytime soon.
True, sooner or later there might be ways to make sure that a picture or video are digitally signed and probably it would be very hard to crack, but theoretically a fake video might still pass for real (though it would require a lot of resources to make that happen)
I doubt these tools will ever get to a level of quality that can confuse a court. They’ll get better, sure, but they’ll never really get there.
A bit dramatic imo. For most of legal history we didn’t actually have perfectly recorded video or audio, and while they are great tools at the present, they are still not the silver-bullet people would expect them to be at trial. (Think Trump and his cucks) Furthermore, most poor people try to avoid being recorded when doing crimes.
It will probably mean that focus will shift to other kinds of evidence and evidence-gathering methods. But definitely not the end of law as we know it, far from it.
Right, but anyone would like to not be in a video implying them in a crime, but I was wondering what would happen if fake videos of said person were to appear implicating a crime that actually did not take place