Ask Lemmy
A Fediverse community for open-ended, thought provoking questions
Please don't post about US Politics.
Rules: (interactive)
1) Be nice and; have fun
Doxxing, trolling, sealioning, racism, and toxicity are not welcomed in AskLemmy. Remember what your mother said: if you can't say something nice, don't say anything at all. In addition, the site-wide Lemmy.world terms of service also apply here. Please familiarize yourself with them
2) All posts must end with a '?'
This is sort of like Jeopardy. Please phrase all post titles in the form of a proper question ending with ?
3) No spam
Please do not flood the community with nonsense. Actual suspected spammers will be banned on site. No astroturfing.
4) NSFW is okay, within reason
Just remember to tag posts with either a content warning or a [NSFW] tag. Overtly sexual posts are not allowed, please direct them to either [email protected] or [email protected].
NSFW comments should be restricted to posts tagged [NSFW].
5) This is not a support community.
It is not a place for 'how do I?', type questions.
If you have any questions regarding the site itself or would like to report a community, please direct them to Lemmy.world Support or email [email protected]. For other questions check our partnered communities list, or use the search function.
Reminder: The terms of service apply here too.
Partnered Communities:
Logo design credit goes to: tubbadu
view the rest of the comments
Maybe each camera could have a unique private key that it could use to watermark keyframes with a hash of the frames themselves.
How would you prove that the camera itself is real, is the only device with access to the private key and isn't falsifying it's video feed?
The sort of case I was thinking of is if different parties present different versions of an image or video and you want to establish which version is altered and which is original.
You still have the same problem though. You can produce a camera in court and reject one of the images, but you still need to prove that the camera wasn't tampered with and it was the one at the scene of the crime.
Leica has one camera that does this, and others are working on them. Just posted this link in another comment
The camera can sign things however it wishes, but that doesn't automatically make the camera trustworthy.
In the same sense, I can sign any number of documents claiming to have seen a crime take place but that doesn't make it sufficient evidence.
In this case, digitally signing an image verifies that the image was generated by a specific camera (not just any camera of that brand) and that the image generated by that camera looks such and such a way. If anyone further edits the image the hash won't match the one from the signature, so it will be apparent it was tampered with.
What it can't do is tell you if someone pasted a printout of some false image over the lens, or in some other sophisticated way presented a doctored scene to the camera. But there's nothing preventing us from doing that today.
The question was about deepfakes right? So this is one tool to address that, but certainly not the only one the legal system would want to use.
I think that’s exactly how it’s going to work - you can’t force all ‘fake’ sources to have signatures- it’s too easy to make one without one for malicious reasons. Instead you have to create trusted sources of real images. Much easier and more secure
Usually I see non-technical people throw ideas like this and they're stupid, but I've been thinking about this for a few minutes and it's actually kinda smart