The ability of artificial intelligence techniques to build synthesized brand new videos or to alter the facial expression of already existing ones has been efficiently demonstrated in the literature. The identification of such new threat generally known as Deepfake, but consisting of different techniques, is fundamental in multimedia forensics. In fact this kind of manipulated information could undermine and easily distort the public opinion on a certain person or about a specific event. Thus, in this paper, a new technique able to distinguish synthetic generated portrait videos from natural ones is introduced by exploiting inconsistencies due to the prediction error in the re-encoding phase. In particular, features based on inter-frame prediction error have been investigated jointly with a Long Short-Term Memory (LSTM) model network able to learn the temporal correlation among consecutive frames. Preliminary results have demonstrated that such sequence-based approach, used to distinguish between original and manipulated videos, highlights promising performances.
Dettaglio pubblicazione
2020, IH and MMSec 2020 - Proceedings of the 2020 ACM Workshop on Information Hiding and Multimedia Security, Pages 97-102
Exploiting Prediction Error Inconsistencies through LSTM-based Classifiers to Detect Deepfake Videos (04b Atto di convegno in volume)
Amerini I., Caldelli R.
ISBN: 9781450370509
Gruppo di ricerca: Computer Vision, Computer Graphics, Deep Learning, Gruppo di ricerca: Cybersecurity, Gruppo di ricerca: Theory of Deep Learning
keywords