scispace - formally typeset
A

Alexandros Haliassos

Researcher at Imperial College London

Publications -  11
Citations -  204

Alexandros Haliassos is an academic researcher from Imperial College London. The author has contributed to research in topics: Computer science & Engineering. The author has an hindex of 1, co-authored 5 publications receiving 5 citations.

Papers
More filters
Posted Content

Lips Don't Lie: A Generalisable and Robust Approach to Face Forgery Detection

TL;DR: Extensive experiments show that this simple approach significantly surpasses the state-of-the-art in terms of generalisation to unseen manipulations and robustness to perturbations, as well as shed light on the factors responsible for its performance.
Proceedings ArticleDOI

Lips Don't Lie: A Generalisable and Robust Approach to Face Forgery Detection

TL;DR: LipForensics as mentioned in this paper targets high-level semantic irregularities in mouth movements, which are common in many generated videos, by first pretraining a spatio-temporal network to perform visual speech recognition (lipreading), thus learning rich internal representations related to natural mouth motion.
Proceedings ArticleDOI

Leveraging Real Talking Faces via Self-Supervision for Robust Forgery Detection

TL;DR: This paper harnesses the natural correspondence between the visual and auditory modalities in real videos to learn temporally dense video representations that capture factors such as facial movements, expression, and identity, and suggests that leveraging natural and unlabelled videos is a promising direction for the development of more robust face forgery detectors.
Proceedings ArticleDOI

SVTS: Scalable Video-to-Speech Synthesis

TL;DR: This work introduces a scalable video-to-speech framework consisting of two components: a video- to-spectrogram predictor and a pre-trained neural vocoder, which converts the mel-frequency spectrograms into waveform audio.
Proceedings ArticleDOI

Auto-AVSR: Audio-Visual Speech Recognition with Automatic Labels

TL;DR: In this article , the authors used publicly available pre-trained ASR models to automatically transcribe unlabeled datasets such as AVSpeech and VoxCeleb2, and trained ASR, VSR and AV-ASR models on the augmented training set, which consists of the LRS2 and LRS3 datasets as well as the additional automatically-transcribed data.