• Corpus ID: 58014139

Enhance the Motion Cues for Face Anti-Spoofing using CNN-LSTM Architecture

  title={Enhance the Motion Cues for Face Anti-Spoofing using CNN-LSTM Architecture},
  author={Xiaoguang Tu and Hengsheng Zhang and M. Xie and Yao Luo and Yuefei Zhang and Z. Ma},
Spatio-temporal information is very important to capture the discriminative cues between genuine and fake faces from video sequences. To explore such a temporal feature, the fine-grained motions (e.g., eye blinking, mouth movements and head swing) across video frames are very critical. In this paper, we propose a joint CNN-LSTM network for face anti-spoofing, focusing on the motion cues across video frames. We first extract the high discriminative features of video frames using the conventional… 

Tables from this paper

Face anti-spoofing with local difference network and binary facial mask supervision

A two-stream network that consists of a convolutional network and a local difference network, aiming for discriminative information mining and information redundancy elimination for face anti-spoofing systems and a simple binary facial mask supervision strategy is advocated.

Two-stream Convolutional Networks for Multi-frame Face Anti-spoofing

An efficient two-stream model is proposed to capture the key differences between live and spoof faces, which takes multi-frames and RGB difference as input respectively and achieves the state-of-the-art results on most of datasets' protocol with much less parameter size.

Enhanced Anisotropic Diffusion-based CNN-LSTM Architecture for Video Face Liveness Detection

  • R. KoshyA. Mahmood
  • Computer Science
    2020 19th IEEE International Conference on Machine Learning and Applications (ICMLA)
  • 2020
A novel deep architecture for face liveness detection on video frames that uses the diffusion of images followed by a deep Convolutional Neural Network and a Long Short-Term Memory to classify the video sequence as real or fake.

FaceBagNet: Bag-Of-Local-Features Model for Multi-Modal Face Anti-Spoofing

This paper proposes a multi-stream CNN architecture called FaceBagNet to make full use of the recently published CASIA-SURF dataset, and designs a Modal Feature Erasing (MFE) operation on the multi-modal features which erases features from one randomly selected modality during training.

Exposing AI-generated videos with motion magnification

This paper introduces a motion discrepancy based method that can effectively differentiate AI-generated fake videos from real ones and shows superior performance contrasted to existing pixel-based fake video forensics approaches.

Enhanced Deep Learning Architectures for Face Liveness Detection for Static and Video Sequences

A novel deep architecture for face liveness detection on video frames that uses the diffusion of images followed by a deep Convolutional Neural Network and a Long Short-Term Memory to classify the video sequence as real or fake.

Networks Ensemble for Multi-modal Cross-ethnicity Face Anti-spoofing

A simple yet effective multi-stream architecture which implements the feature extraction of each modality and the aggregation of the multi-modal features is adopted and the experimental results indicate the effectiveness of the proposed method.

Deep transfer across domains for face antispoofing

This work proposes a deep domain transfer CNN using sparsely unlabeled data from the target domain to learn features that are invariant across domains for face antispoofing, and focuses on improving the generalizability of convolutional neural network (CNN)-based face antisPOofing methods across different kinds of datasets.

DTFA-Net: Dynamic and Texture Features Fusion Attention Network for Face Antispoofing

A dynamic information fusion structure of an interchannel attention block to fuse the magnitude and direction of optical flow to extract facial motion features and an improved Gamma image preprocessing algorithm, which effectively improved the face detection ability.

Image-to-Video Generation via 3D Facial Dynamics

This paper proposes to “imagine” a face video from a single face image according to the reconstructed 3D face dynamics, aiming to generate a realistic and identity-preserving face video, with precisely predicted pose and facial expression.



Learning temporal features using LSTM-CNN architecture for face anti-spoofing

This work proposes a deep neural network architecture combining Long Short-Term Memory (LSTM) units with Convolutional Neural Networks (CNN) that works well for face anti-spoofing by utilizing the LSTM units' ability of finding long relation from its input sequences as well as extracting local and dense features through convolution operations.

Learn Convolutional Neural Network for Face Anti-Spoofing

Instead of designing feature by ourselves, the deep convolutional neural network is relied on to learn features of high discriminative ability in a supervised manner and combined with some data pre-processing, the face anti-spoofing performance improves drastically.

Face anti-spoofing via motion magnification and multifeature videolet aggregation

A new framework for face spoofing detection in videos using motion magnification and multifeature evidence aggregation in a windowed fashion is presented, which yields state-of-the-art performance and robust generalizability with low computational complexity.

Macro- and micro-expression spotting in long videos using spatio-temporal strain

A method that utilizes the strain impacted on the facial skin due to the non-rigid motion caused during expressions to detect and distinguish between large expressions (macro) and rapid and localized expressions (micro).

Real-Time Face Detection and Motion Analysis With Application in “Liveness” Assessment

A robust face detection technique along with mouth localization, processing every frame in real time (video rate), is presented and "liveness" verification barriers are proposed as applications for which a significant amount of computation is avoided when estimating motion.

Transfer Learning Using Convolutional Neural Networks for Face Anti-spoofing

This work presents an approach based on transfer learning using a pre-trained CNN model using only static features to recognize photo, video or mask attacks to be comparable to the state-of-the-art.

Deep face liveness detection based on nonlinear diffusion using convolution neural network

This paper presents an efficient and non-intrusive method to counter face-spoofing attacks that uses a single image to detect spoofing attacks and proposes a specialized deep convolution neural network that can extract the discriminative and high-level features of the input diffused image to differentiate between a fake face and a real face.

Describing Videos by Exploiting Temporal Structure

This work proposes an approach that successfully takes into account both the local and global temporal structure of videos to produce descriptions and proposes a temporal attention mechanism that allows to go beyond local temporal modeling and learns to automatically select the most relevant temporal segments given the text-generating RNN.

Face Liveness Detection from a Single Image with Sparse Low Rank Bilinear Discriminative Model

This paper develops two new extensions to the sparse logistic regression model which allow quick and accurate spoof detection and proposes two strategies to extract the essential information about different surface properties of a live human face or a photograph, in terms of latent samples.

Large-Scale Video Classification with Convolutional Neural Networks

This work studies multiple approaches for extending the connectivity of a CNN in time domain to take advantage of local spatio-temporal information and suggests a multiresolution, foveated architecture as a promising way of speeding up the training.