Multi-modal deception detection in videos
Deception detection has much significance as it has many real-world applications. This project focuses on the verbal and visual modalities for deception detection from videos. The experiments were conducted on the most widely used dataset: Real-Life Trial. In the project, text information, visual in...
Saved in:
Main Author: | |
---|---|
Other Authors: | |
Format: | Final Year Project |
Language: | English |
Published: |
Nanyang Technological University
2022
|
Subjects: | |
Online Access: | https://hdl.handle.net/10356/158188 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Nanyang Technological University |
Language: | English |
id |
sg-ntu-dr.10356-158188 |
---|---|
record_format |
dspace |
spelling |
sg-ntu-dr.10356-1581882023-07-07T19:23:02Z Multi-modal deception detection in videos Jin, Yibing Alex Chichung Kot School of Electrical and Electronic Engineering Wuhan University EACKOT@ntu.edu.sg Engineering::Electrical and electronic engineering Deception detection has much significance as it has many real-world applications. This project focuses on the verbal and visual modalities for deception detection from videos. The experiments were conducted on the most widely used dataset: Real-Life Trial. In the project, text information, visual information, and multimodal cues were considered individually for detecting deception. Moreover, this project used both machine learning and deep learning methods to obtain the best performance on this task. For verbal feature extraction, TF-IDF, N-Grams, and LIWC were used to transfer the text into vectors. These vectors were processed by SVM, Naïve Bayes, Random Forest, and RNN. For visual feature extraction, facial action features and gaze direction features were extracted by OpenFace. The visual features are learned by a machine learning method, SVM. This paper also used a hybrid classification model based on CNN and GRU neural networks. For the multimodal machine learning method, the features from two different modalities were concatenated together after extraction, and then the extracted features were put into SVM for classification. The experimental results suggest that the hybrid model of CNN and GRU performs best among all the methods when only information from one modality is used, and using SVM outperforms other models when two modalities are used. Bachelor of Engineering (Electrical and Electronic Engineering) 2022-05-31T08:44:25Z 2022-05-31T08:44:25Z 2022 Final Year Project (FYP) Jin, Y. (2022). Multi-modal deception detection in videos. Final Year Project (FYP), Nanyang Technological University, Singapore. https://hdl.handle.net/10356/158188 https://hdl.handle.net/10356/158188 en application/pdf Nanyang Technological University |
institution |
Nanyang Technological University |
building |
NTU Library |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
NTU Library |
collection |
DR-NTU |
language |
English |
topic |
Engineering::Electrical and electronic engineering |
spellingShingle |
Engineering::Electrical and electronic engineering Jin, Yibing Multi-modal deception detection in videos |
description |
Deception detection has much significance as it has many real-world applications. This project focuses on the verbal and visual modalities for deception detection from videos. The experiments were conducted on the most widely used dataset: Real-Life Trial. In the project, text information, visual information, and multimodal cues were considered individually for detecting deception. Moreover, this project used both machine learning and deep learning methods to obtain the best performance on this task.
For verbal feature extraction, TF-IDF, N-Grams, and LIWC were used to transfer the text into vectors. These vectors were processed by SVM, Naïve Bayes, Random Forest, and RNN. For visual feature extraction, facial action features and gaze direction features were extracted by OpenFace. The visual features are learned by a machine learning method, SVM. This paper also used a hybrid classification model based on CNN and GRU neural networks. For the multimodal machine learning method, the features from two different modalities were concatenated together after extraction, and then the extracted features were put into SVM for classification.
The experimental results suggest that the hybrid model of CNN and GRU performs best among all the methods when only information from one modality is used, and using SVM outperforms other models when two modalities are used. |
author2 |
Alex Chichung Kot |
author_facet |
Alex Chichung Kot Jin, Yibing |
format |
Final Year Project |
author |
Jin, Yibing |
author_sort |
Jin, Yibing |
title |
Multi-modal deception detection in videos |
title_short |
Multi-modal deception detection in videos |
title_full |
Multi-modal deception detection in videos |
title_fullStr |
Multi-modal deception detection in videos |
title_full_unstemmed |
Multi-modal deception detection in videos |
title_sort |
multi-modal deception detection in videos |
publisher |
Nanyang Technological University |
publishDate |
2022 |
url |
https://hdl.handle.net/10356/158188 |
_version_ |
1772826258946654208 |