Can we trust generative AI?

Generative Artificial Intelligence (GAI) models have demonstrated remarkable capabilities across various domains, yet their robustness remains a significant challenge, especially when exposed to adversarial attacks. This project aims to address the robustness issues in GAI models by evaluating their...

Full description

Saved in:
Bibliographic Details
Main Author: Mu, Zhan
Other Authors: Alex Chichung Kot
Format: Thesis-Master by Coursework
Language:English
Published: Nanyang Technological University 2025
Subjects:
Online Access:https://hdl.handle.net/10356/182919
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Nanyang Technological University
Language: English
id sg-ntu-dr.10356-182919
record_format dspace
spelling sg-ntu-dr.10356-1829192025-03-10T05:37:53Z Can we trust generative AI? Mu, Zhan Alex Chichung Kot School of Electrical and Electronic Engineering Rapid-Rich Object Search (ROSE) Lab EACKOT@ntu.edu.sg Engineering Computer and information science Generative Artificial Intelligence (GAI) models have demonstrated remarkable capabilities across various domains, yet their robustness remains a significant challenge, especially when exposed to adversarial attacks. This project aims to address the robustness issues in GAI models by evaluating their performance under adversarial conditions. Specifically, we utilize the COCO dataset to gen- erate adversarial examples using a variety of attack methods, including PGD, FMA, MULTIMODAL-FUSION, and DI-PGD. BLIP-large serves as the pre- trained model to generate adversarial datasets, which are subsequently used to attack other GAI models, including UniDiffuser, LLaVA, BLIP2-OPT, MiniGPT- 4, and BLIP-base. To assess the impact of adversarial attacks, we adopt eval- uation metrics such as CLIP score and textual cosine similarity to measure the differences between the captions generated on adversarial datasets and the orig- inal captions as well as their alignment with the original image content. This project not only reveals the vulnerabilities of current GAI models under ad- versarial scenarios but also establishes a benchmark for adversarial attacks on the latest GAI models, providing new references for research in this field. The findings of this study lay a solid foundation for future efforts to enhance the robustness of GAI models. Master's degree 2025-03-10T05:37:53Z 2025-03-10T05:37:53Z 2024 Thesis-Master by Coursework Mu, Z. (2024). Can we trust generative AI?. Master's thesis, Nanyang Technological University, Singapore. https://hdl.handle.net/10356/182919 https://hdl.handle.net/10356/182919 en ISM-DISS-04409 application/pdf Nanyang Technological University
institution Nanyang Technological University
building NTU Library
continent Asia
country Singapore
Singapore
content_provider NTU Library
collection DR-NTU
language English
topic Engineering
Computer and information science
spellingShingle Engineering
Computer and information science
Mu, Zhan
Can we trust generative AI?
description Generative Artificial Intelligence (GAI) models have demonstrated remarkable capabilities across various domains, yet their robustness remains a significant challenge, especially when exposed to adversarial attacks. This project aims to address the robustness issues in GAI models by evaluating their performance under adversarial conditions. Specifically, we utilize the COCO dataset to gen- erate adversarial examples using a variety of attack methods, including PGD, FMA, MULTIMODAL-FUSION, and DI-PGD. BLIP-large serves as the pre- trained model to generate adversarial datasets, which are subsequently used to attack other GAI models, including UniDiffuser, LLaVA, BLIP2-OPT, MiniGPT- 4, and BLIP-base. To assess the impact of adversarial attacks, we adopt eval- uation metrics such as CLIP score and textual cosine similarity to measure the differences between the captions generated on adversarial datasets and the orig- inal captions as well as their alignment with the original image content. This project not only reveals the vulnerabilities of current GAI models under ad- versarial scenarios but also establishes a benchmark for adversarial attacks on the latest GAI models, providing new references for research in this field. The findings of this study lay a solid foundation for future efforts to enhance the robustness of GAI models.
author2 Alex Chichung Kot
author_facet Alex Chichung Kot
Mu, Zhan
format Thesis-Master by Coursework
author Mu, Zhan
author_sort Mu, Zhan
title Can we trust generative AI?
title_short Can we trust generative AI?
title_full Can we trust generative AI?
title_fullStr Can we trust generative AI?
title_full_unstemmed Can we trust generative AI?
title_sort can we trust generative ai?
publisher Nanyang Technological University
publishDate 2025
url https://hdl.handle.net/10356/182919
_version_ 1826362243358392320