Can we trust generative AI?
Generative Artificial Intelligence (GAI) models have demonstrated remarkable capabilities across various domains, yet their robustness remains a significant challenge, especially when exposed to adversarial attacks. This project aims to address the robustness issues in GAI models by evaluating their...
Saved in:
Main Author: | |
---|---|
Other Authors: | |
Format: | Thesis-Master by Coursework |
Language: | English |
Published: |
Nanyang Technological University
2025
|
Subjects: | |
Online Access: | https://hdl.handle.net/10356/182919 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Nanyang Technological University |
Language: | English |
id |
sg-ntu-dr.10356-182919 |
---|---|
record_format |
dspace |
spelling |
sg-ntu-dr.10356-1829192025-03-10T05:37:53Z Can we trust generative AI? Mu, Zhan Alex Chichung Kot School of Electrical and Electronic Engineering Rapid-Rich Object Search (ROSE) Lab EACKOT@ntu.edu.sg Engineering Computer and information science Generative Artificial Intelligence (GAI) models have demonstrated remarkable capabilities across various domains, yet their robustness remains a significant challenge, especially when exposed to adversarial attacks. This project aims to address the robustness issues in GAI models by evaluating their performance under adversarial conditions. Specifically, we utilize the COCO dataset to gen- erate adversarial examples using a variety of attack methods, including PGD, FMA, MULTIMODAL-FUSION, and DI-PGD. BLIP-large serves as the pre- trained model to generate adversarial datasets, which are subsequently used to attack other GAI models, including UniDiffuser, LLaVA, BLIP2-OPT, MiniGPT- 4, and BLIP-base. To assess the impact of adversarial attacks, we adopt eval- uation metrics such as CLIP score and textual cosine similarity to measure the differences between the captions generated on adversarial datasets and the orig- inal captions as well as their alignment with the original image content. This project not only reveals the vulnerabilities of current GAI models under ad- versarial scenarios but also establishes a benchmark for adversarial attacks on the latest GAI models, providing new references for research in this field. The findings of this study lay a solid foundation for future efforts to enhance the robustness of GAI models. Master's degree 2025-03-10T05:37:53Z 2025-03-10T05:37:53Z 2024 Thesis-Master by Coursework Mu, Z. (2024). Can we trust generative AI?. Master's thesis, Nanyang Technological University, Singapore. https://hdl.handle.net/10356/182919 https://hdl.handle.net/10356/182919 en ISM-DISS-04409 application/pdf Nanyang Technological University |
institution |
Nanyang Technological University |
building |
NTU Library |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
NTU Library |
collection |
DR-NTU |
language |
English |
topic |
Engineering Computer and information science |
spellingShingle |
Engineering Computer and information science Mu, Zhan Can we trust generative AI? |
description |
Generative Artificial Intelligence (GAI) models have demonstrated remarkable capabilities across various domains, yet their robustness remains a significant challenge, especially when exposed to adversarial attacks. This project aims to address the robustness issues in GAI models by evaluating their performance under adversarial conditions. Specifically, we utilize the COCO dataset to gen- erate adversarial examples using a variety of attack methods, including PGD, FMA, MULTIMODAL-FUSION, and DI-PGD. BLIP-large serves as the pre- trained model to generate adversarial datasets, which are subsequently used to attack other GAI models, including UniDiffuser, LLaVA, BLIP2-OPT, MiniGPT- 4, and BLIP-base. To assess the impact of adversarial attacks, we adopt eval- uation metrics such as CLIP score and textual cosine similarity to measure the differences between the captions generated on adversarial datasets and the orig- inal captions as well as their alignment with the original image content. This project not only reveals the vulnerabilities of current GAI models under ad- versarial scenarios but also establishes a benchmark for adversarial attacks on the latest GAI models, providing new references for research in this field. The findings of this study lay a solid foundation for future efforts to enhance the robustness of GAI models. |
author2 |
Alex Chichung Kot |
author_facet |
Alex Chichung Kot Mu, Zhan |
format |
Thesis-Master by Coursework |
author |
Mu, Zhan |
author_sort |
Mu, Zhan |
title |
Can we trust generative AI? |
title_short |
Can we trust generative AI? |
title_full |
Can we trust generative AI? |
title_fullStr |
Can we trust generative AI? |
title_full_unstemmed |
Can we trust generative AI? |
title_sort |
can we trust generative ai? |
publisher |
Nanyang Technological University |
publishDate |
2025 |
url |
https://hdl.handle.net/10356/182919 |
_version_ |
1826362243358392320 |