Developing an automated wild boar and otter detection and monitoring system through video analysis
As Singapore continues its urban expansion while maintaining its "City in Nature" vision, human-wildlife interactions have become more frequent, leading to potential conflicts and ecological challenges. This study aims to develop an automated wild boar and otter detection and monitoring sy...
Saved in:
主要作者: | |
---|---|
其他作者: | |
格式: | Final Year Project |
語言: | English |
出版: |
Nanyang Technological University
2025
|
主題: | |
在線閱讀: | https://hdl.handle.net/10356/183814 |
標簽: |
添加標簽
沒有標簽, 成為第一個標記此記錄!
|
id |
sg-ntu-dr.10356-183814 |
---|---|
record_format |
dspace |
spelling |
sg-ntu-dr.10356-1838142025-04-17T05:09:22Z Developing an automated wild boar and otter detection and monitoring system through video analysis Wong, Wei Kai Lee Bu Sung, Francis College of Computing and Data Science EBSLEE@ntu.edu.sg Computer and Information Science As Singapore continues its urban expansion while maintaining its "City in Nature" vision, human-wildlife interactions have become more frequent, leading to potential conflicts and ecological challenges. This study aims to develop an automated wild boar and otter detection and monitoring system using motion-activated camera traps and artificial intelligence to support stakeholders such as the National Parks Board (NParks) and the Animal Concerns Research and Education Society (ACRES). Through extensive experimentation, YOLOv10b was selected as the optimal object detection model, achieving an impressive mAP@50 of 0.955, F1-score of 0.915, and an inference time of 13.8ms, outperforming DETR and other YOLOv10 variants with a balance of performance and inference speed. Small vision-language models were evaluated for extracting key attributes and generating captions for images with the detected wildlife. Qwen2-VL-2B was chosen over InternVL2.5-2B after experimentation due to its superior attribute extraction and image captioning capabilities. Prompt engineering via system prompts with added context improved F1-scores for 6 out of 8 key attribute extractions and increased the percentage of relevant image captioning responses from 71% to 96%, resulting in more accurate and informative outputs. A simple rule-based approach was used to classify the image’s need to notify stakeholders achieved an impressive F1-score of 0.912. This system serves as a step towards automated wildlife detection and monitoring systems, enabling stakeholders to respond quickly to human-wildlife encounters and early mitigation efforts to reduce risks to both people and wildlife, supporting Singapore’s vision of a “City in Nature”. Bachelor's degree 2025-04-17T05:09:21Z 2025-04-17T05:09:21Z 2025 Final Year Project (FYP) Wong, W. K. (2025). Developing an automated wild boar and otter detection and monitoring system through video analysis. Final Year Project (FYP), Nanyang Technological University, Singapore. https://hdl.handle.net/10356/183814 https://hdl.handle.net/10356/183814 en CCDS24-0748 application/pdf Nanyang Technological University |
institution |
Nanyang Technological University |
building |
NTU Library |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
NTU Library |
collection |
DR-NTU |
language |
English |
topic |
Computer and Information Science |
spellingShingle |
Computer and Information Science Wong, Wei Kai Developing an automated wild boar and otter detection and monitoring system through video analysis |
description |
As Singapore continues its urban expansion while maintaining its "City in Nature" vision, human-wildlife interactions have become more frequent, leading to potential conflicts and ecological challenges. This study aims to develop an automated wild boar and otter detection and monitoring system using motion-activated camera traps and artificial intelligence to support stakeholders such as the National Parks Board (NParks) and the Animal Concerns Research and Education Society (ACRES). Through extensive experimentation, YOLOv10b was selected as the optimal object detection model, achieving an impressive mAP@50 of 0.955, F1-score of 0.915, and an inference time of 13.8ms, outperforming DETR and other YOLOv10 variants with a balance of performance and inference speed. Small vision-language models were evaluated for extracting key attributes and generating captions for images with the detected wildlife. Qwen2-VL-2B was chosen over InternVL2.5-2B after experimentation due to its superior attribute extraction and image captioning capabilities. Prompt engineering via system prompts with added context improved F1-scores for 6 out of 8 key attribute extractions and increased the percentage of relevant image captioning responses from 71% to 96%, resulting in more accurate and informative outputs. A simple rule-based approach was used to classify the image’s need to notify stakeholders achieved an impressive F1-score of 0.912. This system serves as a step towards automated wildlife detection and monitoring systems, enabling stakeholders to respond quickly to human-wildlife encounters and early mitigation efforts to reduce risks to both people and wildlife, supporting Singapore’s vision of a “City in Nature”. |
author2 |
Lee Bu Sung, Francis |
author_facet |
Lee Bu Sung, Francis Wong, Wei Kai |
format |
Final Year Project |
author |
Wong, Wei Kai |
author_sort |
Wong, Wei Kai |
title |
Developing an automated wild boar and otter detection and monitoring system through video analysis |
title_short |
Developing an automated wild boar and otter detection and monitoring system through video analysis |
title_full |
Developing an automated wild boar and otter detection and monitoring system through video analysis |
title_fullStr |
Developing an automated wild boar and otter detection and monitoring system through video analysis |
title_full_unstemmed |
Developing an automated wild boar and otter detection and monitoring system through video analysis |
title_sort |
developing an automated wild boar and otter detection and monitoring system through video analysis |
publisher |
Nanyang Technological University |
publishDate |
2025 |
url |
https://hdl.handle.net/10356/183814 |
_version_ |
1831146560712343552 |