Learning deep networks for image segmentation

The domain of image processing and computer vision has witnessed significant strides in semantic segmentation, primarily propelled by advancements in Deep Convolutional Networks (DCNN). This paper conducts a comprehensive evaluation of traditional semantic segmentation methods, such as FastSCNN with...

全面介紹

Saved in:
書目詳細資料
主要作者: Akash, T
其他作者: Lin Guosheng
格式: Final Year Project
語言:English
出版: Nanyang Technological University 2024
主題:
在線閱讀:https://hdl.handle.net/10356/175141
標簽: 添加標簽
沒有標簽, 成為第一個標記此記錄!
實物特徵
總結:The domain of image processing and computer vision has witnessed significant strides in semantic segmentation, primarily propelled by advancements in Deep Convolutional Networks (DCNN). This paper conducts a comprehensive evaluation of traditional semantic segmentation methods, such as FastSCNN with its lightweight model and U-Net with its precise localization capabilities, compared with modern approaches like the Segment Anything Model (SAM) and its lightweight alternative, FastSAM. By implementing these varied models on the common benchmarking Cityscapes dataset, we dissect their strengths and weaknesses through various metrics. The study extends to adjusting and optimizing these models' parameters to enhance their performance. Furthermore, the research explores the integration of prompt-guided methodologies into conventional segmentation frameworks to elevate their adaptability and utility more robustly to unseen data. The future objective is to fuse the precision of traditional methods with the versatility of prompt-based techniques to forge models that are not only accurate but also proficient in handling unseen data scenarios.