Learning deep networks for image segmentation
The domain of image processing and computer vision has witnessed significant strides in semantic segmentation, primarily propelled by advancements in Deep Convolutional Networks (DCNN). This paper conducts a comprehensive evaluation of traditional semantic segmentation methods, such as FastSCNN with...
Saved in:
主要作者: | |
---|---|
其他作者: | |
格式: | Final Year Project |
語言: | English |
出版: |
Nanyang Technological University
2024
|
主題: | |
在線閱讀: | https://hdl.handle.net/10356/175141 |
標簽: |
添加標簽
沒有標簽, 成為第一個標記此記錄!
|
總結: | The domain of image processing and computer vision has witnessed significant strides in semantic segmentation, primarily propelled by advancements in Deep Convolutional Networks (DCNN). This paper conducts a comprehensive evaluation of traditional semantic segmentation methods, such as FastSCNN with its lightweight model and U-Net with its precise localization capabilities, compared with modern approaches like the Segment Anything Model (SAM) and its lightweight alternative, FastSAM. By implementing these varied models on the common benchmarking Cityscapes dataset, we dissect their strengths and weaknesses through various metrics. The study extends to adjusting and optimizing these models' parameters to enhance their performance. Furthermore, the research explores the integration of prompt-guided methodologies into conventional segmentation frameworks to elevate their adaptability and utility more robustly to unseen data. The future objective is to fuse the precision of traditional methods with the versatility of prompt-based techniques to forge models that are not only accurate but also proficient in handling unseen data scenarios. |
---|