研究目的
To optimize CNN training for informative context using depth data to enhance semantic segmentation accuracy of RGB-D images.
研究成果
The SCN, guided by super-pixels and depth data, outperforms state-of-the-art methods by about 2% on NYUDv2 and SUN-RGBD datasets, demonstrating effective context representation for semantic segmentation.
研究不足
Relies on low-level image features for super-pixel generation, which may not separate objects accurately in low-contrast images. Future work could make super-pixel generation sensitive to semantic segmentation results.
1:Experimental Design and Method Selection:
The SCN uses super-pixels and depth data to guide context representation construction, with local structural and top-down switchable information propagation.
2:Sample Selection and Data Sources:
Two public datasets, NYUDv2 and SUN-RGBD, are used for training and testing.
3:List of Experimental Equipment and Materials:
RGB-D images, super-pixels generated using a toolkit, HHA images derived from depth data.
4:Experimental Procedures and Operational Workflow:
Input RGB and depth images, generate super-pixels and HHA images, train SCN with SGD, perform multiscale testing, and apply dense CRF for post-processing.
5:Data Analysis Methods:
Mean intersection-over-union (IoU) is used to evaluate segmentation accuracy.
独家科研数据包,助您复现前沿成果,加速创新突破
获取完整内容