TY - JOUR
T1 - RGB-D Salient Object Detection Using Saliency and Edge Reverse Attention
AU - Ikeda, Tomoki
AU - Ikehara, Masaaki
N1 - Publisher Copyright:
© 2013 IEEE.
PY - 2023
Y1 - 2023
N2 - RGB-D salient object detection is a task to detect visually significant objects in an image using RGB and depth images. Although many useful CNN-based methods have been proposed in the past, there are some problems such as blurring of object boundaries and inability to detect important parts of objects, leading to a decrease in detection accuracy. In this paper, we propose RGB-D salient object detection using Saliency and Edge Reverse Attention(SERA), which combines the fusion of saliency and edge features with reverse attention. The reverse attention process has the effect of making it easier to capture the boundaries of objects and undetected objects by inverting the up-sampled saliency features and weighting other saliency features, and the edge reverse attention process has the effect of making salient object regions stand out by inverting the edge features and weighting saliency features. The interaction between the salient object features and the edge features enhances each other's features and refines the information on the boundary of objects and salient object regions. In addition, to make it easier to refer to the global information of an image, we introduced the Multi-Scale Interactive Module(MSIM), which is capable of acquiring information at rich scales by converting feature maps to different resolutions and interacting with them. In addition to the salient object output, supervised learning is applied to multiple edge outputs of each resolution to improve the accuracy of both salient objects and boundary areas. Experimental results on five benchmarks show that the proposed method quantitatively performs better than the conventional method, and qualitatively improves the sharpness of object boundaries and the accuracy of detecting important parts of objects.
AB - RGB-D salient object detection is a task to detect visually significant objects in an image using RGB and depth images. Although many useful CNN-based methods have been proposed in the past, there are some problems such as blurring of object boundaries and inability to detect important parts of objects, leading to a decrease in detection accuracy. In this paper, we propose RGB-D salient object detection using Saliency and Edge Reverse Attention(SERA), which combines the fusion of saliency and edge features with reverse attention. The reverse attention process has the effect of making it easier to capture the boundaries of objects and undetected objects by inverting the up-sampled saliency features and weighting other saliency features, and the edge reverse attention process has the effect of making salient object regions stand out by inverting the edge features and weighting saliency features. The interaction between the salient object features and the edge features enhances each other's features and refines the information on the boundary of objects and salient object regions. In addition, to make it easier to refer to the global information of an image, we introduced the Multi-Scale Interactive Module(MSIM), which is capable of acquiring information at rich scales by converting feature maps to different resolutions and interacting with them. In addition to the salient object output, supervised learning is applied to multiple edge outputs of each resolution to improve the accuracy of both salient objects and boundary areas. Experimental results on five benchmarks show that the proposed method quantitatively performs better than the conventional method, and qualitatively improves the sharpness of object boundaries and the accuracy of detecting important parts of objects.
KW - Deep learning
KW - RGB-D salient object detection
KW - reverse attention
UR - http://www.scopus.com/inward/record.url?scp=85164428126&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85164428126&partnerID=8YFLogxK
U2 - 10.1109/ACCESS.2023.3292880
DO - 10.1109/ACCESS.2023.3292880
M3 - Article
AN - SCOPUS:85164428126
SN - 2169-3536
VL - 11
SP - 68818
EP - 68825
JO - IEEE Access
JF - IEEE Access
ER -