jimaging-07-00187-v2
jimaging-07-00187-v2
Imaging
Article
Detecting Salient Image Objects Using Color Histogram
Clustering for Region Granularity
Seena Joseph and Oludayo O. Olugbara *
Department of Information Technology, Durban University of Technology, Durban 4000, South Africa;
seenaj@dut.ac.za
* Correspondence: oludayoo@dut.ac.za
Abstract: Salient object detection represents a novel preprocessing stage of many practical image
applications in the discipline of computer vision. Saliency detection is generally a complex process to
copycat the human vision system in the processing of color images. It is a convoluted process because
of the existence of countless properties inherent in color images that can hamper performance. Due
to diversified color image properties, a method that is appropriate for one category of images may
not necessarily be suitable for others. The selection of image abstraction is a decisive preprocessing
step in saliency computation and region-based image abstraction has become popular because of
its computational efficiency and robustness. However, the performances of the existing region-
based salient object detection methods are extremely hooked on the selection of an optimal region
granularity. The incorrect selection of region granularity is potentially prone to under- or over-
segmentation of color images, which can lead to a non-uniform highlighting of salient objects. In
this study, the method of color histogram clustering was utilized to automatically determine suitable
homogenous regions in an image. Region saliency score was computed as a function of color contrast,
contrast ratio, spatial feature, and center prior. Morphological operations were ultimately performed
to eliminate the undesirable artifacts that may be present at the saliency detection stage. Thus, we
Citation: Joseph, S.; Olugbara, O.O.
have introduced a novel, simple, robust, and computationally efficient color histogram clustering
Detecting Salient Image Objects
method that agglutinates color contrast, contrast ratio, spatial feature, and center prior for detecting
Using Color Histogram Clustering for
salient objects in color images. Experimental validation with different categories of images selected
Region Granularity. J. Imaging 2021, 7,
187. https://doi.org/10.3390/
from eight benchmarked corpora has indicated that the proposed method outperforms 30 bottom-up
jimaging7090187 non-deep learning and seven top-down deep learning salient object detection methods based on the
standard performance metrics.
Academic Editor: Edoardo Provenzi
Keywords: color contrast; contrast ratio; histogram clustering; region saliency; saliency detection
Received: 1 August 2021
Accepted: 13 September 2021
Published: 16 September 2021
1. Introduction
Publisher’s Note: MDPI stays neutral
Salient object detection is an arduous open research problem aimed at retrieving the
with regard to jurisdictional claims in
most conspicuous visually distinct foreground information from an image in a manner
published maps and institutional affil-
reminiscent of the human vision system [1–8]. It is a challenging task because human
iations.
vision is difficult to mimic by automated systems. Salient object detection methods attempt
to extract points and regions of a visual scene that are more significant to human visual
attention by forming a map that defines how a region stands out from its background and
analyzing image surroundings [1,8,9]. Saliency detection is extensively used to mitigate
Copyright: © 2021 by the authors. the complexity of image analysis and speed up the processing time, and it has gained
Licensee MDPI, Basel, Switzerland.
popular applications in the disciplines of computer vision and artificial intelligence [8,10].
This article is an open access article
The numerous application domains of saliency include image segmentation [11–14], object
distributed under the terms and
detection and recognition [15–17], anomaly detection [18,19], image retrieval [20,21], image
conditions of the Creative Commons
compression [22], object classification [23], object tracking [24], image retargeting, and
Attribution (CC BY) license (https://
summarization [25,26], alpha matting [26], target detection [27], video object segmenta-
creativecommons.org/licenses/by/
4.0/).
tion [28], video summarization [29], user perceptions of digital video contents [30], and
visual tracking [31]. Countless applications of saliency detection have led to the occurrence
of numerous methods for saliency computation. The orthodox saliency detection methods
can be classified into two approaches, top-down and bottom-up, based on the perspective
of information processing [9,32–35]. The top-down approach is task-driven with seman-
tic information, prior knowledge and it focuses on supervised machine learning from a
plethora of training images [8,32,36]. The approach has had great success in salient object
detection with the progress of deep learning methods [8,37–41]. Deep saliency detection
methods are often trained with a large set of finely annotated pixel-level ground truth im-
ages [42–44]. However, the performance of deep learning methods is highly dependent on
the construction of well-annotated training datasets and can be adversely affected [43,45].
The bottom-up approach is data-driven without semantic information but grounded
in the connotation of primitive features such as color, intensity, shape, and texture that are
simple to implement [32,46,47]. The bottom-up methods compute uniqueness in primitive
features of image pixels and surrounding regions. These saliency detection methods have
extensively used different visual rarities to separate foreground and background regions in
images. The visual rarities include color prior [3,48,49], contrast prior [32,50], brightness
prior [11,51], background prior [33,52], boundary prior [4,53], center prior [13,54], shape
prior [55], context prior [25], object position prior [56], and connectivity prior [7,44,53,57].
However, despite the development of several methods for salient object detection, there are
still intrinsic challenges with different categories of images. The presence of cluttered and
non-homogeneous background regions, inter-object dissimilarity, heterogeneous objects
with varying sizes, counts, and positions have led to ambiguous and diverse challenges.
Examples of image categories are salient objects with erratic sizes, positions, and counts,
cluttered backgrounds, and low dissimilarity among regions of heterogeneous foreground
or heterogeneous background. The task of completely highlighting salient objects in
different image categories is still not adequately resolved in most of the existing saliency
methods [58–60]. The other major challenge is the mitigation of computational complexity
because salient object detection is an essential preprocessing stage in computer vision.
This study addresses the problem of automatic selection of optimum homogenous
regions for image abstraction to reduce the computational complexity, improve the ef-
fectiveness, and increase the efficiency of salient objects detection for different classes of
images. The method of color histogram-based clustering has been developed in this current
study for this purpose. A near resolution of detecting salient objects in different images has
been achieved by successfully integrating holistic strategy of color contrast, contrast ratio,
center prior, and regional spatial feature while adhering rigidly to the efficacy requirement
of salient object detection. The idiosyncratic contributions of this study to the existing
research in computer vision are threefold:
• The comprehensive review of related literature on salient object detection methods
and approaches to demonstrate trends, uniqueness, recency, and relevance of the
current study.
• The construction of a novel bottom-up saliency computation method that exploits the
strategy of color contrast, contrast ratio, center prior, and spatial feature to obtain a
robust salient object detection process.
• The intensive experimental comparison with different prominent salient object detec-
tion methods that were reported in the literature to determine the effectiveness of the
proposed method.
The remainder of this paper is succinctly structured as follows. Section 2 gives a
comprehensive review of the related literature. Section 3 describes the proposed salient
object detection method. Section 4 explicates the intensive experimental comparison of
the proposed method against the existing modern methods based on the widely known
benchmarked corpora and performance evaluation metrics. Section 5 provides a discussion
of experimental results and a brief concluding remark.
J. Imaging 2021, 7, 187 3 of 36
2. Review of Literature
A plethora of color image saliency detection methods have been reported in the
literature, strikingly developed in the last two decades. The bottom-up method by Itti
et al. [61] is considered a cornerstone strategy grounded in the biological model for eye-
fixation activities of humans. The method is based on center-surroundedness differences
in color, intensity, and orientation that can detect spatial discontinuities in a scene. It
estimates the locations of visual gaze by computing multi-scale feature maps using a
Gaussian pyramid. The second category of saliency detection methods has emerged from
the works [15,62] where saliency was defined as a binary segmentation problem. The
third wave of saliency has emerged with the introduction of the convolutional neural
network (CNN) to lessen the reliance on center bias knowledge. Neurons in the CNN
model with large receptive fields of global information can enhance the detection of the
most salient region in an image [63]. A plethora of salient region detection methods have
been developed among which bottom-up methods are pervasive because of their simplicity,
elegance, and computational efficiency.
ranking to lessen the constraint of local contrast that highlights the boundaries of objects
rather than the entire region. However, consideration of local relevance among neighbor-
ing regions can lead to incorrect suppression of salient regions, especially in images with
heterogeneous salient object features [76]. A local contrast-based method for detecting
small targets by computing contrast between the targeted small regions and surrounding
regions was proposed [77]. Due to the limited spatial neighborhood consideration in the
local contrast method, large salient regions can be easily excluded [42].
accuracy on saliency detection [3]. Color contrast prior is not sufficient to successively
detect salient objects from images with low color contrast between foreground or back-
ground and complex background or foreground scenes. This restraint emphasizes that even
though a significant improvement has been witnessed, salient object detection remains
a challenging issue because of image diversity, inherent complexity, and uncertainty of
salient regions [32,81,82].
and coarsest abstraction of regions to resolve the granularity problem of superpixels [56].
However, the iteration process increases the computational complexity that can adversely
affect the performance of the saliency detection process in real-time applications.
In summary, emphasizing high-contrast edges while suppressing the interior of salient
regions is a major obstacle of the pixel or patch-based methods. Region-based image
abstractions are considered superior to pixel- or patch-based methods because they can
employ a richer feature representation for saliency detection. Superpixel-based methods
have gained popularity in recent years because of their computational efficiency. However,
finding an optimum superpixels granularity is a challenging task for the superpixel-based
image abstraction process. This is because the efficiency and robustness hallmarks of
saliency detection methods are highly dependent on the granularity of superpixels. This
empathizes the significance of constructing an efficient method that can automatically
detect the number of regions for image abstraction. The proposed color histogram-based
image abstraction can automatically detect the appropriate image region granularity based
on the color distribution of an image as explicated in the subsequent section.
3. Methods
The novel regional color histogram clustering method is introduced in this study for
detecting salient objects in red, green, and blue (RGB) images. The quantized RGB (QRGB)
color image is the input to the histogram-based clustering process to reduce the number of
colors in the input image. The numerous color models used in saliency detection methods
include RGB [67,70,106], hue, saturation, value (HSV) [107], lightness, redness, yellowness
(L*a*b*) [66,68,84,108,109] and combination of color models [67,85,106]. This study has
used the QRGB color image for clustering while the L*a*b* color image was applied for
the extraction of color features because of its perceptual uniformity [50,66,110]. Literature
has shown that color quantization in the RGB color model relatively performed better than
quantization in the L*a*b* color model [111]. The purpose of transforming the original RGB
color image into the L*a*b* color image instead of the QRGB image was to minimize the
effects of quantization error. Consequently, the L*a*b* color model was selected for color
feature extraction in the range of [0, 1] to suppress the effect of any possible dominant colors
and to take the intrinsic advantages of perceptual uniformity of the color model [108,112].
The proposed method exploits the strategy of color contrast, contrast ratio, spatial feature,
and center prior to efficiently compute pixel-level saliency scores. The method is comprised
of three essential steps of input image segmentation into regions, calculation of region
saliency scores, and post-processing of the computed saliency map. The outline of the
proposed method for salient objects detection is depicted in Figure 1.
Figure 1. Flowchart
Figure 1. Flowchart of
of the
the color
color histogram
histogram clustering
clusteringmethod
methodfor
forsalient
salientobjects
objectsdetection.
detection.
The true-color image contains a maximum possibility of 256 3 = 16 ,777 ,216 colors
where wr = 8, w g = 64, and wb = 1 are the weights of R, G, and B colors, respectively. The
that is generally greater than the number of pixels in an image [118,119]. Since extremely
green channel was assigned the highest weight value because the human visual system is
rare colors are not significant for highlighting salient regions, less dominant colors can be
highly sensitive to the green color than other colors [122].
excluded for saliency detection [66]. Color quantization is a widely used technique for
merging less dominant colors into dominant colors to significantly reduce the computa-
tional complexity of image processing [119,120]. The minimum variance method [66] or
J. Imaging 2021, 7, 187 10 of 36
where ( L, a, b ) is the color value of the region in L*a*b* color model, k·k2 indicates the L2
norm, and M is the number of regions automatically detected. There will be a maximum
of 8 colors in an image, assuming each image channel has 2 distinct intensity levels. The
number, M of the possible colors or regions in a quantized color image, will lie in the range
of [8, 512]. The regional weight function W = (W1 , . . . , WM ) is integrated into the region
saliency calculation process. The weight function will account for the contribution of high
saliency by larger regions than for the smaller ones. The weight of a region is calculated
based on the relative probability of the pixels in the region to emphasize the color contrast
of larger regions [50,66], as defined by Equation (3).
fi
Wi = (3)
f
where f i is the frequency of the pixels occupied in each region ri and f is the total number
of pixels in the input image. The spatial contrast function SC(ri ) integrates the global color
contrast with the spatial feature and color ratio of a region (ri ) as follows:
M
∑ Wj φ
SC(ri ) = Wi CC (ri ) + ri , r j ) exp( − DS(ri , r j ) (4)
j =1
In a divergence from the work [66] that utilized regional saliency differences as a
weighting coefficient to suppress the effect of non-salient regions, our method utilizes a
more resilient function, φ(ri , r j ) based on the contrast ratio given by Equation (5). The
contrast ratio is an important aspect of image quality that measures the difference between
the maximum and minimum brightness of an image. In the context of this study, it measures
the difference between the maximum and minimum brightness of regions in an image.
!
CC (ri ) + 0.05
φ (ri , r j ) = (5)
CC (r j ) + 0.05
The significance of center prior in saliency detection as given by Equation (6) has
been highlighted in literature following the fundamental assumption that salient objects
are framed near the image center while background pixels are distributed at the image
borders [36,59,66,68,102]. It is usually formulated and extensively used in literature as
a Gaussian distribution [3,36,66,106,113]. The region saliency score CS(ri ) is obtained in
J. Imaging 2021, 7, 187 11 of 36
terms of spatially weighted color contrast and the Euclidean distance between the region
spatial center and image screen center. This is to integrate the center prior with color
contrast, contrast ratio, and spatial feature using Equation (6).
where S(ri ) is the spatial center of a region and C = (0.5, 0.5) is the image screen center.
Since salient objects are always not positioned at the image center, the concept of center
prior can lead to the exclusion of salient objects located at the image boundary or inclusion
of a background region [66,106,124]. This can occur, especially when an object possesses
multiple colors such that object colors at the image center are different from those in the
background. The parameter α ∈ [0.1, 1.0] is incorporated into the region saliency score
function to strengthen the center prior. Even though the function can compute a low
saliency score for a region around the boundary, an appropriate α value can make adequate
salient objects more salient, regardless of their positions. In addition to the color contrast
features, spatial features play a significant role in human attention, and the use of spatial
coherence in saliency computation is widely accepted by many researchers [36,50,66–68,97].
The spatial distance DS(ri , r j ) between two regions is computed using Equation (7).
y y
DS(ri , r j ) = k Cix , Ci − Cjx , Cj k (7)
2
y
where Cix , Ci is the spatial center of a region ri that is computed by averaging the x and
y coordinates of pixels in the region. The regional saliency score generated is normalized to
the range of [0, 1] before assigning the pixel level saliency. The saliency score of each pixel
is assigned by the saliency score of the respective region to obtain the saliency map, C Map
as shown in Equation (8). The assignment is based on the assertation that pixels belonging
to the same region have the same saliency.
4. Experimental Results
This study has applied the properties of salient objects to categorize various images
into different groups to provide a more comprehensive experimental evaluation of the
proposed saliency detection method. Figure 2 shows these properties to be the location
of salient objects (center or boundary), object sizes (salient objects that overlap center and
boundary regions), number of salient objects (multiple objects), color contrast (low contrast),
and complex background. The performance of the proposed method was validated against
30 modern bottom-up and seven deep-learning-based top-down methods. Since we do not
have access to source codes of the deep learning, and five bottom-up saliency methods,
they were considered for the extended complex scene saliency dataset (ECSSD). The rest of
the methods were included for comparison on six categories of images. The only parameter
that was used in the proposed method is the central bias weight, α, selected experimentally
as α ∈ [0.1, 1.0].
Figure 2. Category of images: (a) boundary; (b) center; (c) complex background; (d) low contrast; (e) overlap; (f) multiple
objects; (g) ECSSD.
4.1. Datasets
Experimental images were selected from different benchmarked datasets of MSRA10K [62],
ASD [48], SED2 [129], ImgSal [130], DUT OMRON [83], ECSSD [131], HKU IS [132], and
SOC [133]. These datasets have been extensively used for evaluating salient object detection
methods [5,38,56,63,66,68,79,80,106,133]. The MSRA10K is a descendent of the Microsoft
Research Asia (MSRA) dataset, where many images in this dataset are often with a single
salient object and simple background [26,57]. The ASD is a subset of the MSRA dataset
with ground truth region annotation, single foreground, and simple background [35].
These two datasets are mainly used for selecting center located, boundary located, fore-
ground or background overlapped, and low contrast images. The SED2, ImgSal, and
DUT-OMRON datasets are known for multiple salient objects with relatively complex
backgrounds [57,102,112,134]. The images with multiple salient objects and complex back-
grounds were selected from these three datasets. The salient objects of the SED2 dataset
exhibit different color, position, and size properties.
In addition, images from the ECSSD were selected for evaluation [131]. The EC-
SSD dataset includes 1000 images that contain salient objects with colors that are af-
fected by background regions, and salient objects with heterogeneous colors, sizes, and
location properties to present huge ambiguity for the methods of salient object detec-
tion. The dataset is fundamentally considered to be complex for performance compar-
isons [5,57,67,68,84,135,136]. The HKU-IS dataset has 4447 complex scenes with multiple
disconnected objects that are highly similar to the background regions with a diverse spatial
distribution [43,132]. Salient objects in clutter (SOC) is a recently introduced dataset [133]
and is a subset of the common objects in contexts (COCO) dataset [137]. SOC is a chal-
lenging dataset of salient objects with attributes reflecting occlusion, cluttered background,
and challenges in real-world scenes developed for evaluating CNN-based salient object
detection methods. The proposed method is not featured for detecting salient objects from
J. Imaging 2021, 7, 187 13 of 36
occluded or cluttered backgrounds and is not based on the CNN approach. It was tested
against 1500 images of the SOC dataset to determine its ability to detect salient objects in
real-world scenes.
In this study, we run the source codes of the methods of AC, BGFG, CNS, DCLC,
DGL, DRFI, GB, GMR, HDCT, IT, MAP, MR, and RPC with their default parameters. The
implementations of salient object detection methods in [63] with default parameters were
employed to obtain the saliency maps of CA, COV, DSR, FES, GR, MC, SEG, SeR, SR, SUN,
and SWD. Since we have no access to the source codes of the remaining methods, they were
J. Imaging 2021, 7, 187 14 of 36
excluded for qualitative comparison, analysis of computational time complexity, and could
not compare with all the selected image categories. The method of FCB was considered for
the category of overlap images and ECSSD dataset based on the saliency results provided
by their authors.
Figure 3. Cont.
J. Imaging 2021, 7, 187 16 of 36
J. Imaging 2021, 7, x 18 of 37
MC
MR
RPC
SEG
SeR
SIM
SR
SUN
SWD
(d) Complex Background (e) Low Contrast (f) Multiple Objects (g) Overlap
1 2 3 1 2 3 1 2 3 1 2 3
Image
GT
OURS
AC
BGFG
CA
CNS
COV
DCL
DGL
Figure 3. Cont.
J. Imaging 2021, 7, 187 17 of 36
J. Imaging 2021, 7, x 19 of 37
DRFI
DSR
Not Not
FCB
available available
FES
GB
GR
HDC
IT
MAP
MC
MR
RPC
SEG
SeR
SIM
Figure3.3.Qualitative
Figure Qualitativeperformance
performanceofofthe
theinvestigated
investigatedmethods
methodsononECSSD
ECSSD and
and selected
selected categories
categories ofof images:
images: (a)(a) ECSSD;
ECSSD; (b)
(b) Boundary; (c) Center; (d) Complex Background; (e) Low Contrast; (f) Multiple Objects; (g) Overlap.
Boundary; (c) Center; (d) Complex Background; (e) Low Contrast; (f) Multiple Objects; (g) Overlap.
Figure 4 shows
Methods the qualitative
that exploited results
center prior of the proposed
performed methodwith
well on images in comparison with
centrally located
the top-performing methods on the challenging HKU-IS and SOC datasets.
salient objects, but they showed challenges in some cases such as Figure 3(c2), where The HKU-IS
is well-known
two for multiple
objects (bowl and disconnected
and strawberry) salient
are centrally objectsHence,
located. that show
thesehigh similarity
methods tendtoto
the background regions. The SOC dataset contains images that are closer
concomitantly detect both objects as a salient region. However, methods that incorporate to real-world
conditions.
color The
contrast qualitative
such as RPC,results
DCLC,are shown
GR, CNS,inand
Figure
HDCT 4 highlight the performances
have managed to highlight of the
real
proposed
salient method
objects. and sixsuch
Methods other
as methods
MC, MR,that MAP,generally
and DGL perform well for the
that exploited category
boundary of
prior
multiple objects. The proposed method shows good results on these two
have failed to detect salient objects in this category of images because they consideredchallenging da-
tasetsboundary
black with the output
regionsalmost resembling
as background the ground
regions truth. In Figure
and incorrectly 4b, forthe
highlighted instance, the
white bowl
proposed method highlighted the salient object as in the ground truth image,
as a salient object. The performance of the proposed method for this category of images while other
methods
is detected all objects
highly commendable becauseonitthe
hastable. Similarly,strength
demonstrated the performance
in detectingof the proposed
salient objects,
method is commendable, regardless of the complexity of these images (Figure
irrespective of variation in sizes as in Figure 3(c1) (small salient object) and Figure 3(c3) 4a,c,d).
J. Imaging 2021, 7, 187 18 of 36
(large salient object). The salient object in Figure 3(c3) also shows heterogeneous properties
in terms of color and appearance; hence, many modern methods such as DCLC, GR, MAP,
MC, MR, and RPC have failed to uniformly highlight salient regions. These methods have
managed to highlight only a portion of salient regions rather than the entire salient regions.
In contrast, the proposed method shows impressive results that are almost like the ground
truth images. The visual results of the existing methods in a complex background are
shown in Figure 3d. The results show the power of the proposed method in detecting
salient objects from complex and heterogeneous backgrounds while all other methods
show lower performance. The performance of DRFI is comparatively better than the rest of
the existing methods because of the inclusion of color and texture features along with the
use of multi-level pre-segmentation maps to detect multi-invariant objects.
Salient objects with low contrast to the background are considered a challenging case
for contrast-based and graph-based methods. The visual representation of this image
category is demonstrated in Figure 3e; it is worth noticing that the performance of all
methods except ours is not remarkable. The DGL method proposed a deformed smoothness
constraint to overcome this challenge of graph-based methods. However, DGL still had
failure cases as in Figure 3(e3) that it cannot effectively handle low contrast objects. The
result in Figure 3(e3) shows that the performance of DRFI is not free from the limitation
of contrast-based cues because of the use of feature extraction by contrast vectors. The
performance of DSR is relatively better than the rest of the methods; nevertheless, the results
are not free from background noise as in Figure 3(e3). The regional contrast-based method
of RPC based on low-level color contrast features also demonstrated poor performance on
low contrast objects. The proposed method has illustrated good results as compared to the
listed modern methods. The ability to uniformly highlight salient objects in the category
of multiple objects is still challenging for many of the modern methods because of the
heterogeneous nature of objects as illustrated in Figure 3f.
The results of Figure 3(f1) have illustrated that many methods such as CNS, DGL,
DSR, MAP, MC, and MR can detect only one object. The proposed method has again
demonstrated its ability in detecting heterogeneous objects from this class of images.
Except for the proposed method, only DRFI shows relatively better results for this category
of images. The images that belong to the overlapped category are generally larger and they
touch the image boundary and image center as shown in Figure 3g. The proposed method
shows an outstanding performance for images in this category like the previous categories.
Moreover, the graph-based methods or diffusion-based methods such as DGL, MR, MC,
and MAP have achieved good performance on the category of overlapped images. In
opposition to the performance of DSR for the category of low contrast images, DSR has
demonstrated poor performance on overlapped images because the method has incorrectly
assigned all image boundaries as a background template. The methods such as COV, FES,
IT, GB, SeR, SUN, SWD, and SIM, as illustrated in Figure 3, generally showed challenges in
highlighting salient objects from all the listed categories of images.
The ECSSD dataset is generally well-known for salient objects with heterogeneous
properties and occluded backgrounds. The proposed method has again demonstrated
remarkable results on images from this dataset. The learning-based methods such as HDCT
and DRFI have shown better performance on images in this dataset. The results indicate
the merits of the proposed method on a wide spectrum of image categories and obviously,
its output is more reliable with results that are almost like the ground truth in comparison
to the existing modern methods.
Figure 4 shows the qualitative results of the proposed method in comparison with
the top-performing methods on the challenging HKU-IS and SOC datasets. The HKU-IS
is well-known for multiple and disconnected salient objects that show high similarity to
the background regions. The SOC dataset contains images that are closer to real-world
conditions. The qualitative results are shown in Figure 4 highlight the performances of
the proposed method and six other methods that generally perform well for the category
of multiple objects. The proposed method shows good results on these two challenging
J. Imaging 2021, 7, 187 19 of 36
datasets with the output almost resembling the ground truth. In Figure 4b, for instance,
the proposed method highlighted the salient object as in the ground truth image, while
J. Imaging 2021, 7, x 20 of 37
other methods detected all objects on the table. Similarly, the performance of the proposed
method is commendable, regardless of the complexity of these images (Figure 4a,c,d).
a
SOC
c
HKU-IS
Figure4.4.Qualitative
Figure Qualitativeperformance
performance of of the
the proposed
proposed method,
method, CNS,
CNS, DCLC,
DCLC, DGL,
DGL,DRFI,
DRFI,DSR
DSRand,
and,GR
GRon
onSOC
SOCand
andHKU-IS
HKU-IS
datasets: (a) Salient object with the heterogeneous background; (b) Salient object surrounded by multiple non-salient ob-
datasets: (a) Salient object with the heterogeneous background; (b) Salient object surrounded by multiple non-salient objects;
jects; (c) Salient objects with illumination change; (d) Multiple salient objects.
(c) Salient objects with illumination change; (d) Multiple salient objects.
4.5. Quantitative
4.5. Quantitative Results
Results
The quantitative comparison
The quantitative comparison of of the
the proposed
proposedmethod
methodagainst
againstother
othermethods
methodsininterms
terms
of the metrics of precision, recall, F-measure, MAE, and OR are revealedininTable
of the metrics of precision, recall, F-measure, MAE, and OR are revealed Table2 2toto
objectively reinforce
objectively reinforce the
the performance
performance of of the
the proposed
proposedmethod
methodonondiverse
diversecategories
categoriesofof
images.
images.
Table 2. The performance statistics for six Objects
4.5.1. Salient categories of images.
Located arrow indicates that a higher value gives better
The up Boundary
at Image
performance, and the down arrow shows that a lower value gives better performance.
Tables 2 and 3 show comprehensive results of the investigated methods based on
(a) Metric OURs AC BGFG
the CA performance
standard CNS COV DCLC
metrics. The DGL show
results DRFIthat DSR
our method FES scored GB the highest
GR
Preci-
sion
0.945 0.698 precision
0.807 (0.945),
0.621 F-measure
0.800 (0.932),0.928
0.580 and OR0.909
(0.844) 0.867
with a slightly
0.846 lower
0.765 recall as compared
0.578 0.924
MAE 0.062 0.133 0.111 0.140 0.071 0.130 0.057 0.073 0.080 0.057 0.110 0.152 0.095
OR 0.844 0.541 The GR used
0.659 0.545a convex
0.700 hull to estimate
0.387 0.832 salient
0.794 objects
0.808 and0.747
centroid of the0.480
0.551 convex 0.837
hull as
Metric HDCT center
IT prior
MAPinstead MCof image
MR centerRPCto favor
SEGthe detection
SeR of
SIM salient objects
SR located
SUN farther
SWD
Preci-
0.878
from
0.532
the image
0.804
center.
0.884
The0.900
DCLC 0.860
ranked saliency
0.873
based on0.562
0.531
foreground
0.536
seeds
0.58
obtained
0.627
by
sion local contrast and performed well in this category unlike other diffusion-based methods
Recall 0.927 0.705 0.799 0.808 0.841 0.797 0.624 0.787 0.529 0.701 0.578 0.648
F-meas-
such as MC and MR, which considered the nodes that touch the image boundaries as
0.888 0.564
background 0.803
seeds.0.865
The SeR0.886
achieved0.844 0.799 precision
the lowest 0.574 of0.554 0.567for all
0.532 and 0.58 0.632
other metrics,
ure
MAE 0.077 SIM showed
0.179 0.092the lowest
0.104 performance.
0.070 0.097 Regardless
0.274 of the use
0.191 0.325of a center
0.154 prior,
0.233appropriate
0.219
OR 0.816 0.417 0.691 0.741 0.785 0.698 0.588 0.452
selection of α value has enabled the proposed method to0.337
produce 0.425
a robust0.408 0.446
detection of
Metric OURs AC BGFG CA CNS COV DCLC DGL DRFI DSR FES GB GR
Preci-
salient objects located far off the image center. Figure 5 demonstrates the average precision,
0.949 0.633 0.854 F-measure,
recall, 0.606 0.819 and
MAE, 0.733
OR on0.910 0.906 of boundary
the category 0.862 0.856
images0.772 0.664investigated
for all the 0.913
sion
Recall 0.889 0.543 methods.
0.850 0.655 0.903 0.699 0.915 0.909 0.933 0.888 0.734 0.765 0.866
F-meas-
0.934 0.610 0.853 0.617 0.837 0.725 0.911 0.906 0.877 0.863 0.763 0.685 0.901
ure
Center (370) 1
MAE 0.067 0.184 0.112 0.204 0.058 0.147 0.063 0.063 0.075 0.062 0.136 0.183 0.122
OR 0.846 0.420 0.727 0.435 0.768 0.520 0.838 0.830 0.804 0.762 0.589 0.511 0.801
Metric HDCT IT MAP MC MR RPC SEG SeR SIM SR SUN SWD
Preci-
0.859 0.549 0.874 0.896 0.908 0.839 0.808 0.505 0.474 0.507 0.500 0.742
sion
Recall 0.925 0.618 0.899 0.906 0.892 0.795 0.568 0.559 0.259 0.521 0.336 0.649
F-meas-
0.873 0.564 0.88 0.898 0.904 0.828 0.736 0.516 0.398 0.510 0.450 0.719
ure
MAE 0.091 0.218 0.063 0.079 0.061 0.109 0.279 0.273 0.381 0.214 0.319 0.230
OR 0.794 0.386 0.796 0.82 0.818 0.686 0.527 0.344 0.171 0.330 0.245 0.475
J. Imaging 2021, 7, 187 20 of 36
Table 2. The performance statistics for six categories of images. The up arrow ↑ indicates that a higher value gives better
performance, and the down arrow ↓ shows that a lower value gives better performance.
(a) Metric OURs AC BGFG CA CNS COV DCLC DGL DRFI DSR FES GB GR
Precision ↑ 0.945 0.698 0.807 0.621 0.800 0.580 0.928 0.909 0.867 0.846 0.765 0.578 0.924
Recall ↑ 0.891 0.692 0.782 0.843 0.825 0.561 0.887 0.859 0.932 0.868 0.677 0.774 0.898
F-measure ↑ 0.932 0.697 0.801 0.661 0.805 0.576 0.918 0.897 0.882 0.851 0.743 0.614 0.918
MAE ↓ 0.062 0.133 0.111 0.140 0.071 0.130 0.057 0.073 0.080 0.057 0.110 0.152 0.095
OR ↑ 0.844 0.541 0.659 0.545 0.700 0.387 0.832 0.794 0.808 0.747 0.551 0.480 0.837
Boundary (350) 1
Metric HDCT IT MAP MC MR RPC SEG SeR SIM SR SUN SWD
Precision ↑ 0.878 0.532 0.804 0.884 0.900 0.860 0.873 0.531 0.562 0.536 0.58 0.627
Recall ↑ 0.927 0.705 0.799 0.808 0.841 0.797 0.624 0.787 0.529 0.701 0.578 0.648
F-measure ↑ 0.888 0.564 0.803 0.865 0.886 0.844 0.799 0.574 0.554 0.567 0.58 0.632
MAE ↓ 0.077 0.179 0.092 0.104 0.070 0.097 0.274 0.191 0.325 0.154 0.233 0.219
OR ↑ 0.816 0.417 0.691 0.741 0.785 0.698 0.588 0.452 0.337 0.425 0.408 0.446
Metric OURs AC BGFG CA CNS COV DCLC DGL DRFI DSR FES GB GR
Precision ↑ 0.949 0.633 0.854 0.606 0.819 0.733 0.910 0.906 0.862 0.856 0.772 0.664 0.913
Recall ↑ 0.889 0.543 0.850 0.655 0.903 0.699 0.915 0.909 0.933 0.888 0.734 0.765 0.866
F-measure ↑ 0.934 0.610 0.853 0.617 0.837 0.725 0.911 0.906 0.877 0.863 0.763 0.685 0.901
MAE ↓ 0.067 0.184 0.112 0.204 0.058 0.147 0.063 0.063 0.075 0.062 0.136 0.183 0.122
Center (370) 1 OR ↑ 0.846 0.420 0.727 0.435 0.768 0.520 0.838 0.830 0.804 0.762 0.589 0.511 0.801
Metric HDCT IT MAP MC MR RPC SEG SeR SIM SR SUN SWD
Precision ↑ 0.859 0.549 0.874 0.896 0.908 0.839 0.808 0.505 0.474 0.507 0.500 0.742
Recall ↑ 0.925 0.618 0.899 0.906 0.892 0.795 0.568 0.559 0.259 0.521 0.336 0.649
F-measure ↑ 0.873 0.564 0.88 0.898 0.904 0.828 0.736 0.516 0.398 0.510 0.450 0.719
MAE ↓ 0.091 0.218 0.063 0.079 0.061 0.109 0.279 0.273 0.381 0.214 0.319 0.230
OR ↑ 0.794 0.386 0.796 0.82 0.818 0.686 0.527 0.344 0.171 0.330 0.245 0.475
Metric OURs AC BGFG CA CNS COV DCLC DGL DRFI DSR FES GB GR
Precision ↑ 0.933 0.404 0.774 0.550 0.768 0.670 0.847 0.875 0.856 0.827 0.629 0.598 0.762
Recall ↑ 0.753 0.317 0.697 0.405 0.747 0.554 0.793 0.810 0.827 0.774 0.595 0.537 0.531
F-measure ↑ 0.885 0.380 0.755 0.508 0.763 0.639 0.834 0.859 0.849 0.814 0.621 0.583 0.692
Complex background MAE ↓ 0.120 0.253 0.179 0.311 0.130 0.195 0.133 0.135 0.138 0.127 0.200 0.259 0.26
(210) 1 OR ↑ 0.710 0.22 0.568 0.295 0.623 0.418 0.700 0.726 0.721 0.657 0.438 0.384 0.49
Metric HDCT IT MAP MC MR RPC SEG SeR SIM SR SUN SWD
Precision ↑ 0.824 0.482 0.828 0.821 0.819 0.695 0.683 0.334 0.287 0.416 0.381 0.708
Recall ↑ 0.780 0.362 0.803 0.769 0.774 0.601 0.310 0.180 0.055 0.273 0.123 0.376
F-measure ↑ 0.814 0.448 0.822 0.808 0.809 0.671 0.535 0.279 0.146 0.371 0.257 0.588
MAE ↓ 0.160 0.303 0.131 0.164 0.139 0.185 0.341 0.439 0.454 0.318 0.430 0.321
OR ↑ 0.666 0.254 0.694 0.669 0.670 0.485 0.285 0.131 0.045 0.197 0.102 0.310
Metric OURs AC BGFG CA CNS COV DCLC DGL DRFI DSR FES GB GR
Precision ↑ 0.908 0.539 0.787 0.614 0.717 0.710 0.844 0.837 0.843 0.814 0.738 0.672 0.792
Recall ↑ 0.715 0.365 0.653 0.510 0.628 0.545 0.715 0.721 0.753 0.710 0.545 0.600 0.501
F-measure ↑ 0.854 0.486 0.751 0.586 0.694 0.663 0.810 0.807 0.820 0.788 0.682 0.654 0.698
MAE ↓ 0.122 0.227 0.178 0.248 0.155 0.193 0.146 0.159 0.148 0.134 0.187 0.224 0.233
Low contrast (165) 1 OR ↑ 0.659 0.278 0.538 0.38 0.516 0.423 0.625 0.631 0.654 0.599 0.445 0.436 0.457
Metric HDCT IT MAP MC MR RPC SEG SeR SIM SR SUN SWD
Precision ↑ 0.805 0.585 0.804 0.827 0.820 0.730 0.740 0.461 0.499 0.537 0.467 0.731
Recall ↑ 0.685 0.491 0.720 0.710 0.720 0.572 0.249 0.370 0.219 0.441 0.238 0.452
F-measure ↑ 0.774 0.560 0.783 0.797 0.795 0.686 0.508 0.437 0.385 0.511 0.382 0.640
MAE ↓ 0.173 0.252 0.156 0.175 0.153 0.182 0.310 0.340 0.388 0.261 0.371 0.274
OR ↑ 0.578 0.348 0.606 0.613 0.61 0.466 0.233 0.257 0.175 0.316 0.198 0.364
Metric OURs AC BGFG CA CNS COV DCLC DGL DRFI DSR FES GB GR
Precision 0.876 0.640 0.735 0.576 0.752 0.537 0.84 0.834 0.807 0.790 0.633 0.556 0.86
Recall 0.786 0.567 0.696 0.592 0.743 0.535 0.748 0.762 0.818 0.759 0.587 0.644 0.666
F-measure 0.853 0.621 0.726 0.580 0.750 0.537 0.812 0.816 0.810 0.783 0.621 0.574 0.806
MAE ↓ 0.836 0.921 0.888 0.958 0.840 0.911 0.850 0.860 0.842 0.839 0.896 0.955 0.909
Multiple objects (160) 1 OR 0.695 0.425 0.528 0.371 0.582 0.331 0.652 0.656 0.663 0.614 0.410 0.382 0.599
Metric HDCT IT MAP MC MR RPC SEG SeR SIM SR SUN SWD
Precision 0.801 0.536 0.741 0.813 0.820 0.741 0.771 0.427 0.422 0.506 0.442 0.583
Recall 0.791 0.586 0.733 0.741 0.714 0.666 0.381 0.469 0.245 0.537 0.259 0.464
F-measure 0.799 0.547 0.739 0.795 0.793 0.723 0.624 0.436 0.362 0.513 0.380 0.550
MAE ↓ 0.864 0.967 0.866 0.878 0.851 0.883 1.032 1.047 1.124 0.960 1.091 1.021
OR 0.638 0.338 0.574 0.619 0.619 0.521 0.352 0.255 0.141 0.314 0.176 0.295
Metric OURs AC BGFG CA CNS COV DCLC DGL DRFI DSR FCB FES GB
Precision ↑ 0.986 0.703 0.969 0.738 0.881 0.815 0.981 0.969 0.975 0.949 0.968 0.853 0.777
Recall ↑ 0.767 0.344 0.593 0.442 0.638 0.395 0.804 0.767 0.756 0.661 0.615 0.478 0.461
F-measure ↑ 0.925 0.567 0.845 0.639 0.810 0.654 0.934 0.913 0.924 0.862 0.855 0.722 0.671
MAE ↓ 0.134 0.313 0.217 0.280 0.148 0.285 0.130 0.105 0.130 0.157 0.140 0.260 0.274
Overlap (250) 1 OR ↑ 0.757 0.313 0.581 0.381 0.609 0.358 0.790 0.755 0.768 0.641 0.603 0.447 0.402
Metric GR HDCT IT MAP MC MR RPC SEG SeR SIM SR SUN SWD
Precision ↑ 0.96 0.967 0.666 0.96 0.963 0.971 0.956 0.782 0.622 0.515 0.644 0.671 0.872
Recall ↑ 0.658 0.747 0.361 0.712 0.702 0.767 0.568 0.216 0.38 0.104 0.365 0.293 0.370
F-measure ↑ 0.868 0.906 0.557 0.889 0.887 0.915 0.826 0.487 0.542 0.269 0.547 0.517 0.664
MAE ↓ 0.178 0.154 0.304 0.136 0.147 0.114 0.225 0.321 0.311 0.389 0.307 0.326 0.287
OR ↑ 0.65 0.728 0.305 0.697 0.690 0.755 0.556 0.214 0.312 0.092 0.304 0.260 0.342
1 Number of images.
J. Imaging 2021, 7, 187 21 of 36
J. Imaging 2021, 7, x 22 of 37
(a)
(b) (c)
Figure
Figure 5.
5. (a)
(a)F-measure;
F-measure; (b)
(b) MAR
MAR and
and (c)
(c) OR
OR on
on image
image category:
category: Boundary.
Boundary.
(a)
(b) (c)
Figure
Figure 6.
6. (a)
(a) F-measure;
F-measure; (b)
(b) MAR
MAR and
and (c)
(c) OR
OR on
on image category: Center.
image category: Center.
4.5.3. Salient
Salient Objects
Objects with
with Complex
Complex Background
The
The results achieved by the investigated salient object detection methods indicated
that performance
performance is generally
generally challenging
challenging forfor this
this image
image category.
category. However,
However, the proposed
method shows
showsitsitscapability
capabilityforfor
precisely detecting
precisely salient
detecting objects,
salient and it and
objects, is theitonly method
is the only
that recorded
method a precision
that recorded score between
a precision score 0.900
betweenand 0.900
1.000 and
with1.000
the highest F-measure
with the of 0.885.
highest F-meas-
The of
ure proposed method
0.885. The also achieved
proposed method thealsobest MAE score
achieved of 0.120.
the best MAESurprisingly,
score of 0.120.DCLC gave
Surpris-
good results for boundary and center image categories but achieved
ingly, DCLC gave good results for boundary and center image categories but achievedunsatisfactory results
for this category
unsatisfactory of images.
results Incategory
for this contrast, of
DGL and DRFI
images. improved
In contrast, DGL their
andperformances
DRFI improved for
this category
their of images.
performances Thecategory
for this deformed of smoothness
images. Theconstraint-based
deformed smoothnessmanifold ranking
constraint-
approach
based used by
manifold the DGL
ranking method
approach hasby
used helped
the DGLto improve
method performance
has helped tofor this image
improve per-
formance for this image category compared to other manifold ranking-based methods
such as MR. As stated in [7], results obtained for MR have demonstrated poor perfor-
mance on complex background images when compared to other categories of images. The
SIM again scored the lowest performance on this category of images. Figure 7 shows the
J. Imaging 2021, 7, 187 23 of 36
category compared to other manifold ranking-based methods such as MR. As stated in [7],
J. Imaging 2021, 7, x 24 of 37
results obtained for MR have demonstrated poor performance on complex background
images when compared to other categories of images. The SIM again scored the lowest
performance on this category of images. Figure 7 shows the average precision, recall,
average precision,
F-measure, MAE, andrecall,
ORF-measure, MAE, and OR
for all the investigated for all The
methods. the investigated
results show methods. The
the capability
results show the capability of the proposed method in the handling of images
of the proposed method in the handling of images with a complex background to exhibitwith a com-
plex background
its superiority to the
over exhibit itsmethods
other superiority over the other methods investigated.
investigated.
(a)
(b) (c)
Figure 7. (a) F-measure; (b) MAE and (c) OR on image category: Complex background.
4.5.4.
4.5.4. Salient
Salient Objects
Objects with
with Low
Low Color
Color Contrast
Contrast to to Background
Background
The proposed method showed strength in effectively
The proposed method showed strength in effectively detecting detecting salient
salient objects
objects fromfrom
the
the low contrast object category like other image categories. It achieved
low contrast object category like other image categories. It achieved the highest scores for the highest scores
for
mostmost performance
performance metrics,
metrics, except
except for recall.
for the the recall. The highest
The highest recallrecall
values values on images
on images from
from this category
this category are between
are between 0.7 and0.7 0.8and
while0.8the
while the proposed
proposed method method
scored ascored a recall
recall value of
value
0.715. of 0.715.
It is It is from
evident evident
thisfrom this research
research that themethods
that the existing existing methods
investigatedinvestigated have
have difficulty
difficulty in effectively
in effectively detecting detecting salientwhen
salient regions regions when an
an object object
shares shares acolor
a similar similar color with
contrast con-
trast with background
background regions. This regions. Thislearning-based
includes includes learning-based methodsthe
methods because because the perfor-
performances of
mances of HDCT and DRFI are not encouraging on images from
HDCT and DRFI are not encouraging on images from this category. Furthermore, contrast this category. Further-
more, contrast
prior-based prior-based
methods such as methods
DCLC, GR, suchCNS,
as DCLC,
and RPC GR,have
CNS,demonstrated
and RPC have thedemon-
lowest
strated the lowest
performances when performances
compared to when
othercompared
categories toofother categories
images. of images.
Like the results Like the
of other
results of other
categories, categories,
SIM again scored SIMtheagain
lowestscored
valuesthe forlowest
all the values for all metrics.
performance the performance
Figure 8
metrics. Figure 8 shows the average precision, recall, F-measure, MAE, and OR of all
methods, wherein the capability of the proposed method in the handling of salient objects
with low color contrast to the background is superior to the existing methods investigated.
J. Imaging 2021, 7, 187 24 of 36
shows the average precision, recall, F-measure, MAE, and OR of all methods, wherein the
J. Imaging 2021, 7, x 25 of 37
capability of the proposed method in the handling of salient objects with low color contrast
to the background is superior to the existing methods investigated.
(a)
(b) (c)
Figure
Figure8.
8.(a)
(a)F-measure;
F-measure;(b)
(b)MAE
MAEand
and(c)
(c)OR
ORon
on image
image category:
category: Low
Lowcontrast.
contrast.
4.5.5.
4.5.5.Multiple
MultipleSalient
SalientObjects
Objects
ItItisishard
hard to
to detect salient objects
objects when
whentheytheyexhibit
exhibitheterogeneous
heterogeneousfeatures featuresinin terms
terms of
of location,
location, color,
color, size,
size, andand count.
count. ThisThis image
image category
category contains
contains multiple
multiple objects
objects withwith var-
varying
ying locations,
locations, sizes,sizes,
counts,counts, and colors.
and colors. However,However, the performance
the performance of the proposed
of the proposed method
is commendable
method with thewith
is commendable bestthe
value
bestfor precision
value (0.876),(0.876),
for precision F-measure (0.853),(0.853),
F-measure MAE (0.836),
MAE
and ORand
(0.836), (0.695). The learning-based
OR (0.695). methodsmethods
The learning-based of DRFI (0.818)
of DRFI and HDCT
(0.818) and(0.791)
HDCT scored the
(0.791)
highestthe
scored recall value,
highest followed
recall value, by the proposed
followed method (0.786).
by the proposed methodThe results
(0.786). Theobtained by
results ob-
the rest of the methods clearly showed difficulty in detecting multiple
tained by the rest of the methods clearly showed difficulty in detecting multiple salient salient objects with
heterogeneous
objects properties. Inproperties.
with heterogeneous this category In of images,
this categoryall methods
of images,showed relatively
all methods poorer
showed
performance
relatively in terms
poorer of MAE.in terms of MAE.
performance
Inaddition
In additionto toour
ourmethod,
method,DGL DGLshowed
showedcomparatively
comparativelygood goodresults
resultswithwiththe
thesecond-
second-
highestvalues
highest valuesforforF-measure
F-measure(0.834)
(0.834)andandOROR (0.656).
(0.656). TheThelimitation
limitationof of COV
COV in in detecting
detecting
multiplesalient
multiple salientobjects
objectsisisclear
clearfrom
fromthese
theseresults
resultsas asititshows
showsaacomparatively
comparativelylow lowperfor-
perfor-
mance when compared to other image categories. This is because
mance when compared to other image categories. This is because of the consideration of of the consideration of
the assumption of spatial coincidence in multiscale saliency computation
the assumption of spatial coincidence in multiscale saliency computation [98]. The SIM [98]. The SIM
methodagain
method againscored
scoredthethelowest
lowestperformance
performanceon onthis
this category
category ofof images.
images. Figure
Figure99 shows
shows
theaverage
the averageprecision,
precision,recall,
recall,F-measure,
F-measure,MAE, MAE, and andOR ORof ofall
allthe
theinvestigated
investigatedmethods
methods on on
image category of multiple objects. The results show the capability of the proposed
method in handling salient objects with heterogeneous properties in terms of position,
count, and size.
J. Imaging 2021, 7, 187 25 of 36
J. Imaging 2021, 7, x image category of multiple objects. The results show the capability of the proposed method
26 of 37
in handling salient objects with heterogeneous properties in terms of position, count, and
size.
(a)
(b) (c)
Figure 9. (a)
Figure 9. (a) F-measure;
F-measure; (b)
(b) MAE
MAE and (c) OR
and (c) OR on
on image
image category:
category: Multiple salient objects.
Multiple salient objects.
4.5.6. Images
Images with
with Foreground
Foreground and Background Overlapped Objects
The
The average
average precision,
precision,recall,
recall,F-measure,
F-measure,MAE,
MAE,and andORORscores
scoresachieved
achievedforforthe
thecate-
cat-
gory
egoryofofoverlapped
overlappedimages
imagesareareillustrated
illustratedinin Figure
Figure 10.
10. In
In this
this category
category of images, the
DCLC obtained
obtained thethebest
bestoverall
overallperformance
performance with
with thethe highest
highest recall
recall (0.804),
(0.804), OR (0.790),
OR (0.790), and
F-measure
and (0.934).
F-measure The The
(0.934). proposed method
proposed achieved
method the highest
achieved precision
the highest valuevalue
precision of 0.986 and
of 0.986
is highly
and competitive
is highly with with
competitive DCLC. Surprisingly,
DCLC. the graph-based
Surprisingly, methods
the graph-based of DGLof(0.105)
methods DGL
and MR
(0.105) (0.114)
and achieved
MR (0.114) the best
achieved MAE
the bestscores, while SIM
MAE scores, whileandSIMSUNandscored inferior
SUN scored MAE
inferior
valuesvalues
MAE of 0.389of and 0.326,
0.389 and respectively. In this category
0.326, respectively. of images
In this category of also,
imagesthealso,
SIM the
method
SIM
recorded the lowest performance.
method recorded the lowest performance.
foreground and backgrounds seed selection methods such as MRBF and FBSS have also
achieved a better MAE score compared to BGFG, which is also based on background and
foreground seed selection. The DCLC that showed superiority in the image category of
overlap declined its performance on the ECSSD dataset. The SIM method showed the
lowest value for most of the performance metrics, except the MAE, while the method of
SUN scored relatively the worst value for MAE. The effectiveness of the proposed method
J. Imaging 2021, 7, x in detecting salient objects from a wide range of image categories has been successfully
27 of 37
proven by experiments.
(a)
(b) (c)
Figure 10.
Figure 10. (a)
(a) F-measure;
F-measure; (b)
(b) MAE
MAE and
and (c)
(c) OR
OR on
on image
image category:
category: Overlapped
Overlapped objects.
objects.
4.5.7.
4.5.8. Comparison with ECSSD Dataset
Deep-Learning-based Top-down Saliency Methods
The results
proposedof the proposed
method method
is not relatedwere further compared
to top-down against all the 30methods.
or deep-learning-based bottom-
up saliencywe
However, methods on the ECSSD
have extended dataset ascomparison
the quantitative in Table 3 to and Figure
seven 11 to evaluate its
deep-learning-based
performance. The ECSSD
top-down saliency dataset
detection methodsis well known
on the ECSSD fordataset
harboring complex images
to demonstrate while the
the superiority
superiority of themethod.
of the proposed proposedRecently,
method is theobvious becauseofitdeep-learning-based
performance has achieved the besttop-down
values of
precision (0.853), some
methods brought F-measure (0.790),
challenges MAE (0.163),
for bottom-up and methods
saliency OR (0.573). The
[140]. learning-based
However, the per-
formanceofofDRFI
method our method has revealed
and graph-based the ability
methods of bottom-up
of DGL, FBSS, and saliency detection
MRBF also achieved methods
better
can compete
results; favorably
however, with
only the deep-learning-based
proposed method managed top-down
to score methods.
precisionTable
above 4 illustrates
0.800. The
the comparison
foreground andof our method seed
backgrounds with selection
deep learning methods
methods such based on F-measure
as MRBF and FBSS and haveMAEalso
values reported
achieved a betterinMAE
the original references.
score compared to Regardless
BGFG, which of the complex
is also basednature of the ECSSD
on background and
dataset, the proposed
foreground method
seed selection. Thehas achieved
DCLC the best F-measure
that showed superiority(0.790)
in the when
imagecompared
category to of
deep-learning-based methods. In terms of MAE, the deep learning method
overlap declined its performance on the ECSSD dataset. The SIM method showed the low- of DS shows a
est value for most of the performance metrics, except the MAE, while the method of SUN
scored relatively the worst value for MAE. The effectiveness of the proposed method in
detecting salient objects from a wide range of image categories has been successfully
proven by experiments.
(). The up arrow, indicates that a higher value gives better performance, and the down arrow shows that a smaller
value gives better performance.
Method Precision Recall F-measure MAE OR Method Precision Recall F-measure MAE OR
OURs 0.853 0.635 0.790 0.163 0.573 GR 0.714 0.391 0.600 0.283 0.348
J. Imaging
AC 2021, 7,0.439
187 0.300 0.396 0.210 0.263 HDCT 0.767 0.640 0.733 0.198 27 of 36
0.519
BGFG 0.723 0.606 0.692 0.208 0.467 IT 0.570 0.406 0.521 0.289 0.285
BPFS 0.660 0.820 0.690 0.166 MAP 0.758 0.661 0.733 0.185 0.534
CA 0.532 0.374 0.485 0.310 0.266 MC 0.768 0.652 0.738 0.202 0.531
relatively best value of 0.160, but the MAE value of the proposed method is 0.163, which is
CNS 0.708 0.600 0.680 0.166 0.480 MCVS 0.780 0.540 0.700 0.170
a very close result. This result shows that the proposed method is even competitive with
COV 0.679 0.527 0.636 0.215 0.388 MR 0.767 0.647 0.736 0.186 0.525
deep-learning-based top-down methods. The F-measure and MAE scores in Tables 3 and 4
CSV 0.760 0.650 0.740 0.210 MRBF 0.780 0.670 0.760 0.177
illustrate that deep-learning-based methods of MSNSD-A and MSNSD, respectively, scored
DCLC 0.769 0.636 0.734 0.182 0.530 RPC 0.629 0.489 0.590 0.218 0.372
DGL 0.785 0.655
the 0.750
second and third best
0.191 0.548
F-measure
SEG
values
0.662
and higher
0.230
than0.462
those of 0.340
other bottom-up
0.212
DRFI 0.794 0.698 methods,
0.769 including
0.170 the 0.572
graph-based SeR and learning-based
0.366 0.207methods listed in0.404
0.311 Table 3. 0.144
In terms
DSR 0.753 0.647 of MAE
0.726 scores, deep
0.171 learning
0.517 methods
SIM of DS and
0.365 LCNN
0.078scored the
0.197 best values
0.433 and showed
0.062
FBSS 0.770 0.560 that0.709
their saliency
0.169maps are close to
SR the ground
0.460 truth. However,
0.302 the
0.411 performances
0.311 of these
0.212
FCB 0.721 0.515 methods
0.660 are highly
0.173 dependent
0.422 on
SUNsupervised
0.384learning based
0.102 on labeled
0.235 training
0.437 data
0.087[44].
FES 0.672 0.545 Due0.638
to the high dependency
0.212 0.404 and sensitivity
SWD of
0.704 deep learning
0.354 methods
0.573 on training
0.318 datasets,
0.283
GB 0.629 0.519 these methods
0.600 are
0.263 restricted
0.364 from using real-time and diverse categories of images [42,94].
0.5
0.4
0.3
0.2
0.1
0.0
AC
OURs
BPFS
CA
CNS
HDCT
IT
RPC
BGFG
DRFI
MAP
DSR
FCB
GB
GR
MR
SEG
SeR
SR
COV
SIM
SWD
CSV
DCLC
DGL
MC
FBSS
FES
MCVS
MRBF
SUN
Method
(a)
0.8
MAE OR
0.7
0.6
0.5
0.4
Score
0.3
0.2
0.1
0.0
SIM
BPFS
CNS
COV
FCB
DSR
SEG
GR
SUN
AC
CA
DCLC
HDCT
MRBF
OURs
IT
MAP
BGFG
CSV
DGL
DRFI
FBSS
FES
GB
MCVS
SWD
MR
SeR
SR
MC
RPC
Method
(b)
Figure
Figure 11.
11. (a)
(a) F-measure;
F-measure; (b)
(b) MAE
MAE and
and OR
OR on
on ECSSD
ECSSD dataset.
dataset.
4.5.8.
Table Comparison
4. Comparisonwith Deep-Learning-based
with deep learning methods inTop-down Saliency
terms of F-measure Methods
and MAE on ECSSD dataset.
The proposed method is not related to top-down or deep-learning-based methods.
Method F-Measure MAE
However, we have extended the quantitative comparison to seven deep-learning-based
MSNSD-A [38] 0.777 0.171
MSNSD [38] 0.774 0.179
DS [92] 0.759 0.160
LCNN [91] 0.715 0.162
[141] 0.430 0.255
TSL [90] 0.737 0.178
MCDL [93] 0.732
OURs 0.790 0.163
J. Imaging 2021, 7, 187 28 of 36
Table 5. Results of precision, recall, F-Measure, MAE and OR on HKU-IS and SOC datasets.
expensive because they have consumed more time in feature extraction. The running
time of the recent method of CNS is also higher and it is mainly influenced by the sample
size parameter used in attention map computation. The DGL is computationally more
expensive than other graph-based methods such as GR, MAP, MC, and MR.
Method OURS AC BGFG CA CNS COV DCLC DGL DRFI DSR FES GB GR
Time (s) 0.23 80.33 5.56 15.15 11.34 4.29 0.47 1.33 6.16 1.82 0.21 0.52 0.36
Method HDCT IT MAP MC MR RPC SEG SeR SIM SR SUN SWD
Time (s) 4.17 0.26 0.21 0.24 0.54 2.08 1.91 0.51 0.39 0.12 2.39 0.12
the limitations of contrast-based methods, regardless of the use of color contrast features.
However, the assumption of a narrow image border as a pseudo background can affect the
performance of DRFI on the category of boundary images. Computational complexity is
another intrinsic drawback of this method. The methods such as DGL, MR, MAP, and MC
that exploited the boundary prior have shown relatively low performance on the category
of boundary images when compared to the category of images with center prior. This
shows the major challenge of boundary prior in treating boundary regions as backgrounds
and is not effective when salient objects are near to the image boundary.
The methods such as CNS, DCLC, RPC, and GR that exploited contrast prior have
demonstrated relatively low performance on the category of low contrast images. This is
because contrast prior works well with images that have distinct color contrast differences
between foreground and background regions. This indicates that performances of the
investigated methods are highly dependent on salient object properties such as count, loca-
tion, size, color contrast, or background complexity. However, the proposed method has
performed well on most categories of images, irrespective of the various object properties
and background complexity. The extended evaluation of the proposed method on HKU-IS
and SOC datasets has further revealed the strength of our method in handling images
from differing datasets. However, the performances of our method and other bottom-up
methods in detecting the salient objects in the cluttered and occluded background were
not achieved with remarkable results. This is because the primitive image features such as
color, contrast, and texture are not adequate to detect the salient objects from cluttered and
occluded images in a meaningful manner [148]. The detection of objects from the cluttered
and occluded background can be enhanced by incorporating high-level features [148,149].
The integration of color contrast, contrast ratio, spatial feature, and center prior
information in the proposed method has provided adequate segregation of salient regions
from non-salient regions and uniformly highlighted salient objects. The accomplishment
of the proposed method makes it nearly universal for detecting salient objects in a wide
spectrum of images. Moreover, the quantitative comparison of the investigated methods
has exhibited the superiority of the proposed method and we were flabbergasted by the
performance of our method against the deep-learning-based top-down methods. Finally,
all region-based methods have shown good performances when compared to the patch
and pixel-wise methods. However, the performances of these methods are completely
dependent on the selection of region granularity. Due to the ability of the proposed
method to automatically detect the optimum number of regions, it has achieved the best
results when compared to the investigated methods. There is always a tradeoff between
computational complexity and accuracy. However, this is not the case with the proposed
method because we have achieved the best performance while upholding an efficient
run time of 0.23 s per image as demonstrated in Table 6. It should be observed that
preprocessing was not considered in the proposed as in the case of most methods and can
be optional.
5.2. Conclusions
This study has enriched the research on salient object detection by proposing a simple,
effective, and efficient method that incorporates histogram-based region formation for
image abstraction. The method has successfully integrated color contrast, contrast ratio,
spatial features, and center prior for achieving an impressive salient object detection
process. The method is capable of accurate and robust detection of salient objects from
a wide gamut of challenging images by uniformly highlighting. This accomplishment is
achieved by the successful integration of color contrast, contrast ratio, spatial feature, and
center prior. Experiments on different image categories have established that our method
has outperformed all 30 bottom-up saliency methods and seven deep-learning-based top-
down saliency methods. The computational efficiency of our method has demonstrated
that it can be exploited in real-time applications such as object segmentation and object
recognition. The proposed method has proven to be effective and efficient for a large set of
J. Imaging 2021, 7, 187 31 of 36
Author Contributions: Conceptualization, S.J. and O.O.O., methodology, S.J. and O.O.O., formal
analysis, S.J., data curation, S.J., writing original draft preparation, S.J., writing review and editing,
O.O.O., supervision, O.O.O. All authors have read and agreed to the published version of the
manuscript.
Funding: This research received no external funding.
Data Availability Statement: The ECSSD dataset is available at https://www.cse.cuhk.edu.hk/
leojia/projects/hsaliency/dataset.html (Accessed on 26 May 2019). The MSRA10K dataset is available
at https://mmcheng.net/msra10k/ (Accessed on 30 May 2019). The SED2 dataset is available at https:
//www.wisdom.weizmann.ac.il/~vision/Seg_Evaluation_DB/dl.html (Accessed on 30 May 2019).
The ASD dataset is available at https://www.epfl.ch/labs/ivrl/research/saliency/frequency-tuned-
salient-region-detection/ (Accessed on 30 May 2019). The DUT OMRON dataset is available at
http://saliencydetection.net/dut-omron/ (Accessed on 15 January 2020). The ImgSal dataset is
available at https://qualinet.github.io/databases/image/imgsal_mcgill_database_for_saliency_
detection/ (Accessed on 20 July 2020). The HKU-IS dataset is available at https://i.cs.hku.hk/
~yzyu/research/deep_saliency.html (Accessed on 15 January 2021). The SOC dataset is available at
http://dpfan.net/SOCBenchmark/ (Accessed on 24 June 2021).
Conflicts of Interest: The authors declare no conflict of interest.
References
1. Ye, L.; Liu, Z.; Li, L.; Shen, L.; Bai, C.; Wang, Y. Salient object segmentation via effective integration of saliency and objectness.
IEEE Trans. Multimed. 2017, 19, 1742–1756. [CrossRef]
2. Cong, R.; Lei, J.; Fu, H.; Cheng, M.-M.; Lin, W.; Huang, Q. Review of visual saliency detection with comprehensive information.
IEEE Trans. Circuits Syst. Video Technol. 2018, 29, 2941–2959. [CrossRef]
3. Ishikura, K.; Kurita, N.; Chandler, D.M.; Ohashi, G. Saliency detection based on multiscale extrema of local perceptual color
differences. IEEE Trans. Image Process. 2018, 27, 703–717. [CrossRef] [PubMed]
4. Ji, Y.; Zhang, H.; Tseng, K.-K.; Chow, T.W.; Wu, Q.J. Graph model-based salient object detection using objectness and multiple
saliency cues. Neurocomputing 2019, 323, 188–202. [CrossRef]
5. Lopez-Alanis, A.; Lizarraga-Morales, R.A.; Sanchez-Yanez, R.E.; Martinez-Rodriguez, D.E.; Contreras-Cruz, M.A. Visual Saliency
Detection Using a Rule-Based Aggregation Approach. Appl. Sci. 2019, 9, 2015. [CrossRef]
6. Xia, C.; Zhang, H.; Gao, X.; Li, K. Exploiting background divergence and foreground compactness for salient object detection.
Neurocomputing 2020, 383, 194–211. [CrossRef]
7. Ma, W.-P.; Li, W.-X.; Sun, J.-C.; Cao, P.-X. Saliency detection via manifold ranking based on robust foreground. Int. J. Autom.
Comput. 2021, 18, 73–84. [CrossRef]
8. Wu, Y.; Jia, T.; Pang, Y.; Sun, J.; Xue, D. Salient object detection via a boundary-guided graph structure. J. Vis. Commun. Image
Represent. 2021, 75, 103048. [CrossRef]
9. Wang, Z.; Xu, G.; Wang, Z.; Zhu, C. Saliency detection integrating both background and foreground information. Neurocomputing
2016, 216, 468–477. [CrossRef]
10. Zhang, X.; Wang, Y.; Chen, Z.; Yan, J.; Wang, D. Saliency detection via image sparse representation and color features combination.
Multimed. Tools Appl. 2020, 79, 23147–23159. [CrossRef]
11. Fan, H.; Xie, F.; Li, Y.; Jiang, Z.; Liu, J. Automatic segmentation of dermoscopy images using saliency combined with Otsu
threshold. Comput. Biol. Med. 2017, 85, 75–85. [CrossRef]
12. Yuheng, S.; Hao, Y. Image Segmentation Algorithms Overview. arXiv 2017, arXiv:1707.02051.
13. Olugbara, O.O.; Taiwo, T.B.; Heukelman, D. Segmentation of melanoma skin lesion using perceptual color difference saliency
with morphological analysis. Math. Probl. Eng. 2018, 2018, 1524286. [CrossRef]
14. Joshi, A.; Khan, M.S.; Soomro, S.; Niaz, A.; Han, B.S.; Choi, K.N. SRIS: Saliency-Based Region Detection and Image Segmentation
of COVID-19 Infected Cases. IEEE Access 2020, 8, 190487–190503. [CrossRef]
15. Achanta, R.; Estrada, F.; Wils, P.; Süsstrunk, S. Salient region detection and segmentation. In Proceedings of the International
Conference on Computer Vision Systems, Santorini, Greece, 12–15 May 2008; pp. 66–75.
16. John, V.; Yoneda, K.; Liu, Z.; Mita, S. Saliency map generation by the convolutional neural network for real-time traffic light
detection using template matching. IEEE Trans. Comput. Imaging 2015, 1, 159–173. [CrossRef]
17. Li, H.; Su, X.; Wang, J.; Kan, H.; Han, T.; Zeng, Y.; Chai, X. Image processing strategies based on saliency segmentation for object
recognition under simulated prosthetic vision. Artif. Intell. Med. 2018, 84, 64–78. [CrossRef]
J. Imaging 2021, 7, 187 32 of 36
18. Liu, W.; Feng, X.; Wang, S.; Hu, B.; Gan, Y.; Zhang, X.; Lei, T. Random selection-based adaptive saliency-weighted RXD anomaly
detection for hyperspectral imagery. Int. J. Remote Sens. 2018, 39, 2139–2158. [CrossRef]
19. Al-Gabalawy, M. Detecting anomalies within unmanned aerial vehicle (UAV) video based on contextual saliency. Appl. Soft
Comput. 2020, 96, 106715. [CrossRef]
20. Tsai, Y.H. Hierarchical salient point selection for image retrieval. Pattern Recognit. Lett. 2012, 33, 1587–1593. [CrossRef]
21. Giouvanakis, E.; Kotropoulos, C. Saliency map driven image retrieval combining the bag-of-words model and PLSA. In
Proceedings of the 2014 19th International Conference on Digital Signal Processing, Hong Kong, China, 23–24 August 2014;
pp. 280–285.
22. Zhu, C.; Huang, K.; Li, G. An innovative saliency guided roi selection model for panoramic images compression. In Proceedings
of the 2018 Data Compression Conference, Snowbird, UT, USA, 27–30 March 2018; p. 436.
23. Li, N.; Zhao, X.; Yang, Y.; Zou, X. Objects classification by learning-based visual saliency model and convolutional neural network.
Comput. Intell. Neurosci. 2016, 2016, 7942501. [CrossRef] [PubMed]
24. Guo, M.; Zhao, Y.; Zhang, C.; Chen, Z. Fast object detection based on selective visual attention. Neurocomputing 2014, 144, 184–197.
[CrossRef]
25. Goferman, S.; Zelnik-Manor, L.; Tal, A. Context-aware saliency detection. IEEE Trans. Pattern Anal. Mach. Intell. 2012, 34,
1915–1926. [CrossRef]
26. Li, R.; Cai, J.; Zhang, H.; Wang, T. Aggregating complementary boundary contrast with smoothing for salient region detection.
Vis. Comput. 2017, 33, 1155–1167. [CrossRef]
27. Wan, M.; Ren, K.; Gu, G.; Zhang, X.; Qian, W.; Chen, Q.; Yu, S. Infrared small moving target detection via saliency histogram and
geometrical invariability. Appl. Sci. 2017, 7, 569. [CrossRef]
28. Lin, G.; Fan, W. Unsupervised video object segmentation based on mixture models and saliency detection. Neural Process. Lett.
2020, 51, 657–674. [CrossRef]
29. Marat, S.; Guironnet, M.; Pellerin, D. Video summarization using a visual attention model. In Proceedings of the 2007 15th
European Signal Processing Conference, Poznań, Poland, 3–7 September 2007; pp. 1784–1788.
30. Adeliyi, T.; Olugbara, O. Detecting salient objects in non-stationary video image sequence for analyzing user perceptions of
digital video contents. Multimed. Tools Appl. 2019, 78, 31807–31821. [CrossRef]
31. Wang, Y.; Wei, X.; Ding, L.; Tang, X.; Zhang, H. A robust visual tracking method via local feature extraction and saliency detection.
Vis. Comput. 2019, 36, 683–700. [CrossRef]
32. Zhang, Y.; Zhang, F.; Guo, L. Saliency detection by selective color features. Neurocomputing 2016, 203, 34–40. [CrossRef]
33. Zhang, Q.; Lin, J.; Tao, Y.; Li, W.; Shi, Y. Salient object detection via color and texture cues. Neurocomputing 2017, 243, 35–48.
[CrossRef]
34. Burt, R.; Thigpen, N.N.; Keil, A.; Principe, J.C. Unsupervised foveal vision neural architecture with top-down attention. Neural
Netw. 2021, 141, 145–159. [CrossRef]
35. Xia, C.; Zhang, H. Unsupervised Salient object detection by aggregating multi-level cues. IEEE Photonics J. 2018, 10, 7801711.
[CrossRef]
36. Zhou, L.; Yang, Z.; Yuan, Q.; Zhou, Z.; Hu, D. Salient region detection via integrating diffusion-based compactness and local
contrast. IEEE Trans. Image Process. 2015, 24, 3308–3320. [CrossRef] [PubMed]
37. Zhang, P.; Wang, D.; Lu, H.; Wang, H.; Ruan, X. Amulet: Aggregating multi-level convolutional features for salient ob-
ject detection. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017;
pp. 202–211.
38. Liang, Y.; Liu, H.; Ma, N. A novel deep network and aggregation model for saliency detection. Vis. Comput. 2019, 2020, 1883–1895.
[CrossRef]
39. Bruckert, A.; Tavakoli, H.R.; Liu, Z.; Christie, M.; Le Meur, O. Deep saliency models: The quest for the loss function. Neurocomputing
2021, 453, 693–704. [CrossRef]
40. Cui, W.; Zhang, Q.; Zuo, B. Deep saliency detection via spatial-wise dilated convolutional attention. Neurocomputing 2021, 445,
35–49. [CrossRef]
41. Li, Z.; Lang, C.; Wang, T.; Li, Y.; Feng, J. Deep spatio-frequency saliency detection. Neurocomputing 2021, 453, 645–655. [CrossRef]
42. Gupta, A.K.; Seal, A.; Prasad, M.; Khanna, P. Salient object detection techniques in computer vision—A survey. Entropy 2020,
22, 1174. [CrossRef]
43. Wang, W.; Lai, Q.; Fu, H.; Shen, J.; Ling, H.; Yang, R. Salient object detection in the deep learning era: An in-depth survey. IEEE
Trans. Pattern Anal. Mach. Intell. 2021, 1–20.
44. Zhang, Q.R.; Wang, Y.F. A multi-cues based approach for visual saliency detection. Int. J. Innov. Comput. Inf. Control 2021, 17,
1435–1446. [CrossRef]
45. Cai, Y.; Dai, L.; Wang, H.; Chen, L.; Li, Y. A novel saliency detection algorithm based on adversarial learning model. IEEE Trans.
Image Process. 2020, 29, 4489–4504. [CrossRef] [PubMed]
46. Ren, Z.; Gao, S.; Chia, L.-T.; Tsang, I.W.-H. Region-based saliency detection and its application in object recognition. IEEE Trans.
Circuits Syst. Video Technol. 2013, 24, 769–779. [CrossRef]
47. Li, N.; Bi, H.; Zhang, Z.; Kong, X.; Lu, D. Performance comparison of saliency detection. Adv. Multimed. 2018, 2018, 9497083.
[CrossRef]
J. Imaging 2021, 7, 187 33 of 36
48. Achanta, R.; Hemami, S.; Estrada, F.; Susstrunk, S. Frequency-tuned salient region detection. In Proceedings of the Computer
Vision and Pattern Recognition, CVPR, Miami, FL, USA, 20–21 June 2009; pp. 1597–1604.
49. Zhang, J.; Ehinger, K.A.; Ding, J.; Yang, J. A prior-based graph for salient object detection. In Proceedings of the Image Processing
(ICIP), 2014 IEEE International Conference, Paris, France, 27–30 October 2014; pp. 1175–1178.
50. Cheng, M.-M.; Mitra, N.J.; Huang, X.; Torr, P.H.; Hu, S.-M. Global contrast based salient region detection. IEEE Trans. Pattern
Anal. Mach. Intell. 2015, 37, 569–582. [CrossRef] [PubMed]
51. Yang, W.; Li, D.; Wang, S.; Lu, S.; Yang, J. Saliency-based color image segmentation in foreign fiber detection. Math. Comput.
Model. 2013, 58, 852–858. [CrossRef]
52. Zhang, Y.; Han, J.; Guo, L. Salient region detection using background priors. Optik 2014, 125, 5872–5877. [CrossRef]
53. Wei, Y.; Wen, F.; Zhu, W.; Sun, J. Geodesic saliency using background priors. In Proceedings of the European Conference on
Computer Vision, Florence, Italy, 7–13 October 2012; pp. 29–42.
54. Tavakoli, H.R.; Rahtu, E.; Heikkilä, J. Fast and efficient saliency detection using sparse sampling and kernel density estimation. In
Proceedings of the Scandinavian Conference on Image Analysis, Ystad, Sweden, 23–25 May 2011; pp. 666–675.
55. Jiang, H.; Wang, J.; Yuan, Z.; Liu, T.; Zheng, N.; Li, S. Automatic salient object segmentation based on context and shape prior. In
Proceedings of the BMVC, Dundee, Scotland, 29 August–2 September 2011; p. 9.
56. Singh, V.K.; Kumar, N.; Singh, N. A hybrid approach using color spatial variance and novel object position prior for salient object
detection. Multimed. Tools Appl. 2020, 79, 30045–30067. [CrossRef]
57. Niu, Y.; Su, C.; Guo, W. Salient object segmentation based on superpixel and background connectivity prior. IEEE Access 2018, 6,
56170–56183. [CrossRef]
58. Liu, Z.; Le Meur, O.; Luo, S.; Shen, L. Saliency detection using regional histograms. Opt. Lett. 2013, 38, 700–702. [CrossRef]
59. Afzali, S.; Al-Sahaf, H.; Xue, B.; Hollitt, C.; Zhang, M. Foreground and background feature fusion using a convex hull based
center prior for salient object detection. In Proceedings of the 2018 International Conference on Image and Vision Computing
New Zealand (IVCNZ), Auckland, New Zealand, 19–21 November 2018; pp. 1–9.
60. Ullah, I.; Jian, M.; Hussain, S.; Guo, J.; Yu, H.; Wang, X.; Yin, Y. A brief survey of visual saliency detection. Multimed. Tools Appl.
2020, 79, 34605–34645. [CrossRef]
61. Itti, L.; Koch, C.; Niebur, E. A model of saliency-based visual attention for rapid scene analysis. IEEE Trans. Pattern Anal. Mach.
Intell. 1998, 20, 1254–1259. [CrossRef]
62. Liu, T.; Yuan, Z.; Sun, J.; Wang, J.; Zheng, N.; Tang, X.; Shum, H.-Y. Learning to detect a salient object. IEEE Trans. Pattern Anal.
Mach. Intell. 2011, 33, 353–367.
63. Borji, A.; Cheng, M.-M.; Jiang, H.; Li, J. Salient object detection: A benchmark. IEEE Trans. Image Process. 2015, 24, 5706–5722.
[CrossRef] [PubMed]
64. Harel, J.; Koch, C.; Perona, P. Graph-based visual saliency. In Proceedings of the Advances in Neural Information Processing
Systems, Vancouver, BC, Canada, 4–7 December 2006; pp. 545–552.
65. Han, J.; Zhang, D.; Wen, S.; Guo, L.; Liu, T.; Li, X. Two-stage learning to predict human eye fixations via SDAEs. IEEE Trans.
Cybern. 2015, 46, 487–498. [CrossRef]
66. Lou, J.; Ren, M.; Wang, H. Regional principal color based saliency detection. PLoS ONE 2014, 9, e112475.
67. Kim, J.; Han, D.; Tai, Y.-W.; Kim, J. Salient region detection via high-dimensional color transform and local spatial support. IEEE
Trans. Image Process. 2016, 25, 9–23. [CrossRef] [PubMed]
68. Liu, G.-H.; Yang, J.-Y. Exploiting color volume and color difference for salient region detection. IEEE Trans. Image Process. 2018,
28, 6–16. [CrossRef] [PubMed]
69. Tang, J.; Yang, G.; Sun, Y.; Xin, J.; He, D. Salient object detection of dairy goats in farm image based on background and foreground
priors. Neurocomputing 2019, 332, 270–282. [CrossRef]
70. Lou, J.; Wang, H.; Chen, L.; Xu, F.; Xia, Q.; Zhu, W.; Ren, M. Exploiting color name space for salient object detection. Multimed.
Tools Appl. 2020, 79, 10873–10897. [CrossRef]
71. Borji, A. What is a salient object? A dataset and a baseline model for salient object detection. IEEE Trans. Image Process. 2014, 24,
742–756. [CrossRef]
72. Umeki, Y.; Funahashi, I.; Yoshida, T.; Iwahashi, M. Salient object detection with importance degree. IEEE Access 2020, 8,
147059–147069. [CrossRef]
73. Seo, H.J.; Milanfar, P. Static and space-time visual saliency detection by self-resemblance. J. Vis. 2009, 9, 15. [CrossRef] [PubMed]
74. Rahtu, E.; Kannala, J.; Salo, M.; Heikkilä, J. Segmenting salient objects from images and videos. In Computer Vision–ECCV 2010;
Daniilidis, K., Maragos, P., Paragios, N., Eds.; Springer: Berlin/Heidelberg, Germany, 2010; pp. 366–379.
75. Murray, N.; Vanrell, M.; Otazu, X.; Parraga, C.A. Saliency estimation using a non-parametric low-level vision model. In
Proceedings of the Computer Vision and Pattern Recognition (CVPR), Washington, DC, USA, 20–25 June 2011; pp. 433–440.
76. Luo, H.; Han, G.; Liu, P.; Wu, Y. Salient region detection using diffusion process with nonlocal connections. Appl. Sci. 2018,
8, 2526. [CrossRef]
77. Lou, J.; Zhu, W.; Wang, H.; Ren, M. Small target detection combining regional stability and saliency in a color image. Multimed.
Tools Appl. 2017, 76, 14781–14798. [CrossRef]
78. Liu, J.; Wang, S. Salient region detection via simple local and global contrast representation. Neurocomputing 2015, 147, 435–443.
[CrossRef]
J. Imaging 2021, 7, 187 34 of 36
79. Zhou, J.; Zhai, J.; Ren, Y.; Lu, A. Background prior-based salient object detection via adaptive figure-ground classification. KSII
Trans. Internet Inf. Syst. 2018, 12, 1264–1286.
80. Jiang, B.; Zhang, L.; Lu, H.; Yang, C.; Yang, M.-H. Saliency detection via absorbing Markov chain. In Proceedings of the IEEE
International Conference on Computer Vision, Sydney, Australia, 1–8 December 2013; pp. 1665–1672.
81. Zhou, L.; Yang, Z. Salient region detection based on spatial and background priors. In Proceedings of the 2014 IEEE International
Conference on Information and Automation (ICIA), Hailar, China, 28–30 July 2014; pp. 262–266.
82. Pahuja, A.; Majumder, A.; Chakraborty, A.; Venkatesh Babu, R. Enhancing salient object segmentation through attention. arXiv
2019, arXiv:1905.11522.
83. Yang, C.; Zhang, L.; Lu, H.; Ruan, X.; Yang, M.-H. Saliency detection via graph-based manifold ranking. In Proceedings of the
IEEE Conference on Computer Vision and Pattern Recognition, Washington, DC, USA, 23–28 June 2014; pp. 3166–3173.
84. Wu, X.; Ma, X.; Zhang, J.; Wang, A.; Jin, Z. Salient object detection via deformed smoothness constraint. In Proceedings of the
2018 25th IEEE International Conference on Image Processing (ICIP), Athens, Greece, 7–10 October 2018; pp. 2815–2819.
85. Jiang, H.; Wang, J.; Yuan, Z.; Wu, Y.; Zheng, N.; Li, S. Salient object detection: A discriminative regional feature integration
approach. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Washington, DC, USA,
23–28 June 2013; pp. 2083–2090.
86. Xu, L.; Zeng, L.; Duan, H.; Sowah, N.L. Saliency detection in complex scenes. EURASIP J. Image Video Process. 2014, 2014, 31.
[CrossRef]
87. Tang, L.; Meng, F.; Wu, Q.; Sowah, N.L.; Tan, K.; Li, H. Salient object detection and segmentation via ultra-contrast. IEEE Access
2018, 6, 14870–14883. [CrossRef]
88. Achanta, R.; Shaji, A.; Smith, K.; Lucchi, A.; Fua, P.; Süsstrunk, S. SLIC superpixels compared to state-of-the-art superpixel
methods. IEEE Trans. Pattern Anal. Mach. Intell. 2012, 34, 2274–2282. [CrossRef]
89. Pang, Y.; Wu, Y.; Wu, C.; Zhang, M. Salient object detection via effective background prior and novel graph. Multimed. Tools Appl.
2020, 79, 25679–25695. [CrossRef]
90. Yan, K.; Wang, X.; Kim, J.; Feng, D. A new aggregation of DNN sparse and dense labeling for saliency detection. IEEE Trans.
Cybern. 2020, 1–14. [CrossRef]
91. Li, H.; Chen, J.; Lu, H.; Chi, Z. CNN for saliency detection with low-level feature integration. Neurocomputing 2017, 226, 212–220.
[CrossRef]
92. Li, X.; Zhao, L.; Wei, L.; Yang, M.-H.; Wu, F.; Zhuang, Y.; Ling, H.; Wang, J. Deepsaliency: Multi-task deep neural network model
for salient object detection. IEEE Trans. Image Process. 2016, 25, 3919–3930. [CrossRef]
93. Zhao, R.; Ouyang, W.; Li, H.; Wang, X. Saliency detection by multi-context deep learning. In Proceedings of the IEEE Conference
on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1265–1274.
94. Wang, H.; Zhu, C.; Shen, J.; Zhang, Z.; Shi, X. Salient object detection by robust foreground and background seed selection.
Comput. Electr. Eng. 2021, 90, 106993. [CrossRef]
95. Tang, Y.; Wu, X. Saliency detection via combining region-level and pixel-level predictions with CNNs. In Proceedings of the
European Conference on Computer Vision, Amsterdam, The Netherlands, 11–14 October 2016; pp. 809–825.
96. Wang, L.; Dong, S.-L.; Li, H.-S.; Zhu, X.-B. A brief survey of low-level saliency detection. In Proceedings of the 2016 International
Conference on Information System and Artificial Intelligence (ISAI), Hong Kong, China, 24–26 June 2016; pp. 590–593.
97. Duan, L.; Wu, C.; Miao, J.; Qing, L.; Fu, Y. Visual saliency detection by spatially weighted dissimilarity. In Proceedings of the
Computer Vision and Pattern Recognition (CVPR), Washington, DC, USA, 20–25 June 2011; pp. 473–480.
98. Erdem, E.; Erdem, A. Visual saliency estimation by nonlinearly integrating features using region covariances. J. Vis. 2013, 13, 11.
[CrossRef]
99. Borji, A.; Itti, L. Exploiting local and global patch rarities for saliency detection. In Proceedings of the 2012 IEEE Conference on
Computer Vision and Pattern Recognition, Washington, DC, USA, 16–21 June 2012; pp. 478–485.
100. Zhang, L.; Tong, M.H.; Marks, T.K.; Shan, H.; Cottrell, G.W. SUN: A Bayesian framework for saliency using natural statistics. J.
Vis. 2008, 8, 32. [CrossRef] [PubMed]
101. Jia, C.; Li, S.; Chen, W.; Kong, F. Image salient object detection based on perceptually homogeneous patch. In Proceedings of the
2019 IEEE 19th International Conference on Communication Technology (ICCT), Xi’an, China, 16–19 October 2019; pp. 1644–1647.
102. Jiang, L.; Zhong, H.; Lin, X. Saliency Detection via Boundary Prior and Center Prior. Int. Robot. Autom. J. 2017, 2, 134–139.
103. Huang, X.; Zhang, Y.-J. 300-FPS salient object detection via minimum directional contrast. IEEE Trans. Image Process. 2017, 26,
4243–4254. [CrossRef] [PubMed]
104. Ren, X.; Malik, J. Learning a classification model for segmentation. In Proceedings of the Ninth IEEE International Conference on
Computer Vision, Nice, France, 13–16 October 2003; p. 10.
105. Filali, I.; Allili, M.S.; Benblidia, N. Multi-scale salient object detection using graph ranking and global–local saliency refinement.
Signal Process. Image Commun. 2016, 47, 380–401. [CrossRef]
106. Li, X.; Lu, H.; Zhang, L.; Ruan, X.; Yang, M.-H. Saliency detection via dense and sparse reconstruction. In Proceedings of the IEEE
International Conference on Computer Vision, Sydney, Australia, 1–8 December 2013; pp. 2976–2983.
107. Feng, L.; Li, H.B.; Gao, Y.K.; Zhang, Y.K. A color image segmentation method based on region salient color and fuzzy C-means
algorithm. Circuits Syst. Signal Process. 2020, 39, 586–610. [CrossRef]
J. Imaging 2021, 7, 187 35 of 36
108. Sun, J.; Lu, H.; Liu, X. Saliency region detection based on Markov absorption probabilities. IEEE Trans. Image Process. 2015, 24,
1639–1649. [CrossRef]
109. Wang, J.; Lu, H.; Li, X.; Tong, N.; Liu, W. Saliency detection via background and foreground seed selection. Neurocomputing 2015,
152, 359–368. [CrossRef]
110. Han, A.; Han, F.; Hao, J.; Yuan, Y. An improved saliency detection method based on non-uniform quantification and channel-
weighted color distance. Multimed. Tools Appl. 2017, 76, 11037–11050. [CrossRef]
111. Zhou, J.-X.; Liu, X.-D.; Xu, T.-W.; Gan, J.-H.; Liu, W.-Q. A new fusion approach for content based image retrieval with color
histogram and local directional pattern. Int. J. Mach. Learn. Cybern. 2018, 9, 677–689. [CrossRef]
112. Nouri, F.; Kazemi, K.; Danyali, H. Salient object detection using local, global and high contrast graphs. Signal Image Video Process.
2018, 12, 659–667. [CrossRef]
113. Yang, C.; Zhang, L.; Lu, H. Graph-regularized saliency detection with convex-hull-based center prior. IEEE Signal Process. Lett.
2013, 20, 637–640. [CrossRef]
114. Hu, Z.; Zhang, Z.; Sun, Z.; Zhao, S. Saliency detection based on salient edges and remarkable discriminating for superpixel pairs.
Multimed. Tools Appl. 2018, 77, 5949–5968. [CrossRef]
115. Jiang, F.; Kong, B.; Li, J.; Dashtipour, K.; Gogate, M. Robust visual saliency optimization based on bidirectional Markov chains.
Cogn. Comput. 2021, 13, 69–80. [CrossRef]
116. Zhang, X.; Wang, Y.; Yan, J.; Chen, Z.; Wang, D. A unified saliency detection framework for visible and infrared images. Multimed.
Tools Appl. 2020, 79, 23147–23159. [CrossRef]
117. Liu, Y.; Liu, G.; Liu, C.; Sun, C. A novel color-texture descriptor based on local histograms for image segmentation. IEEE Access
2019, 7, 160683–160695. [CrossRef]
118. Márquez-de-Silva, S.; Felipe-Riverón, E.; Fernández, L.P.S. A simple and effective method of color image quantization. In
Proceedings of the Iberoamerican Congress on Pattern Recognition, Havana, Cuba, 9–12 September 2008; pp. 749–757.
119. Feng, X.; Guoying, C.; Richang, H.; Jing, G. Camouflage texture evaluation using a saliency map. Multimed. Syst. 2015, 21,
169–175. [CrossRef]
120. Cheng, G.; Wei, J. Color quantization application based on K-means in remote sensing image processing. J. Phys. Conf. Ser. 2019,
1213, 42012. [CrossRef]
121. Olugbara, O.O.; Adetiba, E.; Oyewole, S.A. Pixel intensity clustering algorithm for multilevel image segmentation. Math. Probl.
Eng. 2015, 2015, 649802. [CrossRef]
122. Yu, C.-Y.; Zhang, W.-S.; Wang, C.-L. A saliency detection method based on global contrast. Int. J. Signal Process. Image Process.
Pattern Recognit. 2015, 8, 111–122. [CrossRef]
123. Lee, S.; Xin, J.H.; Westland, S. Evaluation of image similarity by histogram intersection. Color Res. Appl. 2005, 30, 265–274.
[CrossRef]
124. Fareed, M.M.S.; Chun, Q.; Ahmed, G.; Asif, M.R.; Fareed, M.Z. Saliency detection by exploiting multi-features of color contrast
and color distribution. Comput. Electr. Eng. 2018, 70, 551–566. [CrossRef]
125. Vincent, L. Morphological grayscale reconstruction in image analysis: Applications and efficient algorithms. IEEE Trans. Image
Process. 1993, 2, 176–201. [CrossRef]
126. Zhang, J.; Guo, Z.; Jiao, T.; Wang, M. Defect detection of aluminum alloy wheels in radiography images using adaptive threshold
and morphological reconstruction. Appl. Sci. 2018, 8, 2365. [CrossRef]
127. Kavallieratou, E. A binarization algorithm specialized on document images and photos. In Proceedings of the Eighth International
Conference on Document Analysis and Recognition (ICDAR’05), Seoul, Korea, 31 August–1 September 2005; pp. 463–467.
128. Ismail, S.M.; Abdullah, S.N.H.S. Novel binarization method for enhancing ancient and historical manuscript images. In
Proceedings of the Mexican International Conference on Artificial Intelligence, Tuxtla Gutiérrez, Mexico, 16–22 November 2014;
pp. 393–406.
129. Alpert, S.; Galun, M.; Brandt, A.; Basri, R. Image segmentation by probabilistic bottom-up aggregation and cue integration. IEEE
Trans. Pattern Anal. Mach. Intell. 2011, 34, 315–327. [CrossRef]
130. Li, J.; Levine, M.D.; An, X.; Xu, X.; He, H. Visual saliency based on scale-space analysis in the frequency domain. IEEE Trans.
Pattern Anal. Mach. Intell. 2012, 35, 996–1010. [CrossRef]
131. Shi, J.; Yan, Q.; Xu, L.; Jia, J. Hierarchical image saliency detection on extended CSSD. IEEE Trans. Pattern Anal. Mach. Intell. 2015,
38, 717–729. [CrossRef] [PubMed]
132. Li, G.; Yu, Y. Visual saliency based on multiscale deep features. In Proceedings of the IEEE Conference on Computer Vision and
Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 5455–5463.
133. Fan, D.-P.; Cheng, M.-M.; Liu, J.-J.; Gao, S.-H.; Hou, Q.; Borji, A. Salient objects in clutter: Bringing salient object detection to the
foreground. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018;
pp. 186–202.
134. Zhu, X.; Tang, C.; Wang, P.; Xu, H.; Wang, M.; Chen, J.; Tian, J. Saliency detection via affinity graph learning and weighted
manifold ranking. Neurocomputing 2018, 312, 239–250. [CrossRef]
135. Oh, S.J.; Benenson, R.; Khoreva, A.; Akata, Z.; Fritz, M.; Schiele, B. Exploiting saliency for object segmentation from image level
labels. In Proceedings of the IEEE Conference Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017;
pp. 5038–5047.
J. Imaging 2021, 7, 187 36 of 36
136. Oh, K.; Lee, M.; Lee, Y.; Kim, S. Salient object detection using recursive regional feature clustering. Inf. Sci. 2017, 387, 1–18.
[CrossRef]
137. Lin, T.-Y.; Maire, M.; Belongie, S.; Hays, J.; Perona, P.; Ramanan, D.; Dollár, P.; Zitnick, C.L. Microsoft coco: Common objects in
context. In Proceedings of the European Conference on Computer Vision, Zurich, Switzerland, 6–12 September 2014; pp. 740–755.
138. Jiang, P.; Pan, Z.; Tu, C.; Vasconcelos, N.; Chen, B.; Peng, J. Super diffusion for salient object detection. IEEE Trans. Image Process.
2019, 29, 2903–2917. [CrossRef] [PubMed]
139. Hou, X.; Zhang, L. Saliency detection: A spectral residual approach. In Proceedings of the Computer Vision and Pattern
Recognition, CVPR’07, Minneapolis, MN, USA, 17–22 June 2007; pp. 1–8.
140. Ding, M.; Xu, X.; Zhang, F.; Xiao, Z.; Liu, Y.; Geng, L.; Wu, J.; Wen, J.; Wang, M. Saliency detection via background prior and
foreground seeds. Multimed. Tools Appl. 2020, 79, 14849–14870. [CrossRef]
141. Nasiripour, R.; Farsi, H.; Mohamadzadeh, S. Visual saliency object detection using sparse learning. IET Image Process. 2019, 13,
2436–2447. [CrossRef]
142. Niu, Y.; Chen, J.; Guo, W. Meta-metric for saliency detection evaluation metrics based on application preference. Multimed. Tools
Appl. 2018, 77, 26351–26369. [CrossRef]
143. Li, X.; Li, Y.; Shen, C.; Dick, A.; Hengel, A.V.D. Contextual hypergraph modeling for salient object detection. In Proceedings of the
2013 IEEE International Conference on Computer Vision, Sydney, Australia, 1–8 December 2013; pp. 3328–3335.
144. Jiang, B.; He, Z.; Ding, C.; Luo, B. Saliency detection via a multi-layer graph based diffusion model. Neurocomputing 2018, 314,
215–223. [CrossRef]
145. Nawaz, M.; Yan, H. Saliency detection via multiple-morphological and superpixel based fast fuzzy C-mean clustering network.
Expert Syst. Appl. 2020, 161, 113654. [CrossRef]
146. Perazzi, F.; Krähenbühl, P.; Pritch, Y.; Hornung, A. Saliency filters: Contrast based filtering for salient region detection. In
Proceedings of the 2012 IEEE Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, 16–21 June 2012;
pp. 733–740.
147. Sikha, O.; Kumar, S.S.; Soman, K. Salient region detection and object segmentation in color images using dynamic mode
decomposition. J. Comput. Sci. 2018, 25, 351–366. [CrossRef]
148. Borenstein, E.; Ullman, S. Class-specific, top-down segmentation. In European Conference on Computer Vision; Springer:
Berlin/Heidelberg, Germany, 2002; pp. 109–122.
149. Bravo, M.J.; Farid, H. Recognizing and segmenting objects in clutter. Vis. Res. 2004, 44, 385–396. [CrossRef] [PubMed]