Next Article in Journal
Parallel Fish School Tracking Based on Multiple Appearance Feature Detection
Previous Article in Journal
Secure Encapsulation Schemes Using Key Recovery System in IoMT Environments
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

WS-RCNN: Learning to Score Proposals for Weakly Supervised Instance Segmentation

School of Automation Science and Engineering, South China University of Technology, Guangzhou 510641, China
*
Author to whom correspondence should be addressed.
Sensors 2021, 21(10), 3475; https://doi.org/10.3390/s21103475
Submission received: 26 March 2021 / Revised: 29 April 2021 / Accepted: 7 May 2021 / Published: 17 May 2021
(This article belongs to the Topic Scientific Advances in STEM: From Professor to Students)

Abstract

:
Weakly supervised instance segmentation (WSIS) provides a promising way to address instance segmentation in the absence of sufficient labeled data for training. Previous attempts on WSIS usually follow a proposal-based paradigm, critical to which is the proposal scoring strategy. These works mostly rely on certain heuristic strategies for proposal scoring, which largely hampers the sustainable advances concerning WSIS. Towards this end, this paper introduces a novel framework for weakly supervised instance segmentation, called Weakly Supervised R-CNN (WS-RCNN). The basic idea is to deploy a deep network to learn to score proposals, under the special setting of weak supervision. To tackle the key issue of acquiring proposal-level pseudo labels for model training, we propose a so-called Attention-Guided Pseudo Labeling (AGPL) strategy, which leverages the local maximal (peaks) in image-level attention maps and the spatial relationship among peaks and proposals to infer pseudo labels. We also suggest a novel training loss, called Entropic OpenSet Loss, to handle background proposals more effectively so as to further improve the robustness. Comprehensive experiments on two standard benchmarking datasets demonstrate that the proposed WS-RCNN can outperform the state-of-the-art by a large margin, with an improvement of 11.6 % on PASCAL VOC 2012 and 10.7 % on MS COCO 2014 in terms of mAP 50 , which indicates that learning-based proposal scoring and the proposed WS-RCNN framework might be a promising way towards WSIS.

1. Introduction

Instance segmentation [1,2,3,4] refers to the task of jointly localizing, categorizing and segmenting the spatial extents of individual visual objects from a given image. Like many other computer vision tasks, remarkable progress has recently been made on instance segmentation driven by the prosperity of convolutional neural networks (CNN) [2,4,5,6,7]. Nevertheless, CNN-based solutions to vision tasks commonly suffer from the data-hungry nature, i.e., the necessity of a large amount of annotated data for training. This may be particularly infeasible for instance segmentation since pixel-wise annotations at the instance level are extremely labor-intensive. Weakly supervised instance segmentation (WSIS)  [8,9] is one possible way to alleviate the dependency on such strong annotations, which aims to achieve instance segmentation by the use of weaker and thus less labor-intensive annotations [10,11,12], ideally image-level labels only [8] as we are concerned with in the present work.
There recently emerge a couple of attempts on WSIS with image-level labels in the literature [8,9,13,14,15], which can be typically outlined as three major steps: (1) proposal generation, i.e., a number of class-agonistic segment proposals are generated from the given image; (2) proposal scoring, i.e., classification scores are assigned to the proposals; (3) postprocessing, i.e., final results are retrieved from the scored proposals by using non-maximal suppression or certain postprocessing procedures (e.g., applying Mask R-CNN for refinement [14]). It has been well-established that CNN classifiers trained globally at the image level have remarkable ability of spatial localization [16,17], and  the so-called attention maps of a certain form (e.g., Class Activation Maps [18], Occlusion Maps [19], Saliency Maps [20], Excitation Backprop [17]) are utilized to represent such localization cues, where the intensity stands for the possibility of spatial occurrence of visual objects. Taking advantage of this fact, previous works on WSIS mainly put their efforts on the step of proposal scoring, i.e., how to infer reasonable classification scores for the proposals from global attention maps. As a pioneering work on WSIS, Zhou et al. [8] proposed Peak Response Map (PRM) to boost the instance localization ability of CNNs, which then enables better proposal scoring. Zhu et al. [13] presented the Instance Activation Map which aims to enhance the spatial extents of instances in PRM and consequently improves proposal scoring. Ahn et al. [14] deployed the Inter-pixel Relation Network to learn a high-quality proposal generator, as well as pairwise semantic affinities which guide the proposal scoring.
While these wisdoms of making improvements on proposal scoring (or other particular components) can indeed benefit to an extent, previous works on WSIS share one major limitation that they commonly follow a heuristic way to exploit attention maps for proposal scoring, lacking of a unified framework. Concretely, they utilize certain attention maps and hand-crafted scoring rules to assign classification scores to proposals [8,13]. In this way, the attention maps are expected to be aware of the spatial extents of visual objects, such that they can match the shapes of true objects and consequently assign them with high scores. On the other hand, attention maps can only sparsely highlight some sites discriminative for classification, usually at very coarse spatial solutions and unaware of object extents. Such discrepancy explains why previous works all devote their efforts to enhancing the perception of object extents of attention maps. Nevertheless, this is by nature a rather difficult perceptual grouping task (see Figure 1), which may limit the substantial advances on this topic. These observations motivate us to consider if these exist a simpler but more effective way towards WSIS.
In this paper, we introduce a general framework for weakly supervised instance segmentation, called Weakly Supervised R-CNN (WS-RCNN). The underlying idea is very simple and natural: Instead of relying on heuristic strategies for proposal scoring, we deploy a deep network to learn to score proposals. For this purpose, there exists an inherent challenge, i.e., how to acquire proposal-level pseudo labels to enable the learning since only image-level labels are available in our problem setting (the pseudo labeling issue)? This is actually a key issue for any weakly supervised learning problem including ours. In addition, an appropriate training loss function is also well worthy of exploration in our learning task. We propose specific solutions to these key issues and further design a unified framework for WSIS. Since the resulting framework can be conceptually interpreted as a Fast R-CNN model [21], which is a representative model of general object detection among the R-CNN family [2,21,22,23], under the particluar setting of weak supervision, we name our framework as Weakly Supervised R-CNN (WS-RCNN).
As another major contribution, we propose an effective solution to this pseudo labeling issue, termed as Attention-Guided Pseudo Labeling (AGPL). AGPL also utilizes attention maps, however it relies only on local maxima (peaks) in attention maps and the spatial relationship among peaks and proposal regions. Specifically, it admits object instances of a target class by inspecting the positional inclusion relationship between proposals and peaks in the attention maps associated with that particular class (see Section 3.3 for details). The key insight lies in that AGPL leverages very weak information, instead of necessitating to take into account object extents like existing approaches [8,9,13,14], which makes the pseudo labeling procedure simple but robust. We stress that such careful design of AGPL is crucial to the success of the whole WS-RCNN framework (see Section 4.3 for experimental validation).
As our third contribution, we suggest a novel loss function for model training to further boost the performance, called Entropic Open-Set (EOS) Loss. As our method is proposal-based, the proposal generator will inevitably yield a portion of background proposals (those belonging to none of the concerned classes) even with the state-of-the-art method. How to make our model conscious of background proposals to avoid misclassification is important to its robustness. This is by nature an open-set recognition problem [24,25], which has been well studied in the context of robust pattern recognition. We propose to adapt the state-of-the-art method for CNN-based open-set recognition [25] to our background handling. To our knowledge, this is the work which considers the open-set issue in object detection/instance segmentation.
We will carry out comprehensive experiments on standard benchmarking datasets to demonstrate that WS-RCNN can push forward the state-of-the-art on WSIS by a remarkable margin, with an improvement of 11.6 % on PASCAL VOC 2012 and 10.7 % on MS COCO in terms of mAP 50 (see Section 4.2 for details). Such overwhelming superiority in performance suggests that WS-RCNN might be a more appropriate framework for WSIS worthy of further exploration than existing ones which focus on deducing object extents from attention maps for proposal scoring. To sum up, the main contributions of our work are as below:
  • We propose a novel framework called Weakly Supervised R-CNN (WS-RCNN) for weakly supervised instance segmentation. The key insight is to deploy a deep network to learn to score proposals, instead of relying on heuristic proposal scoring strategies, which may provide a new perspective for future exploration on WSIS.
  • We propose a simple but effective strategy for inferring pseudo labels from attention maps, called Attention-Guided Pseudo Labeling (AGPL).
  • We introduce an Entropic Open-Set (EOS) Loss for handling background proposals in model training to further boost the performance.
The remainder of this paper is organized as follows: Section 2 briefly reviews related literature. Section 3 details the proposed WS-RCNN framework, including the AGPL scheme for pseudo labeling and the EOS loss. Section 4 focuses experiments, including comparison, ablation study and related analysis. In Section 5, we make a conclusion and give some remarks on future work.

2. Related Work

In this section, we present a brief review of the literature that are closely related to ours, including those on weakly supervised instance segmentation, weakly supervised semantic segmentation, weakly supervised object detection and open-set recognition and background handling.

2.1. Weakly Supervised Instance Segmentation

Instance segmentation aims to produce accurate masks that can distinguish between instances of specific object classes. It generally requires pixel-wise object masks to train an instance segmentation model, which are extremely expensive. Weakly supervised instance segmentation (WSIS) devotes to conquering this challenge by using certain forms of weaker annotations, most ideally image-level labels only as concerned in this paper. Along the line of WSIS using image-level annotations, PRM [8] is a pioneering work which treats peaks in class response maps obtained by CNN classifiers as indicators of the existence of object instances, and introduces the so-called Peak Response Map (PRM) to score object proposals so as to identify true object instances. In [26], the results obtained by PRM are used as pseudo annotations to train a Mask R-CNN [2] for further refinement. Zhu et al. [13] proposed the Instance Activation Map to enhance PRM [8] by taking better care of spatial extents of objects. Ahn et al. [14] presented the Inter-pixel Relation Network to learn a good proposal generator, and also inter-pixel connections which can guide proposal scoring. In [9], the authors collaboratively combined weakly supervised object detection with WSIS under a unified framework of course learning. As previously stated, WSIS generally remains at the early stage of exploration, and the few existing approaches mainly focus on figuring out the shapes of object instances from attention maps for better proposal scoring, which faces a very difficult perceptual grouping task.

2.2. Weakly Supervised Semantic Segmentation

The mainstream paradigm for weakly supervised semantic segmentation (WSSS) can be summarized as to derive pseudo semantic masks to enable the training of a model of fully supervised semantic segmentation (most typically FCN [27]), and existing approaches can be distinguished by how they acquire the pseudo masks. In [28], discriminative regions are selected from CAMs [18] based on three principles, called “seed, expand and constrain”, as pseudo labels to supervise a segmentation network. Roy et al. [29] utilized a novel deep architecture which fuses bottom-up, top-down, and smoothness cues to acquire pseudo masks. Other representative strategies for enhancing the quality of pseudo masks include the LSE pooling [30], the EM algorithm [31], seeded region growing [32], semantic affinity [33], dilated convolution [34], the anti-erase strategy [35,36], similar region mining [37], the self-erasing strategy [38], visual saliency [39,40], etc. Although WSSS is another task different from WSIS, our work was somewhat inspired by the paradigm of inferring pseudo labels to train fully supervised models.

2.3. Weakly Supervised Object Detection

The task of weakly supervised object detection (WSOD) is similiar to WSIS except for the necessity of yielding instance segmentation masks. Earlier works on WSOD mostly follow the pipeline of multiple instance learning [41,42,43]. Recently CNN-based approaches have attracted more and more research interest. As a milestone work along this direction, WSDDN [44] adopts a two-branch network where softmax operations are performed over proposals and classes respectively in the two branches, and the obtained classification scores are synthesized into an image-level score so as to establish the training loss. Diba et al. [45] introduced a novel cascaded network for WSOD, which performs ROI pooling at multiple levels to boost the performance. Ref. [46] proposes a weakly supervised region proposal network which is trained using only image-level annotations and the proposed region proposal network can be plugged into a WSOD network easily. In [47], Tang et al. introduced the Proposal Clustering Learning which generates proposal clusters to learn refined instance classifiers by an iterative process. Other successful approaches to WSOD include the Min-Entropy Latent Model [48], the  Category-Aware Spatial Constraint [49], etc. While the tasks are essentially different, these works on WSOD are similar to ours in being proposal-based.

2.4. Open-Set Recognition and Background Handling

Traditional recognition systems obey the closed-set assumption, i.e., the training and testing data are drawn from the the same set of semantic classes. However, a more realistic scenario is that data from unseen classes can emerge unexpectedly during testing, which may drastically decrease the robustness of the system. This is well-known as the open-set issue in pattern recognition, which has attracted a lot of research interests [24,25,50]. In two-stage object detectors (like FRCNN [21] and our WS-RCNN), the proposals obtained by the proposal generator will inevitably include a portion of proposals from image background (unseen classes to the model), which is by nature an open-set recognition problem [24]. But surprisingly, no previous authors have addressed this problem from the perspective of open-set recognition to our knowledge. Instead, they usually perform background handling in certain ad-hoc strategies, e.g., adding a dummy background class to the model.

3. Approach

In this section, we start with the problem statement of WSIS, followed by describing the proposed WS-RCNN framework. Then we detail the key components, including the Attention-Guided Pseudo Labeling and the Entropic Open-Set Loss. Finally, we present some remarks.

3.1. Problem Statement

Given a set of classes of interest C = { 1 , , C } and a training set I = { ( I k , y k ) } k = 1 K , where I k is an image and y k { 0 , 1 } C × 1 is the corresponding multi-class label vector, the task of weakly supervised instance segmentation (WSIS) in our work can be roughly stated as to segment, for an input testing image, all the object instances belonging to the classes C . Such a problem setting differs intrinsically from general instance segmentation [2,4] in that no pixel-wise instance annotations but only image-level labels are available for model establishment, which makes the task very challenging.
Like general instance segmentation, WSIS can also follow a proposed-based paradigm [8,13,33], which can be epitomized as a three-step pipeline as aforementioned. These approaches can then be viewed as to retrieve true object instances from a pool of proposals according to the assigned scores, central to which is proposal scoring, i.e., how to appropriately assign classification scores to proposals. One commonly-used strategy for proposal scoring is to make use of the well-established localization ability of CNNs [8,16,18]. Specifically, the training set I with image-level labels are firstly taken to train an image-level CNN classifier, from which a collection of class-specific attention maps are derived to assign classification scores to the proposals. For this purpose, it is desired that these attention maps can preserve object shapes, which is however a difficult perceptual grouping task. In addition, the hand-crafted scoring rules adopted by existing methods are also limited as well. These facts motivates us to propose the WS-RCNN framework.

3.2. The Proposed WS-RCNN Framework

The basic idea of WS-RCNN is to deploy a deep network to learn to score proposals under the special setting of weak supervision, instead of relying on heuristic proposal scoring strategies. To achieve this goal, one major obstacle is the absence of proposal-level labels necessitated for training. To conquer this challenge, we develop an effective strategy, called Attention-Guided Pseudo Labeling (AGPL), to take advantage of the attention maps associated with the image-level CNN classifier to infer proposal-level pseudo labels. Furthermore, we introduce an Entropic Open-Set Loss (EOSL) to handle the background issue in training to further improve the robustness of our framework. In the following, we will first present an overview of WS-RCNN, followed by detailing the AGPL stategy and the EOSL loss.
Network Architecture: The overall network architecture of WS-RCNN is shown in Figure 2. Following the notations above, the input image I sized by H I × W I is first fed into a proposal generator (using the off-the-shelf method [51,52] in our implementation) to obtain the segment proposals { R n } n = 1 N , where each R n is an H I × W I binary mask representing a segment proposal with arbitrary shapes (rather than regular bounding-boxes). The image then goes through a backbone CNN for feature extraction, yielding the feature maps F R H × W × M , where H × W is the size and M the number of the feature maps. Afterwards, the network bifurcates into two branches, i.e., the proposal scoring branch and the pseudo labeling branch. Notice that these two branches share the same backbone CNN.
In the proposal scoring branch, the features corresponding to each individual proposal are extracted. A standard operation for this task is RoIAlign [2], widely used in two-stage object detectors, which however cannot be directly applied to our case since it is designed for bounding-box proposals. Therefore, we modify RoIAlign to adapt to segment proposals, resulting in the SegAlign operation (see details below). For each proposal R n , the corresponding features can be extracted from F and aligned to a canonical grid via SegAlign, denoted by f n R h × w × M (we use h = w = 7; M = 512 in this paper), which is followed by three fully-connected layers (FCs, with the node numbers being 4096, 4096 and C respectively) and a softmax layer to get the proposal-level classification score x n R C × 1 .
The pseudo labeling branch is executed for training only, where the feature maps F are followed by an image-level classifier. Then, a set of class-specific attention maps, denoted by { M c } c = 1 C , are extracted from this classifier, where each M c reflects the spatial probability of occurrence of object instances belonging to the class c. Among possible choices of attention maps, we adopt the Class Peak Responses [8] in our implementation due to its excellent localization ability. These attention maps (as well as the image-level label y ) are then utilized to infer the proposal-level pseudo class labels { z n } n = 1 N , where z n R C × 1 is a one-hot vector z n 0 C × 1 standing for the background class), by the use of AGPL.
Training Strategy: We adopt a two-phase training strategy to train the WS-RCNN model. In the first phase, we train the image-level classifier in the pseudo labeling branch, which is initialized by the model pre-trained on ImageNet. Proposal-level pseudo labels are then inferred from the trained imagelevel classifier using AGPL. In the second phase, we train the proposal scoring branch, where the backbone CNN is reinitialized with the model pre-trained on ImageNet. We will validate the effectiveness of this two-phase training strategy by comparative experiments in Section 4.3. Notice that since there usually exist significantly more background proposals than target-class ones after pseudo labeling, we always make their numbers identical by uniformly sampling background proposals.
Training Loss: For the training of the image-level classifier (the first-phase training), we use the given image-level labels { y k } k = 1 K and the conventional cross-entropy loss function for multi-label classification to establish the training loss.
For the training of the proposal scoring branch (the secondphase training), suppose for the image I k labeled with y k in the training set I , the proposals obtained are { R n k } n = 1 N . For each R n k , let us denote by x n k the classification score predicted by the proposal scoring branch, and by z n k n the pseudo class label inferred by the pseudo labeling branch using AGPL. Given all these, the training loss for the proposal scoring stream can be established by
L p = 1 K k = 1 K 1 N n = 1 N EOSL ( x n k , z n k ) ,
where EOSL is the proposed Entropic Open-Set Loss (see details in Section 3.4).
SegAlign: As shown in Figure 3, following the notations above, suppose for a segment mask R in the W I × H I image I , the corresponding receptive field mapped to the feature maps F R H × W × M is R F . SegAlign extracts from F the features corresponding to R F and maps them to canonical feature maps f R h × w × M , which is basically a modified RoIAlign to adapt to segment masks. Concretely, suppose R F is bounded by the rectangle B , φ is the bilinear transform from the spatial coordinates ( i , j ) f to ( i , j ) B , i.e.,  ( i , j ) = φ ( i , j ) , and  g is the bilinear interpolation function over F . The SegAlign operation can then be defined by
f ( i , j ) = g ( F , φ ( i , j ) ) , if φ ( i , j ) R F , 0 , otherwise .
Note we drop the channel dimension of feature maps above without loss of clarity.

3.3. Attention-Guided Pseudo Labeling

AGPL leverages the localization ability of CNNs and the spatial relationship among proposals to achieve pseudo labeling. As shown in Figure 4, for the image I I , given the class label vector y = [ y 1 , , y C ] T , the segment proposals { R n } n = 1 N and the class-specific attention maps { M c } c = 1 C , AGPL can be outlined as follows:
(1) For each target class c (with y c = 1 ), all the local maxima (peaks) are identified from M c , denoted as { p i c } i = 1 m c , where p i c R 2 stands for pixel coordinates and m c the number of peaks. For each p i c , we pick up all the proposals spatially including this point, which are further averaged and thresholded to get a support mask S i c as follows
A i c = 1 m i c { n | p i c R n } R n ,
( S i c ) p q = ( A i c ) p q , if ( A i c ) p q > β , 0 , otherwise .
where m i c is the number of picked proposals corresponding to p i c , p and q are pixel indices, and the threshold β [ 0 , 1 ] is a parameter (we adopt β = 0.7 in our implementation. See Section 4.4 for parameter study). The resulting peaks { p i c } i = 1 m c and associated support masks { S i c } i = 1 m c are then utilized to admit proposals belonging to the class c.
(2) Sort all the peaks { p i c | c = 1 , , C ; i = 1 , , m c } in the descending order of their values in the attention maps, i.e.,  { M c ( p i c ) } . Then, for each ordered peak p i c and the associated S i c , those proposals which overlap sufficiently with S i c are labeled as the class c, i.e.,  z n = c if
IoU ( S i c , R n ) > 0.5 ,
where IoU stands for the Intersection-over-Union operation. Notice that one proposal is allowed to be exclusively assigned to one class only during the ordered labeling. For clarity, we summarize the AGPL algorithm above in Algorithm 1.

3.4. Entropic Open-Set Loss

Since our WS-RCNN is proposal-based, the proposals after pseudo labeling will unavoidably contain some background proposals, i.e., those labeled as none of the target classes. It is necessary to handle these background proposals in model training, otherwise a model trained only with samples from target classes will be distracted by the unseen background proposals in testing, degrading its robustness. A natural solution is to add a dummy class into the model to accommodate background proposals. However, since the background class is a class of “stuff”, its variance is so large that it is hard to be modeled by any single class.
Algorithm 1 Attention-Guided Pseudo Labeling (AGPL)
 Input: 
The label vector y = [ y 1 , , y C ] { 0 , 1 } C × 1 , the segment proposals { R n } n = 1 N and the class-specific attention maps { M c } c = 1 C (associated with the image I ); the parameter  β .
 Output: 
The one-hot pseudo class labels { z n } n = 1 N for the proposals.
1:
Initialize z n = 0 , Q .
2:
for allc with y c = 1 (target class) do
3:
  Find the local maxima { p i c } i = 1 m c in M c ;
4:
  for  i = 1 , , m c  do
5:
    Calculate the support mask S i c using Equations (3) and (4);
6:
     Q ( p i c , S i c ) ;
7:
  end for
8:
end for
9:
Sort Q in the descending order of the values { M c ( p i c ) } , denoted by Q ˜ the sorted set.
10:
L { 1 , , N } ;
11:
for all ( p i c , S i c ) Q ˜ do
12:
  Find the proposals indexed by I L satisfying Equation (5);
13:
  Set ( z n ) c = 1 , n I ;
14:
   L L \ I ;
15:
end for
To address this issue, we observe that the task of background handling here is by nature an open-set recognition (OSR) problem, which has been well studied in robust pattern recognition. Hence, we propose to introduce the Entropic Open-Set Loss (OSEL), which is a representative method for OSR [25], to address our background handling problem. The basic idea of OSEL is to treat the samples from target and background classes separately in establishing the training loss. For target classes, the standard cross-entropy loss is used, while for the background class, an entropic loss is used to encourage predicting uniformly-distributed classification scores. Since the C scores sum up to 1 (output by the softmax layer), encouraging uniform distribution on background class will make these scores small and therefore suppressed during the Non-maximal Suppression (NMS) procedure. Formally, suppose the predicted score vector of a proposal is x = [ x 1 , , x C ] R C × 1 and the corresponding one-hot pseudo label vector is z = [ z 1 , , z C ] { 0 , 1 } C × 1 , the EOSL is defined by
EOSL ( x , z ) = z c log x c , if z c 0 , i = 1 C log x i , if z = 0 .
To our knowledge, this is the first work which addresses the background handling problem in object detection/instance segmentation from the perspective of open-set recognition.

3.5. Remarks

Despite the proposed WS-RCNN is very natural and simple, we will later experimentally demonstrate that it has overwhelming superiority in performance over the state-of-the-art methods, which is very likely due to two key insights of our approach. First, we adopt a learning-based approach to proposal scoring, which is advantageous in being able to learn to directly map feature representations of proposals to classification scores, rather than depending on hand-crafted scoring rules. Second, AGPL only utilizes very loose information to achieve pseudo labeling, i.e., some sparse points (peaks in attention maps) and their spatial relationship with proposal regions, which makes our approach much easier and thereby more robust to complex scenes in comparison with existing works that involve the consideration of spatial extents of objects. The proposed WSRCNN framework can be simply interpreted as “learning to score proposal under weak supervision”, which may provide a promising new perspective for addressing WSIS.

4. Experiments

In this section, we perform extensive experiments to evaluate the effectiveness of the proposed WS-RCNN, mainly including: (1) comparison with the state-of-the-art WSIS methods, both quantitatively and qualitatively; (2) validation of the effectiveness of some key components by comparing with variant baselines; (3) parameter study; (4) analysis on failure patterns. Our method was implemented in PyTorch on a workstation with 2 Nvidia Titan XP GPUs, Intel Core(TM) i7-8700 3.70 GHz CPU, 32 GB RAM and Ubuntu 18.04 OS.

4.1. Experimental Setup

4.1.1. Datasets

Two well-known benchmarking datasets for instance segmentation, namely PASCAL VOC 2012 (termed as VOC) [53] and Microsoft COCO 2014 (termed as COCO) [54] are adopted for our experiments throughout this paper. VOC may be the most representative one for evaluating WSIS methods [8,13,33], which includes 10,582 images for training (trainset) and 1449 images for validation (valset) from 20 object classes. COCO is a much larger and more challenging dataset, including 82,783 images for training (trainset) and 40,775 images for validation (valset) from 80 object classes. This dataset has rarely been utilized for the task of WSIS before, and we consider it to enable more comprehensive evaluation. For both datasets, we take trainset for training and valset for testing, and no other annotations except for image-level class labels are used for training according to our problem definition of WSIS.

4.1.2. Performance Metrics

We adopt m A P r , the most commonly-used performance metric for instance segmentation [8,13,33,53], for quantitative evaluation and comparison, where r is the IoU threshold utilized to calculate the metric [53,55]. In our experiments on WSIS, we use m A P 50 ( r = 0.5 ) as the major metric for comparison and analysis, but we also report m A P 25 ( r = 0.25 ) and m A P 75 ( r = 0.75 ) for more in-depth evaluation.

4.1.3. Implementation Details

We use VGG-16 [56] pre-trained on ImageNet as the backbone for the proposed WS-RCNN. The number of segment proposals per image is set to be N = 200 following [8,13]. For training WS-RCNN, the SGD optimizer is used with a initial learning rate of 5 × 10 4 for the first 3.5 × 10 4 iterations. In the following 10 4 iterations, the learning rate decreases to 5 × 10 5 . For data augmentation, we use five image scales { 480 , 576 , 688 , 864 , 1200 } (for the shorter side) with horizontal flips for both training and testing.

4.2. Comparison with State-of-the-Art

The study on WSIS is still at its early stage and there have not been many works so far. We consider four state-of-the-art WSIS methods in the literature for our comparative study, termed as PRM [8], IAM [13], Label-PEnet [9] and IRnet [14] respectively. For PRM [8] and IRnet [14], we use the source codes as well as configurations provided by the authors themselves. For IAM [13] and Label-PEnet [9], since no source codes are publicly released, we directly cite the results in the original literature [9,13] wherever available (marked with “-” if the results are unavailable). Besides, we also construct a variant of WSDDN [44], which is a very impactful method for weakly supervised object detection (rather than instance segmentation), termed as WSDDN-seg. For this sake, WSDDN-seg can be easily adapted to our task utilizes segments as the proposals to replace the original bounding-boxes in WSDDN (while keeping everything else unchanged). We consider WSDDN-seg because this method is spiritually similar to ours in that it also deploys a deep network to learn to score proposals. For fair comparison, we use the same method [51,52] to generate segment proposals for all the compared methods except IRnet [14], because the method focuses on improving the component of proposal generation. In addition, some methods perform post-processing (typically taking the obtained instance masks as pseudo annotations to run Mask R-CNN like [14]) to further refine the results while others do not. For fair and comprehensive comparison, we consider the two settings separately, i.e., without and with (marked by “+p”) using Mask R-CNN for refinement.

4.2.1. Results on VOC

The quantitative results obtained by various methods on VOC are reported in Table 1. As can be observed, our WS-RCNN outperforms the state-of-the-art (IRnet [14]) by 11.6 % in terms of the major metric m A P 50 , which also achieves the best performance in terms of other metrics, either without or with refinement. Relative to the pioneering work of PRM [8], all the other methods make an improvement to an extent by enhancing certain aspects, including localization maps (IAM [13] and IRnet [14]), proposal generation (IAM [13]) and combination with object detection (Label-PEnet [9]). Comparatively, our WS-RCNN is able to improve by a far larger margin, demonstrating its effectiveness. Interestingly, one can observe that WSDDN-seg [44] can manage to achieve comparable or even better results than PRM [8]. Notice that WSDDN-seg exploits a CNN-based model to simply learn to score proposals (like ours), even without using any localization map at all. This may suggest that proposal scoring is indeed critical to WSIS with a large room for improvement, and “learning to score” is a promising strategy worthy of further exploration. We also show some comparisons of instance segmentation performance under different supervision on VOC in Table 2, including SDI [10], Mask R-CNN [2] and our WS-RCNN. SDI [10] uses bounding box supervision and Mask R-CNN [2] uses full supervision.
To intuitively justify the motivation and merits of WS-RCNN, we comparatively visualize in Figure 5 two sets of representative intermediate results, namely, the scores of a same set of proposals acquired by PRM [8], WSDDN-seg [44] and our WS-RCNN (shown in the descending order of the proposal scores). One can observe that WSDDN-seg [44] tends to highlight one dominating instance while ignoring the others of the same class, and PRM [8] tends to highly score object parts or adjoining objects undesirably. Comparatively, our method can get more favorable proposal scores, which we argue is likely due to our advantageous learning mechanism. More representative results obtained by the various methods are further visualized in Figure 6.

4.2.2. Results on COCO

Table 3 shows the quantitative results on the COCO dataset. Notice that COCO has not been considered in previous works compared, and the source codes of IAM [13] and Label-PEnet [9] are not publicly released, so the results for these two methods are unavailable. This dataset is far more difficult than VOC, as can be seen by the much worse overall performance of all the methods. However, our WS-RCNN can still outperform the compared methods, with a remarkable margin of 10.7 % in terms of m A P 50 over the state-of-the-art (IRnet [14]), and 4.6 % over WSDDN-seg constructed by ourselves. The relative performance of the compared methods in terms of other metric remains consistent with those on VOC, which further verifies the effectiveness of our approach. Some typical results are presented in Figure 7, which further demonstrate that our WS-RCNN can obtain more favorable instance masks, especially in case of multiple or adjoining instances.

4.3. Validation of Key Components

The superiority of the proposed WS-RCNN, to a large extent, should be attributed to the considerate design of its key components. We further carry out experiments on the VOC dataset to validate this point. For each component of concern, we construct some variants to replace the original one in WS-RCNN while keeping everything else unchanged, and compare the overall performance in terms of m A P 50 .

4.3.1. Pseudo Labeling Strategy

We consider the following three variants of AGPL:
  • PLS-1: For each p i c we assign those proposals which spatially include this point to the class c, i.e., assigning z n = c if p i c R n , without considering the support mask S i c (removing Line 5-8 in Algorithm 1 and meanwhile changing the condition in Equation (5) to be “ p i c R n ” ).
  • PLS-2: For each p i c we thresold M c (using the threshold value of 0.5 ) and take the connected component surrounding this peak as the support mask S i c (changing Line 6 in Algorithm 1).
  • PLS-3: We simply adopt the method of PRM [8] for pseudo labeling, i.e., taking the proposal classification results obtained by the whole pipeline in [8] as the pseudo labels (changing Line 6 in Algorithm 1).
The results are reported in Table 4, which shows all these variants significantly underperform the proposed AGPL. The results validate the merit of AGPL, i.e., relying on peaks in attention maps and the spatial relationship between peak points and proposal regions to achieve pseudo labeling.

4.3.2. Training Strategy

In our WS-RCNN, we adopt a particular two-phase strategy for network training (see Section 3.2. To verify the effectiveness of such a design, we construct two different training strategies for comparative study as follows:
  • TS-coupled: When training the proposal scoring branch in the second phase, the backbone CNN is not reinitialized by the parameters pre-trained on ImageNet, but by those obtained during the training of the image-level classifier in the first phase.
  • TS-joint: The image-level classifier is trained first, and the proposal scoring branch is trained jointly with the image-level classifier, i.e., combining the two training losses when training this branch.
The results in Table 5 demonstrate that the training strategy adopted by our WS-RCNN are more advantageous than the variants compared.

4.3.3. Training Loss

To validate the effectiveness of our proposed EOSL loss, we compare it with a widely-used strategy for background handling as below
  • LOSS-CE: We add a dummy class to accommodate background proposals and adopt the conventional binary cross-entropy to replace the EOSL loss in Equation (2).
As can be observed from Table 6 that our EOSL loss performs much better.

4.4. Parameter Study

Our approach involves specifying two parameters, i.e., the number of proposals N and the threshold value β in AGPL. Here we further conduct experiments to analyze the impact of these parameters on performance.

4.4.1. The Number of Proposals N

We vary the number of proposals to be N = { 50 , 100 , 150 , 200 , 300 , 400 , 500 } , and report the results in Figure 8a. As can be seen, the performance will not increase significantly after N > 200 , and we adopt N = 200 throughout our experiments for the best tradeoff between performance and efficiency.

4.4.2. The Threshold β in AGPL

We vary the threshold value to be β = { 0.4 , 0.5 , 0.6 , 0.7 , 0.8 , 0.9 } , and the results are depicted in Figure 8b. One can observe that the performance is relatively stable when β [ 0.5 , 0.8 ] and we choose a fixed value of β = 0.7 throughout our experiments.

4.5. Failure Cases

Despite the effectiveness of WS-RCNN, WSIS is essentially a challenging task and the overall performance still has much room for improvement. Here we show several typical failure cases of WS-RCNN in Figure 9. Since our method is proposal-based, it relies much on the quality of proposals. If the proposal generator fails to cover the spatial extents of the true instance, our method will fail consequently (see Figure 9a). Another typical failure case is that if there are a number of instances of a class with overlap among or close to each other, our method may pick up those large proposals which cover multiple instances (see Figure 9b) or the small proposals which cover only a part of a instance (Figure 9c) (Notice that the desired proposals are present in (b) and (c).

5. Conclusions

In this paper, we have presented a simple, natural but surprisingly effective framework, termed as Weakly Supervised R-CNN (WS-RCNN), for weakly supervised instance segmentation (WSIS). The basic idea is to deploy a deep network to learn to score proposals under the particular setting of weak supervision. For this sake, a strategy called Attention-Guided Pseudo Labeling is proposed to address the key issue of proposal-level pseudo labeling. And a so-called Entropic Open-Set Loss is introduced for model training to further improve the robustness. Comprehensive experiments on two well-known datasets, i.e., PASCAL VOC 2012 and Microsoft COCO 2014, have demonstrated that the proposed WS-RCNN can significantly outperform the state-of-the-art. Experiments have also been carried out to validate effectiveness of the key components and to study the impacts of some key parameters. In our future work, we will consider integrating a learning-based proposal network to replace the current heuristic proposal generator in WS-RCNN in order to alleviate its dependency on the quality of proposals. We will also explore how to further improve our approach in dealing with complex scenes (e.g., complex background, multiple or adjoining instances of the same class) since there is generally much room for improvement.

Author Contributions

Methodology, J.-R.O.; writing—original draft preparation, J.-R.O., S.-L.D.; writing—review and editing, J.-R.O., S.-L.D., J.-G.Y.; project administration, J.-G.Y.; funding acquisition, J.-G.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Key R&D Program of Guangdong Province under Grant 2018B030339001 and National Natural Science Foundation of China under Grants 62076099, and also in part by the Guangzhou Science and Technology Program under Grant 201904010299.

Institutional Review Board Statement

The study did not involve humans or animals.

Informed Consent Statement

The study did not involve humans.

Data Availability Statement

The study did not report any data.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Hariharan, B.; Arbeláez, P.; Girshick, R.; Malik, J. Simultaneous detection and segmentation. In Proceedings of the European Conference on Computer Vision, Zurich, Switzerland, 6–12 September 2014; pp. 297–312. [Google Scholar]
  2. He, K.; Gkioxari, G.; Dollár, P.; Girshick, R. Mask r-cnn. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 Ocotber 2017; pp. 2961–2969. [Google Scholar]
  3. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You only look once: Unified, real-time object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar]
  4. Liu, S.; Qi, L.; Qin, H.; Shi, J.; Jia, J. Path aggregation network for instance segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 8759–8768. [Google Scholar]
  5. Chen, L.C.; Hermans, A.; Papandreou, G.; Schroff, F.; Wang, P.; Adam, H. Masklab: Instance segmentation by refining object detection with semantic and direction features. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 4013–4022. [Google Scholar]
  6. Chen, K.; Pang, J.; Wang, J.; Xiong, Y.; Li, X.; Sun, S.; Feng, W.; Liu, Z.; Shi, J.; Ouyang, W.; et al. Hybrid task cascade for instance segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019; pp. 4974–4983. [Google Scholar]
  7. Jeong, D.; Kim, B.G.; Dong, S.Y. Deep joint spatiotemporal network (DJSTN) for efficient facial expression recognition. Sensors 2020, 20, 1936. [Google Scholar] [CrossRef] [Green Version]
  8. Zhou, Y.; Zhu, Y.; Ye, Q.; Qiu, Q.; Jiao, J. Weakly supervised instance segmentation using class peak response. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 3791–3800. [Google Scholar]
  9. Ge, W.; Guo, S.; Huang, W.; Scott, M.R. Label-PEnet: Sequential label propagation and enhancement Networks for Weakly Supervised Instance Segmentation. In Proceedings of the IEEE International Conference on Computer Vision, Long Beach, CA, USA, 16–20 June 2019; pp. 3345–3354. [Google Scholar]
  10. Khoreva, A.; Benenson, R.; Hosang, J.; Hein, M.; Schiele, B. Simple does it: Weakly supervised instance and semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Venice, Italy, 22–29 Ocotber 2017; pp. 876–885. [Google Scholar]
  11. Remez, T.; Huang, J.; Brown, M. Learning to segment via cut-and-paste. In Proceedings of the European Conference on Computer Vision, Salt Lake City, UT, USA, 18–23 June 2018; pp. 37–52. [Google Scholar]
  12. Kuo, W.; Angelova, A.; Malik, J.; Lin, T.Y. Shapemask: Learning to segment novel objects by refining shape priors. In Proceedings of the IEEE International Conference on Computer Vision, Long Beach, CA, USA, 16–20 June 2019; pp. 9207–9216. [Google Scholar]
  13. Zhu, Y.; Zhou, Y.; Xu, H.; Ye, Q.; Doermann, D.; Jiao, J. Learning instance activation maps for weakly supervised instance segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019; pp. 3116–3125. [Google Scholar]
  14. Ahn, J.; Cho, S.; Kwak, S. Weakly supervised learning of instance segmentation with inter-pixel relations. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019; pp. 2209–2218. [Google Scholar]
  15. Arun, A.; Jawahar, C.; Kumar, M.P. Weakly supervised instance segmentation by learning annotation consistent instances. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; pp. 254–270. [Google Scholar]
  16. Zhou, B.; Khosla, A.; Lapedriza, A.; Oliva, A.; Torralba, A. Object detectors emerge in deep scene cnns. arXiv 2014, arXiv:1412.6856. [Google Scholar]
  17. Zhang, J.; Bargal, S.A.; Lin, Z.; Brandt, J.; Shen, X.; Sclaroff, S. Top-Down Neural Attention by Excitation Backprop. Int. J. Comput. Vis. 2018, 126, 1084–1102. [Google Scholar] [CrossRef] [Green Version]
  18. Zhou, B.; Khosla, A.; Lapedriza, A.; Oliva, A.; Torralba, A. Learning deep features for discriminative localization. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 2921–2929. [Google Scholar]
  19. Zeiler, M.D.; Fergus, R. Visualizing and understanding convolutional networks. In Proceedings of the European Conference on Computer Vision, Zurich, Switzerland, 6–12 September 2014; pp. 818–833. [Google Scholar]
  20. Simonyan, K.; Vedaldi, A.; Zisserman, A. Deep inside convolutional networks: Visualising image classification models and saliency maps. arXiv 2013, arXiv:1312.6034. [Google Scholar]
  21. Girshick, R. Fast r-cnn. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 13–16 December 2015; pp. 1440–1448. [Google Scholar]
  22. Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Rich feature hierarchies for accurate object detection and semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Zurich, Switzerland, 6–12 September 2014; pp. 580–587. [Google Scholar]
  23. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster r-cnn: Towards real-time object detection with region proposal networks. In Proceedings of the Advances in Neural Information Processing Systems, Montreal, QC, Canada, 7–12 December 2015; pp. 91–99. [Google Scholar]
  24. Geng, C.; Huang, S.j.; Chen, S. Recent advances in open set recognition: A survey. IEEE Trans. Pattern Anal. Mach. Intell. 2020. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  25. Dhamija, A.R.; Günther, M.; Boult, T. Reducing network agnostophobia. In Proceedings of the Advances in Neural Information Processing Systems, Vancouver, BC, Canada, 3–8 December 2018; pp. 9157–9168. [Google Scholar]
  26. Laradji, I.H.; Vazquez, D.; Schmidt, M. Where are the Masks: Instance Segmentation with Image-level Supervision. arXiv 2019, arXiv:1907.01430. [Google Scholar]
  27. Long, J.; Shelhamer, E.; Darrell, T. Fully convolutional networks for semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Santiago, Chile, 13–16 December 2015; pp. 3431–3440. [Google Scholar]
  28. Kolesnikov, A.; Lampert, C.H. Seed, expand and constrain: Three principles for weakly-supervised image segmentation. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 8–16 October 2016; pp. 695–711. [Google Scholar]
  29. Roy, A.; Todorovic, S. Combining bottom-up, top-down, and smoothness cues for weakly supervised image segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 3529–3538. [Google Scholar]
  30. Pinheiro, P.O.; Collobert, R. From image-level to pixel-level labeling with convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1713–1721. [Google Scholar]
  31. Papandreou, G.; Chen, L.C.; Murphy, K.; Yuille, A. Weakly-and semi-supervised learning of a DCNN for semantic image segmentation. arXiv 2015, arXiv:1502.02734. [Google Scholar]
  32. Huang, Z.; Wang, X.; Wang, J.; Liu, W.; Wang, J. Weakly-supervised semantic segmentation network with deep seeded region growing. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7014–7023. [Google Scholar]
  33. Ahn, J.; Kwak, S. Learning pixel-level semantic affinity with image-level supervision for weakly supervised semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 4981–4990. [Google Scholar]
  34. Wei, Y.; Xiao, H.; Shi, H.; Jie, Z.; Feng, J.; Huang, T.S. Revisiting dilated convolution: A simple approach for weakly-and semi-supervised semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7268–7277. [Google Scholar]
  35. Zhang, X.; Wei, Y.; Feng, J.; Yang, Y.; Huang, T.S. Adversarial complementary learning for weakly supervised object localization. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 1325–1334. [Google Scholar]
  36. Li, K.; Wu, Z.; Peng, K.C.; Ernst, J.; Fu, Y. Tell me where to look: Guided attention inference network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 9215–9223. [Google Scholar]
  37. Wang, X.; You, S.; Li, X.; Ma, H. Weakly-supervised semantic segmentation by iteratively mining common object features. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 1354–1362. [Google Scholar]
  38. Hou, Q.; Jiang, P.; Wei, Y.; Cheng, M.M. Self-erasing network for integral object attention. In Proceedings of the Advances in Neural Information Processing Systems, Vancouver, BC, Canada, 3–8 December 2018; pp. 549–559. [Google Scholar]
  39. Fan, R.; Hou, Q.; Cheng, M.M.; Yu, G.; Martin, R.R.; Hu, S.M. Associating inter-image salient instances for weakly supervised semantic segmentation. In Proceedings of the European Conference on Computer Vision, Salt Lake City, UT, USA, 18–23 June 2018; pp. 367–383. [Google Scholar]
  40. Fan, R.; Cheng, M.M.; Hou, Q.; Mu, T.J.; Wang, J.; Hu, S.M. S4Net: Single stage salient-instance segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019; pp. 6103–6112. [Google Scholar]
  41. Zhang, C.; Platt, J.C.; Viola, P.A. Multiple instance boosting for object detection. In Proceedings of the Advances in Neural Information Processing Systems, Vancouver, BC, Canada, 4–7 December 2006; pp. 1417–1424. [Google Scholar]
  42. Ren, W.; Huang, K.; Tao, D.; Tan, T. Weakly supervised large scale object localization with multiple instance learning and bag splitting. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 38, 405–416. [Google Scholar] [CrossRef] [PubMed]
  43. Wang, X.; Zhu, Z.; Yao, C.; Bai, X. Relaxed multiple-instance SVM with application to object discovery. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 13–26 December 2015; pp. 1224–1232. [Google Scholar]
  44. Bilen, H.; Vedaldi, A. Weakly supervised deep detection networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 2846–2854. [Google Scholar]
  45. Diba, A.; Sharma, V.; Pazandeh, A.; Pirsiavash, H.; Van Gool, L. Weakly supervised cascaded convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 914–922. [Google Scholar]
  46. Tang, P.; Wang, X.; Wang, A.; Yan, Y.; Liu, W.; Huang, J.; Yuille, A. Weakly supervised region proposal network and object detection. In Proceedings of the European Conference on Computer Vision, Munich, Germany, 8–14 September 2018; pp. 352–368. [Google Scholar]
  47. Tang, P.; Wang, X.; Bai, S.; Shen, W.; Bai, X.; Liu, W.; Yuille, A. Pcl: Proposal cluster learning for weakly supervised object detection. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 42, 176–191. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  48. Wan, F.; Wei, P.; Jiao, J.; Han, Z.; Ye, Q. Min-Entropy Latent Model for Weakly Supervised Object Detection. IEEE Trans. Pattern Anal. Mach. Intell. 2019, 41, 2395. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  49. Shen, Y.; Ji, R.; Yang, K.; Deng, C.; Wang, C. Category-Aware spatial constraint for weakly supervised detection. IEEE Trans. Image Process. 2019, 29, 843–858. [Google Scholar] [CrossRef] [PubMed]
  50. Bendale, A.; Boult, T.E. Towards open set deep networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 July 2016; pp. 1563–1572. [Google Scholar]
  51. Pont-Tuset, J.; Arbelaez, P.; Barron, J.T.; Marques, F.; Malik, J. Multiscale combinatorial grouping for image segmentation and object proposal generation. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 39, 128–140. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  52. Maninis, K.K.; Pont-Tuset, J.; Arbeláez, P.; Van Gool, L. Convolutional oriented boundaries: From image segmentation to high-level tasks. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 40, 819–833. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  53. Everingham, M.; Eslami, S.A.; Van Gool, L.; Williams, C.K.; Winn, J.; Zisserman, A. The pascal visual object classes challenge: A retrospective. Int. J. Comput. Vis. 2015, 111, 98–136. [Google Scholar] [CrossRef]
  54. Lin, T.Y.; Maire, M.; Belongie, S.; Hays, J.; Perona, P.; Ramanan, D.; Dollár, P.; Zitnick, C.L. Microsoft coco: Common objects in context. In Proceedings of the European Conference on Computer Vision, Zurich, Switzerland, 6–12 September 2014; pp. 740–755. [Google Scholar]
  55. Hariharan, B.; Arbeláez, P.; Bourdev, L.; Maji, S.; Malik, J. Semantic contours from inverse detectors. In Proceedings of the IEEE International Conference on Computer Vision, Barcelona, Spain, 6–13 November 2011; pp. 991–998. [Google Scholar]
  56. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
Figure 1. Illustration of the motivation of our work. Existing proposal-based approaches to WSIS commonly follow a heuristic way to exploit attention maps for proposal scoring. In order to assign high scores to the proposals of true objects, the attention maps are expected to be aware of the spatial extents of objects, which is also the main focus of previous efforts on WSIS. But unfortunately, this is by nature a very difficult perceptual grouping task, since attention maps can only have very coarse resolution sparsely highlighting some discriminative sites (unaware of object extents). For intuition, we show some exemplary results obtained by the poineering approach of PRM [8], where (a) is the original image, (b) is the PRM of the cow class, and (c,d) are two proposals and the obtained classification scores. As can be seen, the more favorable proposal in (c) is undesirably assigned with a lower score.
Figure 1. Illustration of the motivation of our work. Existing proposal-based approaches to WSIS commonly follow a heuristic way to exploit attention maps for proposal scoring. In order to assign high scores to the proposals of true objects, the attention maps are expected to be aware of the spatial extents of objects, which is also the main focus of previous efforts on WSIS. But unfortunately, this is by nature a very difficult perceptual grouping task, since attention maps can only have very coarse resolution sparsely highlighting some discriminative sites (unaware of object extents). For intuition, we show some exemplary results obtained by the poineering approach of PRM [8], where (a) is the original image, (b) is the PRM of the cow class, and (c,d) are two proposals and the obtained classification scores. As can be seen, the more favorable proposal in (c) is undesirably assigned with a lower score.
Sensors 21 03475 g001
Figure 2. Overview of the proposed Weakly Supervised R-CNN (WS-RCNN) framework for weakly supervised instance segmentation. WS-RCNN adapts Fast R-CNN (FRCNN) [21], a representative model for general object detection, to the particular setting of weak supervision, which can be roughly interpretted as to derive pseudo labels from the image-level CNN classifier to initiate the FRCNN. The network mainly consists of two streams, i.e., the proposal scoring stream and the image-level classification stream. The former learns to score proposals, trained by the Entropic Open-Set Loss, and the latter performs pseudo labeling using the Attention-Guided Pseudo Labeling.
Figure 2. Overview of the proposed Weakly Supervised R-CNN (WS-RCNN) framework for weakly supervised instance segmentation. WS-RCNN adapts Fast R-CNN (FRCNN) [21], a representative model for general object detection, to the particular setting of weak supervision, which can be roughly interpretted as to derive pseudo labels from the image-level CNN classifier to initiate the FRCNN. The network mainly consists of two streams, i.e., the proposal scoring stream and the image-level classification stream. The former learns to score proposals, trained by the Entropic Open-Set Loss, and the latter performs pseudo labeling using the Attention-Guided Pseudo Labeling.
Sensors 21 03475 g002
Figure 3. Illustration of the SegAlign operation. SegAlign adapts the widely-used RoIAlign [2] to segment masks.
Figure 3. Illustration of the SegAlign operation. SegAlign adapts the widely-used RoIAlign [2] to segment masks.
Sensors 21 03475 g003
Figure 4. Illustration of Attention-Guided Pseudo Labeling (AGPL). AGPL leverages the localization ability of CNNs and the spatial relationship among proposals to achieve pseudo labeling.
Figure 4. Illustration of Attention-Guided Pseudo Labeling (AGPL). AGPL leverages the localization ability of CNNs and the spatial relationship among proposals to achieve pseudo labeling.
Sensors 21 03475 g004
Figure 5. (a,b) are visualization of two sets of representative proposal scores obtained by various methods. In each set, from top to bottom are the results of PRM [8], WSDDN-seg [44] and our WS-RCNN respectively. In each row, the first column is the final WSIS result, and the second to sixth columns are the scores of the five proposals obtained by the method.
Figure 5. (a,b) are visualization of two sets of representative proposal scores obtained by various methods. In each set, from top to bottom are the results of PRM [8], WSDDN-seg [44] and our WS-RCNN respectively. In each row, the first column is the final WSIS result, and the second to sixth columns are the scores of the five proposals obtained by the method.
Sensors 21 03475 g005
Figure 6. Representative results on VOC obtained by various methods. In each row, from left to right are (a) the input image, (b) the ground truth, and the results of (c) PRM [8], (d) IRnet [14], (e) WSDDN-seg [44] and (f) our WS-RCNN, respectively. Notice that we always output the same number of instances as that in the ground truth for all the methods to enable in-depth comparison.
Figure 6. Representative results on VOC obtained by various methods. In each row, from left to right are (a) the input image, (b) the ground truth, and the results of (c) PRM [8], (d) IRnet [14], (e) WSDDN-seg [44] and (f) our WS-RCNN, respectively. Notice that we always output the same number of instances as that in the ground truth for all the methods to enable in-depth comparison.
Sensors 21 03475 g006
Figure 7. Representative results on COCO obtained by various methods. In each row, from left to right are (a) the input image, (b) the ground truth, and the results of (c) PRM [8], (d) IRnet [14], (e) WSDDN-seg [44] and (f) our WS-RCNN, respectively. Notice that we always output the same number of instances as that in the ground truth for all the methods to enable in-depth comparison.
Figure 7. Representative results on COCO obtained by various methods. In each row, from left to right are (a) the input image, (b) the ground truth, and the results of (c) PRM [8], (d) IRnet [14], (e) WSDDN-seg [44] and (f) our WS-RCNN, respectively. Notice that we always output the same number of instances as that in the ground truth for all the methods to enable in-depth comparison.
Sensors 21 03475 g007
Figure 8. Impacts of varying the key parameters (a) the number of proposals N and (b) the threshold value β in AGPL.
Figure 8. Impacts of varying the key parameters (a) the number of proposals N and (b) the threshold value β in AGPL.
Sensors 21 03475 g008
Figure 9. Typical failure cases of WS-RCNN. (a) Poor proposals. (b) Picking up large proposals covering multiple instances. (c) Picking up small proposals covering only a part of a instance.
Figure 9. Typical failure cases of WS-RCNN. (a) Poor proposals. (b) Picking up large proposals covering multiple instances. (c) Picking up small proposals covering only a part of a instance.
Sensors 21 03475 g009
Table 1. Quantitative results obtained by various WSIS methods on VOC.
Table 1. Quantitative results obtained by various WSIS methods on VOC.
MethodsmAP50 mAP 25 mAP 75 mAP 50 ( + p ) mAP 25 ( + p ) mAP 75 ( + p )
PRM [8]26.844.39.038.052.814.1
IAM [13]28.845.911.9---
Label-PEnet [9]30.249.112.9---
IRnet [14]31.149.210.746.760.515.6
WSDDN-seg [44]27.547.59.843.755.916.9
WS-RCNN (ours)42.757.219.447.362.219.8
Table 2. Instance segmentation performance under different supervision on VOC.
Table 2. Instance segmentation performance under different supervision on VOC.
MethodSupervision mAP 50 mAP 25 mAP 75
SDI [10]Box-Level44.8-16.3
Mask R-CNN [2]Fully-Supervised69.076.752.5
WS-RCNNImage-Level47.362.219.8
Table 3. Quantitative results obtained by various WSIS methods on COCO.
Table 3. Quantitative results obtained by various WSIS methods on COCO.
MethodsmAP50 mAP 25 mAP 75 mAP 50 ( + p ) mAP 25 ( + p ) mAP 75 ( + p )
PRM [8]5.812.11.814.723.76.6
IRnet [14]7.313.92.413.222.24.8
WSDDN-seg [44]14.424.06.023.030.911.5
WS-RCNN (ours)18.027.47.224.232.111.6
Table 4. Comparison with Various Pseudo Labeling Strategies on VOC.
Table 4. Comparison with Various Pseudo Labeling Strategies on VOC.
PLS-1PLS-2PLS-3Ours
24.025.236.242.7
Table 5. Comparison with Various Network Structures and Training Strategies on VOC.
Table 5. Comparison with Various Network Structures and Training Strategies on VOC.
TS-CoupledTS-JointOurs
36.537.642.7
Table 6. Comparison with Different Training Loss for Background Handling on VOC.
Table 6. Comparison with Different Training Loss for Background Handling on VOC.
LOSS-CEOurs
34.542.7
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Ou, J.-R.; Deng, S.-L.; Yu, J.-G. WS-RCNN: Learning to Score Proposals for Weakly Supervised Instance Segmentation. Sensors 2021, 21, 3475. https://doi.org/10.3390/s21103475

AMA Style

Ou J-R, Deng S-L, Yu J-G. WS-RCNN: Learning to Score Proposals for Weakly Supervised Instance Segmentation. Sensors. 2021; 21(10):3475. https://doi.org/10.3390/s21103475

Chicago/Turabian Style

Ou, Jia-Rong, Shu-Le Deng, and Jin-Gang Yu. 2021. "WS-RCNN: Learning to Score Proposals for Weakly Supervised Instance Segmentation" Sensors 21, no. 10: 3475. https://doi.org/10.3390/s21103475

APA Style

Ou, J. -R., Deng, S. -L., & Yu, J. -G. (2021). WS-RCNN: Learning to Score Proposals for Weakly Supervised Instance Segmentation. Sensors, 21(10), 3475. https://doi.org/10.3390/s21103475

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop