Next Article in Journal
Improving Churn Detection in the Banking Sector: A Machine Learning Approach with Probability Calibration Techniques
Previous Article in Journal
Hybrid Control Strategy for LLC Converter Based on Improved Fruit Fly Optimization Algorithm
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Text-Guided Unknown Pseudo-Labeling for Open-World Object Detection

School of Computer Engineering and Science, Shanghai University, Shanghai 200444, China
*
Author to whom correspondence should be addressed.
Electronics 2024, 13(22), 4528; https://doi.org/10.3390/electronics13224528
Submission received: 26 October 2024 / Revised: 16 November 2024 / Accepted: 16 November 2024 / Published: 18 November 2024

Abstract

:
Open-world object detection (OWOD) focuses on training models with partially known class labels, enabling the detection of objects from known classes while concurrently identifying objects from unknown classes. Current models often perform suboptimally in generating pseudo-labels for unknown objects based on objectness scores due to inherent biases towards known classes. To address this issue, we propose a cross-modal learning model named Text-Guided Unknown Pseudo-Labeling for Open-world Object Detection(TGOOD) building on the Featurized Query R-CNN (FQR-CNN) framework. Specifically, we introduce a module called Similarity-Random-Similarity (SRS) to guide the model in detecting unknown objects during training. Additionally, we replace the one-to-one label assignment strategy in FQR-CNN with a one-to-many (OTM) label assignment strategy to provide more supervisory information during training. Moreover, we propose the ROI features Refinement Module (RRM) to enhance the discriminability of all objects. Experimental evaluations on the PASCAL VOC, MS-COCO, and COCO-O benchmarks demonstrate TGOOD’s superior open-world detection capability.

1. Introduction

Deep learning has made significant advancements in object detection. However, conventional object detection methods operate under a closed-set assumption, where all target classes seen during testing must also be present during training. In contrast, real-world scenarios involve a diverse array of objects and the continuous emergence of new categories, complicating object detection tasks. While effective in closed-set frameworks, traditional methods often struggle to handle open-world scenarios. Recently, researchers have made substantial progress in developing methods capable of detecting unseen object classes.
The first open-world object detection (OWOD) method, ORE, introduced by Joseph [1], addresses the critical challenge of enabling models to recognize previously unseen object categories as the “unknown” class while continuing to detect known categories. Recent advancements in OWOD can be classified into two main approaches. The first focuses on learning feature-level distributions for both known and unknown object categories during training [2,3,4,5]. The second involves generating pseudo-labels for unknown class objects during the training phase and treating the unknown class as a distinct “known class” for joint learning [1,6,7,8,9,10,11,12]. The former often requires defining a threshold to classify predictions as the unknown class, while the latter relies on pseudo-labeling unknown class objects to guide model learning during training. However, this method often depends on limited statistics from known classes, such as objectness scores, which can introduce bias. Specifically, as illustrated in Figure 1a, during pseudo-labeling, proposals that include parts of known-class objects are prone to being mislabeled as the unknown class. This mislabeling undermines the model’s ability to differentiate between known and unknown classes, thereby compromising overall detection performance.
Our primary objective is to reduce the bias towards known classes in current pseudo-labeling methods. Leveraging advancements in cross-modal tasks, we can seamlessly integrate information from diverse modalities, including images, text, audio, and other forms of data. We propose that while images provide detailed visual information through shapes, textures, and colors, language provides abstract semantic information, offering more comprehensive guidance. Building on the FQR-CNN framework [13], we have developed a simple yet effective pseudo-labeling strategy for OWOD, termed TGOOD (Text-Guided Unknown Pseudo-Labeling for Open-World Object Detection). During training, we apply a novel module called Similarity-Random-Similarity (SRS) to label candidate boxes as unknown when they do not match any ground truth ( g t ). Additionally, we adopt a one-to-many (OTM) matching strategy during training, combined with Non-Maximum Suppression (NMS) during inference, as a replacement for the original one-to-one algorithm used in FQR-CNN. This modification resolves the problem of insufficient supervision [14,15]. High-quality feature representation of foreground objects is critical for the training of object locators and classifiers. To enhance the model’s ability to recognize objects across all categories, we introduce the ROI features Refinement Module (RRM).
Our main contributions are summarized as follows:
  • TGOOD: We propose an OWOD detector, TGOOD, built upon the FQR-CNN framework [13], which incorporates the SRS module, OTM technique, and RRM module. TGOOD leverages the strengths of both Faster R-CNN-style and DETR-style OWOD detectors, leading to a streamlined and efficient detection approach.
  • Benchmark Performance: Extensive evaluations on OWOD benchmarks, including PASCAL VOC [16] and MS-COCO [17], demonstrate TGOOD’s ability to effectively adapt to open-world environments. It maintains high performance on known classes while exhibiting strong capabilities in detecting unknown objects.
  • Cross-Domain Generalization: Comparative analysis on COCO-O [18], which includes images from diverse real-world domains, reveals that TGOOD not only excels in generalizing across different classes but also performs exceptionally well across diverse domains, underscoring its robustness and superiority.

2. Related Work

2.1. Open-World Object Detection

Joseph pioneered the task of open-world object detection, which involves a model’s ability to both recognize previously unknown objects and progressively acquire the capability to detect new known objects. They introduced ORE [1] as a solution to this complex challenge, building upon Faster R-CNN [19]. A key difficulty in this task is the accurate identification of unknown-class objects due to the lack of labels. The use of pseudo-labels has gained substantial attention and is now a commonly employed technique [1,6,7,8,9,10,12]. Several studies, including ORE [1], OW-DETR [7], RandBox [9], UC-OWOD [10], and SA [11], leverage objectness scores of candidate proposals to generate pseudo-labels for unknown instances. Additionally, other works, such as Open World DETR [12], CAT [8], and RE-OWOD [12], utilize supplementary proposal generation techniques (e.g., selective search [20]) to improve the selection of potential candidates for unknown classes. These well-developed pseudo-label methods for unknowns have demonstrated considerable effectiveness in real world applications. In contrast, there are other methods that do not use pseudo-labels. Like OW-RCNN [2], 2B-OCD [3], OCPL [4], PROB [5] and Ann [21], they try to distinguish between known- and unknown-class objects in the feature space. Our primary focus is on OWOD algorithms that leverage pseudo-labeling, with the goal of developing a simple yet robust approach for generating pseudo-labels.

2.2. Class-Agnostic Object Detection

In open object detection tasks, it is crucial for models to learn to detect unknown objects. The class-agnostic object detection task aims to enhance the capability of object detection models to identify objects without considering their specific class. This paradigm depends on a finite set of known-class training datasets to develop a detector that can recognize all foreground objects within an image, regardless of class distinctions. WACV [22] highlights that in certain real-world scenarios, accurately determining the presence and precise location of objects is more critical than classifying them into specific categories, thus introducing the challenge of class-agnostic object detection. Methods such as OLN [23], SIBGRAPI [24], LDET [25], and GOOD [26] focus on improving the model’s detection capabilities at the image level. MAVL [27] noted that prior methods often lack supervision from easily interpretable semantic signals. To address this, they utilized a multi-modality visual transformer trained on aligned image–text data, leading to improved performance in detecting unknown objects. Being inspired by this, we propose to use semantic information as a guiding signal to assist the model in detecting objects of unknown classes during training.

2.3. Pre-Trained Visual–Language Models

In recent years, pre-trained vision–language models have garnered significant attention in both computer vision and natural language processing. These models, trained on extensive text and image datasets, acquire comprehensive semantic and visual representations, serving as powerful feature extractors for a wide range of downstream tasks. For example, CLIP [28] utilizes contrastive learning on a dataset of 400 million image/text pairs sourced from the internet. By aligning the features of images and their corresponding text in the feature space, CLIP produces highly robust image and text encoders, demonstrating exceptional performance across various downstream tasks, such as semantic segmentation [29], object detection [30], image editing [31], image generation [32], and video comprehension [33]. Contemporary computer vision research primarily employs CLIP-based methodologies, where text features from the CLIP text encoder are used as substitutes for traditional classifiers. However, in our approach, we leverage CLIP to align visual and text features within the feature space. We use embedded text information as high-level semantic guidance, allowing the model to be directed without bias. Consequently, we enhance the model’s detection capabilities across all objects in an image, including both known and novel classes.

3. Method

3.1. Preliminary: Formulation for OWOD

OWOD consists of a series of subtasks, represented as T = { T 1 , T 2 , T 3 , } . The corresponding training data can be divided into D = { D 1 , D 2 , D 3 } . The set of categories for all annotated objects in D is C = { C 1 , C 2 , C 3 } , where C i contains all annotated categories in D i , and it satisfies C i C j = , i j . To maintain consistency with the existing literature, in this study, the sizes of sets T, D, and C are all set to 4 to simulate a realistic open environment. As the training data is sequentially fed, the model learns each subtask step by step. Assuming the initial model is represented as M o d e l 0 , after completing the training for task T 1 on D 1 , the model will be updated to M o d e l 1 , and so on, following this pattern. During the T i phase, the categories in C i are referred to as the currently known classes, which are the classes included in the annotated data during model training. The set C = C 1 , C i 1 is referred to as the previously known classes, while the set C = C i + 1 , is referred to as the currently unknown classes, which will be gradually learned in the subsequent incremental learning process. Assume there are K N annotated objects in an image from D i , with each object’s label containing category information and bounding box coordinates, denoted as { Y 1 , Y 2 , , Y K N } , where Y k n = [ c k n , x k n 1 , y k n 1 , x k n 2 , y k n 2 ] and k n { 1 , 2 , , K N } . Here, c k n C i represents the category to which the object belongs, and x k n 1 , y k n 1 , x k n 2 , y k n 2 indicate the positional coordinates of the object in the image. After completing the learning phase of each subtask T i , we evaluate the current model M o d e l i on a test dataset that contains all categories (both known and unknown). We are constantly working to enhance the model’s ability to generalize, allowing it to accurately detect objects from unknown categories while maintaining exceptional detection performance for known categories in diverse and dynamic real-world detection environments.

3.2. Overall Architecture

TGOOD is implemented using the FQR-CNN framework due to its advantageous trade-off between accuracy and speed, achieved by incorporating the query mechanism from DETR [34] into the R-CNN-like detector. The overall training process follows the standard paradigm of existing OWOD methods like [1]. After each subtask, the model is fine-tuned on a small dataset containing a few samples from previously known classes.
As illustrated in Figure 2, the main training process of TGOOD is divided into the following five steps: (1) Feature extraction: for a given image and its description, the backbone network, QGN [13], and ROI Align [35] modules are used to extract queries (Q) and region of interest (ROI) features from the image. Nouns are extracted from the description, and the text embeddings (E) are obtained by the CLIP text encoder. (2) Feature interaction: the ROI features from step (1) are firstly enhanced through the RRM module, allowing the query features to better capture the discriminative information of the foreground objects. By the Query-based RCNN Head [13], we successfully derive the query features ( Q i m g ), which have undergone interaction with the augmented ROI features ( R a u g ). (3) One-to-many label assignment: in this phase, at least one candidate query is assigned to each g t , ensuring sufficient supervision signals. (4) Text-guided pseudo-labeling: after label assignment for known-class objects, candidate queries with high similarity to novel text embeddings are selected as pseudo-labels for unknown objects in an unbiased manner. (5) Loss calculation: the corresponding box loss and classification loss are computed for known classes, while only the classification loss is calculated for unknown classes.
During the inference stage, text input is not required, nor is it necessary to set separate threshold boundaries for detecting unknown objects. The model treats unknown-class objects as a distinct category, labeled “unknown”, and detects them in the same manner as known-class objects.

3.2.1. SRS: Text-Guided Pseudo-Label Generation Strategy

As shown in Figure 2, part (4), following the label assignment process in part (3), the SRS module is applied to obtain candidate queries for unknown classes from those that do not match any known class g t . Figure 3 illustrates the flowchart of the SRS module. Two pseudo-label generation strategies are explored: a standard version and an enhanced version.
Firstly, the query features Q i m g = { q i m g 1 , q i m g 2 , , q i m g N } , with N representing the number of queries, are mapped to the CLIP text feature space for dimensional alignment. A feature projector P r o j : R d i m g R d t x t , consisting of linear layers and an activation layer, is designed to perform this mapping, as shown in Equation (1). The dimensions of the query features before and after mapping are denoted by d i m g and d t x t , respectively.
q t x t i = P r o j ( q i m g i ) , i 1 , 2 , , N
Then, the queries projected into the CLIP text feature space are divided into two groups: (1) the queries matching the known-class targets (the yellow hollow circles in Figure 3), denoted as Q t x t m , with a quantity of M; (2) the remaining queries, denoted as Q t x t u , with a quantity of U, where M + U = N . For Q t x t m , a loss function combining cosine similarity and Euclidean distance is applied to minimize the distance between the query and the corresponding text embedding in the feature space. As shown in Equation (2), e i represents the text embedding feature corresponding to the known class. The first term of L c l o s e represents the Euclidean distance, while the second term represents the cosine distance.
L c l o s e = i = 1 M { 1 d t x t q t x t i e i 2 + [ 1 s i m ( q t x t i , e i ) ] }
For Q t x t u , we use it to generate pseudo candidates for unknown classes. After removing information related to known-class objects from the image caption, the remaining nouns are likely to represent unknown-class objects. These nouns are referred to as novel texts, and the corresponding text embedding features are expressed as E n o v e l = { e 1 , e 2 , , e O } , where O represents the number of novel texts. It is hypothesized that if the features of candidate queries show similarity above a certain threshold to any novel text feature, they may correspond to potential candidate boxes for unknown objects. Therefore, the similarity matrix between Q t x t u and E n o v e l is computed, resulting in a matrix M a t r i x R U × O . The maximum value along the last dimension of the matrix is then used to represent the similarity score S i m i l a r i t y R U .
  • The standard version:
    In the standard version, the pseudo-labeling process is conducted according to the rule outlined in Equation (3), where b g means background, and l q i denotes the label of the i-th candidate query, which will be regarded as an unknown class if its similarity score exceeds the threshold δ 1 and ranks highly among all similarity scores. S i m i l a r i t y t o p K R K contains the topK scores from S i m i l a r i t y R U . Empirically, the value of K was set to 5.
    l q i = u n k n o w n , S i m i l a r i t y t o p K [ i ] > δ 1 b g , S i m i l a r i t y t o p K [ i ] δ 1 , i 1 , 2 , , K
  • The enhanced version:
    Since the text encoder used in this study is trained on 400 million image/text pairs, while the dataset employed in our experiments consists of only 80 common categories, the CLIP text features and unknown text features may exhibit some similarity. This could introduce a degree of interference in the labeling process, potentially leading to the generation of false labels. To mitigate this, we propose an enhanced pseudo-labeling strategy combined with random de-biasing. The overall framework of this strategy aligns with the standard version, but the decision-making process based on the similarity score is divided into three steps: (1) initially applying a small threshold δ 1 to filter out most candidate queries containing b g class; (2) randomly selecting r candidate queries from the remaining pool; and (3), finally, using a larger threshold δ 2 to filter out low-quality candidate queries and generate the final unknown-class pseudo-labels.

3.2.2. OTM: One-to-Many Label Assignment

The FQR-CNN framework we utilize adopts a decoding structure similar to DETR. It employs the Hungarian algorithm [36] for label matching, which follows a one-to-one matching mode. While this retains the end-to-end detection advantages by eliminating post-processing operations like NMS, it often lacks sufficient supervisory information. To address this issue, we implement a one-to-many matching strategy using the optimal transport algorithm [37]. During inference, NMS is applied as a post-processing step. Since the number of candidate boxes in the FQR-CNN framework is considerably smaller than that in traditional R-CNN-based frameworks, the computational cost of applying NMS remains minimal.
Following OTA [38], we employ the concept of optimal transport to achieve one-to-many label assignment. For each image, let S represents the labels (both known and b g classes) as suppliers, where each supplier s i can provide k i labels. The N candidate query boxes act as demanders, with each requiring a label. The value of k i for the known class is dynamically determined based on the overlap between the target box and all candidate query boxes, as in OTA [38]. The total number of labels provided by all known classes is k i , while the b g class provides N k i labels. The element c i j in the cost matrix represents the cost of assigning one label (one of the k i labels) from supplier s i to the j-th query, which is calculated as shown in Equation (4), where g t represents the target label and α is a moderator. If it is a known class, c i j is composed of the classifier’s predicted loss and the locator’s predicted loss; for the b g class, c i j only includes the classifier’s predicted loss.
c i j = L c l s ( p j , g t i ) + α · L b o x ( p j , g t i ) , i f g t i s k n o w n c l a s s L c l s ( p j , g t i ) , i f g t i s b g
The goal of the optimal transport algorithm is to find an optimal assignment plan π R S × N that minimizes the overall matching cost. The symbolic expression is provided in Equation (5).
m i n π i = 1 S j = 1 N c i j · π i j , s . t . i = 1 S π i j = d j , j = 1 N π i j = s i , i = 1 S s i = j = 1 N d i j , π i j 0
The classical solution for finding π involves multiple iterations using the Sinkhorn–Knopp algorithm [39]. However, to enhance computational efficiency, we did not employ this iterative algorithm to find the optimal π . Instead, after determining k i and c i j , the labels of the known class are assigned to the k i candidate queries with the lowest matching cost. If a candidate query is matched with multiple g t labels, it is assigned to the g t with the lower matching cost.
Suppose an image contains two known target objects, which can provide k 1 = 2 and k 2 = 1 labels, respectively, and the number of candidate queries is 5. This means the b g class provides 5 2 1 = 2 labels. The pairs formed by the one-to-one matching method and the one-to-many matching method are illustrated in Figure 4. The one-to-one matching algorithm only assigns Q 1 to G T 1 . In the one-to-many matching mode, 2, 1, and 2 candidate queries are assigned to G T 1 , G T 2 , and the b g class, respectively. For G T 1 , both Q 1 and Q 4 are candidate queries with a high matching degree, thereby providing a more comprehensive supervisory signal.

3.2.3. RRM: ROI Features Refinement Module

To enable the query to better capture information related to foreground objects, allowing the classifier and locator to make more accurate decisions, we introduce an ROI feature refinement module based on the attention mechanism. The rationale is that a single ROI feature conveys limited information. By taking a more holistic view and incorporating relevant nearby information, the representational capacity of the ROI features can be enhanced. Specifically, an improved self-attention mechanism is applied to the ROI features before their interaction with the queries. While the attention mechanism effectively enhances the relevant information near the ROI features, directly applying attention to these high-dimensional features can result in excessive computational complexity. To address this, we employ a pooling operation to retain key information while reducing the dimensionality of the ROI features.
Figure 5 illustrates the overall process of foreground feature enhancement using an attention mechanism. As shown in Equation (6), the original ROI features, R O I o r i , are processed through max pooling and flattened into a one-dimensional vector, which serves as the input Q for the attention module. Meanwhile, R O I o r i are average pooled and flattened into a one-dimensional vector, and used as the inputs K and V for the attention module. Due to the inevitable introduction of noise when absorbing information from others, Ref. [40] used the Weights Normalized Convolutional kernel to reduce noise around the object. Inspired by this, we propose applying an average pooling operation to the attention matrix before performing the softmax operation (Equation (6)) to retain only the most salient information, thus minimizing noise. After passing through the attention module, the results are restored to the original dimensionality of R O I o r i using rearrangement and upsampling operations. Finally, they are subsequently fused with R O I o r i via residual connections to get the refinement ROI features R O I a u g .
Q = F l a t t e n ( M a x P o o l ( R O I o r i ) ) , K , V = F l a t t e n ( A v g P o o l ( R O I o r i ) ) , R O I a u g = R O I o r i + U p s a m p l e ( R e s h a p e ( A t t e n t i o n ( Q , K , V ) ) )

4. Experiments and Results

4.1. Datasets and Metrics

4.1.1. Datasets

The datasets used in this study include MS-COCO [17], Pascal VOC [16], and COCO-O [18]. The current mainstream OWOD dataset configurations include OWOD SPLIT, proposed in ORE [1], and MS-COCO SPLIT, proposed in OW-DETR [7]. These two configurations address open-world object detection with category increments but overlook the fact that many real-world scenarios involve significant distribution shifts—referred to as domain increments. To account for this, we propose a third dataset configuration, COCO-O SPLIT, which evaluates the model’s domain generalization capability under these conditions. The following sections introduce these three dataset configurations in detail:
OWOD SPLIT: In this setup, the MS-COCO [17] and Pascal VOC [16] datasets were first combined, with all objects belonging to categories in Pascal VOC designated as known categories for task 1. The remaining 60 categories in MS-COCO [17] were then divided into three groups, with each group containing 20 categories of objects, which were designated as known categories for tasks 2, 3, and 4. Since the images in Pascal VOC do not include description information, meanwhile, BLIP [41], a large model pre-trained on a vast amount of data, demonstrates strong performance across multiple visual–language tasks. So, we used BLIP to randomly generate a caption for each image in Pascal VOC.
MS-COCO SPLIT: Superclass data leakage occurs in OWOD SPLIT, with task 1 predominantly featuring classes from the vehicle and animal categories, while task 2 introduces related subclasses like truck, elephant, bear, zebra, and giraffe. The MS-COCO SPLIT partitioning method was proposed by Gupta et al. [7] to address the issue of superclass information leakage that may occur in the OWOD SPLIT configuration. It uses only the MS-COCO [17] dataset and organizes the known categories for each task based on a superclass partitioning strategy, with nearly 20 classes in each task.
COCO-O SPLIT: COCO-O SPLIT utilizes distribution shift data from COCO-O [18] to assess the model’s performance in diverse domains. COCO-O, proposed by [18], is a test dataset derived from the MS-COCO validation set and includes six natural distribution shifts: weather, painting, handmade, cartoon, tattoo, and sketch. These six types of data are regarded as unknown domain data derived from the original images in MS-COCO.

4.1.2. Metrics

The evaluation metrics we used align with the mainstream evaluation criteria of existing methods [1]. For the detection performance of known classes, mean Average Precision (mAP) is used, including metrics for previously known classes (pre mAP), the current task’s known classes (cur mAP), and all known classes up to the current task (both mAP). To assess the detection ability of unknown-class objects, the Unknown Recall (UR) is the key metric. Additionally, Absolute Open Set Error (A-OSE) and the Wilderness Impact (WI) factor, proposed by [1], are used to evaluate the model’s overall ability to detect both known- and unknown-class objects. A-OSE represents the absolute number of errors where the model mistakenly classifies unknown objects as known classes, while WI quantifies how the model’s ability to detect unknown-class objects impacts its detection performance on known classes.

4.2. Implementation Details

The implementation is carried out using Python 3.7, with model training and inference performed in the PyTorch deep learning framework on two GeForce RTX 3090 GPUs. The backbone network is ResNet-50, initialized with ImageNet [42] pre-trained weights. The number of queries, N, is set to 100 with a feature dimension of 256. During training, the number of pseudo-candidate boxes randomly selected for unknown classes, r, is set to 5. Two similarity thresholds δ 1 and δ 2 are 0.5 and 0.8, respectively. In the inference stage, the threshold of NMS is set to 0.6 by default, aligning with the standard practice reported in [38].

4.3. Main Results

This section presents a detailed comparative analysis of TGOOD against various OWOD algorithms. To maintain fairness in comparison, we follow prior work by removing the energy evaluation module from ORE, denoting the modified version as ORE-EBUI. For OWOD SPLIT dataset, we categorize those models into two groups: those that utilize pseudo-labels and those that do not. Table 1 and Table 2 summarize the comparison results for these two categories. Table 3 presents the performance comparison between TGOOD and existing models on the MS-COCO SPLIT dataset. Table 4 shows the performance comparison of TGOOD with classic OWOD methods relying on objectness scores to select pseudo-labels on the COCO-O SPLIT. A downward arrow (↓) indicates that lower values signify better model performance, while an upward arrow (↑) indicates that higher values are preferred. Best results are marked in red, with the next results indicated in blue.
Table 1 compares TGOOD with existing pseudo-label-based OWOD methods under the OWOD SPLIT.
Among all the comparative methods, ORE-EBUI, OW-DETR, SA, UC-OWOD, and RandBox all employ pseudo-label strategies based on objectness scores. A common deficiency of these methods is their weak detection capability for unknown-class objects, as indicated by their low UR values. Particularly, RandBox, despite achieving an advantage in detection accuracy for known classes with its diffusion model-based detection framework, only achieved recall rates of 10.6%, 6.3%, and 7.8% for unknown classes in tasks 1, 2, and 3, respectively. We attribute the insufficiency of these models in detecting unknown classes primarily to the mechanism of pseudo-label generation. Since the learning of objectness scores is based on labeled known-class objects, models tend to misjudge candidates containing parts of known classes as unknown-class objects when selecting labels based on objectness scores, thereby diminishing the model’s ability to distinguish between known- and unknown-class objects and resulting in an inability to efficiently identify true unknown-class objects. In contrast, TGOOD, through its SRS module, generates unbiased pseudo-labels for unknown classes under textual guidance, significantly enhancing the model’s detection capability for unknown-class objects, achieving recall rates of 23.1%, 18.7%, and 22.1% for unknown classes in tasks 1, 2, and 3, respectively.
Additionally, RE-OWOD and CAT, which employ additional pseudo-label generation strategies such as selective search [20], also demonstrate decent detection performance. Compared to these methods, TGOOD shows comparable or superior performance. Although RE-OWOD’s detection accuracy for known classes is slightly higher than that of TGOOD in tasks 2, 3, and 4, its recall rate for unknown-class objects is almost half that of TGOOD. Overall, TGOOD has achieved a better balance between the detection performance of known and unknown classes among all comparative methods, especially excelling in the recall rate of unknown classes. However, there is still room for improvement in the model’s performance on the WI and A-OSE metrics, indicating that there is further research potential in enhancing the model’s ability to distinguish between known- and unknown-class objects, which will become one of the focal points of future research.
Table 2 presents a performance comparison between TGOOD and existing OWOD methods that do not use pseudo-labels under the OWOD SPLIT. Among those methods, OCPL aims to reduce the overlap between known- and unknown-class distributions in the feature space to construct more discriminative feature representations; 2B-OCD employs an object-centered calibrator to identify candidate boxes with scores above a certain threshold in the b g as unknown classes; PROB utilizes class-agnostic Gaussian distributions to model object features; Ann adopts a label transfer learning paradigm to decouple features of known- and unknown-class objects. The essence of these methods is to enhance the separability of known- and unknown-class objects in the feature space. Compared to the classic objectness score-based pseudo-label methods, approaches that do not rely on pseudo-labels indeed achieve higher recall rates for unknown classes. This indirectly confirms the bias problem towards known classes in objectness score-based pseudo-label methods. However, the proposed TGOOD in this paper, with the synergistic effect of multiple modules, has achieved optimal or near-optimal results in both the detection accuracy of known classes and the recall rate of unknown classes in all task phases.
Table 3 compares TGOOD with existing OWOD methods under the MS-COCO SPLIT. The experimental results demonstrate that TGOOD also significantly enhances the model’s detection performance on the more challenging MS-COCO SPLIT dataset. Specifically, TGOOD excels in detecting unknown classes, achieving recall rates of 29.4%, 29.0%, and 35.1% for tasks 1, 2, and 3, respectively, surpassing all comparative methods. However, we observed that in task 1, although TGOOD showed substantial improvement in the detection accuracy of known classes compared to ORE-EBUI, its performance was still behind methods employing the Deformable DETR framework [43], such as OW-DETR, PROB, and CAT. We attribute this gap to the basic framework used. The FQR-CNN framework adopted by TGOOD is more lightweight compared to Deformable DETR. However, as the tasks progress, TGOOD demonstrated superior performance in both known-class detection accuracy and unknown-class recall in tasks 2, 3, and 4. Notably, a significant issue common to comparative methods is that the model’s detection capability for current known classes is markedly lower than for previously known classes, with the cur-mAP value significantly lower than the pre-mAP value. TGOOD effectively reduced this discrepancy and achieved the optimal value in overall detection accuracy (both mAP). We believe this is due to our proposed RRM module, which enables the model to learn more discriminative foreground object features, thereby enhancing the model’s generalization and robustness across different tasks.
Table 4 compares the detection performance of TGOOD with traditional methods, ORE-EBUI and OW-DETR, which use objectness scores for pseudo-labeling, on the COCO-O SPLIT dataset. The average performance across the six domain datasets shows that TGOOD’s ability to detect known classes is comparable to or slightly better than ORE-EBUI and OW-DETR. Notably, TGOOD maintains superior unknown-class recognition capabilities across unseen domains, outperforming the comparison methods. This is attributed to the effective guidance provided by text information containing abstract semantics. The general nature of semantic information in the text helps TGOOD sustain high generalization performance even with detection data from various domain fields.

4.4. Ablation Study

Extensive ablation experiments were conducted to validate the effectiveness of TGOOD. Unless stated otherwise, these experiments were performed on task 1 using the OWOD SPLIT dataset setting.

4.4.1. Components of TGOOD

To evaluate the effectiveness of each module in TGOOD, we conducted ablation experiments for each module individually. The results are presented in Table 5. The baseline, shown in the first row, is the FQR-CNN model without any enhancements. This baseline highlights that the basic FQR-CNN detection model lacks the capability to detect objects of unknown classes.
“Obj1” refers to selecting a candidate query with the highest objectness score as the unknown-class object, similar to the pseudo-labeling method used in ORE-EBUI. The results in the second row demonstrate that “Obj1” allows our basic detection model to exhibit some capability in detecting unknown-class objects. However, it is insufficient for effectively distinguishing between known and unknown categories, as indicated by the relatively high A-OSE score. “Obj5” involves selecting the top five candidate queries with the highest objectness scores as unknown-class objects, akin to the pseudo-labeling method used in OW-DETR. The third row of results shows that increasing the number of pseudo-labeled candidates enhances the model’s ability to identify unknown-class objects, with the UR value improving from 10.0% to 13.7%. Nevertheless, this approach significantly degrades the accuracy of known-class object detection, with the mAP decreasing from 56.42% to 55.87%.
“SRS” refers to the pseudo-label generation module proposed in this paper. The results in the fourth row demonstrate that this module significantly enhances the recall rate for unknown-class objects, with a UR rate 2.28 times higher than that of “Obj1”. Additionally, “SRS” effectively mitigates the issue of mistakenly identifying unknown-class objects as known-class objects, reducing the A-OSE score from 77,927 to 15,169 compared to “Obj1”. “OTM” stands for replacing the one-to-one label matching algorithm with a one-to-many label matching algorithm. As illustrated in the fifth row, assigning multiple candidate queries to each g t during training strengthens the supervisory signal, thereby improving the model’s performance in detecting known-class objects. Furthermore, the integrated “RRM” module enhances the prominent features of foreground objects. As shown in the sixth row, this module further improves the model’s ability to detect all foreground objects, including both known- and unknown-class objects. In Figure 6, we can intuitively perceive the influence of TGOOD’s various components on the model’s performance in terms of mAP and UR metrics.

4.4.2. The Versatility of SRS

For the pseudo-label generation module SRS proposed in this paper, we evaluated the effectiveness of its various substructures through comparative experiments. Figure 7 provides an overview of these results: Item A demonstrates the standard version of the text-guided pseudo-label generation strategy. It achieves a UR value of 19.4%, significantly outperforming existing pseudo-label methods that rely on objectness scores. Item B shows that incorporating the random de-biasing scheme further improves the UR value by an additional 3.5%. This indicates that random selection effectively reduces the model’s bias towards known-class objects. Item C illustrates the enhanced version of the text-guided pseudo-label generation strategy. Applying the secondary filtering with a higher similarity threshold results in higher-quality pseudo-labels. This enhancement improves the model’s detection performance for both known and unknown classes.
Additionally, since the basic target detection framework used in this paper differs from traditional OWOD methods based on objectness scores (e.g., ORE-EBUI), we tested the cross-framework effectiveness of the proposed SRS module by applying it to ORE-EBUI. During training, for candidate boxes that did not match any g t , those with extremely low objectness scores were excluded, and the fifty candidate boxes with the highest objectness scores were retained. This step aimed to reduce noise in the pseudo-labeling process. Subsequently, the SRS module was applied.
Table 6 presents the comparison results. “ORE-EBUI+SRS” denotes the integration of SRS with ORE-EBUI, while “ORE-EBUI+Obj5” represents the replacement of “Obj1” with “Obj5” in the original ORE-EBUI. The results indicate that simply increasing the number of pseudo-labels for unknown classes negatively impacts the detection performance for known classes, with “Obj5” causing a drop in known-class detection accuracy from 56.00% to 18.31%. In contrast, the SRS module not only maintains but also enhances the detection performance for known-class objects, and improves the recall rate for unknown-class objects from 4.9% to 7.6%. Furthermore, the SRS module significantly improves the WI and A-OSE indicators, which assess the model’s comprehensive detection capabilities in open environments.

4.4.3. Different Methods for ROI Refinement

To assess the effectiveness of the RRM module proposed in this paper, we compared it with the “RoIAttn” module from [44] and two other commonly used feature enhancement strategies. The comparison results are presented in Table 7, where “TGOOD-RRM” refers to the TGOOD model without any enhancement module. One comparison method is an LSTM network. In this study, all ROI features are treated as a sequence, and the average of the encoded bidirectional LSTM sequences is taken to generate the enhanced ROI features. The results of this approach are shown in the “BiLstm” row. Additionally, we adopted a graph convolutional neural network (GCN) as another comparison method. In this method, the ROI features are treated as nodes in a graph, with edges formed based on the cosine similarity between the ROI features. To reduce computational complexity, the enhanced features are obtained after performing a single graph convolution update, and the results are shown in the “GCN” row.
The use of an enhancement module positively impacts the model’s performance compared to not using any ROI feature enhancement module. However, our RRM module consistently demonstrates superior overall performance. The BiLSTM and GCN methods increase mAP by 0.26% and 0.1%, and UR by 0.1% and 0.2%, respectively. In contrast, the RRM module proposed in this paper achieves higher improvements, increasing mAP by 0.35% and UR by 0.2%. Notably, when detecting known-class objects, the “RoIAttn” module negatively affects the model’s performance. This issue is likely due to inherent flaws in the “RoIAttn” approach, which involves two additional memory units for clustering operations. In an open environment, where unknown-class objects lack labels, this process is prone to noise interference. In contrast, the RRM module enhances foreground object features by leveraging the similarity between ROI features, thus avoiding this drawback.

4.4.4. Hyperparameter Analysis

In the SRS module proposed in this study, two key hyperparameters, δ 1 and δ 2 , were evaluated through ablation experiments, with results shown in Figure 8. To optimize the model’s ability to detect both known- and unknown-class objects, the final values of δ 1 and δ 2 were set to 0.5 and 0.8, respectively, as these values provided the best performance enhancement.

4.5. Visualization

To provide a more intuitive demonstration of TGOOD’s effectiveness, we present several test results in Figure 9. After training on task 1, the test results of ORE-EBUI, OW-DETR, and TGOOD are shown in the first, second, and third row, respectively.
Superior Performance: TGOOD demonstrates superior performance in detecting both known and unknown objects within images. First, TGOOD excels at accurately localizing and predicting the categories of known objects with high confidence, even in cases of severe occlusion. For example, TGOOD successfully identifies and classifies a chair leg in the top-right corner of the images in the third column and a small, heavily occluded car in the top-left corner of the fourth column. In contrast, both ORE-EBUI and OW-DETR fail to detect these objects. Second, TGOOD shows a strong ability to detect unknown-class objects. For instance, in the first column, ORE-EBUI does not detect any unknown-class objects, while OW-DETR incorrectly identifies the b g as an unknown object, generating two nearly identical bounding boxes. TGOOD, however, correctly identifies unknown objects such as a skateboard, shoes, and a board in the image. Although OW-DETR detects more unknown objects in the third column (e.g., broccoli, carrot), it fails to delineate the boundaries of these objects accurately and misses known-class objects. Furthermore, it incorrectly labels two objects that do not belong to any known classes. In contrast, TGOOD accurately locates unknown objects (such as a cup and rice) and correctly identifies all known-class objects (dining table and chair).
Limitations: While the proposed method demonstrates superior object detection capabilities for both known and unknown categories compared to existing methods relying on objectness scores, it faces a limitation in the confidence level when identifying unknown-category objects. The confidence is generally low, around 20%. This limitation arises from the use of a single pseudo-label mechanism, which assigns the same label to all unknown-category objects. As a result, the purity of the pseudo-labels may be compromised. One of the key challenges for future research in open-environment object detection is how to generate high-purity pseudo-labels for unknown objects that lack labels.

5. Discussion and Conclusions

In this paper, we propose TGOOD, a cross-modal learning object detection method for open environments. The core idea behind TGOOD is to guide the model in generating high-quality pseudo-labels for unknown classes during training through the use of text containing high-level semantic information. This approach substantially reduces the model’s bias towards known classes compared to traditional open-environment detection methods based on objectness scores (e.g., ORE [1] and OW-DETR [7]). Specifically, we propose three main improvement modules: (1) SRS, a pseudo-label generation module combining text guidance with random de-biasing to address the bias of existing strategies towards known classes; (2) OTM, a one-to-many label matching strategy that enriches supervisory signals during the learning process of the query-based object detection model; and (3) RRM, an ROI feature enhancement module that enhances the discriminability of foreground objects’ ROI features using an advanced attention mechanism.
Comparative experiments were conducted on TGOOD using established evaluation benchmarks, such as OWOD SPLIT, MS-COCO SPLIT, and the COCO-O domain generalization benchmark newly proposed in this paper. These experiments assess the model’s performance in detecting objects in open environments. These experiments evaluate the model’s performance in detecting objects in open environments. While TGOOD demonstrates superior performance in detecting unknown-class objects, the prediction scores for these unknown classes remain lower than desired, as outlined in Section 4. Future research will focus on generating higher-purity pseudo-labels to further improve the model’s detection capabilities for unknown-class objects in open environments.

Author Contributions

Conceptualization, X.W. and D.X.; methodology, X.W.; software, X.W.; validation, X.W. and D.X.; formal analysis, X.W. and D.X.; investigation, X.W.; resources, X.W.; data curation, X.W.; writing—original draft preparation, X.W. and D.X.; writing—review and editing, X.W. and D.X.; visualization, X.W.; supervision, D.X.; project administration, D.X. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The data that support the findings of this study are available from the corresponding author upon reasonable request.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
R real space
Etext feature
Qquery feature
ROIregion of interest
SRSSimilarity-Random-Similarity, text-guided pseudo-label generation strategy
OTMone-to-many matching strategy
RRMROI features Refinement Module
TGOODText-Guided Unknown Pseudo-Labeling for Open-world Object Detection
mAPmean Average Precision
URUnknown Recall
A-OSEAbsolute Open Set Error
WIWilderness Impact
Obj1selecting 1 candidate query with the highest objectness score as the unknown-class object
Obj5selecting 5 candidate queries with the highest objectness scores as the unknown-class objects

References

  1. Joseph, K.; Khan, S.; Khan, F.S.; Balasubramanian, V.N. Towards open world object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 5830–5840. [Google Scholar]
  2. Pershouse, D.; Dayoub, F.; Miller, D.; Sünderhauf, N. Addressing the Challenges of Open-World Object Detection. arXiv 2023, arXiv:2303.14930. [Google Scholar]
  3. Wu, Y.; Zhao, X.; Ma, Y.; Wang, D.; Liu, X. Two-branch objectness-centric open world detection. In Proceedings of the 3rd International Workshop on Human-Centric Multimedia Analysis, Lisboa, Portugal, 10 October 2022; pp. 35–40. [Google Scholar]
  4. Yu, J.; Ma, L.; Li, Z.; Peng, Y.; Xie, S. Open-world object detection via discriminative class prototype learning. In Proceedings of the 2022 IEEE International Conference on Image Processing (ICIP), Bordeaux, France, 16–19 October 2022; pp. 626–630. [Google Scholar]
  5. Zohar, O.; Wang, K.C.; Yeung, S. Prob: Probabilistic objectness for open world object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 11444–11453. [Google Scholar]
  6. Dong, N.; Zhang, Y.; Ding, M.; Lee, G.H. Open world DETR: Transformer based open world object detection. arXiv 2022, arXiv:2212.02969. [Google Scholar]
  7. Gupta, A.; Narayan, S.; Joseph, K.; Khan, S.; Khan, F.S.; Shah, M. OW-DETR: Open-world Detection Transformer. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022. [Google Scholar]
  8. Ma, S.; Wang, Y.; Wei, Y.; Fan, J.; Li, T.H.; Liu, H.; Lv, F. Cat: Localization and identification cascade detection transformer for open-world object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 19681–19690. [Google Scholar]
  9. Wang, Y.; Yue, Z.; Hua, X.S.; Zhang, H. Random Boxes Are Open-world Object Detectors. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Paris, France, 2–3 October 2023; pp. 6233–6243. [Google Scholar]
  10. Wu, Z.; Lu, Y.; Chen, X.; Wu, Z.; Kang, L.; Yu, J. UC-OWOD: Unknown-classified open world object detection. In Proceedings of the European Conference on Computer Vision, Tel Aviv, Israel, 23–27 October 2022; Springer: Berlin/Heidelberg, Germany, 2022; pp. 193–210. [Google Scholar]
  11. Yang, S.; Sun, P.; Jiang, Y.; Xia, X.; Zhang, R.; Yuan, Z.; Wang, C.; Luo, P.; Xu, M. Objects in Semantic Topology. In Proceedings of the International Conference on Learning Representations, Virtual, 25–29 April 2022. [Google Scholar]
  12. Zhao, X.; Ma, Y.; Wang, D.; Shen, Y.; Qiao, Y.; Liu, X. Revisiting open world object detection. IEEE Trans. Circuits Syst. Video Technol. 2023, 34, 3496–3509. [Google Scholar] [CrossRef]
  13. Zhang, W.; Cheng, T.; Wang, X.; Chen, S.; Zhang, Q.; Liu, W. Featurized query r-cnn. arXiv 2022, arXiv:2206.06258. [Google Scholar]
  14. Chen, Q.; Chen, X.; Wang, J.; Zhang, S.; Yao, K.; Feng, H.; Han, J.; Ding, E.; Zeng, G.; Wang, J. Group detr: Fast detr training with group-wise one-to-many assignment. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Paris, France, 2–3 October 2023; pp. 6633–6642. [Google Scholar]
  15. Jia, D.; Yuan, Y.; He, H.; Wu, X.; Yu, H.; Lin, W.; Sun, L.; Zhang, C.; Hu, H. Detrs with hybrid matching. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 19702–19712. [Google Scholar]
  16. Everingham, M.; Gool, L.V.; Williams, C.K.I.; Winn, J.M.; Zisserman, A. The Pascal Visual Object Classes (VOC) Challenge. Int. J. Comput. Vis. 2010, 88, 303–338. [Google Scholar] [CrossRef]
  17. Lin, T.Y.; Maire, M.; Belongie, S.J.; Hays, J.; Perona, P.; Ramanan, D.; Dollár, P.; Zitnick, C.L. Microsoft COCO: Common Objects in Context. In Proceedings of the European Conference on Computer Vision, Zurich, Switzerland, 6–12 September 2014. [Google Scholar]
  18. Mao, X.; Chen, Y.; Zhu, Y.; Chen, D.; Su, H.; Zhang, R.; Xue, H. COCO-O: A Benchmark for Object Detectors under Natural Distribution Shifts. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Paris, France, 2–3 October 2023; pp. 6339–6350. [Google Scholar]
  19. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster r-cnn: Towards real-time object detection with region proposal networks. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 1137–1149. [Google Scholar] [CrossRef] [PubMed]
  20. Uijlings, J.R.R.; van de Sande, K.E.A.; Gevers, T.; Smeulders, A.W.M. Selective Search for Object Recognition. Int. J. Comput. Vis. 2013, 104, 154–171. [Google Scholar] [CrossRef]
  21. Ma, Y.; Li, H.; Zhang, Z.; Guo, J.; Zhang, S.; Gong, R.; Liu, X. Annealing-Based Label-Transfer Learning for Open World Object Detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 11454–11463. [Google Scholar]
  22. Jaiswal, A.; Wu, Y.; Natarajan, P.; Natarajan, P. Class-agnostic object detection. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Virtual, 5–9 January 2021; pp. 919–928. [Google Scholar]
  23. Kim, D.; Lin, T.Y.; Angelova, A.; Kweon, I.S.; Kuo, W. Learning open-world object proposals without learning to classify. IEEE Robot. Autom. Lett. 2022, 7, 5453–5460. [Google Scholar] [CrossRef]
  24. Gonçalves, G.R.; Sena, J.; Schwartz, W.R.; Caetano, C.A. Pixel-level Class-Agnostic Object Detection using Texture Quantization. In Proceedings of the 2022 35th SIBGRAPI Conference on Graphics, Patterns and Images (SIBGRAPI), Natal, Brazil, 24–27 October 2022; Volume 1, pp. 31–36. [Google Scholar]
  25. Saito, K.; Hu, P.; Darrell, T.; Saenko, K. Learning to detect every thing in an open world. In Proceedings of the European Conference on Computer Vision, Tel Aviv, Israel, 23–27 October 2022; Springer: Berlin/Heidelberg, Germany, 2022; pp. 268–284. [Google Scholar]
  26. Huang, H.; Geiger, A.; Zhang, D. Good: Exploring geometric cues for detecting objects in an open world. arXiv 2022, arXiv:2212.11720. [Google Scholar]
  27. Maaz, M.; Rasheed, H.; Khan, S.; Khan, F.S.; Anwer, R.M.; Yang, M.H. Class-agnostic Object Detection with Multi-modal Transformer. In Proceedings of the 17th European Conference on Computer Vision (ECCV), Tel Aviv, Israel, 23–27 October 2022; Springer: Berlin/Heidelberg, Germany, 2022. [Google Scholar]
  28. Radford, A.; Kim, J.W.; Hallacy, C.; Ramesh, A.; Goh, G.; Agarwal, S.; Sastry, G.; Askell, A.; Mishkin, P.; Clark, J.; et al. Learning transferable visual models from natural language supervision. In Proceedings of the International Conference on Machine Learning, PMLR, Virtual, 18–24 July 2021; pp. 8748–8763. [Google Scholar]
  29. Rao, Y.; Zhao, W.; Chen, G.; Tang, Y.; Zhu, Z.; Huang, G.; Zhou, J.; Lu, J. Denseclip: Language-guided dense prediction with context-aware prompting. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 18082–18091. [Google Scholar]
  30. Zhao, S.; Zhang, Z.; Schulter, S.; Zhao, L.; Vijay Kumar, B.; Stathopoulos, A.; Chandraker, M.; Metaxas, D.N. Exploiting unlabeled data with vision and language models for object detection. In Proceedings of the European Conference on Computer Vision (ECCV), Tel Aviv, Israel, 23–27 October 2022; Springer: Berlin/Heidelberg, Germany, 2022; pp. 159–175. [Google Scholar]
  31. Wei, T.; Chen, D.; Zhou, W.; Liao, J.; Tan, Z.; Yuan, L.; Zhang, W.; Yu, N. Hairclip: Design your hair by text and reference image. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 18072–18081. [Google Scholar]
  32. Patashnik, O.; Wu, Z.; Shechtman, E.; Cohen-Or, D.; Lischinski, D. Styleclip: Text-driven manipulation of stylegan imagery. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 2085–2094. [Google Scholar]
  33. Xu, H.; Ghosh, G.; Huang, P.Y.; Okhonko, D.; Aghajanyan, A.; Metze, F.; Zettlemoyer, L.; Feichtenhofer, C. Videoclip: Contrastive pre-training for zero-shot video-text understanding. arXiv 2021, arXiv:2109.14084. [Google Scholar]
  34. Carion, N.; Massa, F.; Synnaeve, G.; Usunier, N.; Kirillov, A.; Zagoruyko, S. End-to-end object detection with transformers. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; Springer: Berlin/Heidelberg, Germany, 2020; pp. 213–229. [Google Scholar]
  35. He, K.; Gkioxari, G.; Dollár, P.; Girshick, R. Mask r-cnn. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2961–2969. [Google Scholar]
  36. Kuhn, H.W. The Hungarian method for the assignment problem. Nav. Res. Logist. Q. 1955, 2, 83–97. [Google Scholar] [CrossRef]
  37. Peyré, G.; Cuturi, M. Computational optimal transport: With applications to data science. Mach. Learn. 2019, 11, 355–607. [Google Scholar] [CrossRef]
  38. Ge, Z.; Liu, S.; Li, Z.; Yoshie, O.; Sun, J. Ota: Optimal transport assignment for object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 303–312. [Google Scholar]
  39. Cuturi, M. Sinkhorn distances: Lightspeed computation of optimal transport. In Proceedings of the 26th International Conference on Neural Information Processing Systems—Volume 2, Lake Tahoe, NV, USA, 3–6 December 2012; pp. 2292–2300. [Google Scholar]
  40. Kim, J.; Choi, J.; Choi, H.J.; Kim, S.J. Shepherding Slots to Objects: Towards Stable and Robust Object-Centric Learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 19198–19207. [Google Scholar]
  41. Li, J.; Li, D.; Xiong, C.; Hoi, S. Blip: Bootstrapping language-image pre-training for unified vision-language understanding and generation. In Proceedings of the International Conference on Machine Learning, PMLR, Baltimore, MD, USA, 17–23 July 2022; pp. 12888–12900. [Google Scholar]
  42. Ridnik, T.; Ben-Baruch, E.; Noy, A.; Zelnik-Manor, L. Imagenet-21k pretraining for the masses. arXiv 2021, arXiv:2104.10972. [Google Scholar]
  43. Zhu, X.; Su, W.; Lu, L.; Li, B.; Wang, X.; Dai, J. Deformable detr: Deformable transformers for end-to-end object detection. arXiv 2020, arXiv:2010.04159. [Google Scholar]
  44. Liang, X.; Song, P. Excavating roi attention for underwater object detection. In Proceedings of the 2022 IEEE International Conference on Image Processing (ICIP), Bordeaux, France, 16–19 October 2022; pp. 2651–2655. [Google Scholar]
Figure 1. Comparison of pseudo-labeling methods using the objectness score (a) and TGOOD (b). Green boxes represent ground-truth objects of known classes, while red boxes denote candidate boxes that the model identifies as potentially containing an unknown object. The symbol # indicates the objectness score, and @ denotes the similarity score of ROI features with the embedding of the word “cat”. The method using the objectness score tends to misclassify boxes containing fractional-known-class objects as unknown. In contrast, TGOOD provides more accurate labels of unknown-class objects guided by text.
Figure 1. Comparison of pseudo-labeling methods using the objectness score (a) and TGOOD (b). Green boxes represent ground-truth objects of known classes, while red boxes denote candidate boxes that the model identifies as potentially containing an unknown object. The symbol # indicates the objectness score, and @ denotes the similarity score of ROI features with the embedding of the word “cat”. The method using the objectness score tends to misclassify boxes containing fractional-known-class objects as unknown. In contrast, TGOOD provides more accurate labels of unknown-class objects guided by text.
Electronics 13 04528 g001
Figure 2. The architecture of our proposed TGOOD. The symbol ⊕ represents an addition operation. TGOOD consists of five main steps: (1) image/text feature extraction and queries generation; (2) the interaction between queries and enhanced ROI features; (3) one-to-many label assignment; (4) text-guided pseudo-labeling; (5) loss calculation.
Figure 2. The architecture of our proposed TGOOD. The symbol ⊕ represents an addition operation. TGOOD consists of five main steps: (1) image/text feature extraction and queries generation; (2) the interaction between queries and enhanced ROI features; (3) one-to-many label assignment; (4) text-guided pseudo-labeling; (5) loss calculation.
Electronics 13 04528 g002
Figure 3. Diagram of text-guided pseudo-label generation strategy.
Figure 3. Diagram of text-guided pseudo-label generation strategy.
Electronics 13 04528 g003
Figure 4. One−to−one vs. one−to−many matching mechanism. GT: ground truth, BG: background, Q: query.
Figure 4. One−to−one vs. one−to−many matching mechanism. GT: ground truth, BG: background, Q: query.
Electronics 13 04528 g004
Figure 5. Feature enhancement module based on attention mechanism.
Figure 5. Feature enhancement module based on attention mechanism.
Electronics 13 04528 g005
Figure 6. mAP and UR metrics under different configurations of TGOOD components.
Figure 6. mAP and UR metrics under different configurations of TGOOD components.
Electronics 13 04528 g006
Figure 7. Ablation of substructures in SRS. (A) denotes the standard version of the SRS, (B) signifies the enhanced version of the SRS, excluding step (3), and (C) indicates the enhanced version of the SRS.
Figure 7. Ablation of substructures in SRS. (A) denotes the standard version of the SRS, (B) signifies the enhanced version of the SRS, excluding step (3), and (C) indicates the enhanced version of the SRS.
Electronics 13 04528 g007
Figure 8. Ablation of hyperparameters in SRS.
Figure 8. Ablation of hyperparameters in SRS.
Electronics 13 04528 g008
Figure 9. Visualization of TGOOD comparison with ORE-EBUI and OW-DETR. “unknown” in the figure represents unknown-class object in current stage. The red boxes indicates objects of known classes predicted by the model, while the blue boxes signifies objects of unknown classes predicted by the model.
Figure 9. Visualization of TGOOD comparison with ORE-EBUI and OW-DETR. “unknown” in the figure represents unknown-class object in current stage. The red boxes indicates objects of known classes predicted by the model, while the blue boxes signifies objects of unknown classes predicted by the model.
Electronics 13 04528 g009
Table 1. Performance comparison of TGOOD vs. methods with pseudo-labels (OWOD SPLIT).
Table 1. Performance comparison of TGOOD vs. methods with pseudo-labels (OWOD SPLIT).
Task IDsTask 1Task 2Task 3Task 4
WIAOSEmAP (↑)URWIAOSEmAP (↑)URWIAOSEmAP (↑)URmAP (↑)
(↓)(↓)Cur(↑)(↓)(↓)PreCurBoth(↑)(↓)(↓)PreCurBoth(↑)PreCurBoth
ORE-EBUI0.062110,45956.004.90.028210,44552.7026.0039.402.90.0211799038.2012.7029.703.929.6012.4025.30
OW-DETR0.057110,24059.207.50.0278844153.6033.5042.906.20.0156680338.3015.8030.805.731.4017.1027.80
SA0.0417488956.201.90.0213254653.3926.4939.940.80.0146212038.0412.8129.630.130.1113.3125.91
UC-OWOD0.0136929450.662.40.0117560233.1330.5431.843.40.0073380128.8016.3424.658.725.5715.8823.14
RandBox0.0240449861.8010.60.00781880--45.306.30.00541452--39.407.8--35.40
RE-OWOD0.0449-59.709.10.0330-54.1137.2645.649.90.0241-43.0624.6437.5911.437.9928.6635.66
CAT0.0581707059.9021.80.0263590254.0033.6043.8018.60.0177518942.1019.8034.7023.935.1017.1030.60
TGOOD (ours)0.0620499560.3123.10.0253259851.9234.3143.1218.70.0157197841.0623.1035.0722.135.0619.0231.05
Table 2. Performance comparison of TGOOD vs. methods without pseudo-labels (OWOD SPLIT).
Table 2. Performance comparison of TGOOD vs. methods without pseudo-labels (OWOD SPLIT).
Task IDsTask 1Task 2Task 3Task 4
WIAOSEmAP (↑)URWIAOSEmAP (↑)URWIAOSEmAP (↑)URmAP (↑)
(↓)(↓)Cur(↑)(↓)(↓)PreCurBoth(↑)(↓)(↓)PreCurBoth(↑)PreCurBoth
OCPL0.0423567056.648.30.0220569050.6527.5439.107.70.0162516638.6314.7430.6711.930.7514.4226.67
2B-OCD0.0480-56.3712.10.0160-51.5725.3438.469.40.0137-37.2413.2329.2411.730.0613.2825.82
PROB0.0569519559.5019.40.0344645255.7032.2044.0017.40.0151264143.0022.2036.0019.635.7018.9031.50
Ann0.0604833256.6712.80.0269945451.9629.1340.555.00.0157663540.8214.5632.079.831.6813.0927.03
TGOOD (ours)0.0620499560.3123.10.0253259851.9234.3143.1218.70.0157197841.0623.1035.0722.135.0619.0231.05
Table 3. Performance comparison of TGOOD with open-environment object detection methods (MS-COCO SPLIT).
Table 3. Performance comparison of TGOOD with open-environment object detection methods (MS-COCO SPLIT).
Task IDsTask 1Task 2Task 3Task 4
WIAOSEmAP (↑)URWIAOSEmAP (↑)URWIAOSEmAP (↑)URmAP (↑)
(↓)(↓)Cur(↑)(↓)(↓)PreCurBoth(↑)(↓)(↓)PreCurBoth(↑)PreCurBoth
ORE-EBUI--61.401.5--56.5026.1040.603.9--38.7023.7033.703.633.6026.3031.80
OW-DETR0.045819,81571.505.70.049919,74962.8027.5043.806.20.0248923345.2024.9038.506.938.2028.1033.10
CAT0.0234212670.6524.50.0330444165.8335.5450.6822.20.0208354551.0932.8245.0025.045.4834.9042.84
PROB0.0196191573.8517.30.0307340066.1536.1950.4222.10.0170155247.7230.2741.9124.542.8031.7240.03
TGOOD (ours)0.0443249063.8129.40.0244136754.2348.8751.5529.00.0174140549.9643.0047.6435.147.4645.0246.85
Table 4. Performance comparison of TGOOD with classic OWOD methods (COCO-O SPLIT).
Table 4. Performance comparison of TGOOD with classic OWOD methods (COCO-O SPLIT).
Task IDsTask 1Task 2Task 3Task 4
WIAOSEmAP (↑)URWIAOSEmAP (↑)URWIAOSEmAP (↑)URmAP
(↓)(↓)Cur(↑)(↓)(↓)PreCurBoth(↑)(↓)(↓)PreCurBoth(↑)PreCurBoth
ORE-EBUI0.1115151820.7715.30.0482152519.1914.8917.0411.80.0252106614.937.4712.4410.412.428.8511.53
OW-DETR0.124555904.6828.90.062540764.130.552.3428.80.030722431.723.542.3223.90.051.070.30
TGOOD (ours)0.138067819.0752.80.051840114.4219.2816.8548.50.027422114.879.6613.1355.014.1612.7313.80
Table 5. Ablation experiments of TGOOD components.
Table 5. Ablation experiments of TGOOD components.
Line IDBaselineObj1Obj5SRSOTMRRMWI (↓)AOSE (↓)mAP (↑)UR (↑)
1 0.071879,51657.060
2 0.076177,92756.4210.0
3 0.075572,35855.8713.7
4 0.075615,16956.3322.8
5 0.0637514359.9622.8
6 0.0620499560.3123.1
Table 6. Framework-agnostic validation of SMS on ORE-EBUI.
Table 6. Framework-agnostic validation of SMS on ORE-EBUI.
StrategiesWI (↓)AOSE (↓)mAP (↑)UR (↑)
ORE-EBUI0.062110,45956.004.9
ORE+Obj50.048017,34518.317.1
ORE+SMS0.052812,12056.037.6
Table 7. Comparative experiments of different ROI feature enhancement modules.
Table 7. Comparative experiments of different ROI feature enhancement modules.
StrategiesWI (↓)AOSE (↓)mAP (↑)UR (↑)
TGOOD-RRM0.0637514359.9622.8
BiLstm0.0637506860.2222.9
GCN0.0634507460.0623.0
RoIAttn0.0633516159.8923.1
TGOOD0.0620499560.3123.1
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wang, X.; Xu, D. Text-Guided Unknown Pseudo-Labeling for Open-World Object Detection. Electronics 2024, 13, 4528. https://doi.org/10.3390/electronics13224528

AMA Style

Wang X, Xu D. Text-Guided Unknown Pseudo-Labeling for Open-World Object Detection. Electronics. 2024; 13(22):4528. https://doi.org/10.3390/electronics13224528

Chicago/Turabian Style

Wang, Xuefei, and Dong Xu. 2024. "Text-Guided Unknown Pseudo-Labeling for Open-World Object Detection" Electronics 13, no. 22: 4528. https://doi.org/10.3390/electronics13224528

APA Style

Wang, X., & Xu, D. (2024). Text-Guided Unknown Pseudo-Labeling for Open-World Object Detection. Electronics, 13(22), 4528. https://doi.org/10.3390/electronics13224528

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop