Next Article in Journal
A Survey on GAN Techniques for Data Augmentation to Address the Imbalanced Data Issues in Credit Card Fraud Detection
Previous Article in Journal
Painting the Black Box White: Experimental Findings from Applying XAI to an ECG Reading Setting
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Skew Class-Balanced Re-Weighting for Unbiased Scene Graph Generation

School of Electrical Engineering, Korea Advanced Institute of Science and Technology (KAIST), Daejeon 34141, Republic of Korea
*
Author to whom correspondence should be addressed.
Mach. Learn. Knowl. Extr. 2023, 5(1), 287-303; https://doi.org/10.3390/make5010018
Submission received: 27 January 2023 / Revised: 17 February 2023 / Accepted: 7 March 2023 / Published: 10 March 2023
(This article belongs to the Topic Applications in Image Analysis and Pattern Recognition)

Abstract

:
An unbiased scene graph generation (SGG) algorithm referred to as Skew Class-Balanced Re-Weighting (SCR) is proposed for considering the unbiased predicate prediction caused by the long-tailed distribution. The prior works focus mainly on alleviating the deteriorating performances of the minority predicate predictions, showing drastic dropping recall scores, i.e., losing the majority predicate performances. It has not yet correctly analyzed the trade-off between majority and minority predicate performances in the limited SGG datasets. In this paper, to alleviate the issue, the Skew Class-Balanced Re-Weighting (SCR) loss function is considered for the unbiased SGG models. Leveraged by the skewness of biased predicate predictions, the SCR estimates the target predicate weight coefficient and then re-weights more to the biased predicates for better trading-off between the majority predicates and the minority ones. Extensive experiments conducted on the standard Visual Genome dataset and Open Image V4 and V6 show the performances and generality of the SCR with the traditional SGG models.

1. Introduction

Currently, computer vision [1,2,3,4,5,6] has been prevalent in various fields: material [7], chemical [8,9], and medical science [10]. Of great interest are tasks related to classifying and detecting objects by classical neural networks [11,12,13,14,15]. With the development of computer vision, graph neural networks [16,17,18,19,20,21,22] and multimodal text and image models [23,24,25,26,27,28] have also been intensely investigated. However, there is less information on Scene Graph Generation (SGG) models in the literature. For this reason, SGG has recently been receiving increased attention for improving image comprehension, divided into two approaches: one-stage [29,30] and two-stage [31,32,33,34,35,36,37,38,39,40,41,42,43,44,45]. The core building blocks of SGG are the objects in the image. There can be diverse relationships among the objects [46,47]. All relationships can be represented as triplets subject, relation, object, which can be used for generating the scene graph. The precise interpretation of an image depends on the core relations between the subject–object pairs. For example, given an image of dog, woman, kite objects, the image can be interpreted in multiple ways such as dog , has , leg , dog , near , woman , and woman , holding , kite .
In general, however, the real-world large data sets often have shown long-tailed predicate distributions as depicted in Figure 1a—the predicate proportion of the Visual Genome [31], containing 51 predicates. One cause of long-tail distribution is the background predicates. The background predicates account for more than 80 % in total predicates since it is not trivial to annotate all possible relationships, leading to possibly far more subject-object pairs without annotating ground truth than ones with ground truth labels. Another reason is the visual event frequency, as shown in Figure 1b. The majority predicates more frequently occur than the minority ones given a visual subject–object pair in the real-world scene. For example, given a man-board pair, the possible predicates—backgrounds, on, holding, riding, etc. represents a long-tail distribution. The predicate on is more frequently observed than holding and riding .
This long-tailed label distribution causes the trained model to tend to be biased toward the majority predicates. The SGG models [43,48,49,50,51,52] focus on increasing the minority predicate performances for the unbiased scene graph generation, i.e., increasing mean recall scores. However, these methods lead to deteriorating performances of the majority predicates, drastically dropping recall scores, i.e., losing the majority predicate performances. Recently, the issue has been alleviated through resampling images and object instances [49]. However, this method needs more computing power and time to train the model on the resampled samples. The current state-of-the-art models focus on re-balancing biased loss [53] or correcting noisy labels [54] to acquire an unbiased SGG model. Nevertheless, the prior works have not yet adequately described and analyzed the trade-off performances between majority and minority predicates for learning SGG models based on the given imbalanced datasets.
In this paper, the Skew Class-Balanced Re-Weighting (SCR) loss function is considered for alleviating the issues and acquiring the best trade-off performances in the multiple SGG tasks. Leveraged by the skewness of predicate predictions, the SCR estimates its weight coefficients and then reweights more to biased predicate samples to adaptively be unbiased SGG models. The extensive experimental results show that the SCR loss function gives more generalized performances than priors in the multiple SGG tasks on the Visual Genome dataset [55] and the Open Images datasets [56].
  • Contributions of our SCR learning scheme to unbiased SGG models:
  • Leveraged by the skewness of biased predicate predictions, the Skew Class-Balanced Re-Weighting (SCR) loss function is firstly proposed for the unbiased scene graph generation (SGG) models.
  • The SCR is applied to the current state-of-the-art SGG models to show its effectiveness, leading to more generalized performances: the SCR outperforms the prior reweighted methods on both mean recall and recall measurements in the multiple SGG tasks.
This paper is organized as follows. The Related Work section provides discussions on unbiased scene graph generation. The unbiased SGG section presents scene graph generation. In the Skew Class-Balanced Re-Weighting (SCR) section, the SCR loss function is depicted in detail. In the experimental section, the results of scene graph generation on the Visual Genome and Open Image V4 and V6 dataset are examined, along with an analysis and ablation study on the SCR with the current state-of-the-art SGG models. Finally, we conclude.

2. Related Works

Unbiased Scene Graph Generation (SGG). Predicate distribution is much more long-tailed than object distribution. For N objects and R predicates, the model has to address the fundamental challenge of learning O ( N 2 R ) relations with few [57,58]. To overcome the limited training dataset, the linguistic external knowledge [38,57,59] was used by Yu et al. [60], regularizing the deep neural network; using linguistic knowledge, the probabilistic model has also alleviated the semantical ambiguity of visual relationships [61]. Furthermore, to alleviate the imbalanced relationship distribution, Yin et al. [38] reformulated the conventional one-hot classification as a n-hot multiclass hierarchical recognition via novel Intra-Hierarchical trees (IH-trees) for each label set in the triplet s u b j e c t , p r e d i c a t e , o b j e c t . Recently, unbiased SGG [43,47,48,49,50,51,52,53,54,62,63,64,65,66,67,68,69,70,71] has drawn unprecedented interest for more generalized SGG models. Occurrence-based Node Priority Sensitive (NPS)-loss [47] was used for balancing predictions; the Total Direct Effect (TDE) method has proposed firstly for unbiased learning by Tang et al. [43], which directly separates the bias from biased predictions through the counterfactual methodologies on causal graphs; CogTree [48] addressed the debiasing issue based on the coarse-to-fine structure of the relationships from the cognition viewpoint; Li et al. [49] improved the context modeling for tail categories by designing the bipartite graph network and message propagation on resampled objects and images. Lastly, the Predicate Probability Distribution based Loss (PPDL) [53] has proposed to train the biased SGG models, which measure the semantic predicate representation to re-balance the biased training loss. In this work, the Skew Class-Balanced Re-Weighting (SCR) loss function is proposed for alleviating biased predicate predictions, leading to the most generalized SGG models through the novel adaptive re-weighting learning scheme.
Re-Weighting Based Unbiased SGG. Overall unbiased SGG models can be categorized into the re-balancing strategy of re-weighting [47,48,50,53] and re-sampling [49] and biased model-based strategy [43,51,52]. For unbiased SGG modes, Tang et al. [43] first investigated the re-weighting learning algorithm. However, they observed that the performances of the majority predicates were drastically dropped, resulting in low recall scores while with high mean recall scores. This shows the general tendency that there is a trade-off performance between majority predicates and minority ones. To alleviate the trade-off issue, Yu et al. [48] proposed CogTree based on the coarse-to-fine structure of the relationships from the cognition viewpoint. Recently, the Predicate Probability Distribution (PPD) [53] re-balances the biased training loss according to the similarity between the predicted probability distribution and the estimated one. However, it has not yet correctly analyzed the trade-off performances between the majority and minority classes in various SGG tasks. In this paper, we measure the sample skew score based on the sample estimates for bias toward the majority classes to assign the sample weight correctly. The sample skewness is computed as the Fisher–Pearson coefficient of skewness on its sample mean value [72]. However, since the mean value tends to be biased toward the majority predicates, we measure the sample skew score fairly on its target logit instead of its mean value. Based on the sample skew score, the SCR assigns the sample weights adaptively—if there is no bias, we assign fewer weights to the sample (majority). If it is biased to one side, we assign larger weights to the samples (minority). Such that the SGG models with SCR show superior performances and generality on the multiple SGG tasks.

3. Unbiased Scene Graph Generation

In this section, we discuss the general scene graph generation model of the object and predicate predictions and depict the predicate sample estimates for measuring its skewness of biased predictions.

3.1. Scene Graph Generations

Given an image I, a scene graph model generates a graph G = ( V , E ) , where V and E are the sets of nodes and edges, respectively. Each node o i V is represented by a bounding box v b b o x R 4 and a corresponding class label o y C o b j . Each edge r i , j E represents the predicate between the subject node v i and the object node v j . The corresponding predicate label is r y C r e l .
We depict the proposed unbiased SGG framework as shown in Figure 2. Object detector outputs its predictions (a): man and hourse. Given man-horse label prediction pair, the traditional SGG model predicts (d) predicate predictions ( R ^ see Equation (3)) through a fusion of (b) non-visual FREQ prior predicates and (c) visual predicate predictions ( R ^ v i s ). For unbiased SGG model training, our SCR estimates (e) a sample size of the possible predicate candidates through FREQ ( R f r e q or R e m b ) and measures (f) the target label skew score and then calculates (g) the training target sample weight for the adaptive re-weighted loss. In (f) and (g), red lines around the target label indicate predicate skew ( S s k e w i see Equation (9)) and re-weight scores ( W i see Equation (7)), respectively, where they have an approximately inverse relationship.

3.1.1. Object Predictions

Following [43], the node features are derived from the object detector. In particular, for each bounding box v i b b o x , the detector returns a RoI-Align feature x i R o I and an object label embedding l i . In general SGG model, the N number of node features are constructed by vector concatenation X = { [ x i R o I ; l i ; b i ] } i = 1 N , where b i is the embedded box feature from the box coordinate v i b b o x .
X ˜ = f c o b j ( X ) R N × D o b j ,
where all f c * denote a fully connected layer for linear transformations or logits, and the object feature dimension D o b j depends on the SGG model. As depicted in Figure 2a, the predicted object label of O ^ R N × | C o b j | is estimated by object logits f c ( X ˜ ) as follows:
O ^ = f c ( X ˜ ) .

3.1.2. Predicate Predictions

Inspired by Tang et al. [43], Zhu et al. [73], predicate predictions can be made by employing multiple logits from visual and non-visual features. We follow the sum over all outputs to generate the final predicate prediction. As illustrated in Figure 2d, the combined predicate logit R ^ is estimated based on the summation of the visual logits and the non-visual logits as follows:
R ^ = R ^ v i s R ^ f r e q R ^ e m b ,
where R ^ R N ( N 1 ) × | C r e l | ; ⊕ is an element-wise sum; as shown in Figure 2c, R ^ v i s is the predicate logits from visual feature F v i s such as a D v i s dimensional union feature and a subject–object pair feature, which also depends on the SGG model,
R ^ v i s = f c v i s ( F v i s ) , F v i s R N ( N 1 ) × D v i s , R ^ f r e q = Sigmoid ( R f r e q ) R N ( N 1 ) × | C r e l | , R ^ e m b = f c e m b ( L e m b ) , L e m b R N ( N 1 ) × 400 .
The FREQ [74] as a non-visual feature, R f r e q (Figure 2b) looks up the empirical distribution over relationships between subject o ^ i and object o ^ j as computed in the training set where o ^ i , o ^ j O ^ . However, since FREQ does not consider any image representations when predicting predicates, it tends to lead to biased predicate predictions due to its imbalanced predicate distribution. To minimize the biased effects, we use the Sigmoid-activated FREQ predicate logits. In addition, for acquiring the more smoothness of the empirical distribution, the concatenated subject–object embedding R ^ e m b is added to the predicate predictions where L e m b = { [ l i ; l j ] } i , j N .

3.2. Sample Estimates

Non-visual predicate features tend to be more biased than visual features due to an imbalanced training set. If the degree of biased predictions can be measured, we can leverage its value to learn the SGG models without bias. According to Brown [72], the degree of bias prediction can generally be measured in a skew score. In general, if there is no bias, the skew score is close to 0, and if it is biased to one side, the skew score is over either 1 or + 1 .
In SCR, we need to estimate how many predicate samples are biased to measure the predicate skew scores. To approximate the biased sample numbers, we use two non-visual prior predicate distributions of FREQ, R ^ f r e q and subject–object label embedding R ^ e m b as described in Figure 2e. Based on the two predicate distributions, at first, we define the skew logit R ^ s k e w with the Sigmoid activation function as follows:
R ^ s k e w = Sigmoid ( R f r e q R ^ e m b ) .
To estimate the predicate sample weights properly, the SCR approximates skewness through the predicate sample estimates R ^ s k e w in Equation (5). We investigate the best predicate sample estimates through several experiments with a combination of non-visual predictions as follows:
  • SCR of EMB: R ^ s k e w = Sigmoid ( R ^ e m b ) ;
  • SCR of FREQ: R ^ s k e w = Sigmoid ( R f r e q ) ;
  • SCR of FREQ+EMB: R ^ s k e w = Sigmoid ( R f r e q R ^ e m b ) .
Then, the predicate sample estimates are acquired as follows:
M = i N ( N 1 ) r ^ i ,
where r ^ i R ^ s k e w ; the estimated m y M is the number of yth predicate sample size and m y [ 0 , N ( N 1 ) ) .
Moreover, the skew predicate R ^ s k e w serves to estimate the predicate candidates that a subject–object pair can have fairly since the Sigmoid activation function suppresses the more extensive biased predictions. For example, a Man-Horse pair may have predicates such as riding, on, with, etc. The Sigmoid activation function amplifies the frequency of the minority predicates while squeezing that of the majority ones, as shown in Figure 2b,e, which is used to calculate Skew Class-Balanced Effective Number that we depict in the following section.
Note in Equation (5), the Sigmoid activation function was applied to estimate the predicate sample sizes (Equation (6)) and measure skewness (Equation (9)), given a subject–object pair. The main reason was that it provides smoothly activated outputs from 0 to 1. We assumed that an output value of 1 is regarded as one predicate.

4. Skew Class-Balanced Re-Weighting (SCR)

In this section, leveraged by the biased predictions deduced by the predicate sample estimates, the Skew Class-Balanced Re-Weighting (SCR) performs sample weight estimates for learning unbiased SGG models.

4.1. Skew Class-Balanced Effective Number

The Skew Class-Balanced Effective Number approximates the mini-batch class-balanced re-weighting coefficients E m y based on the predicate skew logit R ^ s k e w in Equation (5). The ith predicate sample E m y i is defined as follows as [75,76]:
E m y i = 1 β i m y ( 1 β i ) ,
where β i = ( m y 1 ) / m y ; the effective number satisfies the following properties that E m y i = 1 if β i = 0 ( m y = 1 ) ; E m y i m y as β i 1 ( m y ) such that β i controls how fast E m y i grows as the target predicate sample size m y increases.
To estimate the ith predicate effective number E m y i , we adaptively estimate the β i [ 0 , 1 ) by using the entropy H s k e w i and skew score function S s k e w i of the R ^ s k e w as shown in Algorithm 1: if S s k e w i > S t h then the β i [ 0 , 1 ) that assigns more weights to the minority predicate samples than the majority ones; otherwise, S s k e w i S t h then β i = 0 that re-weights uniformly over the entire class sample loss, i.e., conventional cross-entropy loss. The threshold S t h is determined as follows:
S t h = S ¯ s k e w δ ,
where S ¯ s k e w and δ = 0.7 are the mean value of S s k e w and the hyper-parameter used, respectively, in all experiments.

4.2. Skew Measures

In training, the predicate sample skewness depends on the predicate label. In other words, according to Brown [72], some majority predicate skew values tend to be greater than zero while others tend to be zero or negative. This is an unfair skew measure, leading the SCR Algorithm 1 to more weights either in the majority predicates or in the minority ones. In this paper, as portrayed in Figure 2f, to measure the skew value of all target labels fairly, we use the target skew logit instead of the mean value. Therefore, the ith predicate sample skew S s k e w i is firstly measured by the following equation given the target label index y as follows:
S s k e w i = 1 | C r e l | r ^ i , j R ^ s k e w i ( r ^ i , j r ^ i , y ) 3 1 | C r e l | r ^ i , j R ^ s k e w i ( r ^ i , j r ^ i , y ) 2 3 / 2 .
Then, we use the uniformness to determine the β . The uniformness provides confidence in the predicate sample estimates. To calculate the uniformness of the ith predicate sample, we estimate the entropy H s k e w i as follows:
H s k e w i = λ s k e w r ^ i , j R ^ s k e w i p ( r ^ i , j ) log | C r e l | p ( r ^ i , j ) ,
where the number of predicates | C r e l | is used as the base of the logarithm, which H s k e w i [ 0 , 1 ] ; the r ^ i R ^ s k e w have uniform distributions when H s k e w i is close to 1; otherwise, the predicate sample may have either skew distributions or correct distributions and the λ s k e w = 0.06 is the coefficient of S s k e w . The following section depicts the relationship between skew and entropy in detail.
Algorithm 1 Skew Class-Balanced Effective Number.
Require: 
Dataset D , SGG Model f θ
1:
for t = 0 , 1 , 2 , , T do
2:
     B Minibatch( D )
3:
     R ^ s k e w S k e w _ L o g i t s ( R ^ f r e q , R ^ e m b ; B , f θ )             (Equation (5))
4:
     M S a m p l e _ E s t i m a t e s ( R ^ s k e w )                              (Equation (6))
5:
     S s k e w S k e w ( R ^ s k e w , R )                                               (Equation (9))
6:
     H s k e w E n t r o p y ( R ^ s k e w )                                             (Equation (10))
7:
    if  S s k e w > S t h  then
8:
         β = 1 . 0 H s k e w
9:
    else
10:
         β = 0 . 0
11:
    end if
12:
     E m y i = 1 β i m y 1 β i ; m y M , i 0 , N ( N 1 )         (Equation (7))
13:
end for

4.3. Target Sample Weights

The interpretation of the relationship between skew and weight is depicted based on biased predicate prediction as shown in Figure 3. In Figure 3a, to understand the sample weight estimate from the predicate sample estimates, we assume that the predicate biased predictions R ^ s k e w are given by the predicate sample frequency M 1 / 5 (see Figure 1a) and the predicate entropy H s k e w , i.e., the more frequent sample is the more biased prediction is; the more biased the prediction is the less entropy is. The skew S s k e w measures not only the degree of true/false prediction but also biased predictions as shown in Figure 3c. The following simple equation summarizes the degree of true/false predictions:
S s k e w i > 0 if arg max j r ^ i , j r y S s k e w i < 0 otherwise .
where S s k e w i 1 for the true prediction; S s k e w i = 0 for the uniform prediction and S s k e w i 0 for the false and biased prediction. Moreover, the predicate target label-wise skew has more discrete at high entropy. The final ith sample weight w y i = 1 / E m y i (Figure 2g) is acquired by the criteria of skew and entropy as shown in Figure 3b. The minority predicates have larger weights than the majority when S s k e w i 0 .

4.4. Learning with SCR

Except for object loss, all traditional SGG models are learned by the skew class-balanced re-weighting cross-entropy loss functions. The conventional cross-entropy loss for the objects is computed, given object predictions O ^ and the object ground-truth o y i O as follows:
L o b j ( O ^ , O ) = i N γ o b j · w y i log exp ( o ^ y ) o ^ j O ^ i exp ( o ^ j ) ,
where, given N of objects, ith object sample weight is w y i = 1 s.t. | C o b j | = j | C o b j | w j i ; here, the sample weight denominator γ o b j = 1 / i N w y i is used for the mean object cross-entropy loss. The skew class-balanced cross-entropy loss computes the skew predicate-balanced cross-entropy, based on the predicate predictions R ^ and the predicate ground-truth r y i R :
L r e l ( R ^ , R ) = i N ( N 1 ) γ r e l · w y i log exp ( r ^ y ) r ^ j R ^ i exp ( r ^ j ) ,
where, given N ( N 1 ) of object-subject pairs, the ith predicate sample weight is w y i = 1 E m y i = 1 β i 1 β i m y s.t. | C r e l | = j | C r e l | w j i ; here, we set the predicate sample normalizer as γ r e l = 1 / i N ( N 1 ) w y i for mean predicate loss. In summary, the total objective loss function L t o t a l for unbiased SGG learning can be formulated as follows:
L t o t a l = L o b j + L r e l .

5. Experiments

The proposed SCR is evaluated with the traditional SGG models on the Visual Genome benchmark datasets [55], and the performances of the SCR are compared with others in the multiple SGG tasks. The source code is available at https://github.com/ihaeyong/Unbiased-SGG (accessed on 26 January 2023).

5.1. Visual Genome

We used Visual Genome (VG) [55] dataset to train and evaluate our models, which is composed of 108k images across 75k object categories and 37k predicate categories. We followed the widely adopted VG split [31,42,74] containing the most frequent 150 object categories and 50 predicate categories. The original split only has a training set ( 70 % ) and a test set ( 30 % ). We followed Zellers et al. [74] to sample a 5k validation set from the training set for parameter tuning.

5.2. Open Images

The Open Images dataset [56] is a large-scale dataset proposed by Google recently. Compared with the Visual Genome dataset, it has a superior annotation quality for the scene graph generation. In this work, we conduct experiments on Open Images V 4 & V 6 , following similar data processing and evaluation protocols in [47,56,77]. The Open Images V4 is introduced as a benchmark for scene graph generation by Zhang et al. [77] and Lin et al. [47], which has 53,953 and 3234 images for the train and validation sets, 57 objects categories, and 9 predicate categories in total. The Open Images V6 has 126,368 images used for training, 1813, and 5322 images for validation and testing, respectively, with 301 object categories and 31 predicate categories. This dataset has a comparable amount of semantics categories with the VG.

5.3. Experiments Configurations

State-of-the-Art Comparisons. For fair comparisons, all the compared SGG models should use the FREQ [74], which looks up the empirical distribution over relationships between subject prediction o ^ i and object ones o ^ j . To evaluate the effectiveness of the SCR learning algorithm, we follow the same experimental settings as the CogTree [48]. We also set the current state-of-the-art SGG models as the baseline: MOTIFS [74], VCTree [41] and SG-Transformer [48] which contains 3 O 2 O blocks and 2 R 2 O blocks with 12 attention heads, and Bipartite-Graph [49] without resampling layers and compare the performance with the state-of-the-art debiasing approach TDE [43], CogTree [48], PCPL [50], DLFE [51], BPL-SA [52], PPDL [53], and NICE [54].
Implementation. Following the previous works [43,48,49], the object detector is the pre-trained Faster R-CNN [78] with ResNeXt-101-FPN [79]. In bi-level resampling [49], we also set the repeat factor t = 0.07 , instances drop rate γ d = 0.7 , and weight of fusion the entities features ρ = 5 . The α , β are initialized as 2.2 and 0.025, respectively.

5.4. Evaluations

Our SCR has the following two evaluations:
Relationship Retrieval (RR) contains three sub-task: (1) Predicate Classification (PredCls): taking ground truth bounding boxes and labels as inputs, (2) Scene Graph Classification (SGCls): using ground truth bounding boxes without labels, (3) Scene Graph Detection (SGDet): detecting SGs from scratch. The conventional metric of RR is Recall@K (R@K), included in this paper even though the biased prediction is reported by Misra et al. [80] for the performance of the SCR. Moreover, to evaluate the general performances, we adopted mean Recall@ K (mR@K) that retrieves each predicate separately and then averages R@K for all predicates.
Zero-Shot Relationship Retrieval (ZSRR). The Zero-Shot Recall@K was firstly evaluated on the VG dataset in [43], which reports the R@K of those subject-predicate-object triplets that have never been observed in the training set. ZSRR also has three sub-tasks as RR.

5.5. Quantitative Results

Visual Genome. The SCR is compared with others on the two evaluation tasks: RR and ZSRR, which are the same as shown in Table 1 and Table 2. The SCR achieves the best and second best performances over the previous methods: TDE, PCPL, Cogtree, DLFE, BPL-SA PPDL, and NICE, demonstrating its generality and effectiveness on the two measures of RR task. Moreover, the SCR shows the best trade-off performances on the ZSRR task as shown in Figure 2.
The Best Trade-off performances. To analyze the trade-off between majority and minority predicate performances, firstly, we need to understand the two measurements, mRR and RR. The mRR was introduced to calculate the recall on each predicate category independently and then to average the results by Tang et al. [41]. This is because of RR’s bias, which was reported by Misra et al. [80]. The higher the recall scores (RR) are, the more biased the majority categories are. Therefore, a higher mRR is a more unbiased SGG model; a higher RR is a more biased SGG model. Traditional SGG models show the tendency that if mRR is higher than RR, RR performances are low, i.e., MOTIFS, VCTREE, and SG-Transformer show the best Recall but low mRR@100 in Table 1. However, our SCR shows the best trade-off in terms of mRR and RR measurements; MOTIFS with SCR shows the best performances in mRR and the second-best performances in RR except for MOTIFS. So do the other models’ results. Figure 4 supports our analysis of the best trade-off since overall predicate performances were better than the traditional SGG model (SG-Transformer).
Open Image V4 and V6. To show the effectiveness of SCR, we set BGNN as the baseline, as shown in Table 3. On Open Images Dataset V4, SCR outperformed BGNN except for s c o r e w t d measurement. Mainly, SCR shows outstanding performance in terms of phrase evaluation. Moreover, on Open Images Dataset V6, SCR outperformed all baselines such as BGNN, GPS-Net, Etc., showing a good trade-off between mean recall@50 and recall@50. These results proved that we could have a good trade-off performance in long-tailed predicated distributions if properly assigning weights in training.

5.6. Ablation Study

We investigate the predicate-biased prediction and the best hyper-parameter settings of the SCR loss function for the better generalized SGG models.
Predicate Bias. To estimate the predicate bias and assign the proper sample weights, we define the predicate sample estimates R ^ s k e w of R ^ f r e q and R ^ e m b in Equation (5). To investigate the proper predicate sample estimates, we examined the effectiveness of the predicate sample estimates-SCR of EMB, SCR of FREQ, and SCR of FREQ+EMB with the fixed predicate predictions (Equation (3)) as shown in Table 4. In the experiments, the SCR of FREQ+EMB leads to more generalized performances over others. In summary, the previous Re-Weight methods worsen the recall performances, while the SCR does not.
Hyper-parameters. The hyper-parameters of SCR control the weight of the SGG loss. We investigate the best hyper-parameter settings as shown in Table 5. The best hyper-parameter settings δ = 0.7 and λ s k e w = 0.06 show the generalized performances of the SGG tasks. The smaller λ s k e w is, the higher mean Recall scores are, while the higher λ s k e w is, the higher Recall scores are, i.e., the λ s k e w controls the trade-off between the majority predicates and the minority ones since the smaller H s k e w tend to assign more weights to the minority predicates in the SCR. The δ = 0.7 shows the best proportion of Re-weighting the predicate samples.

5.7. Qualitative Examples

To demonstrate the effectiveness of the sample-wise SCR Re-weighting, we show the comparison of Recall @100 on PredCls of all predicates based on the SG-Transformer [48] as shown in Figure 4. The SCR of FREQ+EMB achieves a significant performance gain on the overall predicate categories. Moreover, the skew predicate R ^ s k e w serves to estimate not only the predicate candidates that a subject–object pair can have but also the target skew S s k e w which determines the sample weight. Figure 5 includes a subject–object pair which have possible predicates, its predicate prediction R ^ i , and the target skew S s k e w i deduced by the predicate sample estimate R ^ s k e w i . The higher the sample estimate is, the lower the target skew is. To be specific, when one woman-  S s k e w wearing -coat pair have possible predicates such as backgrounds, has, wearing, in, etc., we show predicate predictions R ^ i with its target skew S s k e w i deduced by its sample estimates R ^ s k e w i . The larger R ^ s k e w i R ^ s k e w i tends to be a smaller S s k e w i , when compared to another hill- S s k e w covered in  -snow pair in a SCR-trained SGG model.

6. Conclusions

In this paper, the unbiased Scene Graph Generation (SGG) algorithm, referred to as Skew Class-Balanced Re-Weighting (SCR), was proposed for considering the unbiased predicate prediction caused by the long-tailed distribution. The prior works focus mainly on alleviating the deteriorating performances of the minority predicate predictions, showing drastic dropping recall scores, i.e., forgetting the majority predicate class. It has not yet properly analyzed the trade-off performances between majority and minority predicates in the given SGG datasets. In this paper, to address the issues leveraged by the skewness of biased predicate predictions, firstly, the SCR estimated the predicate re-weighting coefficient and then re-weighted more to the biased predicates for the better trading-off performances between the majority and the minority predicates. Extensive experiments conducted on the standard Visual Genome dataset and Open Image V4 and V6 showed the SCR’s effectiveness and generality with the traditional SGG models.

Author Contributions

Writing—original draft, H.K.; Writing—review & editing, C.D.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Institute of Information & communications Technology Planning & Evaluation (IITP) grant funded by the Korean government (MSIT) (No. 2022-0-00951, Development of Uncertainty-Aware Agents Learning by Asking Questions), and partly supported by the National Research Foundation of Korea (NRF) grant funded by the Korea government (MSIT) (No. 2022R1A2C2012706).

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Jarvis, R.A. A perspective on range finding techniques for computer vision. IEEE Trans. Pattern Anal. Mach. Intell. 1983, PAMI-5, 122–139. [Google Scholar] [CrossRef] [PubMed]
  2. Forsyth, D.A.; Ponce, J. Computer Vision: A Modern Approach; Prentice Hall Professional Technical Reference: Hoboken, NJ, USA, 2002. [Google Scholar]
  3. Voulodimos, A.; Doulamis, N.; Doulamis, A.; Protopapadakis, E. Deep learning for computer vision: A brief review. Comput. Intell. Neurosci. 2018, 2018, 7068349. [Google Scholar] [CrossRef] [PubMed]
  4. Kang, J.-S.; Kang, J.; Kim, J.J.; Jeon, K.W.; Chung, H.J.; Park, B.H. Neural Architecture Search Survey: A Computer Vision Perspective. Sensors 2023, 23, 1713. [Google Scholar] [CrossRef] [PubMed]
  5. Høye, T.T.; Ärje, J.; Bjerge, K.; Hansen, O.L.; Iosifidis, A.; Leese, F.; Mann, H.M.; Meissner, K.; Melvad, C.; Raitoharju, J. Deep learning and computer vision will transform entomology. Proc. Natl. Acad. Sci. USA 2021, 118, e2002545117. [Google Scholar] [CrossRef] [PubMed]
  6. Guo, M.H.; Xu, T.X.; Liu, J.J.; Liu, Z.N.; Jiang, P.T.; Mu, T.J.; Zhang, S.H.; Martin, R.R.; Cheng, M.M.; Hu, S.M. Attention mechanisms in computer vision: A survey. Comput. Vis. Media 2022, 8, 331–368. [Google Scholar] [CrossRef]
  7. Sethian, J.A. Level Set Methods and Fast Marching Methods: Evolving Interfaces in Computational Geometry, Fluid Mechanics, Computer Vision, and Materials Science; Cambridge University Press: Cambridge, UK, 1999; Volume 3. [Google Scholar]
  8. Scheuerman, M.K.; Hanna, A.; Denton, E. Do datasets have politics? Disciplinary values in computer vision dataset development. Proc. Acm Hum. Comput. Interact. 2021, 5, 1–37. [Google Scholar] [CrossRef]
  9. Verma, T.; De, A.; Agrawal, Y.; Vinay, V.; Chakrabarti, S. Varscene: A deep generative model for realistic scene graph synthesis. In Proceedings of the International Conference on Machine Learning, Guangzhou, China, 18–21 February 2022; PMLR; pp. 22168–22183. [Google Scholar]
  10. Esteva, A.; Chou, K.; Yeung, S.; Naik, N.; Madani, A.; Mottaghi, A.; Liu, Y.; Topol, E.; Dean, J.; Socher, R. Deep learning-enabled medical computer vision. NPJ Digit. Med. 2021, 4, 5. [Google Scholar] [CrossRef]
  11. Andriyanov, N.A.; Dementiev, V.E.; Tashlinskiy, A. Detection of objects in the images: From likelihood relationships towards scalable and efficient neural networks. Comput. Opt. 2022, 46, 139–159. [Google Scholar] [CrossRef]
  12. Dutordoir, V.; van der Wilk, M.; Artemev, A.; Hensman, J. Bayesian Image Classification with Deep Convolutional Gaussian Processes. In Proceedings of the Twenty Third International Conference on Artificial Intelligence and Statistics, Online, 26–28 August 2020; pp. 1529–1539. [Google Scholar]
  13. Papakostas, G.; Diamantaras, K.; Papadimitriou, T. Parallel pattern classification utilizing GPU-based kernelized Slackmin algorithm. J. Parallel Distrib. Comput. 2017, 99, 90–99. [Google Scholar] [CrossRef]
  14. Joseph, K.J.; Khan, S.; Khan, F.S.; Balasubramanian, V.N. Towards open world object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 5830–5840. [Google Scholar]
  15. Wang, C.Y.; Bochkovskiy, A.; Liao, H.Y.M. YOLOv7: Trainable bag-of-freebies sets new state-of-the-art for real-time object detectors. arXiv 2022, arXiv:2207.02696. [Google Scholar]
  16. Zhang, Q.; Wipf, D.; Gan, Q.; Song, L. A Biased Graph Neural Network Sampler with Near-Optimal Regret. arXiv 2021, arXiv:2103.01089. [Google Scholar] [CrossRef]
  17. Zhang, X.; Bosselut, A.; Yasunaga, M.; Ren, H.; Liang, P.; Manning, C.D.; Leskovec, J. GreaseLM: Graph REASoning Enhanced Language Models. In Proceedings of the International Conference on Learning Representations, Online, 25–29 April 2022. [Google Scholar]
  18. Wu, Y.X.; Wang, X.; Zhang, A.; He, X.; Chua, T.S. Discovering Invariant Rationales for Graph Neural Networks. In Proceedings of the International Conference on Learning Representations, Online, 25–29 April 2022. [Google Scholar]
  19. Gao, J.; Ribeiro, B. On the Equivalence Between Temporal and Static Equivariant Graph Representations. In Proceedings of the 39th International Conference on Machine Learning, Baltimore, MD, USA, 17–23 July 2022; pp. 7052–7076. [Google Scholar]
  20. Yoon, M.; Wu, Y.; Palowitch, J.; Perozzi, B.; Salakhutdinov, R. Scalable and Privacy-enhanced Graph Generative Model for Graph Neural Networks. arXiv 2022, arXiv:2207.04396. [Google Scholar] [CrossRef]
  21. Andriyanov, N. Application of Graph Structures in Computer Vision Tasks. Mathematics 2022, 10, 4021. [Google Scholar] [CrossRef]
  22. Zhou, J.; Cui, G.; Hu, S.; Zhang, Z.; Yang, C.; Liu, Z.; Wang, L.; Li, C.; Sun, M. Graph neural networks: A review of methods and applications. AI Open 2020, 1, 57–81. [Google Scholar] [CrossRef]
  23. Chen, S.; Li, B. Multi-Modal Dynamic Graph Transformer for Visual Grounding. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 15534–15543. [Google Scholar]
  24. Ding, Y.; Yu, J.; Liu, Y.; Hu, Y.; Cui, M.; Wu, Q. MuKEA: Multimodal Knowledge Extraction and Accumulation for Knowledge-Based Visual Question Answering. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 5089–5098. [Google Scholar]
  25. Lou, C.; Han, W.; Lin, Y.; Zheng, Z. Unsupervised Vision-Language Parsing: Seamlessly Bridging Visual Scene Graphs with Language Structures via Dependency Relationships. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 15607–15616. [Google Scholar]
  26. Walmer, M.; Sikka, K.; Sur, I.; Shrivastava, A.; Jha, S. Dual-Key Multimodal Backdoors for Visual Question Answering. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 15375–15385. [Google Scholar]
  27. Koh, J.Y.; Salakhutdinov, R.; Fried, D. Grounding Language Models to Images for Multimodal Generation. arXiv 2023, arXiv:2301.13823. [Google Scholar] [CrossRef]
  28. Iwamura, K.; Kasahara, J.L.; Moro, A.; Yamashita, A.; Asama, H. Image Captioning Using Motion-CNN with Object Detection. Sensors 2021, 21, 1270. [Google Scholar] [CrossRef]
  29. Liu, H.; Yan, N.; Mortazavi, M.; Bhanu, B. Fully convolutional scene graph generation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 11546–11556. [Google Scholar]
  30. Cong, Y.; Yang, M.Y.; Rosenhahn, B. RelTR: Relation Transformer for Scene Graph Generation. arXiv 2022, arXiv:2201.11460. [Google Scholar] [CrossRef]
  31. Xu, D.; Zhu, Y.; Choy, C.B.; Li, F.-F. Scene graph generation by iterative message passing. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 5410–5419. [Google Scholar]
  32. Dai, B.; Zhang, Y.; Lin, D. Detecting visual relationships with deep relational networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 3076–3086. [Google Scholar]
  33. Li, Y.; Ouyang, W.; Zhou, B.; Wang, K.; Wang, X. Scene graph generation from objects, phrases and region captions. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 1261–1270. [Google Scholar]
  34. Li, Y.; Ouyang, W.; Wang, X. Vip-cnn: A visual phrase reasoning convolutional neural network for visual relationship detection. arXiv 2017, arXiv:1702.07191. [Google Scholar]
  35. Hwang, S.J.; Ravi, S.N.; Tao, Z.; Kim, H.J.; Collins, M.D.; Singh, V. Tensorize, factorize and regularize: Robust visual relationship learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 1014–1023. [Google Scholar]
  36. Li, Y.; Ouyang, W.; Zhou, B.; Shi, J.; Zhang, C.; Wang, X. Factorizable net: An efficient subgraph-based framework for scene graph generation. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 335–351. [Google Scholar]
  37. Yang, J.; Lu, J.; Lee, S.; Batra, D.; Parikh, D. Graph r-cnn for scene graph generation. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 670–685. [Google Scholar]
  38. Yin, G.; Sheng, L.; Liu, B.; Yu, N.; Wang, X.; Shao, J.; Change Loy, C. Zoom-net: Mining deep feature interactions for visual relationship recognition. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 322–338. [Google Scholar]
  39. Woo, S.; Kim, D.; Cho, D.; Kweon, I.S. LinkNet: Relational Embedding for Scene Graph. In Proceedings of the Advances in Neural Information Processing Systems, Montreal, QC, Canada, 3–8 December 2018; pp. 558–568. [Google Scholar]
  40. Wang, W.; Wang, R.; Shan, S.; Chen, X. Exploring Context and Visual Pattern of Relationship for Scene Graph Generation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 8188–8197. [Google Scholar]
  41. Tang, K.; Zhang, H.; Wu, B.; Luo, W.; Liu, W. Learning to compose dynamic tree structures for visual contexts. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 155–20 June 2019; pp. 6619–6628. [Google Scholar]
  42. Chen, L.; Zhang, H.; Xiao, J.; He, X.; Pu, S.; Chang, S.F. Counterfactual Critic Multi-Agent Training for Scene Graph Generation. In Proceedings of the IEEE International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 4613–4623. [Google Scholar]
  43. Tang, K.; Niu, Y.; Huang, J.; Shi, J.; Zhang, H. Unbiased scene graph generation from biased training. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 3716–3725. [Google Scholar]
  44. Desai, A.; Wu, T.Y.; Tripathi, S.; Vasconcelos, N. Learning of visual relations: The devil is in the tails. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 15404–15413. [Google Scholar]
  45. Suhail, M.; Mittal, A.; Siddiquie, B.; Broaddus, C.; Eledath, J.; Medioni, G.; Sigal, L. Energy-Based Learning for Scene Graph Generation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 13936–13945. [Google Scholar]
  46. Chen, T.; Yu, W.; Chen, R.; Lin, L. Knowledge-Embedded Routing Network for Scene Graph Generation. In Proceedings of the Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–17 June 2019. [Google Scholar]
  47. Lin, X.; Ding, C.; Zeng, J.; Tao, D. Gps-net: Graph property sensing network for scene graph generation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 3746–3753. [Google Scholar]
  48. Yu, J.; Chai, Y.; Hu, Y.; Wu, Q. CogTree: Cognition Tree Loss for Unbiased Scene Graph Generation. arXiv 2020, arXiv:2009.07526. [Google Scholar]
  49. Li, R.; Zhang, S.; Wan, B.; He, X. Bipartite Graph Network With Adaptive Message Passing for Unbiased Scene Graph Generation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 11109–11119. [Google Scholar]
  50. Yan, S.; Shen, C.; Jin, Z.; Huang, J.; Jiang, R.; Chen, Y.; Hua, X.S. Pcpl: Predicate-correlation perception learning for unbiased scene graph generation. In Proceedings of the 28th ACM International Conference on Multimedia, Online, 12–16 October 2020; pp. 265–273. [Google Scholar]
  51. Chiou, M.J.; Ding, H.; Yan, H.; Wang, C.; Zimmermann, R.; Feng, J. Recovering the unbiased scene graphs from the biased ones. In Proceedings of the 29th ACM International Conference on Multimedia, Chengdu, China, 20–24 October 2021; pp. 1581–1590. [Google Scholar]
  52. Guo, Y.; Gao, L.; Wang, X.; Hu, Y.; Xu, X.; Lu, X.; Shen, H.T.; Song, J. From general to specific: Informative scene graph generation via balance adjustment. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 16383–16392. [Google Scholar]
  53. Li, W.; Zhang, H.; Bai, Q.; Zhao, G.; Jiang, N.; Yuan, X. PPDL: Predicate Probability Distribution Based Loss for Unbiased Scene Graph Generation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 19447–19456. [Google Scholar]
  54. Li, L.; Chen, L.; Huang, Y.; Zhang, Z.; Zhang, S.; Xiao, J. The Devil is in the Labels: Noisy Label Correction for Robust Scene Graph Generation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 18869–18878. [Google Scholar]
  55. Krishna, R.; Zhu, Y.; Groth, O.; Johnson, J.; Hata, K.; Kravitz, J.; Chen, S.; Kalantidis, Y.; Li, L.J.; Shamma, D.A.; et al. Visual genome: Connecting language and vision using crowdsourced dense image annotations. Int. J. Comput. Vis. 2017, 123, 32–73. [Google Scholar] [CrossRef] [Green Version]
  56. Kuznetsova, A.; Rom, H.; Alldrin, N.; Uijlings, J.; Krasin, I.; Pont-Tuset, J.; Kamali, S.; Popov, S.; Malloci, M.; Kolesnikov, A.; et al. The open images dataset v4. Int. J. Comput. Vis. 2020, 128, 1956–1981. [Google Scholar] [CrossRef] [Green Version]
  57. Zhan, Y.; Yu, J.; Yu, T.; Tao, D. On Exploring Undetermined Relationships for Visual Relationship Detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 5128–5137. [Google Scholar]
  58. Sadeghi, M.A.; Farhadi, A. Recognition using visual phrases. In Proceedings of the CVPR 2011, Colorado Springs, CO, USA, 20–25 June 2011; IEEE: Piscataway, NJ, USA, 2011; pp. 1745–1752. [Google Scholar]
  59. Gu, J.; Zhao, H.; Lin, Z.; Li, S.; Cai, J.; Ling, M. Scene graph generation with external knowledge and image reconstruction. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 1969–1978. [Google Scholar]
  60. Yu, R.; Li, A.; Morariu, V.I.; Davis, L.S. Visual relationship detection with internal and external linguistic knowledge distillation. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 1974–1982. [Google Scholar]
  61. Yang, G.; Zhang, J.; Zhang, Y.; Wu, B.; Yang, Y. Probabilistic Modeling of Semantic Ambiguity for Scene Graph Generation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 12527–12536. [Google Scholar]
  62. Lyu, X.; Gao, L.; Guo, Y.; Zhao, Z.; Huang, H.; Shen, H.T.; Song, J. Fine-Grained Predicates Learning for Scene Graph Generation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 19467–19475. [Google Scholar]
  63. Dong, X.; Gan, T.; Song, X.; Wu, J.; Cheng, Y.; Nie, L. Stacked Hybrid-Attention and Group Collaborative Learning for Unbiased Scene Graph Generation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 19427–19436. [Google Scholar]
  64. Goel, A.; Fernando, B.; Keller, F.; Bilen, H. Not All Relations are Equal: Mining Informative Labels for Scene Graph Generation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 15596–15606. [Google Scholar]
  65. Li, Y.; Yang, X.; Xu, C. Dynamic Scene Graph Generation via Anticipatory Pre-Training. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 13874–13883. [Google Scholar]
  66. Teng, Y.; Wang, L. Structured sparse r-cnn for direct scene graph generation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 19437–19446. [Google Scholar]
  67. Zhang, A.; Yao, Y.; Chen, Q.; Ji, W.; Liu, Z.; Sun, M.; Chua, T.S. Fine-Grained Scene Graph Generation with Data Transfer. arXiv 2022, arXiv:2203.11654. [Google Scholar]
  68. Lin, X.; Ding, C.; Zhang, J.; Zhan, Y.; Tao, D. RU-Net: Regularized Unrolling Network for Scene Graph Generation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 19457–19466. [Google Scholar]
  69. Yang, J.; Ang, Y.Z.; Guo, Z.; Zhou, K.; Zhang, W.; Liu, Z. Panoptic Scene Graph Generation. In Proceedings of the European Conference on Computer Vision; Springer: Berlin/Heidelberg, Germany, 2022; pp. 178–196. [Google Scholar]
  70. Deng, Y.; Li, Y.; Zhang, Y.; Xiang, X.; Wang, J.; Chen, J.; Ma, J. Hierarchical Memory Learning for Fine-Grained Scene Graph Generation. arXiv 2022, arXiv:2203.06907. [Google Scholar]
  71. He, T.; Gao, L.; Song, J.; Li, Y.F. Towards open-vocabulary scene graph generation with prompt-based finetuning. In Proceedings of the European Conference on Computer Vision; Springer: Berlin/Heidelberg, Germany, 2022; pp. 56–73. [Google Scholar]
  72. Brown, S. Measures of shape: Skewness and kurtosis. Retrieved August 2011, 20, 2012. [Google Scholar]
  73. Zhu, Z.; He, X.; Qi, G.; Li, Y.; Cong, B.; Liu, Y. Brain tumor segmentation based on the fusion of deep semantics and edge information in multimodal MRI. Inf. Fusion 2023, 91, 376–387. [Google Scholar] [CrossRef]
  74. Zellers, R.; Yatskar, M.; Thomson, S.; Choi, Y. Neural motifs: Scene graph parsing with global context. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 5831–5840. [Google Scholar]
  75. Cui, Y.; Jia, M.; Lin, T.Y.; Song, Y.; Belongie, S. Class-balanced loss based on effective number of samples. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 9268–9277. [Google Scholar]
  76. Kang, H.; Vu, T.; Yoo, C.D. Learning Imbalanced Datasets With Maximum Margin Loss. In Proceedings of the 2021 IEEE International Conference on Image Processing (ICIP), Anchorage, AK, USA, 19–22 September 2021; pp. 1269–1273. [Google Scholar]
  77. Zhang, J.; Shih, K.J.; Elgammal, A.; Tao, A.; Catanzaro, B. Graphical contrastive losses for scene graph parsing. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 155–20 June 2019; pp. 11535–11543. [Google Scholar]
  78. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster r-cnn: Towards real-time object detection with region proposal networks. In Proceedings of the Advances in neural information processing systems, Montreal, QC, Canada, 7–12 December 2015; pp. 91–99. [Google Scholar]
  79. Lin, T.Y.; Dollár, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature pyramid networks for object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2117–2125. [Google Scholar]
  80. Misra, I.; Lawrence Zitnick, C.; Mitchell, M.; Girshick, R. Seeing through the human reporting bias: Visual classifiers from noisy human-centric labels. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26 June–1 July 2016; pp. 2930–2939. [Google Scholar]
Figure 1. The long-tailed predicate distribution: (a) the predicate proportion of the visual genome [31], where backgrounds are possibly far more than others and (b) the possible predicates of a man-board pair.
Figure 1. The long-tailed predicate distribution: (a) the predicate proportion of the visual genome [31], where backgrounds are possibly far more than others and (b) the possible predicates of a man-board pair.
Make 05 00018 g001
Figure 2. Skew class-balanced re-weighting (SCR) for unbiased SGG models. Based on (a) object detection, the SGG model predicts (d) predicates of (b) FREQ and (c) visual predictions; unbiased SGG is trained by (g) target sample weight, which is estimated by (f) skew measures of (e) possible sample estimates.
Figure 2. Skew class-balanced re-weighting (SCR) for unbiased SGG models. Based on (a) object detection, the SGG model predicts (d) predicates of (b) FREQ and (c) visual predictions; unbiased SGG is trained by (g) target sample weight, which is estimated by (f) skew measures of (e) possible sample estimates.
Make 05 00018 g002
Figure 3. The biased predicate weights: (a) the predicate biased logits R ^ s k e w of label frequency M 1 / 5 , depending on entropy H s k e w , (b) the sensitivity of sample weights W , according to entropy H s k e w , and (c) the skew measures S s k e w of a target label r y .
Figure 3. The biased predicate weights: (a) the predicate biased logits R ^ s k e w of label frequency M 1 / 5 , depending on entropy H s k e w , (b) the sensitivity of sample weights W , according to entropy H s k e w , and (c) the skew measures S s k e w of a target label r y .
Make 05 00018 g003
Figure 4. The Recall@100 on PredCls: we compare the SCR of FREQ+EMB with CogTree using the SG-Transformer [48].
Figure 4. The Recall@100 on PredCls: we compare the SCR of FREQ+EMB with CogTree using the SG-Transformer [48].
Make 05 00018 g004
Figure 5. The predicate predictions and sample estimates: in a SCR-trained SGG model, the larger R ^ s k e w i tends to be a smaller S s k e w i , i.e., woman- S s k e w wearing -coat (upper) < hill- S s k e w covered in  -snow (lower).
Figure 5. The predicate predictions and sample estimates: in a SCR-trained SGG model, the larger R ^ s k e w i tends to be a smaller S s k e w i , i.e., woman- S s k e w wearing -coat (upper) < hill- S s k e w covered in  -snow (lower).
Make 05 00018 g005
Table 1. The SGG performances of relationship retrieval on mean Recall@K and Recall@K. SCR denotes SCR of FREQ+EMB. Note the best and second best methods under each setting are marked according to format.
Table 1. The SGG performances of relationship retrieval on mean Recall@K and Recall@K. SCR denotes SCR of FREQ+EMB. Note the best and second best methods under each setting are marked according to format.
PredClsSGClsSGDet
ModelmR@20/50/100R@20/50/100mR@20/50/100R@20/50/100mR@20/50/100R@20/50/100
IMP+ [31] -.-/  9.8/10.552.7/59.3/61.3  -.-/ 5.8/ 6.031.7/34.6/35.4  -.-/ 3.8/ 4.814.6/20.7/24.5
FREQ [74] 8.3/13.0/16.053.6/60.6/62.2 5.1/ 7.2/ 8.529.3/32.3/32.9 4.5/ 6.1/ 7.120.1/26.2/30.1
KERN [46]  -.-/17.7/19.2  -.-/65.8/67.6  -.-/ 9.4/10.0  -.-/36.7/37.4  -.-/ 6.4/ 7.3  -.-/27.1/29.8
MOTIFS [74]10.8/14.0/15.358.5/65.2/67.1 6.3/ 7.7/ 8.232.9/35.8/36.5 4.2/ 5.7/ 6.621.4/27.2/30.3
VCTree [41]14.0/17.9/19.460.1/66.4/68.1 8.2/10.1/10.835.2/38.1/38.8 5.2/ 6.9/ 8.022.0/27.9/31.3
MSDN [33]  -.-/15.9/17.5  -.-/64.6/66.6  -.-/  9.3/  9.7  -.-/38.4/39.8 -.-/  6.1/  7.2  -.-/31.9/36.6
G-RCNN [37]  -.-/16.4/17.2  -.-/64.8/66.7  -.-/  9.0/  9.5  -.-/38.5/37.0 -.-/  5.8/  6.6  -.-/29.7/32.8
BGNN [49]  -.-/30.4/32.9  -.-/59.2/61.3  -.-/14.3/16.5  -.-/37.4/38.5 -.-/10.7/12.6  -.-/31.0/35.8
DT2-ACBS [44]  -.-/35.9/39.7  -.-/23.3/25.6  -.-/24.8/27.5  -.-/16.2/17.6 -.-/22.0/24.4  -.-/15.0/16.3
MOTIFS [74]11.5/14.6/15.859.5/66.0/67.9 6.5/ 8.0/ 8.535.8/39.1/39.9 4.1/ 5.5/ 6.825.1/32.1/36.9
  + TDE [43]18.5/25.5/29.133.6/46.2/51.4 9.8/13.1/14.921.7/27.7/29.9 5.8/ 8.2/ 9.812.4/16.9/20.3
  + PCPL [50]   -.-/24.3/26.1   -.-/54.7/56.5   -.-/12.0/12.7   -.-/35.3/36.1  -.-/10.7/12.6   -.-/27.8/31.7
  + CogTree [48]20.9/26.4/29.031.1/35.6/36.812.1/14.9/16.119.4/21.6/22.27.9/10.4/11.815.7/20.0/22.1
  + DLFE [51]22.1/26.9/28.8   -.-/52.5/54.212.8/15.2/15.9   -.-/32.3/33.18.6/11.7/13.8   -.-/25.4/29.4
  + BPL-SA [52]24.8/29.7/31.7   -.-/50.7/52.514.0/16.5/17.5   -.-/30.1/31.010.7/13.5/15.6   -.-/23.0/26.9
  + PPDL [53]   -.-/32.2/33.3   -.-/47.2/47.6   -.-/17.5/18.2   -.-/28.4/29.3  -.-/11.4/13.5   -.-/21.2/23.9
  + NICE [54]   -.-/29.9/32.3   -.-/55.1/57.2   -.-/16.6/17.9   -.-/33.1/34.0  -.-/12.2/14.4   -.-/27.8/31.8
  + SCR (ours)25.9/31.5/33.651.0/57.9/60.114.2/17.1/18.227.1/31.0/32.3 9.6/13.5/15.918.1/25.1/29.5
VCTree [41]11.7/14.9/16.159.8/66.2/68.1 6.2/ 7.5/ 7.937.0/40.5/41.4 4.2/ 5.7/ 6.924.7/31.5/36.2
  + TDE [43]18.4/25.4/28.736.2/47.2/51.6 8.9/12.2/14.019.9/25.4/27.9 6.9/ 9.3/11.114.0/19.4/23.2
  + PCPL [50]   -.-/22.8/24.5   -.-/56.9/58.7   -.-/15.2/16.1   -.-/40.6/41.7 -.-/10.8/12.6   -.-/26.6/30.3
  + CogTree [48]22.0/27.6/29.739.0/44.0/45.415.4/18.8/19.927.8/30.9/31.77.8/10.4/12.114.0/18.2/20.4
  + DLFE [51]20.8/25.3/27.1   -.-/51.8/53.515.8/18.9/20.0   -.-/33.5/34.68.6/11.8/13.8   -.-/22.7/26.3
  + BPL-SA [52]26.2/30.6/32.6   -.-/50.0/51.817.2/20.1/21.2   -.-/34.0/35.010.6/13.5/15.7   -.-/21.7/25.5
  + PPDL [53]   -.-/33.3/33.8   -.-/47.6/48.0   -.-/14.3/15.7   -.-/32.1/33.0  -.-/11.3/13.3   -.-/20.1/22.9
  + NICE [54]   -.-/30.7/33.0   -.-/55.0/56.9   -.-/19.9/21.3   -.-/37.8/39.0  -.-/11.9/14.1  -.-/27.0/30.8
  + SCR (ours)27.7/33.5/35.549.7/56.4/58.315.4/18.9/20.126.7/30.6/31.910.3/13.8/16.318.1/25.0/29.4
SG-Transformer [48]14.8/19.2/20.558.5/65.0/66.7 8.9/11.6/12.635.6/38.9/39.8 5.6/ 7.7/ 9.024.0/30.3/33.3
  + CogTree[48]22.9/28.4/31.034.1/38.4/39.713.0/15.7/16.720.8/22.9/23.4 7.9/11.1/12.715.1/19.5/21.7
  + SCR (ours)27.0/32.2/34.545.3/52.7/55.014.9/17.7/18.725.1/28.9/30.210.4 /13.4/15.017.7/23.2/26.2
Table 2. The SGG Performances of zero-shot relationship retrieval on Recall@K. SCR denotes SCR of FREQ+EMB. The SGG models re-implemented under our codebase are denoted by the superscript * .
Table 2. The SGG Performances of zero-shot relationship retrieval on Recall@K. SCR denotes SCR of FREQ+EMB. The SGG models re-implemented under our codebase are denoted by the superscript * .
Zero-Shot Relationship RetrievalPredClsSGClsSGDet
ModelMethodR@50R@100R@50R@100R@50R@100
MOTIFS [43]baseline [43]
Reweight [43]
TDE [43]
CogTree [48] *
SCR (ours)
10.9
0.7
14.4
2.4
18.0
14.5
0.9
18.2
4.0
21.1
2.2
0.1
3.4
0.9
5.1
3.0
0.1
4.5
1.5
5.9
0.1
0.0
2.3
0.3
2.4
0.2
0.0
2.9
0.6
3.8
VCTree [43]Baseline [43]
TDE [43]
CogTree [48] *
SCR (ours)
10.8
14.3
3.3
17.6
14.3
17.6
5.0
20.4
1.9
3.2
2.1
4.5
2.6
4.0
2.6
5.2
0.2
2.6
0.4
2.5
0.7
3.2
0.6
3.5
SG-Transformer [48] * Baseline *
CogTree [48] *
SCR (ours)
4.1
5.2
16.4
6.3
7.3
19.6
1.6
2.3
4.6
2.3
3.0
5.3
0.2
0.3
2.0
0.5
0.5
3.2
BGNN [49]BGNN *
CogTree [48] *
SCR (ours)
15.0
13.4
16.3
18.0
16.1
19.5
4.5
5.0
4.9
5.4
5.7
5.9
4.5
0.5
1.9
5.3
0.8
3.0
Table 3. The performances of open images dataset. ∗ denotes results reproduced by Li et al. [49]. SCR denotes SCR of FREQ+EMB.
Table 3. The performances of open images dataset. ∗ denotes results reproduced by Li et al. [49]. SCR denotes SCR of FREQ+EMB.
DatasetModelsmR@50R@50   wmAPscore wtd
relphr
V4RelDN [77] * 70.4075.6636.1339.9145.21
GPS-Net [47] * 69.5074.6535.0239.4044.70
BGNN [49]72.1175.4637.7641.7046.87
BGNN+SCR (ours)72.2075.4838.6445.0145.01
V6RelDN [77] * 33.9873.0832.1633.3940.84
VCTree [41] * 33.9174.0834.1633.1140.21
MOTIFS [74] * 32.6871.6329.9131.5938.93
TDE [43] * 35.4769.3030.7432.8039.27
GPS-Net [47] * 35.2674.8132.8533.9841.69
BGNN [49]40.4574.9833.5134.1542.06
BGNN+SCR (ours)42.4375.2133.9835.1342.66
Table 4. The ablation study for the predicate sample estimates on Recall@100. The underbar represents the predicate sample estimates for better generality.
Table 4. The ablation study for the predicate sample estimates on Recall@100. The underbar represents the predicate sample estimates for better generality.
Relationship RetrievalPredClsSGClsSGDet
ModelMethodmRRZSRRRRmRRZSRRRRmRRZSRRRR
VCTreeSCR of FREQ
SCR of EMB
SCR of FREQ+EMB
31.6
33.3
33.7
22.2
21.1
20.1
60.0
60.1
60.0
16.9
17.9
18.0
6.2
5.9
6.1
35.4
35.2
33.3
13.3
15.4
14.4
3.5
3.5
3.5
31.6
30.0
31.9
Table 5. The ablation study for the scr hyper-parameters on Recall@100. The underbar represents the best trade-off performances between mRecall and Recall. SCR denotes SCR of FREQ+EMB.
Table 5. The ablation study for the scr hyper-parameters on Recall@100. The underbar represents the best trade-off performances between mRecall and Recall. SCR denotes SCR of FREQ+EMB.
Relationship RetrievalPredClsSGClsSGDet
Model λ s k e w δ mRRRRmRRRRmRRRR
VCTree
+SCR
0.03
0.06
0.08
0.7
0.7
0.7
35.5
33.7
32.8
58.3
60.0
60.4
20.1
18.0
17.2
31.9
33.3
34.2
16.3
14.4
13.5
29.4
31.9
32.6
0.06
0.06
0.06
0.6
0.7
0.8
33.5
33.7
33.3
60.6
60.0
59.3
17.8
18.0
17.6
34.3
33.3
33.0
14.2
14.4
14.0
32.0
31.9
30.6
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Kang, H.; Yoo, C.D. Skew Class-Balanced Re-Weighting for Unbiased Scene Graph Generation. Mach. Learn. Knowl. Extr. 2023, 5, 287-303. https://doi.org/10.3390/make5010018

AMA Style

Kang H, Yoo CD. Skew Class-Balanced Re-Weighting for Unbiased Scene Graph Generation. Machine Learning and Knowledge Extraction. 2023; 5(1):287-303. https://doi.org/10.3390/make5010018

Chicago/Turabian Style

Kang, Haeyong, and Chang D. Yoo. 2023. "Skew Class-Balanced Re-Weighting for Unbiased Scene Graph Generation" Machine Learning and Knowledge Extraction 5, no. 1: 287-303. https://doi.org/10.3390/make5010018

APA Style

Kang, H., & Yoo, C. D. (2023). Skew Class-Balanced Re-Weighting for Unbiased Scene Graph Generation. Machine Learning and Knowledge Extraction, 5(1), 287-303. https://doi.org/10.3390/make5010018

Article Metrics

Back to TopTop