Next Article in Journal
Smooth Coverage Path Planning for UAVs with Model Predictive Control Trajectory Tracking
Previous Article in Journal
The Current Research Status of AI-Based Network Security Situational Awareness
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Zero-Shot Learning with Joint Generative Adversarial Networks

1
School of Integrated Circuits and Electronics, Beijing Institute of Technology, Beijing 100081, China
2
Beijing Institute of Technology Chongqing Center for Microelectronics and Microsystems, Chongqing 401332, China
3
Yangtze Delta Region Academy of Beijing Institute of Technology, Jiaxing 314019, China
*
Author to whom correspondence should be addressed.
Electronics 2023, 12(10), 2308; https://doi.org/10.3390/electronics12102308
Submission received: 6 March 2023 / Revised: 8 May 2023 / Accepted: 17 May 2023 / Published: 19 May 2023
(This article belongs to the Section Artificial Intelligence)

Abstract

:
Zero-shot learning (ZSL) is implemented by transferring knowledge from seen classes to unseen classes through embedding space or feature generation. However, the embedding-based method has a hubness problem, and the generation-based method may contain considerable bias. To solve these problems, a joint model with multiple generative adversarial networks (JG-ZSL) is proposed in this paper. Firstly, we combined the generation-based model and the embedding-based model to build a hybrid ZSL framework by mapping the real samples and the synthetic samples into the embedding space for classification, which alleviates the problem of data imbalance effectively. Secondly, based on the original generation-method model, a coupled GAN is introduced to generate semantic embeddings, which can generate semantic vectors for unseen classes in embedded space to alleviate the bias of mapping results. Finally, semantic-relevant self-adaptive margin center loss was used, which can explicitly encourage intra-class compactness and inter-class separability, and it can also guide coupled GAN to generate discriminative and representative semantic features. All the experiments on the four standard datasets (CUB, AWA1, AWA2, SUN) show that the proposed method is effective.

1. Introduction

Supervised classification has achieved great success in the research, but in this kind of classification, each class needs enough labeling training, and the learned classifier cannot deal with unseen classes [1]. To solve the above problems, the methods of few/one-shot learning [2,3,4], open set recogniton [5], cumulative learning [6], class-incremental [7] and open world [8] have been put forward. However, in the above methods, if unseen classes with no available tag instance appear in the test stage, the classifier still cannot determine their class tag. Therefore, zero-shot learning (ZSL) is proposed [9]. With the help of auxiliary information that contains descriptions of seen and unseen classes and the knowledge learned from training sets that belong to seen classes, sufficient labeled instances are provided [10]. ZSL methods can generate predictions for instances that belong to unseen classes despite that the seen and unseen classes are disjointed [11]; that is, given labeled training instances belonging to the seen classes, zero-shot learning aims to learn a classifier which can classify testing instances belonging to the unseen classes. From this definition, we can see that the general idea of zero-shot learning is to transfer the knowledge contained in the training instances to the task of testing instance classification. The label spaces covered by the training and the testing instances are disjoint. Thus, zero-shot learning is a subfield of transfer learning. In transfer learning [12], knowledge contained in the source domain and source task is transferred to the target domain for learning the model in the target task [13].
Since its birth, ZSL has become a fast-developing field in machine learning and has a wide range of applications in computer vision, natural language processing, and ubiquitous computing [13]. Previous works for ZSL learn a space embedding function to implement the classification. According to the choice of embedding space, embedding-based methods can be divided into three categories: semantic space embedding methods, visual space embedding methods, and common space embedding methods [14]. They directly estimate the conditional distribution or mapping between visual features and their corresponding attributes. Semantic space embedding methods map visual features to semantic space directly. DeViSE [15] is one of the most representative models; it learns a linear mapping between image and semantic space using an efficient ranking loss formulation, and it is evaluated on the large-scale ImageNet dataset. However, using the semantic space as the embedding space means that the visual feature vectors need to be projected into the semantic space, which will shrink the variance of the projected data points and thus aggravate the hubness problem [16,17]. To alleviate the hubness problem, Li et al. [18] proposed a novel deep neural network-based embedding model (DEM). Although DEM uses the output visual feature space of a CNN subnet as the embedding space, which can alleviate the hubness problem to a certain extent, the inconsistency between the manifold of visual features and semantic features leads to the semantic gap. To solve the above-mentioned problem, Min et al. proposed a domain-specific embedding network (DSEN) [19] model, which considers the problem of semantic consistency and prevents the semantic relationship from being destroyed in the embedded space. Although the embedding-based method has been used and developed for a long time and is a very competitive zero-shot image classification method, due to the extreme imbalance in the number of training samples between seen and unseen classes, most of the existing methods still have great limitations.
Recent works mainly focus on synthesizing image features with a generative model, and generation-based methods have become a hot research topic [20,21]. These methods fall into the data augmentation-based category. The basic assumption of approaches in this category is that the intra-class cross-sample relationship learned from seen classes can be applied to unseen classes. Once the cross-sample relationship is modeled and learned from seen classes, it can be applied on the unlabeled samples of unseen class to hallucinated new samples, and unsupervised learning is transformed into supervised learning using synthesized new samples [22].
Depending on the different generation models, the existing generation-based methods mainly include GAN-based methods, VAE-based methods, and normalizing flow-based methods [23,24,25]. The normalizing flow-based methods build complex distributions by mapping a simple distribution through invertible functions, and they allow exact likelihood calculation while being efficiently parallelizable, but they have not been widely studied due to the particularity of the architecture [25]. Most of the VAE-based methods are unidirectional alignment. This method captures the low-dimensional potential features of visual features and then realizes unidirectional alignment between the generated pseudo-visual features and semantic attributes through decoding and reconstruction of the formula. SE-GZSL [26] adopts the VAE-based structure, and the generation model is composed of the probabilistic encoder and conditional decoder. At the same time, the feedback drive mechanism is introduced, which can improve the reliability of the generator. Although VAE is capable of generating pseudo-visual features stably to effectively avoid pattern collapse, the semantic information contained in the generated pseudo-visual features is very limited. In order to overcome the above problems, the GAN-based methods are proposed; this method can generate high-quality pseudo-visual features after the model is trained. VERMA et al. [27] proposed a meta-learning model ZSML based on the class attribute condition setting. The generator module and discriminator module with a classifier were associated with the meta-learning agent, respectively, and the model could be trained only by inputting a few visible class samples. Xian et al. [28] use the generative adversarial network to make the classification based on semantic features and Gaussian noise to generate unseen visual features, transforming the zero-shot learning problem into a supervised classification problem. The result of generation-based methods is better than embedding-based methods, and it is also the mainstream method at present.
In the latest work in 2022, both embedding-based and generation-based methods have been further explored and updated. Xu et al. [29] propose a Visually Grounded Semantic Embeddings model (VGSE), which learns visual clusters from seen classes and automatically predicts the semantic embeddings for each category by building the relationship between seen and unseen classes given unsupervised external knowledge sources. In terms of generation-based methods, to generate high-quality and diverse image features, Yu et al. [12] proposes a new generative model that adds a semantic constraint module and introduces a Euclidean distance loss for constraining feature generation. Although the above methods can solve the problem of the existence of zero-shot learning, it also introduces a new problem: previous work on the generation-based methods only used one generative adversarial network to simulate the visual features of unseen classes and ignored the distribution of these generative features in the mapping space. This may make the semantic mapping point of the generated feature closer to the semantic prototype of the seen class in the semantic space, resulting in the final classification result still having a bias toward the seen class.
To obtain the best of both worlds and solve the new problem mentioned above, we first propose a hybrid model, which can implement both the space embedding-based method and the generation-based method. Second, we introduce a generation adversarial network to simulate the mapping point of unseen class features in the embedding space. Although the model with multiple GAN cascaded has been fully proven and used in supervised learning, it has not been applied to zero-shot learning. In this paper, a multilevel GAN stack structure is introduced for the first time in zero-shot learning to optimize the problem of data imbalance. Third, we propose a semantic-relevant self-adaptive margin center loss for the coupled GAN. This loss can encourage intra-class compactness and inter-class separability and realizes that the coupled GAN can better generate representative and differentiated semantic features. We evaluate our method on four benchmark datasets, and the experimental results show that our approach is competitive with other methods.
The contributions of this paper are summarized as follows:
  • A hybrid model with joint generative adversarial networks (JG-ZSL) combining the embedding-based method and the generation-based method is proposed to improve model sensitivity and specificity.
  • A GAN for generating semantic features is introduced to generate mapping points in embedding space, which can generate semantic vectors for unseen classes in semantic space to alleviate the bias of mapping results.
  • Semantic-relevant self-adaptive margin center loss (SEMC-loss) is designed for the semantic generated GAN to ensure the generated mapping points in semantic embedding space are not biased to other categories and realize that the whole model can better distinguish between different classes.
  • We evaluate our model on four benchmarks, and the experimental results show that our proposed method can achieve high accuracy.

2. Materials and Methods

2.1. Problem Definition

We have two disjoint sets of classes in both ZSL and GZSL: the seen class set S = { c i s | i = 1 , , N s } , where c i s is a seen class which provides labeled instances for training, and unseen class set U = { c i u | i = 1 , , N u } contains unlabeled instances for testing. Note that S U = . These instances have different visual features, but for instances from the same class, their labels and semantic descriptions are the same. Denote the visual feature as x, class label as y, and semantic description, which is the attribute in this article as a. Then, each class can be represented as a set C i = { ( x i j , y i j , a i ) | i = 1 , , N S + N u ; j = 1 , , n } , n is the number of instances the class contains; we can infer the semantic descriptor a for an instance x from its class label y.
ZSL aims to learn a classifier that can categorize the testing instances x u belonging to the unseen classes U, f z s l : x u U . Under the more challenging generalized zero-shot learning (GZSL) setting, the testing instances x come from both seen class S and unseen classes U because people are also concerned with the ability to classify instances on seen and unseen classes. GZSL aims to learns a classifier f g z s l : x S U .
Zero-shot learning is divided into three learning settings by Wang [13] et al. according to whether unlabeled testing instances and the class description information of the unseen class are used in model learning, as shown in Table 1. In this paper, unlabeled testing instances are not used when training the generators, but the classifier is trained using average visual features of unlabeled testing instances and synthetic features that the generator generates based on the attribute descriptions of the unseen classes. According to Wang’s definition, our method belongs to the Class-Transductive Instance-Transductive (CTIT) Setting.

2.2. Hybrid Framework Introduction

The proposed joint GAN cascaded for ZSL (JG-ZSL) is illustrated in Figure 1. Specifically, the network consists of an embedded network that maps visual features to semantic space, a GAN that generates visual features based on attributes, and a GAN network that generates semantic space mapping points based on visual features.

2.2.1. Mapping Net

Human beings can summarize the attributes of the observed objects according to the visual features seen by the naked eye and deduce the categories of the observed objects according to the attributes. For example, if a child learns from watching a horse, a panda, and a tiger that they are “horse-like”, “black-white”, and “striped”, he or she can easily distinguish a zebra from a variety of animals after being told that a zebra is a horse with black and white stripes [30]. This ability to recognize objects without any visual samples, only prior knowledge, is zero-shot learning. It is very necessary to ensure machines have the zero-shot learning ability: first, in real life, the object categories to be recognized usually follow the long-tail distribution, some of which have rich training samples, while others have few or no available training samples. Zero-shot learning can not only get rid of the dependence on a large number of manual labeling samples but also have high commercial value in some applications lacking labeling samples. Hence, in order to enable machines to have this capability, ref. [9] introduced a manually defined attribute layer for the first time. Through this attribute layer, the classifier based on low-dimensional image features is transformed into a classifier based on high-dimensional semantic features (attribute layer) so that the trained classifier has broader classification ability and the ability to break through category boundaries. For example, in an animal identification problem in an image, attributes can be a body color (for example, “gray”, “brown”, and “yellow”) or habitat (for example, “coastal”, “desert”, and “forest”). These attributes are then used to construct semantic spaces.
Semantic embedding (SE) in conventional ZSL aims to learn an embedding function E that maps a visual feature x into the semantic embedding space denoted as h = E ( x ) . The embedding function E is usually a linear transformation consisting of two liner layers, whose input dimension is set to the dimension of the visual feature and output dimension is set to the dimension of the semantic feature. At the same time, h = E ( x ) is also called linear semantic space because it is composed of fully connected layers. These commonly used semantic embedding methods rely on a structured loss function proposed in [15]. According to the dot product similarity in the embedding space, the structured loss requires that the embedding of x is closer to the semantic embedding a of its ground-truth class than the other class embeddings. Specifically, the structured loss formula is as follows:
L S E ( E ) = E p ( x , a ) [ max ( 0 , Δ a T E ( x ) + ( a ) T E ( x ) ) ]
where p ( x , a ) is the empirical distribution of the training samples of seen classes, a is a random selection semantic descriptor of the other categories except a, and Δ > 0 and is a margin parameter to make E more robust.
On the basis of the traditional embedding function, Chen et al. [31] found that adding a non-linear projection head H in embedding space as z = H ( h ) can better constrain the original linear embedding space h = E ( x ) , because they showed experimentally that more information can be formed and maintained in h through this non-linear projection. In the same way that h = E ( x ) is called linear space because E is composed of fully connected layers, we called z = H ( h ) a non-linear space because the projection H actually is a ReLU non-linearity. We follow Chen’s strategy in our model; the difference is, Chen set H and E with the same output dimensionality (e.g., 2048-d), while we change the output dimension of E to the dimension of the semantic descriptor of the dataset (e.g., for dataset CUB, 312-d); then, the linear space can be limited to the semantic embedding space.
For the non-linear space z = H ( h ) , we follow the strategy in [32] to perform the ( K + 1 ) -way classification on z i to learn the embedding h i , where K is the number of negative examples h i , which refers to the samples whose class label is different from the class label of h i , while the only one positive example is h i + . Concretely, the cross-entropy loss of this (K + 1)-way classification problem is calculated as follows:
L S E ( H ) = log exp ( z i T z + / τ e ) exp ( z i T z + / τ e ) + k = 1 K exp ( z i T z k / τ e )
where τ e is a constant called the temperature parameter, which is manually set to adjust the degree of attention paid to negative samples. The smaller the temperature parameter is, the more attention is paid to separating this sample from other samples that are most similar.

2.2.2. Feature Generation Nets

The main disadvantage of embedding-based methods is that they suffer from the bias problem. This means that since the projection function is learned using only seen classes during training, it will be biased to predict with seen class labels as output; this bias problem is caused by a serious data imbalance between seen and unseen class data.
In supervised learning, the problem of data imbalance refers to the huge difference in the number of samples in each category of the dataset. Take the binary classification problem as an example: assuming that the number of samples of the positive class is much larger than that of negative class, in this case, the data are called unbalanced data. In zero-shot learning, this problem is even more extreme; that is, part of the class samples as unseen classes are completely missing and cannot participate in the model training process. Therefore, in supervised learning, the method of repeatedly sampling categories with fewer samples (over-sampling) or reducing sampling for categories with more samples (under-sampling) to achieve data balance is not applicable to zero-shot learning. After all, no samples can be collected from unseen classes. Therefore, unseen class data generation has become a hot research topic, which can generate pseudo-samples for unseen classes, so that both seen and unseen classes have training samples and transform unsupervised learning into supervised learning. Generative Adversarial Networks [23] are particularly appealing as they allow generating realistic and sharp images conditioned, for instance, on object categories. Previous work on generation-based methods learn a generation network to produce the unseen sample. However, in previous work on generation-based methods, the synthesized instances are usually assumed to follow some distributions (usually Gaussian distribution) [13], which also leads to a large deviation between the generated sample and the real sample, and it cannot truly represent the real data situation of the unseen class. The idea of stacking multilevel generation networks has been proven to be effective in improving the quality of generation quality, but it has not been used in the ZSL field. In this paper, two conditional GANs ( G f and G a ) are stacked to solve the problem of data imbalance from different aspects.
G f , the GAN for generating visual feature: The network based on traditional GAN takes random noise as the prior information input, and the inherent randomness of the deep neural network makes the quality of the image generated by it unstable. To solve this problem, conditional GAN is proposed. By adding conditional information to the network model, it guides the network model to generate pseudo-samples matching the conditions. We extend the GAN to a conditional GAN by integrating the class embedding to both the generator G 1 and the discriminator D 1 . Given the training data of seen classes, G 1 takes random Gaussian noise ε and semantic embedding a y as its inputs and outputs a CNN image feature x ˜ of class y. Once the generator G 1 learns to generate CNN features of seen class images, i.e., x, conditioned on the seen class embedding a s , it can also generate x ˜ of any unseen class via its class embedding a u . The objective function can be expressed as:
min G 1 max D 1 V ( D 1 , G 1 ) = E [ log D 1 ( x , a y ) ] + E [ ( 1 log D 1 ( x ˜ , a y ) ) ]
However, the adversarial nature of GANs makes them notoriously difficult to train, and the Jenson–Shannon divergence optimized by the original GAN leads to instability issues. To cure the unstable training issues of GANs, Wasserstein-GAN (WGAN) [33] is proposed, which optimizes an efficient approximation of the Wasserstein distance [25]. While WGAN attains better theoretical properties than the original GAN, it still suffers from vanishing and exploding gradient problems due to weight clipping to enforce the 1-Lipschitz constraint on the discriminator. So, we use the improved variant of WGAN, that is, WGAN-GP [34], which can enforce the Lipschitz constraint through gradient penalty. We extend the original WGAN-GP to a conditional WGAN-GP by integrating the class embedding a y to both the generator and the discriminator.
The loss is,
L W G A N f e a t u r e = E [ D 1 ( x , a y ) ] E [ D 1 ( x ˜ , a y ) ] λ E [ ( | | x ^ D 1 ( x ˜ , a y ) | | 2 1 ) 2 ]
where x ˜ = G 1 ( a y , ε ) , x ^ = α x + ( 1 α ) x ˜ with α U ( 0 , 1 ) , and λ is the penalty coefficient. In contrast to the traditional GAN, the discriminative network here eliminates the sigmoid layer and outputs a real value. Instead of optimizing the log-likelihood in Equation (3), the first two terms in Equation (4) approximate the Wasserstein distance, and the third term is the gradient penalty which enforces the gradient of D 1 to have a unit norm along the straight line between pairs of real and generated points.
G a , the GAN for generating semantic embedding: The embedding-based method obtains labeled instances of unseen classes by mapping instances in feature space and attribute prototypes in semantic space into the same space. Feature space contains labeled training instances of seen classes, and semantic space contains attribute prototypes of seen and unseen classes. Both spaces are real number spaces in which instance and attribute prototypes are vectors, respectively. By projecting the instance vectors from these two spaces into a common space, we can obtain labeled instances of unseen classes and classify them in the mapping space. However, in the embedding-based method, for every unseen class c u , it has no labeled instance in the feature space; thus, its attribute prototype a u in semantic space is the only labeled instance belonging to the unseen class. That is, only one labeled instance is available for each unseen class. Therefore, since there are few label instances of unseen classes, the feature generation methods are proposed to solve the problem of data imbalance by generating visual features for unseen classes in feature space. However, in semantic space, labeled instances of the unseen class are still scarce. Especially in the GZSL setting, the mapping results are still biased toward the seen class. Therefore, appropriately adding semantic vectors of unseen classes in semantic space can alleviate the bias of mapping results.
Active learning is similar to zero-shot learning to some extent. Both of them are designed to reduce the dependence on large-scale labeling data and are targeted at scenarios where labeled data are rare or the “cost” of labeling is high. The difference is that zero-shot learning aims to realize knowledge transfer in the absence of labeled samples, while active learning aims to maximize model performance by actively selecting the most valuable samples for labeling. Therefore, some techniques in active learning can enlighten us. In active learning, Parvaneh et al. proposed the feature mixing method: compute the average visual representation x ¯ of the labeled samples per class and call it an anchor. The anchors for all classes form the anchor set x ¯ and serve as representatives of the labeled instances [17]. Inspired by their method, we take the average visual feature as a representation of one class and generate the semantic embedding e ˜ of how this class might be mapped, as shown in Figure 2. The generated semantic embedding e ˜ should have the following two characteristics. First, by generating the different semantic embeddings that may be mapped from the same class, we extend the original unique semantic discriptor of each category in the semantic space into a semantic discriptor a set S i = { a i , e ˜ 1 , , e ˜ n } , where a i is the real semantic discritptor of category i provided by the dataset, while e ˜ 1 to e ˜ n are the synthetic pseudo-semantic-discriptors just like extending the unique evaluation criteria to establish a qualifying interval. Second, generated semantic embeddings should be representative and authentic, which are similar to the semantic embeddings of the real existence mapped by the visual feature, and they can truly simulate the possible mapping situation without deviating from reality. We formulate our assumption for the pseudo-semantic embedding generation method as follows:
e ˜ i = G 2 ( ε , x ¯ i )
x ¯ i = 1 n j = 1 n x i j
where n is the number of visual features instances the class i contains, and x i is the set of visual features contained in class i.
We still select the condition WGAN by integrating the visual feature average x ¯ to both the generator and the discriminator. The loss is,
L W G A N a t t = E [ D 2 ( e i , x ¯ i ) ] E [ D 2 ( e ˜ i , x ¯ i ) ] λ E [ ( | | e ^ i D 2 ( e ˜ i , x ¯ i ) | | 2 1 ) 2 ]
where e i is, corresponding to the synthetic semantic embedding e ˜ i , the real semantic embedding obtained by inputting the average visual features x i of category i into the mapping net.

2.3. Loss Design

2.3.1. Semantic-Relevant Self-Adaptive Margin Center Loss

To encourage G a t t to generate more representative semantic embedding for an unseen class, we used the idea of building a distance metric in metric learning. Metric learning aims to learn such a distance metric for a type of input data that conforms to semantic distance measures between the data instances [35]; this point has been explored and applied in both few-shot learning [35] and zero-shot learning [36,37]. Inspired by previous work, we propose the semantic-relevant self-adaptive margin center loss ( S E M C l o s s , L S E M C ) to constraint G a t t . By narrowing the distance between the generated semantic vector and the real semantic vector in the semantic space, intra-class compactness and inter-class separability are encouraged. It has the advantages of the center loss [38] and triplet loss [39] as well as learning intra-class compactness and inter-class separability. L S E M C is formulated as:
L S E M C = max ( 0 , Δ + γ | | e ˜ i a i | | 2 2 ( 1 γ ) | | e ˜ i a i | | 2 2 )
where a i is the i th (the label of seen visual feature x) class center of semantic embedding, a i is the i th (a randomly selected class label other than i) class center, Δ represents the margin that i controls the distance between intra- and inter-class pairs, e ˜ i is the synthesized semantic embedding of the ith class generated by G a t t and γ [ 0 , 1 ] is used for balancing the inter-class separability and intra-class compactness, which are adaptable to various datasets. The sensitivity of intra-class compactness and inter-class separability to different datasets (coarse-grained datasets and fine-grained datasets) can be satisfied by using balance factors to balance intra-class separability and intra-class compactibility adaptively. We use a large γ for fine-grained datasets (e.g., CUB [40], SUN [41]) and a small γ for coarse-grained datasets (e.g., AWA1 [9], AWA2 [42]). For fine-grained datasets, we can more easily distinguish them by encouraging intra-class compactness, and for the coarse-grained datasets, we can effectively separate them by enlarging the inter-class separability.

2.3.2. Total Loss

In our hybrid framework, we map both the real features and the synthetic features into the semantic embedding space, where we perform the final GZSL classification. Notably, we formulate L S E ( E ) only using the semantic descriptors of seen classes. Therefore, Equation (1) should be extended to:
L S E ( E ) = E p ( x , a ) [ max ( 0 , Δ a T E ( x ) + ( a ) T E ( x ) ) ] + E p G f ( x ˜ , a ) [ max ( 0 , Δ a T E ( G f ( a , ε ) ) + ( a ) T E ( G f ( a , ε ) ) ) ]
where p ( x , a ) is the empirical distribution of the real training samples of seen classes, and p G f ( x ˜ , a ) = p G f ( x ˜ | a ) p ( a ) is the joint distribution of a synthetic feature and its corresponding semantic descriptor.
The total loss of mapping net takes the form of:
L ( G 1 , E , H ) = L S E ( E ) + L S E ( H )
Thus, the total loss of our final hybrid framework is formulated as:
L t o t a l = L ( G 1 , E , H ) + L W G A N f e a t u r e + L W G A N a t t + L S E M C

2.4. Classification

First, given the average visual representation x ¯ of the unlabeled samples per unseen class, we generate semantic features for each unseen class c u by the feature generator network G 2 , which uses the average visual representation x ¯ and Gaussian noise as input and output synthetic features: e ˜ u = G 2 ( x ¯ , ε ) . Second, to keep the inputs of the classifer in the same model, we use the G 1 to generate visual features for each pseudo-semantic embedding, that is, G 1 uses real semantic features and generated semantic features, respectively, to synthesize visual features, which are denoted as x ˜ = G 1 ( a u , ε ) and x ˜ = G 1 ( e ˜ u , ε ) . Then, we can obtain a synthetic training feature set U t r = { x ˜ x ˜ } .
In the end, we map the synthetic training feature set U t r and the given training features of seen classes in S t r into the same embedding space h i = E ( x i ) and utilize the real seen samples and the synthetic unseen samples in the embedding space to train a softmax model as the final classifier. The whole process is shown in Figure 3.

3. Experimental Results

3.1. Datasets

We evaluate our method on four benchmark datasets for ZSL: Animals with Attributes 1 and 2 (AWA1 [9] and AWA2 [42]), Caltech-UCSD Birds-200-2011 (CUB) [40], and SUN Attribute (SUN) [41]. An example of the contents of each dataset is shown in Figure 4, all datasets and their statistics are summarized in Table 2.
AwA1 is a coarse-grained image dataset, containing 30,475 animal pictures in 50 categories, 40 of which are the seen class and 10 of which are the unseen class, and 85-dimensional class level attribute vectors are used. AWA2 is a fixed version of AWA1; they have the same category, category division way, and class-level attribute dimension, except that 37,322 coarse-grained animal pictures are used, and they do not overlap with AwA1 image instances.
CUB is a fine-grained image dataset, including 11,788 bird pictures in 200 classes, of which 150 classes are in the seen class and 50 classes are in the unseen class. CUB also provides an instance-level attribute vector; however, only 312-dimensional class-level attribute vectors are used in this work. The class-level attribute descriptor space is shown in Figure 5.
SUN is a scene dataset; this dataset is also a fine-grained one, which contains 14,340 pieces of 717 scenes. Here, 645 classes are used for training, and 72 classes are used for testing. Each class is annotated with a 102-dimensional attribute vector.

3.2. Implementation Details

We evaluated our method under the new split setting provided by [42], and more details on the settings can be found in [42]. We use strictly the 2048-dimensional feature of each image extracted from the pre-trained ResNet-101 provided by [42] similar to the others, and only the attribute vectors provided by each dataset are used.
We implement our method with PyTorch. We set the dimension of embedding h to the class-level attribute vector, 85 for AWA1 & AWA2, 312 for CUB, and 102 for SUN. The dimension of the non-linear projection’s output z is set to 512. We set a random mini-batch size of 4096 for AWA1 and AWA2, 2048 for CUB, and 1024 for SUN. Our generator and discriminator both contain a 4096-unit hidden layer with LeakyReLU activation. The classification part contains one fully connected layer, which will be utilized in making predictions. The numbers of input and output units follow the dimension of attribute vectors and the number of classes provided by each dataset.
For the hyperparameter, we set the temperature parameter τ e in Equation (3) according to [26]: τ e = 0.1 for AWA1, CUB and SUN, and τ e = 10.0 for AWA2. For the parameter in Equation (8), we use a large γ = 0.8 for fine-grained datasets (CUB and SUN) and a small γ = 0.1 for coarse-grained datasets (AWA1 and AWA2), referring to [43].

3.3. Experiments on Different Datasets

3.3.1. Performance of Different Methods under Comparison

Under the conventional ZSL scenario, we only evaluate the per-class Top-1 accuracy on unseen classes. The average per-class T1 accuracy is measured as follows, where y represents the number of unseen classes and c represents the serial number of each class:
a c c y = 1 | | y | | c = 1 | | y | | c o r r e c t p r e d i c t i o n s i n c s a m p l e s i n c
To show the effectiveness of the proposed method, we compared the simulated results with six other algorithms, and all the results are cited directly from their published papers. To provide a fair comparison, we adopt the experiment settings provided by [42], i.e., the datasets and their splits, and all the algorithms we compared adopt the same experiment settings. Table 3 shows that our method achieved a high value for CUB and the second-best position for AWA1. On the AWA1 dataset, the MG-ZSL yields a Top-2 accuracy of 70.6%, while the best Top-1 accuracy is 73.5% (yielded by ZMSL). It is worth noting that the MG-ZSL yields Top-1 accuracy higher than 70% on the CUB datasets, which is 0.7% higher than the second-banked algorithms. These results show that the method presented in this paper has achieved remarkable results.
In general, the experimental results of this paper have considerable performance with the current best case and are significantly better than previous mapping-based methods, such as DeViSE and DEM. Moreover, it also surpasses SE-GZSL, which is one of the state-of-the-art generation-based methods for all datasets and is comparable to the ZSML. Thus, the MG-ZSL model is very competitive.
The conventional ZSL scenario has been criticized as a restrictive setup because it is based on a strong assumption that the instances used in the test stage only come from unseen classes, which is less realistic. Therefore, GZSL was proposed, which is more realistic in practice. In the GZSL setting, the instances for evaluation may come from seen and unseen classes, so we choose the harmonic mean as our main evaluation indicator instead of the arithmetic mean, because considerably high-class accuracy will significantly affect the overall results with the latter. The harmonic mean can be computed by the following function:
H = 2 × a c c u × a c c s a c c u + a c c s
where a c c s is the average per-class top-1 (T1) accuracy of the test images from the seen classes and a c c u is average per-class top-1 (T1) accuracy of the unseen classes. Both of them are computed by Equation (12). For the GZSL setting, we add more recent models for comparison, and the results are presented in Table 4.
We compute the harmonic accuracy H, corresponding train accuracy a c c s , and test accuracy a c c u of our algorithm on all four of the above-mentioned datasets. The results are recorded in Table 4, and all results are cited directly from their published papers.
Table 4 shows that our method achieved high value in both H m e a n and a c c u for CUB. Our method shows a significant improvement of 2.9% compared to the second one, and for a c c u , we lead the second place by 8.1%. We also achieve the best position for AWA2 on a c c u , which leads the second place by 2.5%, and we achieve the second-best position for AWA2 on H m e a n , while the best Top-1 H m e a n is yielded by IZF. For SUN, we achieve the second-best position on a c c u and H m e a n and were significantly ahead of the third-best result. These results show that the method presented in this paper has achieved remarkable results.
In addition, it is worth noting that although compared with IZF, the current best SoTA model, our results cannot exceed it in all indicators, the IZF model, as acknowledged by its authors, is based on generative flows and has extremely high complexity, requiring a large number of computational resources and complex computational processes, and it takes human experience and trial and error to obtain the optimal combination of parameters. In contrast, our proposed model is lightweight, simple and easy to train. Similar results can be achieved while consuming far less computing resources than IZF.

3.3.2. Ablation Studies

In this paper, we employ a hybrid model combining the generation-based method and embedding-based method and two independent generative networks to synthesize the visual features for each unseen class. While testing, the two generative networks alleviate the problem of data imbalance by synthesizing visual features and semantic embedding, respectively.
In order to illustrate the effects of the multiple generative adversarial networks, we conduct the following experiments on ZSL and GZSL tasks: (1) experiments with only semantic embedding net ( S E ); (2) experiments with semantic embedding net and visual feature generation net ( G f ); (3) experiments with semantic embedding net, visual feature generation net and semantic embedding generation net ( G a ); (4) experiments with the whole JG-ZSL. The results are presented in Table 5 and Table 6, respectively.
From Table 5, it can be seen that the networks have different effects on the datasets, and the JG-ZSL yields the best results on most datasets. For the AWA2 dataset, the accuracy of T 1 per class on S E + G f + G a is higher than the whole JG-ZSL, while the performance is different on the other dataset. Furthermore, from Figure 6, it can be seen that compared with S E O n l y and the visual feature generate-only, the JG-ZSL also outperforms all the networks and settings on all the datasets.
From Table 6, the JG-ZSL yields the best results of harmonic mean on all the datasets. However, because the unseen accuracy U is seriously below the seen accuracies S, the accuracy of harmonic mean H is mostly up to unseen accuracy U. Therefore, the key to improving the harmonic mean H is to improve the unseen accuracy.
Furthermore, from Figure 7, it can be seen that compared with S E + G f and S E + G f + G a , the generated network does not improve the accuracy of the seen class as much as the unseen class, but the JG-ZSL still outperforms all the networks and settings on all the datasets because of the great enhancement to the unseen class.

3.3.3. Hyperparameter Analysis

We study the balance factor γ in Equation (8) to determine its influence on the module, γ was set as 0.1, 0.5 and 0.8 in turn, and the ablation results of CUB and AWA2 were shown in Table 7.
As shown in Figure 8, as γ grows, S, U and H gain consistent improvement on the fine-grained datasets (e.g., CUB), while coarse-grained datasets (e.g., AWA2) do the opposite. This result may reflect that the increase of intra-class compactness can improve the precision of fine-grained datasets, while for coarse-grained datasets, it is necessary to increase the inter-class separability for ambiguous classes.
We then uniformly set the number of generated semantic features to 2 and use generated semantic features and real semantic features to synthesize visual features. Assuming that the total number of synthesized visual features is N, the two generated semantic features and real semantic features generate 1/3*N synthesized visual features, respectively, and they contrast the effects by varying the number of visual features generated. The ablation results of CUB and AWA2 were as follows:
The number of generated samples is an important part for generative methods, so we also implement detailed experiments in this area. Our method achieves the best results on AWA1, AWA2, CUB, and SUN when we synthesize 1800, 2400, 600 and 90 examples per unseen classes, respectively. Figure 9 shows part of the experimental results, and there is an obvious phenomenon here that the number of generated samples for unseen classes is positively correlated with the values of U and H, which shows that the data-imbalance problem has been relieved by the generation model in our framework. However, with the large increase of unseen generated samples, the classification accuracy of seen classes also decreased significantly, which is one of the future directions to explore.
In this paper, the generalization of the G f and the specificity of the G a are combined not only to improve ZSL performance but also to alleviate the data imbalance problem and reduce the gap between seen accuracy S and unseen accuracy U and improve GZSL performance.

4. Conclusions

In this paper, we propose a joint model with multiple generative adversarial networks combining the embedding-based method and the generation-based method to synthesize the visual features and the semantic embedding points which realized the data enhancement of zero-shot learning in two ways, and it is also verified in the more challenging generalized zero-order learning setting. Inspired by the ideas of active learning and generative adversarial networks, the coupled generative networks work cooperatively to synthesize visual features of unseen classes under the constraint of semantic-relevant self-adaptive margin center loss. In addition, we compare the model with the current advanced methods, and the experimental results outperform the state-of-the-art embedding-based method and are competitive with the current generation-based method.
However, there are still some limitations in this paper. For example, all categories use the same way to generate semantic features which are not targeted enough, and there is no attempt to use VAE and other models to generate semantic features for comparison. Making full use of the pseudo-semantic features generated by images and comparing them with more generation models is the direction of future exploration. In addition to the above problem, exploring the more appropriate number of generated semantic features and different proportions of generated samples synthesized by generated semantic features and real semantic features are also problems that can be explored. In future work, we will further explore more efficient pseudo-semantic features generation methods and explore more obvious ways to improve the effect for unseen class classification and conduct experiments on larger datasets to improve the generalization abilities.

Author Contributions

Conceptualization, M.Z. and Y.S.; methodology, M.Z.; investigation and validation, M.Z., X.W. and Y.S.; data curation and formal analysis, M.Z., X.W. and W.W.; writing—original draft preparation, M.Z. and Y.S.; writing—review and editing, S.R., X.W. and W.W.; funding acquisition, S.R. and W.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The dataset used in this paper is publicly available at [42].

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. El Mrabet, M.A.; El Makkaoui, K.; Faize, A. Supervised Machine Learning: A Survey. In Proceedings of the 2021 4th International Conference on Advanced Communication Technologies and Networking (CommNet), Rabat, Morocco, 3–5 December 2021; pp. 1–10. [Google Scholar] [CrossRef]
  2. Tyukin, I.Y.; Gorban, A.N.; Alkhudaydi, M.H.; Zhou, Q. Demystification of Few-shot and One-shot Learning. In Proceedings of the International Joint Conference on Neural Networks (IJCNN), Shenzhen, China, 18–22 July 2021; pp. 1–7. [Google Scholar] [CrossRef]
  3. Jiao, Q.; Liu, Z.; Li, G.; Ye, L.; Wang, Y. Fine-Grained Image Classification with Coarse and Fine Labels on One-Shot Learning. In Proceedings of the 2020 IEEE International Conference on Multimedia and Expo Workshops (ICMEW), London, UK, 6–10 July 2020; pp. 1–6. [Google Scholar] [CrossRef]
  4. Wang, Y.; Yao, Q.; Kwok, J.; Ni, L.M. Generalizing from a Few Examples: A Survey on Few-Shot Learning. Assoc. Comput. Mach. 2021, 53, 63. [Google Scholar] [CrossRef]
  5. Scheirer, W.J.; de Rezende Rocha, A.; Sapkota, A.; Boult, T.E. Toward open set recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 35, 1757–1772. [Google Scholar] [CrossRef] [PubMed]
  6. Fei, G.; Wang, S.; Liu, B. Learning cumulatively to become more knowledgeable. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016; pp. 1565–1574. [Google Scholar]
  7. Rebuffi, S.-A.; Kolesnikov, A.; Sperl, G.; Lampert, C.H. iCaRL: Incremental classifier and representation learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 5533–5542. [Google Scholar]
  8. Leng, Q.; Ye, M.; Tian, Q. A Survey of Open-World Person Re-Identification. IEEE Trans. Circuits Syst. Video Technol. 2020, 30, 1092–1108. [Google Scholar] [CrossRef]
  9. Lampert, C.H.; Nickisch, H.; Harmeling, S. Learning to detect unseen object classes by between-class attribute transfer. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009. [Google Scholar]
  10. Palatucci, M.; Hinton, G.E.; Pomerleau, D.; Mitchell, T.M. Zero-shot learning with semantic output codes. Adv. Neural Inf. Process. Syst. 2009, 22, 1410–1418. [Google Scholar]
  11. Larochelle, H.; Erhan, D.; Bengio, Y. Zero-data learning of new tasks. In Proceedings of the Twenty-Third AAAI Conference on Artificial Intelligence (AAAI’08), Chicago, IL, USA, 13–17 July 2008; pp. 646–651. [Google Scholar]
  12. Zhuang, F.; Qi, Z.; Duan, K.; Xi, D.; Zhu, Y.; Zhu, H.; Xiong, H.; He, Q. A Comprehensive Survey on Transfer Learning. Proc. IEEE 2021, 109, 43–76. [Google Scholar] [CrossRef]
  13. Wang, W.; Zheng, V.W.; Yu, H.; Miao, C. A survey of zero-shot learning: Settings, methods, and applications. ACM Trans. Intell. Syst. Technol. 2019, 10, 13. [Google Scholar] [CrossRef]
  14. Li, J.; Lan, X.; Long, Y.; Liu, Y.; Chen, X.; Shao, L.; Zheng, N. A Joint Label Space for Generalized Zero-Shot Classification. IEEE Trans. Image Process. 2020, 29, 5817–5831. [Google Scholar] [CrossRef] [PubMed]
  15. Frome, A.; Corrado, G.S.; Shlens, J.; Bengio, S.; Dean, J.; Ranzato, M.A.; Mikolov, T. DeViSE: A deep visual-semantic embedding model. In Proceedings of the 26th International Conference on Neural Information Processing Systems-Volume 2 (NIPS’13), Lake Tahoe, NV, USA, 5–10 December 2013; Curran Associates Inc.: Red Hook, NY, USA, 2013; pp. 2121–2129. [Google Scholar]
  16. Dinu, G.; Lazaridou, A.; Baroni, M. Improving zero-shot learning by mitigating the hubness problem. arXiv 2014, arXiv:1412.6568. [Google Scholar]
  17. Radovanović, M.; Nanopoulos, A.; Ivanović, M. Hubs in space: Popular nearest neighbors in high-dimensional data. J. Mach. Learn. Res. 2010, 11, 2487–2531. [Google Scholar]
  18. Zhang, L.; Xiang, T.; Gong, S. Learning a Deep Embedding Model for Zero-Shot Learning. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 3010–3019. [Google Scholar]
  19. Min, S.; Yao, H.; Xie, H.; Zha, Z.-J.; Zhang, Y. Domain-Specific Embedding Network for Zero-Shot Recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019. [Google Scholar]
  20. Chawla, N.V.; Bowyer, K.W.; Hall, L.O.; P, W. Smote: Synthetic minority over-sampling technique. J. Artif. Intell. Res. 2002, 16, 321–357. [Google Scholar] [CrossRef]
  21. Chen, Q.; Koltun, V. Photographic image synthesis with cascaded refinement networks. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017. [Google Scholar]
  22. Li, K.; Zhang, Y.; Li, K.; Fu, Y. Adversarial Feature Hallucination Networks for Few-Shot Learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020. [Google Scholar]
  23. Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial nets. In Proceedings of the Annual Conference on Neural Information Processing Systems 2014: Advances in Neural Information Processing Systems 27, Montreal, QC, Canada, 8–13 December 2014. [Google Scholar]
  24. Kingma, D.P.; Welling, M. Auto-Encoding Variational Bayes. arXiv 2013, arXiv:1312.6114. [Google Scholar]
  25. Bond-Taylor, S.; Leach, A.; Long, Y.; Willcocks, C.G. Deep Generative Modelling: A Comparative Review of VAEs, GANs, Normalizing Flows, Energy-Based and Autoregressive Models. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 44, 7327–7347. [Google Scholar] [CrossRef] [PubMed]
  26. Verma, V.K.; Arora, G.; Mishra, A.; Rai, P. Generalized zero-shot learning via synthesized examples. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018. [Google Scholar]
  27. Verma, V.K.; Brahma, D.; Rai, P. Meta-learning for generalized zero-shot learning. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; pp. 6062–6069. [Google Scholar]
  28. Xian, Y.; Lorenz, T.; Schiele, B.; Akata, Z. Feature Generating Networks for Zero-Shot Learning. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 5542–5551. [Google Scholar]
  29. Xu, W.; Xian, Y.; Wang, J.; Schiele, B.; Akata, Z. VGSE: Visually-Grounded Semantic Embeddings for Zero-Shot Learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022. [Google Scholar]
  30. Fu, Z.; Xiang, T.; Kodirov, E.; Gong, S. Zero-shot object recognition by semantic mani- fold distance. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 2635–2644. [Google Scholar]
  31. Chen, T.; Kornblith, S.; Norouzi, M.; Hinton, G. A simple framework for contrastive learning of visual representations. In Proceedings of the International Conference on Machine Learning, Virtual Event, 13–18 July 2020. [Google Scholar]
  32. Han, Z.; Fu, Z.; Chen, S.; Yang, J. Contrastive Embedding for Generalized Zero-Shot Learning. In Proceedings of the International Conference on Machine Learning, Virtual, 18–24 July 2021. [Google Scholar]
  33. Arjovsky, M.; Chintala, S.; Bottou, L. Wasserstein GAN. arXiv 2017, arXiv:1701.07875. [Google Scholar]
  34. Gulrajani, I.; Ahmed, F.; Arjovsky, M.; Dumoulin, V.; Courville, A. Improved Training of Wasserstein Gans; Curran Associates Inc.: Red Hook, NY, USA, 2017; pp. 5769–5779. [Google Scholar]
  35. Jung, D.; Kang, D.; Kwak, S.; Cho, M. Few-shot Metric Learning: Online Adaptation of Embedding for Retrieval. In Proceedings of the Asian Conference on Computer Vision, Macau, China, 4–8 December 2022. [Google Scholar]
  36. Bucher, M.; Herbin, S.; Jurie, F. Improving Semantic Embedding Consistency by Metric Learning for Zero-Shot Classification. In Proceedings of the Computer Vision–ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016. [Google Scholar]
  37. Kuznetsova, A.; Hwang, S.J.; Rosenhahn, B.; Sigal, L. Exploiting view-specific appearance similarities across classes for zero-shot pose prediction: A metric learning approach. In Proceedings of the Thirtieth AAAI Conference on Artificial Intelligence (AAAI’16), Phoenix, AZ, USA, 12–17 February 2016; pp. 3523–3529. [Google Scholar]
  38. Wen, Y.; Zhang, K.; Li, Z.; Qiao, Y.U. A discriminative feature learning approach for deep face recognition. In Proceedings of the Computer Vision—ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016. [Google Scholar]
  39. Schroff, F.; Kalenichenko, D.; Philbin, J. Facenet: A unified embedding for face recognition and clustering. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015. [Google Scholar]
  40. Wah, C.; Branson, S.; Welinder, P.; Perona, P.; Belongie, S. The Caltech-Ucsd Birds-200–2011 Dataset; California Institute of Technology: Pasadena, CA, USA, 2011. [Google Scholar]
  41. Patterson, G.; Hays, J. Sun attribute database: Discovering, annotating, and recognizing scene attributes. In Proceedings of the 2012 IEEE Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, 16–21 June 2012. [Google Scholar]
  42. Xian, Y.; Lampert, C.H.; Schiele, B.; Akata, Z. Zero-shot learning-a comprehensive evaluation of the good, the bad and the ugly. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 41, 2251–2265. [Google Scholar] [CrossRef] [PubMed]
  43. Shen, Y.; Qin, J.; Huang, L.; Zhu, F.; Shao, L. Invertible zero-shot recognition flows. In Proceedings of the Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020. [Google Scholar]
  44. Shen, F.; Zhou, X.; Yu, J.; Yang, Y.; Liu, L.; Shen, H.T. Scalable zero-shot learning via binary visual-semantic embeddings. IEEE Trans. Image Process 2019, 28, 3662–3674. [Google Scholar] [CrossRef] [PubMed]
  45. Zhang, H.; Koniusz, P. Zero-shot kernel learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 10–23 June 2018. [Google Scholar]
  46. Han, J.; Pang, Y.; Ji, Z.; Wang, J.; Yu, Y. Class-specific synthesized dictionary model for zero-shot learning. Neurocomputing 2019, 329, 339–347. [Google Scholar]
  47. Li, C.; Ye, X.; Yang, H.; Han, Y.; Li, X.; Jia, Y. Generalized Zero Shot Learning via Synthesis Pseudo Features. IEEE Access 2019, 7, 87827–87836. [Google Scholar] [CrossRef]
  48. Jiang, H.; Wang, R.; Shan, S.; Chen, X. Transferable contrastive network for generalized zero-shot learning. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019. [Google Scholar]
  49. Han, Z.; Fu, Z.; Yang, J. Learning the redundancy-free features for generalized zero-shot object recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020. [Google Scholar]
  50. Liu, J.; Bai, H.; Zhang, H.; Liu, L. Near-Real Feature Generative Network for Generalized Zero-Shot Learning. In Proceedings of the 2021 IEEE International Conference on Multimedia and Expo (ICME), Shenzhen, China, 5–9 July 2021. [Google Scholar]
  51. Niu, C.; Shang, J.; Huang, J.; Yang, J.; Song, Y.; Zhou, Z.; Zhou, G. Unbiased feature generating for generalized zero-shot learning. J. Vis. Commun. Image Represent. 2022, 89, 103657. [Google Scholar] [CrossRef]
  52. Zhang, J.; Zhang, H.; Hu, B. Dual Prototype Relaxation for Generalized Zero Shot Learning. In Proceedings of the 2021 IEEE International Conference on Multimedia and Expo (ICME), Shenzhen, China, 5–9 July 2021. [Google Scholar]
Figure 1. Scheme of the proposed joint GANs (JG-ZSL) network.
Figure 1. Scheme of the proposed joint GANs (JG-ZSL) network.
Electronics 12 02308 g001
Figure 2. Scheme of pseudo-semantic embedding.
Figure 2. Scheme of pseudo-semantic embedding.
Electronics 12 02308 g002
Figure 3. Scheme of classification.
Figure 3. Scheme of classification.
Electronics 12 02308 g003
Figure 4. Scheme of classification.
Figure 4. Scheme of classification.
Electronics 12 02308 g004
Figure 5. Example of attribute space for the CUB dataset.
Figure 5. Example of attribute space for the CUB dataset.
Electronics 12 02308 g005
Figure 6. Comparison results with different network options on four benchmark datasets in ZSL.
Figure 6. Comparison results with different network options on four benchmark datasets in ZSL.
Electronics 12 02308 g006
Figure 7. Comparison results with different network options on four benchmark datasets in GZSL.
Figure 7. Comparison results with different network options on four benchmark datasets in GZSL.
Electronics 12 02308 g007
Figure 8. The effectiveness of the balance factor γ .
Figure 8. The effectiveness of the balance factor γ .
Electronics 12 02308 g008
Figure 9. The influence of different numbers of the synthesized samples for each unseen class.
Figure 9. The influence of different numbers of the synthesized samples for each unseen class.
Electronics 12 02308 g009
Table 1. Zero-Shot learning setting.
Table 1. Zero-Shot learning setting.
whether unlabeled testing instances are usedyesInstance-Transductive
noInstance-Inductive
whether description information of unseen class are usedyesClass-Transductive
noClass-Inductive
Table 2. Statistics of the four benchmark datasets used in our experiments.
Table 2. Statistics of the four benchmark datasets used in our experiments.
DatasetSeen/Unseen ClassAttributeTrain SeenTest SeenTest UnseenTotal Instance
AWA140, 108519,8234958568530,475
AWA240, 108523,5275882791337,322
CUB150, 5031270571764296711,788
SUN645, 7210210,3202580144014,340
Table 3. Results of conventional ZSL. The results are reported in %.
Table 3. Results of conventional ZSL. The results are reported in %.
MethodAWA1AWA2CUBSUN
Embedding-basedDeViSE [15]54.259.752.056.5
DEM [18]68.467.151.761.9
DSEN [19]72.371.862.2
Generation-basedSE-GZSL [26]69.569.559.663.4
ZSM L [27]73.576.169.660.2
f-CLSWGAN [28]68.257.360.8
Our JG-ZSL70.669.472.560.3
Table 4. Results of GZSL on five standard datasets, U: T1 per-class accuracy a c c u on unseen class set U, and S: T1 per-class accuracy a c c s on seen class set S, H = harmonic mean. The results report Top-1 accuracy in %, and the best results are marked in bold.
Table 4. Results of GZSL on five standard datasets, U: T1 per-class accuracy a c c u on unseen class set U, and S: T1 per-class accuracy a c c s on seen class set S, H = harmonic mean. The results report Top-1 accuracy in %, and the best results are marked in bold.
MethodAWA1AWA2CUBSUN
USHUSHUSHUSH
BZSL [44]19.923.921.7---18.925.120.917.317.617.4
ZSKL [45]18.379.329.818.982.730.824.263.935.1213125.1
DEM [18]32.884.747.330.586.445.119.65413.420.534.325.6
CSSD [46]34.787.149.6---19.162.729.3---
SPF-GZSL [47]48.559.853.652.460.956.330.263.440.932.259.041.6
TCN [48]49.476.560.061.265.863.452.652.052.331.237.334.0
SE-GZSL [26]56.367.861.558.368.162.841.553.346.740.930.534.9
RFF-GZSL [49]59.875.166.5---52.656.654.645.738.641.9
IZF [43]61.380.569.660.677.568.052.768.059.452.757.054.8
NereNet [50]56.270.162.4---51.056.553.645.738.141.6
UFG [51]59.366.062.5---45.256.850.435.846.040.2
DPR [52]54.781.965.6---48.966.656.48.135.540.7
Our JG-ZSL57.963.460.563.168.365.660.863.962.350.237.943.2
Table 5. Comparison results with different network options during the testing phase in ZSL. The results are reported in %.
Table 5. Comparison results with different network options during the testing phase in ZSL. The results are reported in %.
MethodAWA1AWA2CUBSUN
S E O n l y 54.358.558.450.1
S E + G f 65.968.167.653.7
S E + G f + G a 68.371.270.956.3
S E + G f + G a + L S M A C 70.669.472.560.3
Table 6. Comparison results with different network options during the testing phase in GZSL. The results are reported in %.
Table 6. Comparison results with different network options during the testing phase in GZSL. The results are reported in %.
MethodAWA1AWA2CUBSUN
USHUSHUSHUSH
SE-Only21.655.731.121.059.731.136.344.239.919.027.122.3
SE + G f 50.360.554.950.662.355.952.259.355.535.123.728.3
SE + G f + G a 54.761.357.354.469.361.057.763.360.751.336.142.3
SE + G f + G a + L S M A C 57.963.460.563.168.365.660.863.962.350.237.943.2
Table 7. The effectiveness of the balance factor γ . The results are reported in %.
Table 7. The effectiveness of the balance factor γ . The results are reported in %.
γ CUBAWA2
USHUSH
0.153.457.254.763.168.365.6
0.557.960.160.060.867.764.1
0.860.863.962.360.166.563.1
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, M.; Wang, X.; Shi, Y.; Ren, S.; Wang, W. Zero-Shot Learning with Joint Generative Adversarial Networks. Electronics 2023, 12, 2308. https://doi.org/10.3390/electronics12102308

AMA Style

Zhang M, Wang X, Shi Y, Ren S, Wang W. Zero-Shot Learning with Joint Generative Adversarial Networks. Electronics. 2023; 12(10):2308. https://doi.org/10.3390/electronics12102308

Chicago/Turabian Style

Zhang, Minwan, Xiaohua Wang, Yueting Shi, Shiwei Ren, and Weijiang Wang. 2023. "Zero-Shot Learning with Joint Generative Adversarial Networks" Electronics 12, no. 10: 2308. https://doi.org/10.3390/electronics12102308

APA Style

Zhang, M., Wang, X., Shi, Y., Ren, S., & Wang, W. (2023). Zero-Shot Learning with Joint Generative Adversarial Networks. Electronics, 12(10), 2308. https://doi.org/10.3390/electronics12102308

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop