Next Article in Journal
Landslide Susceptibility Using Climatic–Environmental Factors Using the Weight-of-Evidence Method—A Study Area in Central Italy
Next Article in Special Issue
Robust Image Inpainting Forensics by Using an Attention-Based Feature Pyramid Network
Previous Article in Journal
Semi-Supervised Learning Matting Algorithm Based on Semantic Consistency of Trimaps
Previous Article in Special Issue
LogBD: A Log Anomaly Detection Method Based on Pretrained Models and Domain Adaptation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

SLNER: Chinese Few-Shot Named Entity Recognition with Enhanced Span and Label Semantics

1
College of Information Science and Engineering, Xinjiang University, Urumqi 830049, China
2
Xinjiang Key Laboratory of Signal Detection and Processing, Urumqi 830049, China
3
Xinjiang Uygur Autonomous Regin Product Quality Supervision and Inspection Institute, Urumqi 830049, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2023, 13(15), 8609; https://doi.org/10.3390/app13158609
Submission received: 8 June 2023 / Revised: 12 July 2023 / Accepted: 25 July 2023 / Published: 26 July 2023

Abstract

:
Few-shot named entity recognition requires sufficient prior knowledge to transfer valuable knowledge to the target domain with only a few labeled examples. Existing Chinese few-shot named entity recognition methods suffer from inadequate prior knowledge and limitations in feature representation. In this paper, we utilize enhanced Span and Label semantic representations for Chinese few-shot Named Entity Recognition (SLNER) to address the problem. Specifically, SLNER utilizes two encoders. One encoder is used to encode the text and its spans, and we employ the biaffine attention mechanism and self-attention to obtain enhanced span representations. This approach fully leverages the internal composition of entity mentions, leading to more accurate feature representations. The other encoder encodes the full label names to obtain label representations. Label names are broad representations of specific entity categories and share similar semantic meanings with entities. This similarity allows label names to offer valuable prior knowledge in few-shot scenarios. Finally, our model learns to match span representations with label representations. We conducted extensive experiments on three sampling benchmark Chinese datasets and a self-built food safety risk domain dataset. The experimental results show that our model outperforms the F1 scores of 0.20–6.57% of previous state-of-the-art methods in few-shot settings.

1. Introduction

Named entity recognition (NER) aims to identify specific, meaningful entities from text, such as LOCATION and ORGANIZATION, and classify them into predefined categories, as shown in Figure 1. NER is an essential prerequisite task for many natural language processing task, such as information extraction [1], question-answering systems [2], and machine translation [3].
In recent years, neural-network-based techniques have been widely applied in NER [4,5]. However, neural networks are data-driven machine learning methods, and the quantity of training data often limits their performance. Unfortunately, annotated data used for training are often scarce and expensive, especially in specific domains (e.g., the food safety risk domain). Therefore, there has been widespread interest in a challenging yet practical research field: few-shot NER.
One of the challenges of few-shot NER is how to accurately incorporate prior knowledge to effectively classify unseen entity types when confronted with a few examples. Recently, similarity-based methods such as prototype networks have been extensively studied and achieved great success for few-shot learning [6,7,8]. The core idea is to classify input examples from a new domain based on the similarity between their representations and those of each class in the support set. However, this approach experiences a significant drop in performance in a few-shot setting due to the limited representativeness of the data. The prompt-based approach [9,10], by manually or automatically adding prompt words to sentences, guides the model to learn more quickly and accurately while reducing the gap between pretraining and fine tuning. This approach has shown remarkable performance in few-shot learning. However, these methods do not directly leverage the rich prior knowledge contained in label semantics.
In addition, Chinese NER is more challenging than English NER due to the relatively ambiguous entity mentions in Chinese, which limits feature representation and affects the accuracy of NER. Zhang and Yang [11] addressed this issue by using lattice LSTM to represent the entity in sentences and by incorporating the potential lexical information into a character-based LSTM-CRF model. While this character-based representation effectively solves the segmentation error problem, it requires the introduction of a complex external lexicon. Some more recent attempts have switched to span-based feature representations for Chinese NER [12,13], explicitly utilizing span-level information to address token-wise label dependency and better handle nested entities. However, these span-based feature representations only perform simple concatenation of the start and end positions of the span without fully exploiting the internal information of the span, which limits the feature representation of the named entity. For example, when the named entity “伊河谷食品科技有限公司 (Yihegu Food Technology Limited Company, Urumqi, China)” has the internal information “科技有限公司 (Technology Limited Company)”, it becomes easier to classify it as an ORGANIZATION entity.
In response to these challenges, we propose a model called SLNER with enhanced span and label semantic representations to tackle the challenges of Chinese few-shot NER. Specifically, SLNER utilizes two encoders. One encoder is used to encode the text and its spans. This module captures the head and tail information of spans using the biaffine attention mechanism and incorporates self-attention to capture the internal information of spans. Ultimately, these representations are fused to obtain enhanced span representations. This approach fully utilizes the internal composition of entity mentions to achieve more accurate feature representations. In contrast to traditional span-based methods that concatenate the start and end positions of entity mentions, our approach fully exploits the information of each token within entity mentions, providing sufficient and essential clues for entity recognition.
The other encoder is used to encode full label names. Label names are highly generalized specific entity categories and exhibit similar semantics to entities, which can provide additional prior knowledge in few-shot scenarios. Compared to traditional similarity-based methods (e.g., prototype networks), label semantics provide more generic similarity representations, especially in situations where the target domain has a scarcity of samples.
Ultimately, our model learns to match span representations with label representations. We employ a two-stage training strategy using source and target domains, enabling the model to transfer knowledge from the high-resource source domain to the low-resource target domain.
Furthermore, to promote research and applications in low-resource domains, we developed an NER dataset named RISK, which was specifically designed for the food safety risk domain. RISK comprises 5 coarse-grained and 20 fine-grained entity types, each labeled and organized in a hierarchical structure of coarse-grained + fine-grained. We also conducted a performance evaluation of our model on the RISK dataset, and the experimental results demonstrate the challenging nature of the RISK dataset. Constructing an NER dataset in the food safety domain can drive advancements in related research areas such as food traceability and food safety regulation. Additionally, this dataset can serve as a foundation for the development of applications in food safety, including food safety warning systems and food recall management.
We have documented the experimental results on three sampling benchmark Chinese NER datasets and a self-built food safety risk domain dataset. Our contributions can be summarized as follows:
  • We propose a simple and effective model named SLNER, which leverages enhanced span representations and label semantics to address the issues of inadequate prior knowledge and limitations in feature representation in Chinese few-shot named entity recognition;
  • We created a challenging food safety risk domain dataset, RISK, which is divided into 5 coarse-grained and 20 fine-grained entity categories. This dataset provides data support for the development of named entity recognition applications in the domain of food safety;
  • Our proposed model achieved promising performance on the four sampling Chinese NER datasets (including our self-built dataset). Specifically, our model outperformed previous works with F1 scores ranging from 0.20% to 6.57% in different few-shot settings (following the settings of PCBERT) on the Ontonotes, MSRA and Resume datasets. It also achieved promising F1 scores on our self-built RISK dataset.

2. Related Work

2.1. Few-Shot NER

Compared to traditional standard NER [14,15], few-shot NER aims to learn how to recognize named entities with limited data. Currently, research on few-shot named entity recognition based on the traditional standard NER framework primarily focuses on two aspects: incorporating prior knowledge (external information) at the data level and enhancing model generalization at the model level.
In the first aspect, Tong et al. [16] proposed a model, Mining Undefined Classes from Other-class (MUCO), that can automatically induce different undefined classes from the other class to improve few-shot NER. On the other hand, Cui et al. [9], Lee et al. [17], and Chen et al. [18], using a prompt-based approach, treated named entity recognition as a generation task. By manually or automatically adding templates to the training data, the model is trained to predict the [MASK] positions within the templates. This method reduces the gap between pretraining and fine tuning, allowing the model to perform better, even with limited data. Based on multitask instructions, Wang et al. [19] enhanced the knowledge of the training data by adding different auxiliary tasks and corresponding instructions and options; this allowed the model to accurately identify entity information, even with few samples. Additionally, Chen et al. [20] effectively leveraged illustrative instances to precisely transfer knowledge from external resources by describing both entity types and mentions using a universal concept set.
For the second aspect, methods based on meta-learning [21] aim to enable machines to “learning to learn”. The essence of these methods is to train the network model to learn a more robust initialization, allowing it to quickly generalize to new tasks with only a few examples. Methods based on contrastive learning also focus on learning shared features among instances of the same class and distinguishing differences between instances of different classes. Das et al. [22] trained an encoder that produces similar encodings for instances of the same class while ensuring that the encodings of different classes are as dissimilar as possible. This approach reduces the limitations of generalization in the target domain, thereby improving performance in the few-shot setting [23], which augments the distribution of entity labels by assigning k-nearest neighbors retrieved from the training set. This strategy makes the model more capable of handling long-tail cases, along with better few-shot learning abilities.

2.2. Span-Based Method

Previously, the majority of research work treated NER as a sequence-labeling task, where entity classification is performed at the token level. These methods often employ the BIO tagging scheme, where each character’s category is determined to achieve entity recognition for Chinese NER tasks, as shown in Figure 2. As a representative example, Huang et al. [24] utilized BiLSTM as an encoder to learn contextual representations, then employed a conditional random field (CRF) as a decoder to label tokens. Additionally, leveraging the power of pretrained language models such as ELMo [25] and BERT [5] significantly improved the performance of NER.
While the aforementioned token-based NER methods have achieved considerable success, they suffer from two inherent limitations: token-wise label dependency and difficulty handling nested entities. As shown in Figure 2, this method may fail to correctly classify nested entities such as “Taojie greengrocer” and “Hami Melon”. Additionally, if errors occur during the labeling process, incorrect beginning tags (B) or internal tags (I) may affect subsequent labeling, resulting in the erroneous tagging of the entire entity.
Therefore, recent work has increasingly embraced span-based approaches to address NER tasks. This method involves partitioning all possible spans in a sentence into predefined types (e.g., PER or LOC) and determining whether a given text span belongs to a particular category, as shown in Figure 3. Yu et al. [26] adopted a biaffine attention model to assign scores to all potential spans and achieved state-of-the-art performance on both flat and nested English NER datasets. Shen et al. [27] also employed a span-based framework for Chinese NER datasets, effectively addressing the issue of nested entities. Yu et al. [28] and Wang et al. [29] proposed span-level metric learning to bypass the token-wise label dependency problem while still explicitly utilizing phrase representations. Wang et al. [30] introduced a span-based prototype network and a global boundary matrix to learn explicit span boundary information.
In this study, we also utilize a span-based approach for named entity recognition and extensively explore the internal information of entity word spans to obtain enhanced span feature representations.

2.3. Label Semantics

In the task of few-shot text classification, Luo et al. [31] directly appended the label name to the text input in BERT, obtaining feature vectors with more enriched semantic information, thus demonstrating the effectiveness of label semantics in few-shot scenarios. Cui et al. [9] employed a similar approach by reconstructing the input text template, reframing NER as a cloze task, and using a sequence-to-sequence model to fill the entity label names in predefined templates, achieving few-shot named entity recognition through a prompt-based method. Ma et al. [32] abandoned the template construction process while retaining the word prediction paradigm of pretrained models to predict class-related pivot words (or label words) at entity positions. Inspired by prompt-tuning methods, Zhong et al. and Ye et al. [33,34] initialized markers in the NER task not with random initialization but with meaningful words (such as label names), resulting in a certain degree of improvement in model performance. Our proposed model leverages the connection between label semantics and entity spans (Figure 4) to learn aligning span representations with label representations, exhibiting promising results.

3. Method

This chapter first formalizes the problem of few-shot named entity recognition (Section 3.1). Then, we propose our model, SLNER, for Chinese few-shot NER (Section 3.2). The model consists of two encoders: one for encoding the text and its span to obtain better feature representations (Section 3.3.1) and another for encoding the full label name to capture additional prior knowledge (Section 3.3.2). Additionally, we adopt a two-stage training strategy using source and target domains (Section 3.4). The details are outlined as follows.

3.1. Few-Shot NER Task Formalization

For the few-shot NER task, assume that we have a resource-rich source domain NER dataset, Ɗ R = T 1 , L 1 , , T n , L n , where T i = t 1 , t l represents the i -th text ( i 1 , n ), t i represents the i -th token ( i 1 , l ), and L i represents the labels corresponding to the entity spans in the i -th text ( i 1 , n ). We use R to denote the label set of the source domain dataset. Then, given a resource-scarce target domain dataset, Ɗ T = T 1 , L 1 , , T m , L m , the number of texts in the target domain dataset is limited (i.e., m n ), and the label types in the target domain may differ from those in the source domain (i.e., R T ). We aim to leverage the knowledge from the source domain dataset to improve the model’s performance on the target domain dataset.

3.2. Overall Structure

The overall architecture of the SLNER model is illustrated in Figure 5. For span representation, given a sentence ( T = t 1 , , t l ) of length l , we use BERT as our encoder, which encodes the context of the i -th token in the sentence as follows:
h i = B E R T t i               t i , t i T
where d h is the hidden dimension of the encoder, and the output dimension after passing the original sentence through the encoder is l × d h .
To further enhance the modelling of the sequential order of the text, the embedding representation obtained from BERT is then passed through a bidirectional LSTM layer. The forward LSTM network captures the hidden forward states (historical features), while the backward LSTM network captures the hidden backward states (future features), resulting in a context-aware encoding representation:
x i =   L S T M h i , x i 1   ;     L S T M h i , x i + 1
At this stage, the output dimension of the original sentence through the bidirectional LSTM layer is l × d x .
We predefine an n-gram value ( w ), which represents the maximum length of spans that can be formed in a text. The number of possible spans that can be formed in a sentence of length l is given by:
N s =         w l w 1 + i = 1 w 1 i         l w i = 1 w 1 i                                                                       l < w
Next, the token-embedding sequence obtained from the LSTM layer is used to construct the span-level feature vector representation ( h s p a n ) through a span extractor (details in Section 3.3.1). The dimension of h s p a n is finally expanded to N s × d h through a feed-forward layer.
For label representation (details in Section 3.3.2), we manually define the appropriate full label name for each label. Similarly, we use BERT as the encoder and directly encode the label using Equation 1 to obtain the global semantic feature ( h C L S ). The difference from span encoding is that we further pass h C L S through a pooler layer to obtain the semantic feature vector representation ( h l a b e l ), which serves as the final representation of the label:
h l a b e l = T a n h W p × h C L S + b p
where W p represents the weight parameters in the pooler layer, b p represents the bias parameters, and T a n h is the activation function. The dimension of h l a b e l needs to be expanded to d h × N c , where N c is the number of entity categories.
According to our approach, there is a correlation between labels and spans appearing in the text. Therefore, we capture this correlation through the dot product:
H = h L S T × h s p a n
where the similarity matrix ( H ) has dimensions of N s × N c . We use a standard linear classifier with a softmax function to predict the entity type for each span, resulting in the final predicted output ( y ^ s p a n ):
y ^ s p a n = S o f t m a x W c × h s p a n + b c
where W c is the trainable parameter of the classifier, and b c is the bias. Finally, we use the cross-entropy loss function to compute the loss, which measures the difference between the predicted results and the ground truth labels:
L P S N N E R = y s p a n log y ^ s p a n

3.3. Specific Structure

3.3.1. Enhanced Span Representation

In previous span-based NER models [35], it was common practice to concatenate the embedding information of the start-position token and the end-position token of the entity (referred to as the “outer span”) to represent the span of that entity, which is then used for the final classification decision:
h s p a n = x s t a r t ; x e n d
This approach lacks interaction between the start and end tokens and fails to fully utilize the informative content within the span. Moreover, this span representation is coarse-grained. To address these limitations, Yu et al. [26] proposed a biaffine decoder that utilizes two fully connected layers to enable interaction between the start and end tokens while simultaneously predicting the span type. However, in this biaffine method, the information within the span is still ignored.
To fully utilize the informative content within the span, we employ enhanced span to generate the final span representation (as shown in Figure 6). Specifically, we pass the token-embedding information through the outer and inner span modules. The outer span module, similar to the biaffine decoder method, utilizes the biaffine attention mechanism to obtain the outer span representation:
h o u t e r   s p a n = h s t a r t T U h e n d + W h s t a r t h e n d + b
where h s t a r t and h e n d represent the start and end token embedding of spans in a text, respectively; U and W are learnable parameters; and b is the bias. The dimension of h o u t e r   s p a n is expanded to N s × d x .
We designed the inner span module to capture the token-level information within the span. For this purpose, we use linear attention to generate information interaction for each token. Specifically, this module uses span and the start- and end-position indices of the span as input. It starts by applying a feed-forward neural network (FFNN) to non-linearly transform the input representations, obtaining context-aware representations. Then, it computes normalized scores for each position. Finally, these representations are fed into a self-attention layer, which combines the representations of each position with those of other positions, weighted by the attention scores. This allows the model to capture potential relationships between the tokens within the span. The result is the inner span representation:
a i = W f × x i + b f
s i = exp a i k = n m exp a k
h i n n e r   s p a n = i = n m s i × x i
where x i represents the hidden representation from the bidirectional LSTM, and W f and b f are the learnable weights and biases of the feed-forward neural network, respectively. The indices i n , n + 1 , , m correspond to the token indices within the span, where n and m represent the start and end indices of the span, respectively. When n = m (indicating a span of length 1), we do not extract additional features and simply use the hidden representation ( x i ).
To predict the entity type, we integrate the outer span representation and inner span representation in a gate network to obtain the weight coefficient ( γ ) (as shown in Figure 7):
γ = σ U g h o u t e r   s p a n ; h i n n e r   s p a n + b g
where U g and b g are trainable parameters of the gate network, and σ represents the sigmoid function. The dimension of γ is N s × 1 .
The final enhanced span representation is obtained by weighting the inner span representation and outer span representation using γ :
h s p a n = γ h o u t e r   s p a n + 1 γ h i n n e r   s p a n
where represents element-wise multiplication, and the resulting h s p a n has dimensions of N s × d x .

3.3.2. Label Representation

We believe that label semantics can provide additional prior knowledge. Label semantics carry the semantic information of entities in the same category, as this information is manually summarized and induced from a large amount of data. Therefore, when data are limited, especially in small-sample scenarios, we can introduce label semantics to allow our model to make generalizations from the available data. Furthermore, full label names themselves are mentions that appear in various contexts within the text. Their frequencies are synchronized to some extent with the corresponding entity words of their respective categories. Thus, there exists a semantic correlation between label names and the span tokens appearing in the text, and this correlation can be leveraged and utilized.
Considering that our label encoder is based on BERT and incorporates prior knowledge from pretraining, our label representation module allows any form of text to be used as input. This design not only enables easy and rapid expansion to unseen label sets in low-resource domains but also prevents the model from forgetting prior knowledge. We experimented with different label forms and analyzed their effects (Section 5.1). Table 1 presents the final forms of full label names used in this study (for the non-entity type in each dataset, we uniformly use “其他 (other)” as the label name).

3.4. Training Strategy

Compared to the previous work of traditional NER neural architecture, our model does not require a new randomly initialized top-layer classifier for new datasets with new unseen label names. Therefore, our model allows domain transfer for different label categories, which is very beneficial for few-shot learning. On this basis, we adopt a two-stage training program. In the first stage, we pretune our model on the source dataset to obtain a prior knowledge-rich source domain model. In the second stage, we fine tune the source model from the previous stage as the initial model on the target domain dataset. During model training, two encoders are updated at each iteration of the two stages, which helps align the span-embedding space with the label-embedding space.

4. Experiments

4.1. Datasets

  • Target Domain Datasets
To validate the effectiveness of our model, we used three benchmark Chinese datasets and a self-built dataset as target domain datasets:
Ontonotes 4.0 [36]: This dataset comprises corpora from the news and broadcast domains, covering four entity types.
MSRA [37]: This dataset comprises corpora from the news domain and includes three entity types.
Resume [11]: This dataset consists of abstracts from resumes of senior managers in publicly listed companies. It encompasses eight entity types.
RISK: This dataset was created by us and focuses on food safety risk-related corpora. We hierarchically categorized the dataset into 5 coarse-grained and 20 fine-grained entity types, as shown in Figure 8. The entity count for each fine-grained category is illustrated in Figure 9. The distribution of entity categories is not uniform, making RISK challenging at the fine-grained level.
The detailed statistics of the target domain datasets are presented in Table 2. To ensure the comparability of experimental results, we followed the sampling approach of PCBERT [10] to simulate a few-shot scenario (see Section 4.4 for details). Compared to the N-way K-shot setting, we believe that this sampling approach is more representative of realistic few-shot scenarios. We further divided the datasets into scenarios with even fewer examples and tested our model, demonstrating its effectiveness even in scenarios with fewer data.
  • Source Domain Dataset
We adopted the source domain dataset used in the PCBERT to ensure experimental consistency and reliability. This high-resource dataset is a combination of multiple datasets, including CLUENE [38], CNERTA [39], RenMinRiBao [40], and datasets from unknown sources.

4.2. Implementation Details

We adopted a two-stage training approach, starting with 5 epochs of pretraining on the source domain dataset, followed by 100 epochs of fine tuning using the pretrained model as the initial model in the target domain. For the upstream (BERT) part, we used the standard BERT-wwm-ext pretrained language model with 768-dimensional hidden representations for character embeddings and the learning rate set to 1 × 10−5. For the downstream model, the learning rate was set to 1 × 10−4. The hidden state size of the BiLSTM encoder was set to 200. We used the Adam optimizer [41]. We performed three different random samplings for all datasets to reduce randomness and reported the average performance as the final experimental result. All experiments were conducted on a computer with an Intel Core i9 13900K/F CPU®5.8 GHz and a GeForce RTX 3090 GPU with 24 GB of memory. The models were implemented using the PyTorch 1.12.0 framework.

4.3. Baseline

BERT [5] is the fundamental BERT-based NER method that adds a token classifier layer to the downstream of BERT.
BERT-LC is an effective baseline method for handling Chinese NER. It extends the regular BERT by adding a BiLSTM layer and employs a CRF layer as a decoder to predict token types.
Lattice LSTM [11] is a character-based Chinese NER method that introduces lexicon information by using a lattice-structured LSTM model.
FLAT [42] is a lattice-structured NER method based on a transformer. It constructs a flat-structured transformer to fully utilize lattice information and leverage parallel computing on GPUs.
LEBERT [43] addresses Chinese sequence-labeling tasks using lexicon-enhanced BERT. It incorporates lexical information into the encoding process of BERT’s underlying layers using lexicon adapter layers.
LEBERT-LC is an extension of LEBERT that further adds a BiLSTM layer after the BERT output layer to facilitate a comparison with our proposed SLNER model.
PCBERT [10] is a prompt-based Chinese few-shot NER model. It consists of the P-BERT component and the C-BERT component, integrating lexical features and implicit label features.

4.4. Experimental Results

Our model was thoroughly evaluated on four sampled Chinese NER datasets. Specifically, we sampled K = 250, K = 500, K = 1000, and K = 1350 examples from the Ontonotes, MSRA, and Resume datasets, respectively, as training sets to cover different levels of data scarcity and comprehensively evaluate the training effectiveness and robustness of the model. The standard F1 score is used as the evaluation metric to measure the final performance of the model.
Table 3 shows the experimental results of our proposed model and the baseline models on three benchmark NER datasets. The results indicate that, except for the 1350-shot setting in the resume dataset, our method outperforms the state-of-the-art models by 0.2–6.57 percent, demonstrating the excellent performance of our method in few-shot settings.
Table 4 presents the experimental results of our proposed model on the RISK dataset. Compared to coarse-grained classification, the model shows a maximum decrease of 3.64–6.11 percent in F1 score under fine-grained classification. This indicates that the dataset is more challenging in fine-grained classification.
In addition, to further validate the feasibility of our model in a low-resource setting, we also sampled data under K = 10, K = 20, and K = 50 settings and evaluated the effectiveness of our model with even fewer examples. The results are shown in Table 5. The experimental results demonstrate that our model can still achieve good results, even with as few as a few dozen samples. We analyze this phenomenon in Section 5.3.

5. Analysis

5.1. Ablation Study

To validate the impact of enhanced span representation and label representation on SLNER, we conducted extensive ablation experiments on the Ontonotes, MSRA, Resume, and RISK datasets under different sampling settings (K = 250, 500, 1000, and 1350) as shown in Table 6. Specifically, we first removed the enhanced span representation module and observed a decrease in performance across all datasets with different sampling sizes. This demonstrates the effectiveness of the module, which we analyze in Section 5.2. Next, we removed the label representation module. It is worth mentioning that, due to the removal of the label representation module, we had to introduce a top-layer classifier, which resulted in the inability to use the source-domain-initialized model. As a result, the performance of our model significantly dropped, confirming our hypothesis that the label semantics carry rich prior knowledge. This also indicates the importance of prior knowledge from the pre-fine-tuning stage in low-resource environments. We provide a detailed analysis of this in Section 5.3.
To further analyze the impact of label representation, we experimented with different definitions: full label name (the one used in our model), misleading label name, and indistinguishable label name. Misleading label name refers to randomly shuffling the full label name, for example, changing the label name corresponding to “PER” from “人名 (person name)” to “组织 (organization)”. Indistinguishable label name refers to unifying all different label names into a single label name, such as using “人名 (person name)” for all labels. We conducted a comparative experiment on the Ontonotes dataset, as shown in Figure 10. The results indicate that a misleading label name slightly negatively affects the model, but an indistinguishable label name has a significant negative impact. This suggests that the selection of label names must be category-related and that a label name carries entity-specific prior knowledge.

5.2. The Impact of Enhanced Span Representation

From Table 6, it can be observed that removing the inner span module results in a decrease in F1 scores. Particularly, when there are fewer samples (K = 250), the impact of removing the inner span module is more significant. This is because in the two-stage training process, during the source domain training, the lack of accurate span representations leads to a decrease in NER accuracy. However, when transitioning to the target domain, this inaccuracy is amplified, and the effect is more pronounced with fewer samples. Furthermore, during the ablation experiments, it was found that removing the inner span module leads to slower convergence of the model, indicating that without intra-span information, entity span localization becomes more challenging.

5.3. The Impact of Label Representation

The experimental results in Table 6 indicate that the impact of label name becomes more pronounced as the number of samples decreases (compared to K = 1350, the model’s performance drops more significantly when label name is removed at K = 250). This is because as the number of training samples decreases, the number of words representing the same type of entity also decreases, resulting in less specific entity descriptions. On the other hand, label name itself is a broad concept that can represent entity categories, and its effectiveness increases when the number of samples is reduced. This leads to improved generalization ability.

6. Conclusions

In this study, we propose the SLNER model with enhanced span and label semantics to address the issues of inadequate prior knowledge and feature representation limitations in Chinese few-shot named entity recognition (NER). To tackle the feature representation limitations, we employ the biaffine attention mechanism and self-attention to obtain enhanced span representations, fully leveraging the internal composition of entity mentions. For the problem of inadequate prior knowledge, we introduce label semantics, which provide a highly abstract representation of specific entity categories with similar semantics and provide additional prior knowledge in few-shot scenarios. Our model learns to match span representations and label semantics to achieve entity recognition. Additionally, we constructed the RISK dataset in the food safety risk domain, which consists of 5 fine-grained and 20 coarse-grained entity types, providing a data foundation for NER development in low-resource domains. We extensively validated our proposed model on three benchmark datasets and our self-built dataset, and the results demonstrate the effectiveness of our model in addressing the issues of Chinese few-shot NER.

Author Contributions

Conceptualization, Z.R. and X.Q.; methodology, Z.R.; software, Z.R.; validation, Z.R. and X.Q.; formal analysis, Z.R.; investigation, Z.R. and W.R.; resources, W.R.; data curation, Z.R.; writing—original draft preparation, Z.R.; writing—review and editing, Z.R. and X.Q.; supervision, X.Q. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Major science and technology special projects of Xinjiang Uygur Autonomous Region (2020A03001) and its subprogram, Key technology development and application demonstration of integrated food data supervision platform in Xinjiang region (2020A03001-2).

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Lu, Y.; Liu, Q.; Dai, D.; Xiao, X.; Lin, H.; Han, X.; Sun, L.; Wu, H. Unified Structure Generation for Universal Information Extraction. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics, Dublin, Ireland, 22–27 May 2022; Volume 1, pp. 5755–5772. [Google Scholar]
  2. Mollá, D.; Van Zaanen, M.; Smith, D. Named entity recognition for question answering. In Proceedings of the Australasian Language Technology Workshop 2006, Sydney, Australia, 11 November 2006; pp. 51–58. [Google Scholar]
  3. Stahlberg, F. Neural Machine Translation: A Review. J. Artif. Intell. Res. 2020, 69, 343–418. [Google Scholar] [CrossRef]
  4. Lample, G.; Ballesteros, M.; Subramanian, S.; Kawakami, K.; Dyer, C. Neural architectures for named entity recognition. In Proceedings of the 2016 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, San Diego, CA, USA, 12–17 June 2016; Association for Computational Linguistics: Stroudsburg, PA, USA, 2016; pp. 260–270. [Google Scholar]
  5. Devlin, J.; Chang, M.W.; Lee, K.; Toutanova, K. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. In Proceedings of the NAACL-HLT 2019, Minneapolis, MN, USA, 2–7 June 2019. [Google Scholar]
  6. Vinyals, O.; Blundell, C.; Lillicrap, T.; Wierstra, D. Matching networks for one shot learning. In Proceedings of the Advances in Neural Information Processing Systems 29 (NIPS 2016), Barcelona, Spain, 5–10 December 2016. [Google Scholar]
  7. Snell, J.; Swersky, K.; Zemel, R. Prototypical networks for few-shot learning. In Proceedings of the Advances in Neural Information Processing Systems 30 (NIPS 2017), Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
  8. Hou, Y.; Che, W.; Lai, Y.; Zhou, Z.; Liu, Y.; Liu, H.; Liu, T. Few-shot slot tagging with collapsed dependency transfer and label-enhanced task-adaptive projection network. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, Online, 5–10 July 2020; Association for Computational Linguistics: Stroudsburg, PA, USA, 2020; pp. 1381–1393. [Google Scholar]
  9. Cui, L.; Wu, Y.; Liu, J.; Yang, S.; Zhang, Y. Template-based named entity recognition using BART. In Proceedings of the Findings of the Association for Computational Linguistics, ACL/IJCNLP 2021, Online, 1–6 August 2021; Association for Computational Linguistics: Stroudsburg, PA, USA, 2021; pp. 1835–1845. [Google Scholar]
  10. Lai, P.; Ye, F.; Zhang, L.; Chen, Z.; Fu, Y.; Wu, Y.; Wang, Y. PCBERT: Parent and Child BERT for Chinese Few-shot NER. In Proceedings of the 29th International Conference on Computational Linguistics, Gyeongju, Republic of Korea, 12–17 October 2022; pp. 2199–2209. [Google Scholar]
  11. Zhang, Y.; Yang, J. Chinese NER using lattice LSTM. In Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics, Melbourne, Australia, 15–20 July 2018; Volume 1, pp. 1554–1564. [Google Scholar]
  12. Dong, X.; Xin, X.; Guo, P. Chinese NER by Span-Level Self-Attention. In Proceedings of the 2019 15th International Conference on Computational Intelligence and Security (CIS), Macao, China, 13–16 December 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 68–72. [Google Scholar]
  13. Li, X.; Feng, J.; Meng, Y.; Han, Q.; Wu, F.; Li, J. A Unified MRC Framework for Named Entity Recognition. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, Online, 5–10 July 2020. [Google Scholar]
  14. Chiu, J.P.; Nichols, E. Named entity recognition with bidirectional LSTM-CNNs. Trans. Assoc. Comput. Linguist. 2016, 4, 357–370. [Google Scholar] [CrossRef]
  15. Cui, L.; Zhang, Y. Hierarchically refined label attention network for sequence labeling. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), Hong Kong, China, 3–7 November 2019; Association for Computational Linguistics: Stroudsburg, PA, USA, 2019; pp. 4115–4128. [Google Scholar]
  16. Tong, M.; Wang, S.; Xu, B.; Cao, Y.; Liu, M.; Hou, L.; Li, J. Learning from Miscellaneous Other-Class Words for Few-Shot Named Entity Recognition. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing, Virtual, 1–6 August 2021; Volume 1, pp. 6236–6247. [Google Scholar]
  17. Lee, D.H.; Kadakia, A.; Tan, K.; Agarwal, M.; Feng, X.; Shibuya, T.; Mitani, R.; Sekiya, T.; Pujara, J.; Ren, X. Good Examples Make A Faster Learner: Simple Demonstration-based Learning for Low-resource NER. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics, Dublin, Ireland, 22–27 May 2022; Volume 1, pp. 2687–2700. [Google Scholar]
  18. Chen, X.; Li, L.; Deng, S.; Tan, C.; Xu, C.; Huang, F.; Si, L.; Chen, H.; Zhang, N. LightNER: A Lightweight Tuning Paradigm for Low-resource NER via Pluggable Prompting. In Proceedings of the 29th International Conference on Computational Linguistics, Gyeongju, Republic of Korea, 12–17 October 2022; pp. 2374–2387. [Google Scholar]
  19. Wang, L.; Li, R.; Yan, Y.; Yan, Y.; Wang, S.; Wu, W.; Xu, W. Instructionner: A multi-task instruction-based generative framework for few-shot ner. arXiv 2022, arXiv:2203.03903. [Google Scholar]
  20. Chen, J.; Liu, Q.; Lin, H.; Han, X.; Sun, L. Few-shot Named Entity Recognition with Self-describing Networks. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics, Dublin, Ireland, 22–27 May 2022; Volume 1, pp. 5711–5722. [Google Scholar]
  21. Ma, T.; Jiang, H.; Wu, Q.; Zhao, T.; Lin, C.Y. Decomposed Meta-Learning for Few-Shot Named Entity Recognition. In Proceedings of the Findings of the Association for Computational Linguistics: ACL 2022, Dublin, Ireland, 22–27 May 2022; pp. 1584–1596. [Google Scholar]
  22. Das, S.S.S.; Katiyar, A.; Passonneau, R.J.; Zhang, R. CONTaiNER: Few-Shot Named Entity Recognition via Contrastive Learning. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics, Dublin, Ireland, 22–27 May 2022; Volume 1. [Google Scholar]
  23. Wang, S.; Li, X.; Meng, Y.; Zhang, T.; Ouyang, R.; Li, J.; Wang, G. kNN-NER: Named Entity Recognition with Nearest Neighbor Search. arXiv 2022, arXiv:2203.17103. [Google Scholar]
  24. Huang, Z.; Wei, X.; Kai, Y. Bidirectional lstmcrf models for sequence tagging. Computer Science arXiv 2015, arXiv:1508.01991. [Google Scholar]
  25. Peters, M.E.; Neumann, M.; Iyyer, M.; Gardner, M.; Clark, C.; Lee, K.; Zettlemoyer, L. Deep contextualized word representations. In Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, New Orleans, LA, USA, 1–6 June 2018; Association for Computational Linguistics: Stroudsburg, PA, USA, 2018; Volume 1, pp. 2227–2237. [Google Scholar]
  26. Yu, J.; Bohnet, B.; Poesio, M. Named entity recognition as dependency parsing. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, Online, 5–10 July 2020; Association for Computational Linguistics: Stroudsburg, PA, USA, 2020; pp. 6470–6476. [Google Scholar]
  27. Shen, Y.; Ma, X.; Tan, Z.; Zhang, S.; Wang, W.; Lu, W. Locate and label: A two-stage identifier for nested named entity recognition. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics, Virtual, 1–6 August 2021. [Google Scholar]
  28. Yu, D.; He, L.; Zhang, Y.; Du, X.; Pasupat, P.; Li, Q. Few-shot intent classification and slot filling with retrieved examples. In Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Online, 6–11 June 2021; Association for Computational Linguistics: Stroudsburg, PA, USA, 2021; pp. 734–749. [Google Scholar]
  29. Wang, P.; Xu, R.; Liu, T.; Zhou, Q.; Cao, Y.; Chang, B.; Sui, Z. An enhanced span-based decomposition method for few-shot sequence labeling. arXiv 2021, arXiv:2109.13023. [Google Scholar]
  30. Wang, J.; Wang, C.; Tan, C.; Qiu, M.; Huang, S.; Huang, J.; Gao, M. SpanProto: A Two-stage Span-based Prototypical Network for Few-shot Named Entity Recognition. arXiv 2022, arXiv:2210.09049. [Google Scholar]
  31. Luo, Q.; Liu, L.; Lin, Y.; Zhang, W. Don’t miss the labels: Label-semantic augmented meta-learner for few-shot text classification. In Proceedings of the Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021, Online, 1–6 August 2021; pp. 2773–2782. [Google Scholar]
  32. Ma, R.; Zhou, X.; Gui, T.; Tan, Y.; Li, L.; Zhang, Q.; Huang, X. Template-free Prompt Tuning for Few-shot NER. In Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Stroudsburg, PA, USA, 10–15 July 2022; pp. 5721–5732. [Google Scholar]
  33. Zhong, Z.; Chen, D. A Frustratingly Easy Approach for Entity and Relation Extraction. In Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Online, 6–11 June 2021; pp. 50–61. [Google Scholar]
  34. Ye, D.; Lin, Y.; Li, P.; Sun, M. Packed Levitated Marker for Entity and Relation Extraction. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics, Dublin, Ireland, 22–27 May 2022; Volume 1, pp. 4904–4917. [Google Scholar]
  35. Bekoulis, G.; Deleu, J.; Demeester, T.; Develder, C. Joint entity recognition and relation extraction as a multi-head selection problem. Expert Syst. Appl. 2018, 114, 34–45. [Google Scholar] [CrossRef] [Green Version]
  36. Weischedel, R.; Pradhan, S.; Ramshaw, L.; Palmer, M.; Xue, N.; Marcus, M.; Taylor, A.; Greenberg, C.; Hovy, E.; Houston, A.; et al. Ontonotes Release 4.0. LDC2011T03; Linguistic Data Consortium: Philadelphia, PA, USA, 2011. [Google Scholar]
  37. Gina-Anne, L. The third international Chinese language processing bakeoff: Word segmentation and named entity recognition. In Proceedings of the Fifth SIGHAN Workshop on Chinese Language Processing, Sydney, Australia, 22–23 July 2006; pp. 108–117. [Google Scholar]
  38. Xu, L.; Dong, Q.; Yu, C.; Tian, Y.; Liu, W.; Li, L.; Zhang, X. Cluener2020: Fine-grained name entity recognition for Chinese. arXiv 2020, arXiv:2001.04351. [Google Scholar]
  39. Sui, D.; Tian, Z.; Chen, Y.; Liu, K.; Zhao, J. A large-scale Chinese multimodal NER dataset with speech clues. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing, Virtual, 1–6 August 2021; Volume 1. [Google Scholar]
  40. Xia, Y.; Yu, H.; Nishino, F. The Chinese named entity categorization based on the people’s daily corpus. Int. J. Comput. Linguist. Chin. Lang. Process. 2005, 10, 533–542. [Google Scholar]
  41. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  42. Li, X.; Yan, H.; Qiu, X.; Huang, X. Flat: Chinese ner using flat-lattice transformer. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, Online, 5–10 July 2020; pp. 6836–6842. [Google Scholar]
  43. Liu, W.; Fu, X.; Zhang, Y.; Xiao, W. Lexicon enhanced Chinese sequence labelling using bert adapter. arXiv 2021, arXiv:2105.07148. [Google Scholar]
Figure 1. Example of an NER task. The entities to be recognized are highlighted within dashed boxes, and different colors represent different entity types.
Figure 1. Example of an NER task. The entities to be recognized are highlighted within dashed boxes, and different colors represent different entity types.
Applsci 13 08609 g001
Figure 2. Traditional sequence-labeling NER method.
Figure 2. Traditional sequence-labeling NER method.
Applsci 13 08609 g002
Figure 3. Span-based NER method.
Figure 3. Span-based NER method.
Applsci 13 08609 g003
Figure 4. Different entity labels. (a) The label ‘Location’ and its associated entities. (b) The label ‘Organization’ and its associated entities.
Figure 4. Different entity labels. (a) The label ‘Location’ and its associated entities. (b) The label ‘Organization’ and its associated entities.
Applsci 13 08609 g004
Figure 5. The overall structure of SLNER. The grey module on the left learns span representations, while the grey module on the right learns label representations. The model’s final predictions are calculated through distance matching. N s represents the number of spans, and N c represents the number of entity categories.
Figure 5. The overall structure of SLNER. The grey module on the left learns span representations, while the grey module on the right learns label representations. The model’s final predictions are calculated through distance matching. N s represents the number of spans, and N c represents the number of entity categories.
Applsci 13 08609 g005
Figure 6. Different ways of span representation. (a) Simple start- and end-token concatenation as a span representation. (b) Span representation using the biaffine decoder method, which allows for information interaction between the start and end tokens (indicated by blue arrows). (c) Span representation method used in our model, which incorporates the interaction between the tokens within the span (indicated by yellow arrows).
Figure 6. Different ways of span representation. (a) Simple start- and end-token concatenation as a span representation. (b) Span representation using the biaffine decoder method, which allows for information interaction between the start and end tokens (indicated by blue arrows). (c) Span representation method used in our model, which incorporates the interaction between the tokens within the span (indicated by yellow arrows).
Applsci 13 08609 g006
Figure 7. Enhanced span representation.
Figure 7. Enhanced span representation.
Applsci 13 08609 g007
Figure 8. Hierarchy chart of the RISK dataset.
Figure 8. Hierarchy chart of the RISK dataset.
Applsci 13 08609 g008
Figure 9. Statistical chart of the number of various entity types in the RISK dataset.
Figure 9. Statistical chart of the number of various entity types in the RISK dataset.
Applsci 13 08609 g009
Figure 10. Different definitions of label names.
Figure 10. Different definitions of label names.
Applsci 13 08609 g010
Table 1. The full label name format of the four datasets we used.
Table 1. The full label name format of the four datasets we used.
DatasetEntity LabelFull Label Name
OntonotesPER人名 (person name)
ORG组织 (organization)
LOC位置 (location)
GPE地名 (geographic name)
MSRANS位置 (location)
NR姓名、名字 (person name)
NT组织 (organization)
ResumeCONT国籍 (nationality)
EDU教育背景、学历 (educational background)
LOC位置、地名 (location)
NAME姓名、名字 (person name)
ORG组织 (organization)
PRO专业 (profession)
RACE民族 (race)
TITLE职称、职业 (title and occupation)
RISKLOC-PROV省 (province)
LOC-PREF市、区 (city, district)
LOC-COUNT县 (county)
FOOD-VEG蔬菜 (vegetable)
FOOD-FRUIT水果 (fruit)
FOOD-MEAT肉 (meat)
FOOD-GRAIN粮食 (foodstuff)
FOOD-DAIRY奶制品、饮品 (dairy products and beverages)
CO-PROC加工生产方式 (processing and production methods)
CO-PROD公司、厂 (company, factory)
CO-TRAD售卖商店 (sales store)
CO-CATE饭店 (hotel)
CO-MATE超市 (supermarket)
ORG-REGU监督局 (supervisory authority)
ORG-ADMI管理部门 (management department)
ORG-LEGA法律机构 (legal agency)
ORG-SOCI社会组织 (social organization)
RISK-HIGH病毒 (virus)
RISK-MID化学物质、细菌 (chemicals, bacteria)
RISK-LOW添加剂 (additive)
Table 2. Detailed statistical table for the target domain dataset.
Table 2. Detailed statistical table for the target domain dataset.
DatasetTrainDevTestEntity Types
Ontonotes15.7 k4.3 k4.3 k4
MSRA41.7 k4.6 k4.3 k3
Resume3.8 k0.46 k0.48 k8
RISK1.2 k0.26 k0.26 k5 (20) 1
1 5 (20) representing 5 coarse-grained and 20 fine-grained entity types of the RISK dataset.
Table 3. F1 scores for the four Chinese sampling NER datasets. The best results are highlighted in bold.
Table 3. F1 scores for the four Chinese sampling NER datasets. The best results are highlighted in bold.
DatasetMethodK = 250K = 500K = 1000K = 1350
OntonotesBERT63.8569.5071.3372.42
BERT-LC65.6973.5474.9777.19
Lattice LSTM39.7145.4654.5457.48
FLAT49.0146.3549.3457.44
LEBERT69.4869.0173.7874.84
LEBERT-LC70.2669.8973.8376.01
PCBERT74.4275.6278.3381.52
SLNER (ours)77.6680.4681.5382.23
MSRABERT68.4472.2881.2182.28
BERT-LC79.0183.1387.8489.32
Lattice LSTM54.6963.6174.2776.31
FLAT59.6270.2080.7964.95
LEBERT79.1185.1887.7789.35
LEBERT-LC80.9286.0988.1188.70
PCBERT81.0885.2587.8889.72
SLNER (ours)87.6589.3089.6790.08
ResumeBERT53.8062.6469.3670.65
BERT-LC92.2694.6695.1696.41
Lattice LSTM85.6389.6092.0193.13
FLAT84.6290.7792.9787.79
LEBERT89.1592.5694.0295.19
LEBERT-LC91.6093.0395.4095.16
PCBERT93.4294.0194.9695.97
SLNER (ours)94.0294.8695.9996.36
Table 4. The results on the RISK dataset, showing the F1 scores for different sampling sizes in both coarse-grained and fine-grained categories.
Table 4. The results on the RISK dataset, showing the F1 scores for different sampling sizes in both coarse-grained and fine-grained categories.
DatasetK-ShotCoarse-GrainedFine-Grained
RISKK = 25069.6763.97
K = 50071.7365.62
K = 100072.7068.06
Full 172.6268.98
1 Full represents K = 1218 (the maximum size of RISK).
Table 5. Results with fewer samples.
Table 5. Results with fewer samples.
K-ShotDataset
OntonotesMSRAResumeRISK
K = 1065.9073.1877.2359.50 (44.61) 1
K = 2070.9281.1981.7067.57 (45.02)
K = 5074.9186.3490.5568.87 (57.31)
1 Parentheses represent the fine-grained results of the RISK dataset.
Table 6. The results of the ablation experiments. ESR, enhanced span representation; LR, label representation; SD, source domain dataset.
Table 6. The results of the ablation experiments. ESR, enhanced span representation; LR, label representation; SD, source domain dataset.
DatasetMethodK = 250K = 500K = 1000K = 1350
OntonotesSLNER77.6680.4681.5382.23
-ESR w/SD75.6380.0281.0481.91
-LR w/o SD72.9075.6277.9878.09
MSRASLNER87.6589.3089.6790.08
-ESR w/SD87.9989.6089.4289.80
-LR w/o SD80.7284.0684.8784.54
ResumeSLNER94.0294.8695.9996.36
-ESR w/SD93.5294.4695.3895.95
-LR w/o SD87.6494.1795.4895.35
RISKSLNER63.9765.6268.06N/A 1
-ESR w/SD63.2765.2667.29N/A
-LR w/o SD55.6758.9461.23N/A
1 Since the maximum size of the RISK dataset is less than 1350, we did not conduct experiments with K = 1350 for this dataset.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Ren, Z.; Qin, X.; Ran, W. SLNER: Chinese Few-Shot Named Entity Recognition with Enhanced Span and Label Semantics. Appl. Sci. 2023, 13, 8609. https://doi.org/10.3390/app13158609

AMA Style

Ren Z, Qin X, Ran W. SLNER: Chinese Few-Shot Named Entity Recognition with Enhanced Span and Label Semantics. Applied Sciences. 2023; 13(15):8609. https://doi.org/10.3390/app13158609

Chicago/Turabian Style

Ren, Zhe, Xizhong Qin, and Wensheng Ran. 2023. "SLNER: Chinese Few-Shot Named Entity Recognition with Enhanced Span and Label Semantics" Applied Sciences 13, no. 15: 8609. https://doi.org/10.3390/app13158609

APA Style

Ren, Z., Qin, X., & Ran, W. (2023). SLNER: Chinese Few-Shot Named Entity Recognition with Enhanced Span and Label Semantics. Applied Sciences, 13(15), 8609. https://doi.org/10.3390/app13158609

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop