Next Article in Journal
Saliency-Enabled Coding Unit Partitioning and Quantization Control for Versatile Video Coding
Next Article in Special Issue
Automatic Sarcasm Detection: Systematic Literature Review
Previous Article in Journal
Investigation into Phishing Risk Behaviour among Healthcare Staff
Previous Article in Special Issue
Traditional Chinese Medicine Word Representation Model Augmented with Semantic and Grammatical Information
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Systematic Review

Automatic Text Summarization of Biomedical Text Data: A Systematic Review

eVida Research Group, University of Deusto, Avda/Universidades 24, 48007 Bilbao, Spain
*
Author to whom correspondence should be addressed.
Information 2022, 13(8), 393; https://doi.org/10.3390/info13080393
Submission received: 30 June 2022 / Revised: 24 July 2022 / Accepted: 16 August 2022 / Published: 19 August 2022

Abstract

:
In recent years, the evolution of technology has led to an increase in text data obtained from many sources. In the biomedical domain, text information has also evidenced this accelerated growth, and automatic text summarization systems play an essential role in optimizing physicians’ time resources and identifying relevant information. In this paper, we present a systematic review in recent research of text summarization for biomedical textual data, focusing mainly on the methods employed, type of input data text, areas of application, and evaluation metrics used to assess systems. The survey was limited to the period between 1st January 2014 and 15th March 2022. The data collected was obtained from WoS, IEEE, and ACM digital libraries, while the search strategies were developed with the help of experts in NLP techniques and previous systematic reviews. The four phases of a systematic review by PRISMA methodology were conducted, and five summarization factors were determined to assess the studies included: Input, Purpose, Output, Method, and Evaluation metric. Results showed that 3.5 % of 801 studies met the inclusion criteria. Moreover, Single-document, Biomedical Literature, Generic, and Extractive summarization proved to be the most common approaches employed, while techniques based on Machine Learning were performed in 16 studies and Rouge (Recall-Oriented Understudy for Gisting Evaluation) was reported as the evaluation metric in 26 studies. This review found that in recent years, more transformer-based methodologies for summarization purposes have been implemented compared to a previous survey. Additionally, there are still some challenges in text summarization in different domains, especially in the biomedical field in terms of demand for further research.

1. Introduction

The accelerated growth of textual data for academia, research, and industry from diverse sources of information such as news, books, journals, scientific articles, databases, and medical records, among others, has become a significant landmark in the further development of different technologies aimed at extracting the most meaningful information in different application domains [1].
The growth of large amounts of textual information has also become evident in the last few years, particularly in the biomedical field [2,3,4,5]. Health care professionals face reviewing a huge amount of texts on a daily basis [6,7] due to the meaningful knowledge and information this data includes, such as symptoms, treatment, usage of drugs, and reactions, among others [8]. This information may be represented as biomedical literature, clinical notes, or reports in electronic health records (EHRs), while textual data as biomedical literature refers to text data found in databases of scientific articles and EHR includes electronic records of patients’ health information [9].
Natural Language Processing (NLP) techniques have become indispensable mechanisms for several applications in the biomedical domain, as they allow either illnesses based on narrative medical histories or useful information from clinical notes to be identified or even patient phenotypes from available text resources determined [10,11,12,13]. Recently, one of the most relevant applications has been the search for explicit and implicit knowledge in biomedical literature, since these techniques may provide the basis for better understanding of human diseases and enhance the quality of disease diagnosis, prevention, and therapy [14]. Therefore, NLP-based approaches as a purpose of study provide a basis for improvement in further specific applications that fulfill real-world requirements, such as the extraction of significant information from biomedical textual data. Automatic text summarization is one NLP task method that aims to achieve this goal.
The study of different applications in text summarization has been subject to considerable attention in recent years [15,16,17,18,19], although such applications in biomedical textual data tend to prevail more than in other research fields [20]. Some of the main benefits of using these systems concern the time saving and assistance provided to people in the selection of relevant information. Automatic text summarization aims to offer a subset of the source text that highlights the most significant points with the minimal possibility of redundancy [15]. Thus, text summarizing reduces data to make it easier for users to find and process important information more quickly and accurately. These methods have tended to be employed in the biomedical area mainly for summarizing research publications or EHR data [21], although data extraction of clinical trial descriptions has also been included in recent research [22,23,24]. As a result, text summarizing has become a significant tool in helping physicians and researchers organize their information and expertise [21].
In this sense, there is a requirement for working on and exploring existing techniques in the state of the art, mainly for automatic text summarization tasks that focus on problem-solving in biomedical domains where, besides growth in terms of a large amount of existing textual information, professionals also provide crucial information about new health care findings. Thus, to contribute to the development and prevention of health, these studies would in turn contribute to scientific advances that focus on human well-being.
This survey is mainly aimed at identifying the most recent methods, main areas of application, most common evaluation metrics, and types of data, including the most used datasets that are frequently used when text summarization approaches are employed in different biomedical fields. The article is structured as follows. First, we explain how the methods and methodologies were used to develop the systematic review. Then, we provide the results and analysis of the records included. Last, the discussion and conclusion provide recommendations for further research.

2. Materials and Methods

The main goal of our systematic review is to identify current challenges that summarization systems have to deal with in order to process biomedical text information. The knowledge found in this study will be used as a starting point for future research in this area. In Table 1, the research questions and objectives that are addressed throughout this study are shown.
The review methodology was based on the Systematic Review in the biomedical domain performed by Mishra et al. [25] in conjunction with PRISMA guidelines [26]. In the following subsection, we will cover the steps suggested to identify and screen the studies in our systematic review.

2.1. Data Selection

The data collected in this study were obtained from three databases: WoS, IEEE, and ACM digital libraries. The time window was limited from January 1st 2014 to March 15th 2022, with the aim of preventing overlap with past systematic reviews [21,25], while searches mainly focused on text summarization techniques for health care via an engineering approach. The search strategies were developed with the help of experts in Natural Language Processing techniques and the previous systematic review of text summarization in the biomedical domain by Mishra et al. [25]. The screening of records was attached employing the four phases of a systematic review according to PRISMA methodology and are described below.

2.1.1. Identification

The main aim was to gather articles that include text summarization terms that focus on the medical domain, with the medical and technical keywords determined in our search strategy being shown in Table 2. The search did not use any population filter—on the contrary, all fields were explored. 801 records were found in total in accordance with the search strategy and the three databases proposed.

2.1.2. Screening

The screening process involved applying the first filter with the aim of removing duplicate records. 68 texts were repeated in total among three databases and 733 corresponded to the remaining records.

2.1.3. Eligibility

This phase was conducted in order to find original research that included development or evaluation methods in the medical domain capable of summarizing either biomedical literature or Electronic Health Records EHR documents. Therefore, at this point, the text selection process was carried out in two stages. First, a title review was conducted to exclude studies that were not articles written in English, with opinion papers or their contents not being based on summarization techniques in the biomedical domain. Following this step, 641 papers were discarded, leaving 92 in total remaining.
In the second stage, the abstract screening process focused mainly on being in accordance with eligibility criteria.
  • Studies or summarization tools that describe the evaluation component (metric) and method(s) used.
  • Related Natural Language Processing techniques that can be used as text summarization methods (e.g., text mining, text generation).
In accordance with the aforementioned, 55 texts were excluded following the application of eligibility criteria, resulting in 37 remaining articles to be read completely.

2.1.4. Included

According to two previous phases, the inclusion and exclusion criteria for screening records can be defined in Table 3. Following a complete reading of the 37 remaining documents, 28 were then selected for inclusion in the study.
Figure 1 summarizes the whole process of our systematic review in text summarization for the biomedical domain as a flow chart. Here, we can note the different phases suggested by PRISMA methodology.

2.2. Summarization Factors

In order to evaluate and develop summarization systems, Mani and Maybury in [27] and Jones K. [28] determine that most summarization factors are hard to define and difficult to capture in order to guide the summarization process in particular cases. Therefore, to emphasize the range and varieties in terms of summarizing, context factors can mainly be categorized as: input, purpose, and output. Additionally, the method and summary evaluation metrics used are fundamental factors as comparison criteria, since the objectives set out by this review seek to evaluate the most commonly used methods (Q1) and metrics (Q4) when automatic summaries in the biomedical field are generated. As an added factor, the fact of whether an evaluation has been carried out by human experts will also be determined. In Figure 2, the five classifications used for summarization factors are shown and described below.

2.2.1. Input

This category is considered as a unit input parameter of source, which indicates the number of documents to be summarized according to system, either single-document or multiple-document [28,29,30]. Some studies evaluated their proposed summarization systems with single and multiple input texts, hence single and multiple classification was also included [31,32]. In addition, as Mishra et al. in [21] proposed, another input category of documents should be taken into account according to the nature of the biomedical field as Biomedical literature, which refers to text data found in databases of scientific articles or EHR that includes electronic records of patient health-related information (e.g., demographics, progress notes, problems, medication, vital signs, past medical history, immunizations, laboratory data, and radiology reports) [9]. Furthermore, clinical trials were considered as EHR input medical text [33].

2.2.2. Purpose

This criteria concerns the information that a system needs to identify in order to produce a summary. Thus, it is divided into generic, where the summary is created based on all the information in document(s), and query-based, also known as user-oriented, where the summary generated depends on the specific information needed by the user [21,25].

2.2.3. Output

Output factors usually refer to assessment of the information included in summaries generated, whether extracted from the original input (e.g., sentences, paragraphs, phrases) or an abstraction, which generates new text based on the original document with the most salient concepts [15,25]. One of the methodologies also proposes the generation of the summaries based on the two types of output. First, an extractive phase is used to determine the ideas with the highest score according to different ranking steps; then, the summary is composed under the application of an abstractive method [32]. This method was classified as extractive and abstractive output category.

2.2.4. Method

With a view to determining the approaches used by summarization systems, we have classified the methods into three general categories, Mathematical-Statistical, Machine Learning, and Hybrid. Statistical approaches use statistical features of text, usually use sentence ranking based on a mathematical formula that ranks each sentence according to some factors as keywords, term frequency or sentence location, and then the sentence with the best score is selected for inclusion in the summary [21,34]. Therefore, it is said that Statistical techniques are based on the Edmundsonian paradigm [35]. Graph-based models were included as mathematical approaches, providing a text representation that may be used to rank sentences according to significance. Usually, a graph-based summary is based on a single relevance indicator obtained from the centrality of phrases in its graphic representation [15], while the Machine Learning category refers to the methods that need to learn the weight of each indicator from a corpus of text data. These techniques provide a great deal of flexibility because the number of important indicators is almost limitless [15]. The different models based on deep learning techniques such as transformers were also included in this category.

2.2.5. Evaluation Metrics

The evaluation of summarization system performance is categorized as either extrinsic or intrinsic [36]. The extrinsic technique assesses the influence of summarizing on the quality of specific activities that rely on summaries generated, such as success rate, time-to-completion, and decision-making accuracy, among others [21,37]. On the other hand, intrinsic metrics assess summarizing skills using measurements that evaluate the quality of summaries (e.g., informativeness, accuracy, relevance, comprehensiveness, readability, precision, recall, F-measure) [37,38,39]. Since we intend to assess the most common evaluation metrics of summarization systems in the biomedical field (Q4), we will focus on intrinsic evaluation metrics.

3. Results

3.1. Study Frequency According Geographical Distribution, Years, and Type of Publication

The first stage of eligibility according to title resulted in 92 remaining records, which will subsequently be evaluated based on abstract screening. At this point, a frequency study of articles was carried out, the goal being to categorize the number of studies according to geographical distribution, years, and type of publication. The results are shown in Table 4 and Figure 3. As we can see, most text summarization studies in the biomedical domain have been written in Asia, accounting for 57.61 % of the 92 records in the time window selected, with a greater contribution ( 32.61 % ) from the south of the continent (Figure 3a). It is also shown that in the period from 2014 to 2022 (March), research in this field was the subject of attention within the scientific community, and the steady increase in the number of articles published over these years is evident in Figure 3b. Last, in Figure 3c, a greater contribution from articles published in conferences than in journals is shown.

3.2. Study Frequency according to Summarization Factors

Table 5 and Figure 4 show the number and frequency of studies according to the summarization parameters previously mentioned for 28 records that met inclusion criteria. We can see that the main application is biomedical literature and most summarization systems were applied to single document text. Only occasionally was summarization of multidocument in biomedical literature carried out, and never applied to EHR (Figure 4a). The information that systems required to produce summaries is mostly generic, with only 10 % basing their summaries on information provided by the user query-based (Figure 4b). 75 % of the summary system outputs were text extracted from data input, 21 % were an abstraction generated from the input document, and only one system produced an extractive and abstractive output (Figure 4c). Sixteen studies ( 57 % ) generated summaries using Machine Learning approaches and, in conjunction with Mathematical-statistical models ( 8 ; 28 % ) , were the most common summarization methods, with only four ( 14 % ) studies using both approaches (Figure 4d). The performance of 92 % of systems ( 26 ) was evaluated using at least one Rouge metric, with hardly any of the 7 % of studies evaluated having used a different metric (Figure 4e). Last, only some of the studies ( 7 ; 25 % ) incorporated a human evaluation of the summaries generated (Figure 4f).
Finally, Table 6 shows the features and descriptions of the five summarization factors for each record that met the inclusion criteria defined in Table 3. The geographical distribution, year, and type of publication of studies are included.
To determine the most relevant studies in this review, the number of citations in Google Scholar to date was evaluated. This evaluation was developed for each type of method used—Mathematical-Statistical, Machine Learning, and Hybrid—and the type of technique (Output) used is also included. The results are shown in Table 7.

4. Discussion

As previously stated, due to the accelerated growth of biomedical text data from diverse sources of information, the study of summarization techniques has been subject to considerable attention in recent years [15,16,17,18,60]. That is the reason why summarization techniques play a crucial role in this domain and some systematic reviews have been carried out [21,25,55,61]. In this review, we mainly focus on identifying five main parameters in recent summarization systems, which will be discussed below.
Regarding the Input factor (question Q2), both single and multiple document summarization methods have been researched. As the systematic review by Afantenos et al. [25] explains, the notion of using extractive methods mainly for single-document summarization remains. On the other hand, multiple-document summarization is based on abstractive techniques [32,51], and aside from the biomedical field, this is also evident in other fields [62,63,64,65]. Unlike the review conducted by Mishra et al. [21], the study of techniques in single-document summarization has been subject to more attention in the last few years. Moreover, most studies concern summarizing biomedical literature. This may be due to the following factors: first, the exponential growth in recent years of scientific literature published in different databases, and second, the ease of access to this information in contrast to accessibility of patient clinical records EHR, as we can see in Table A1; among the most common resources used by the studies included in this review, we found seven open access datasets for biomedical literature, versus two for EHR. In addition, there is a great advantage in using biomedical literature as input text, since abstracts or titles of records can serve as ground truth summaries [42,44], resulting in PubMed and BioMed being the most common dataset used for NLP tasks in the biomedical environment (Table A1). Most summarization systems in which EHR have been employed as input have tended to base their techniques on machine learning approaches [22,23,40,54,55]. In terms of purpose factors, although most methods deal with generic approaches ( 89 % ) , some (≈11%) require information taken from different user sources such as clinicians, researchers, and patients [32,43,56].
Regarding question Q4, the performance of systems was mostly evaluated using intrinsic and automatic metrics as (Rouge). This is one of the most widely used metrics for summarization tasks and involves counting the number of matching n —grams between ground truth and model summaries [66]. The rouge-based evaluation metrics most commonly used by the studies were Rouge-N, Rouge-L, and Rouge-SU of which the Rouge-L F1 measure was the most widely used, perhaps due to the correlation existing between this and the evaluation carried out by humans [22,66]. Nevertheless, it is possible to use other evaluation metrics based on biomedical expert tools, such as the UMLS Metathesaurus and MeSH, introduced in [42]. Furthermore, depending on how the summarization problem is presented, it is possible to use classification-based metrics, since selecting relevant sentences is represented as a binary classification issue with all sentences in the input divided into summary and nonsummary sentences [15,58,67]. Although Rouge is the most commonly used metric for evaluating summary systems, there is still a gap in determining which one is the most appropriate.
Additionally, there is also a limitation in determining an evaluation that is not based on word matching but on context, since the use of synonyms in the construction of abstracts is one of the main problems when using this type of evaluation metric. Therefore, it is evident that the study of context-based summarization evaluation techniques should be considered. Currently, this problem is starting to be addressed, using metrics based on embeddings. It would be relevant to the incursion of this approach in the biomedical field, since the semantic understanding of clinical text is a challenge in NLP tasks. This because analyzing clinical unstructured text presents a grammatical challenge in itself [8], the material is often illegible due to the few full sentences, the absence of entire phrases, a high usage of acronyms and abbreviations, and the use of confusing terms [68]. Thus, evaluation metrics that focus on these particular cases would be valuable for the development of the NLP technique in this field. Similarly, another weakness that the evaluation metrics face is the type of summarization technique that the systems use is not taken into account, whether abstractive or extractive; although the techniques have the same purpose, a summary generated using methods that extract information from the text will be completely different from one based on abstract methods. For this reason, the question arises as to whether rouge-based evaluations are the most appropriate for evaluating abstract or extractive generated texts or, in general, if they are suitable for evaluating any generated text by summarization systems. This limitation opens the way for future studies along these lines. Last, quality evaluations are carried out with the aim of ensuring a broader understanding of how well the system is working. Unlike automatic evaluation, system output is not compared to the reference; rather, in this case, an expert human reviewer manually scores the summary based on its quality [37]. According to Table 6, human evaluation was carried out by experts such as radiologists [40,55], mental health professionals [43], doctors [47,54], and also human reviewers [22,51].
Concerning Q1, according to Rouge L and Rouge F1-L scores, the highest metrics were obtained for extractive-based systems, although in the case of abstractive approaches, the best metrics were obtained using Transformers-based models as pointer–generator network [40,42] and encoder–decoder–attention [55]. The above may support the previous research conducted in [69], where it was determined that pointer–generator models improve Rouge scores and the fluidity of summaries, especially those based on abstractive techniques. Regarding extractive techniques, the methods that performed positively were AlphaBERT [47] and GPT2 [53], obtaining scores of 75.1 and 78.22 in Rouge L and Rouge F1-L respectively. Once again, transformer-based methodologies obtained the highest scores in evaluation metrics among the studies in which these values were reported. It is worth noting that the report that used a Bayesian summarization method obtained the highest score according to Rouge-1. In general, as has already been demonstrated among other NLP tasks, transformers have been shown to exceed the prediction accuracy of Recurrent Neural Networks (RNNs) and become the industry standard for NLP applications [70]. In particular for biomedical summarization approaches for the period between January 2014 and March 2022, transformers received greater study attention on the part of the scientific community due to their success in performing satisfactorily. With the accelerated growth in the development of methodologies based on transfer learning, recent models have offered the development of novel architecture such as transformers, which are used to handle sequential input data such as natural language. Unlike RNNs, transformers process the full input at once, and any place in the input sequence is given context by the attention mechanism. This allows for greater parallelization than RNNs, resulting in faster training. A transformer can capture the relationships between words regardless of their position, and so the sequential aspect is not important any more. Therefore, these models can help overcome challenges faced by models that deal with natural language such as long text documents, gradient vanishing, larger training steps, and sequential computation, among others [70].
It is worth noting that the Bayesian statistical method that develops an extractive summarization technique had the highest number of citations to date [7], with respect to all the studies included in the review. For methods based on the machine learning approach, the pointer–generator network presented the highest number of citations, developing an abstractive technique [40], and concerning the hybrid approach, the combination of the BERT transformer-based model together with a clustering technique has been the most relevant [45]. Under this criterion, these summarization techniques applied to biomedical text data have greater recognition, which would indicate the importance of their proposed techniques. Therefore, future studies that are interested in determining the method of summarization could begin by investigating these proposals that have shown a good level of approval. In short, these models are the most suitable and prevalent of the summarization techniques used to date.
Last, in relation to question Q3, about 71 % of studies reviewed developed text summarization based on Biomedical Literature, most of them using scientific articles including the title, abstract, and/or corpus of the document from PubMed or Biomed Central datasets (Table A1). Among these studies, the biomedical fields that have been subject to study in the application of automatic summarization techniques in recent years are: nutrition [51], radiation therapy [52], COVID-19 [53,56], cervical cancer [58], and even input texts concerned with health research ethics protocol [57]. Other methods (≈29%) that use EHR as Input in systems have tended to focus on three main categories: general radiology reports [40], including those aimed at breast cancer [55]; clinical trial descriptions [22,23,24]; and general clinical diagnostics [47,54] with applications in Mental Health [43].
In this sense, summarization systems have been facing some challenges not only in the biomedical field but also in different domains that should lead to future research in this area. Text mining scientists face evident challenges, mainly in the biomedical domain. With regard to performance evaluation, 93 % of the studies included used a metric based on Gisting Evaluation (ROUGE), which would indicate that a common metric has been established to evaluate systems. Another issue that summarization systems have been facing is the length of input texts, as we previously discussed—some of the studies included in this review involve transformer-based techniques. Owing to the advantages of using these models, we could say that this problem is being addressed and solutions are emerging, although this issue is not completely resolved. As for the quality of the input information for the purpose of training and testing systems, studies are more inclined to use biomedical literature due to the ease of access to available information and the data structure where, for example, scientific articles provide both text input for the systems and the references as abstracts needed to evaluate their performance. Furthermore, other methods were developed using the title as reference [44]. In contrast, when using clinical records, summarization systems are facing greater challenges that require further and constant study, as they usually evidence many gaps in their phrases, leading to texts with linguistic mistakes, and new terms and acronyms also emerge over time in the biomedical domain [71]. In addition, in the present review, the lack of biomedical text resources in other languages is clearly noticeable. For this reason, future guidelines should focus on the gathering and divulgation of biomedical texts in different languages, since there is also a large amount of valuable information that would also allow the inclusion and evaluation of the different summarization techniques. In short, more research is required to allow and provide publicly accessible summarizing corpora and reference standards in order to assist in the development of summarization technologies in a variety of applications, especially in biomedical text data such as EHR.

5. Conclusions

In this study, we present a systematic review of current literature on medical text summarization. The review found that systems mainly focus on Single Document and Biomedical Literature with a generic purpose and extractive approach. Among the methods used in relation to Machine Learning, Transfer Learning methodologies based on transformers have been subject to increased interest in research in recent years compared to previous surveys, obtaining positive results; additionally, graph-based models as a representation of textual information and statistics-based techniques remain subject to study for the purpose of biomedical text summarization.
Several issues are being addressed, although it is essential to continue researching into possible ways of providing huge improvements, since we can still consider the technology to be a work in progress. Due to the variety of evaluation methodologies and metrics, a meta-analysis still has its limitations, and the absence of common-standard evaluation methodologies could be a sign of the field’s immaturity in comparison to other similar fields in NLP. Defining a favorable method to evaluate the task of summarization is still subject to study; therefore, different methodologies have emerged that, compared to gisting evaluation or n —grams, are based on, for instance, word representations of texts in a vectorial form, whereby words that are close in the vector space are likely to have comparable meanings. This evaluation is known as BERTScore, and the token similarity is estimated [72,73] using contextual embeddings. This methodology is being used to evaluate the performance of different natural language systems, and so would therefore be of interest when exploring the embeddings in biomedical environments with a view to performing a more specific analysis and assessment. In summarization systems, it would be of great interest to use the vectorial word representation in a particular domain in certain biomedical areas, or also with the different types of input text (EHR or biomedical literature).
Other necessary factors and trends that could have been a focus of attention during the review may have been missing, owing to the fact that we focused on five main features, although as we specified before, there are more summarization factors that could provide essential information about system performance. In addition, we consider that some summarization systems may not have been assessed, this being due to the exclusion criteria we used that involved eliminating articles not authored in English. Nevertheless, the studies included served to achieve the proposed objectives in this review.

Author Contributions

Conceptualization, A.C., C.K. and B.G.-Z.; methodology, A.C. and C.K.; formal analysis, A.C.; investigation, A.C.; writing—original draft preparation, A.C.; writing—review and editing, A.C., C.K. and B.G.-Z.; supervision, B.G.-Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

The authors would like to thank Mario Jojoa for his help and insights in the search strategy and methodology of this review. The authors also express their gratitude to the eVida Research Group from the University of Deusto, recognised by the Basque Government in accordance with code IT1536-22, for their commitment in providing us with the resources necessary to complete the study.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

Table A1. Datasets of biomedical resources for summarization tasks.
Table A1. Datasets of biomedical resources for summarization tasks.
ResourceType of Input TextDescription
PubMed Central (PMC)Biomedical LiteratureMore than 7 million full-text records of biomedical and life sciences journal literature at the U.S. National Institutes of Health’s National Library of Medicine (NIH/NLM). Open access [74]
CRAFT: The Colorado Richly Annotated Full Text CorpusBiomedical LiteratureIt is a manually annotated corpus consisting of 67 full-text biomedical journal articles. Each article is a member of the PMC subset. Open access [75,76]
BioASQ Task-6aBiomedical LiteratureContains 13 million citations from PubMed dataset, and each citation contains the title and abstract. Open access [77]
PubMedBiomedical LiteratureContains more than 34 million citations and abstracts supporting the search and retrieval of biomedical and life sciences literature. Open access [78]
BioMed Central (BMC)Biomedical Literature300 peer-reviewed journals in science, technology, engineering, and medicine. Open access [79]
MEDLINEBiomedical LiteratureThis database contains more than 29 million references to journal articles in life sciences with a concentration on biomedicine. The records are indexed with NLM Medical Subject Headings (MeSH). Open access [80,81]
MEDIQA-AnSBiomedical LiteratureThe dataset includes 156 questions with related documents as the answers for each. Each answer also has an extractive and an abstractive single-answer summaries and multidocument extractive and abstractive summary considering the information presented in all of the answers. [82]
CORD-19: The Covid-19 Open Research DatasetBiomedical LiteratureIt is a resource of scientific papers on COVID-19 and related historical coronavirus research. Open access [83]
Radiology ReportsEHR41,066 real-world radiology reports from MedStar Georgetown University Hospital. Each report describes clinical findings about a specific diagnostic case, and an impression summary [40]
DIAC-WoZ datasetEHRClinical interviews designed to support the diagnosis of psychological distress conditions created by the Institute for Creative Technologies at the University of Southern California. Open access [84,85]
NTUH-iMDEHRThe corpus contains 258,050 discharge diagnoses obtained from the National Taiwan University Hospital Integrated Medical Database and the highlighted extractive summaries written by experienced doctors [47]
Clinical trialsEHRDataset generation of 101,016 records usable for the summarization task from clinical trials. Open access [86]

References

  1. Aggarwal, C.C.; Zhai, C. An Introduction to Text Mining. In Mining Text Data; Aggarwal, C.C., Zhai, C., Eds.; Springer: Berlin/Heidelberg, Germany, 2012; pp. 1–10. [Google Scholar] [CrossRef]
  2. Davidoff, F.; Miglus, J. Delivering clinical evidence where it’s needed: Building an information system worthy of the profession. JAMA 2011, 305, 1906–1907. [Google Scholar] [CrossRef] [PubMed]
  3. Smith, R. Strategies for coping with information overload. BMJ 2010, 341, c7126. [Google Scholar] [CrossRef] [PubMed]
  4. Nadif, M.; Role, F. Unsupervised and self-supervised deep learning approaches for biomedical text mining. Briefings Bioinform. 2021, 22, 1592–1603. [Google Scholar] [CrossRef] [PubMed]
  5. Dash, S.; Acharya, B.R.; Mittal, M.; Abraham, A.; Kelemen, A. Deep Learning Techniques for Biomedical and Health Informatics; Springer: Berlin/Heidelberg, Germany, 2020. [Google Scholar]
  6. Mallick, C.; Das, A.; Nayak, J.; Pelusi, D.; Shanmuganathan, V. Evolutionary Algorithm based Ensemble Extractive Summarization for Developing Smart Medical System. Interdiscip. Sci. Comput. Life Sci. 2021, 13, 229–259. [Google Scholar] [CrossRef] [PubMed]
  7. Moradi, M.; Ghadiri, N. Different approaches for identifying important concepts in probabilistic biomedical text summarization. Artif. Intell. Med. 2018, 84, 101–116. [Google Scholar] [CrossRef]
  8. Johnsi, R.; Kumar, G.B.; Sariki, T.P. A Concise Survey on Datasets, Tools and Methods for Biomedical Text Mining. Int. J. Appl. Eng. Res. 2022, 17, 200–217. [Google Scholar] [CrossRef]
  9. Menachemi, N.; Collum, T.H. Benefits and drawbacks of electronic health record systems. Risk Manag. Healthc. Policy 2011, 4, 47. [Google Scholar] [CrossRef]
  10. Buchan, K.; Filannino, M.; Uzuner, Ö. Automatic prediction of coronary artery disease from clinical narratives. J. Biomed. Inform. 2017, 72, 23–32. [Google Scholar] [CrossRef]
  11. Zhou, L.; Baughman, A.W.; Lei, V.J.; Lai, K.H.; Navathe, A.S.; Chang, F.; Sordo, M.; Topaz, M.; Zhong, F.; Murrali, M.; et al. Identifying patients with depression using free-text clinical documents. In MEDINFO 2015: eHealth-Enabled Health; IOS Press: Amsterdam, The Netherlands, 2015; pp. 629–633. [Google Scholar]
  12. Topaz, M.; Lai, K.; Dowding, D.; Lei, V.J.; Zisberg, A.; Bowles, K.H.; Zhou, L. Automated identification of wound information in clinical notes of patients with heart diseases: Developing and validating a natural language processing application. Int. J. Nurs. Stud. 2016, 64, 25–31. [Google Scholar] [CrossRef]
  13. Spasić, I.; Livsey, J.; Keane, J.A.; Nenadić, G. Text mining of cancer-related information: Review of current status and future directions. Int. J. Med. Inform. 2014, 83, 605–623. [Google Scholar] [CrossRef]
  14. Ye, Z.; Tafti, A.P.; He, K.Y.; Wang, K.; He, M.M. SparkText: Biomedical Text Mining on Big Data Framework. PLoS ONE 2016, 11, 1–15. [Google Scholar] [CrossRef] [PubMed]
  15. Nenkova, A.; McKeown, K. A survey of text summarization techniques. In Mining Text Data; Springer: Berlin/Heidelberg, Germany, 2012; pp. 43–76. [Google Scholar]
  16. Widyassari, A.P.; Rustad, S.; Shidik, G.F.; Noersasongko, E.; Syukur, A.; Affandy, A.; Setiadi, D.R.I.M. Review of automatic text summarization techniques & methods. J. King Saud Univ.-Comput. Inf. Sci. 2022, 34, 1029–1046. [Google Scholar] [CrossRef]
  17. Bui, D.D.A.; Del Fiol, G.; Hurdle, J.F.; Jonnalagadda, S. Extractive text summarization system to aid data extraction from full text in systematic review development. J. Biomed. Inform. 2016, 64, 265–272. [Google Scholar] [CrossRef] [PubMed]
  18. Bhatia, N.; Jaiswal, A. Automatic text summarization and it’s methods—A review. In Proceedings of the 2016 6th International Conference—Cloud System and Big Data Engineering (Confluence), Noida, India, 14–15 January 2016; pp. 65–72. [Google Scholar] [CrossRef]
  19. Rahul; Adhikari, S.; Monika. NLP based Machine Learning Approaches for Text Summarization. In Proceedings of the 2020 Fourth International Conference on Computing Methodologies and Communication (ICCMC), Erode, India, 11–13 March 2020; pp. 535–538. [Google Scholar] [CrossRef]
  20. Gong, L. Application of biomedical text mining. Artif. Intell. Emerg. Trends Appl. 2018, 417. [Google Scholar]
  21. Mishra, R.; Bian, J.; Fiszman, M.; Weir, C.R.; Jonnalagadda, S.; Mostafa, J.; Del Fiol, G. Text summarization in the biomedical domain: A systematic review of recent research. J. Biomed. Inform. 2014, 52, 457–467. [Google Scholar] [CrossRef]
  22. Gulden, C.; Kirchner, M.; Schüttler, C.; Hinderer, M.; Kampf, M.; Prokosch, H.U.; Toddenroth, D. Extractive summarization of clinical trial descriptions. Int. J. Med. Inform. 2019, 129, 114–121. [Google Scholar] [CrossRef]
  23. Cintas, C.; Ogallo, W.; Walcott, A.; Remy, S.L.; Akinwande, V.; Osebe, S. Towards neural abstractive clinical trial text summarization with sequence to sequence models. In Proceedings of the 2019 IEEE International Conference on Healthcare Informatics (ICHI), Xi’an, China, 10–13 June 2019; pp. 1–3. [Google Scholar] [CrossRef]
  24. Reddy, S.M.; Miriyala, S. Exploring Multi Feature Optimization for Summarizing Clinical Trial Descriptions. In Proceedings of the 2020 IEEE Sixth International Conference on Multimedia Big Data (BigMM), New Delhi, India, 24–26 September 2020; pp. 341–345. [Google Scholar] [CrossRef]
  25. Afantenos, S.; Karkaletsis, V.; Stamatopoulos, P. Summarization from medical documents: A survey. Artif. Intell. Med. 2005, 33, 157–177. [Google Scholar] [CrossRef]
  26. Liberati, A.; Altman, D.G.; Tetzlaff, J.; Mulrow, C.; Gøtzsche, P.C.; Ioannidis, J.P.; Clarke, M.; Devereaux, P.J.; Kleijnen, J.; Moher, D. The PRISMA statement for reporting systematic reviews and meta-analyses of studies that evaluate health care interventions: Explanation and elaboration. J. Clin. Epidemiol. 2009, 62, e1–e34. [Google Scholar] [CrossRef]
  27. Maybury, M. Advances in Automatic Text Summarization; MIT Press: Cambridge, MA, USA, 1999. [Google Scholar]
  28. Jones, K.S. Automatic summarising: Factors and directions. Adv. Autom. Text Summ. 1999. [Google Scholar] [CrossRef]
  29. Li, L.; Zhou, K.; Xue, G.R.; Zha, H.; Yu, Y. Enhancing Diversity, Coverage and Balance for Summarization through Structure Learning. In Proceedings of the 18th International Conference on World Wide Web, Madrid, Spain, 20–24 April 2009; Association for Computing Machinery: New York, NY, USA, 2009; pp. 71–80. [Google Scholar] [CrossRef]
  30. Ouyang, Y.; Li, W.; Li, S.; Lu, Q. Applying regression models to query-focused multi-document summarization. Inf. Process. Manag. 2011, 47, 227–237. [Google Scholar] [CrossRef]
  31. Moradi, M. CIBS: A biomedical text summarizer using topic-based sentence clustering. J. Biomed. Inform. 2018, 88, 53–61. [Google Scholar] [CrossRef] [PubMed]
  32. Nguyen, Q.A.; Duong, Q.H.; Nguyen, M.Q.; Nguyen, H.S.; Le, H.Q.; Can, D.C.; Thanh, T.D.; Tran, M.V. A Hybrid Multi-answer Summarization Model for the Biomedical Question-Answering System. In Proceedings of the 2021 13th International Conference on Knowledge and Systems Engineering (KSE), Bangkok, Thailand, 10–12 November 2021; pp. 1–6. [Google Scholar] [CrossRef]
  33. Bertagnolli, M.M.; Anderson, B.; Quina, A.; Piantadosi, S. The electronic health record as a clinical trials tool: Opportunities and challenges. Clin. Trials 2020, 17, 237–242. [Google Scholar] [CrossRef] [PubMed]
  34. Munot, N.; Govilkar, S. Comparative Study of Text Summarization Methods. Int. J. Comput. Appl. 2014, 102, 33–37. [Google Scholar] [CrossRef]
  35. Mani, I. Automatic Summarization; John Benjamins Publishing: Amsterdam, The Netherlands, 2001; Volume 3. [Google Scholar]
  36. Jones, K.S.; Galliers, J.R. Evaluating Natural Language Processing Systems: An Analysis and Review; Springer: Berlin/Heidelberg, Germany, 1996. [Google Scholar]
  37. Saziyabegum, S.; Sajja, P. Review on text summarization evaluation methods. Indian J. Comput. Sci. Eng. 2017, 8, 497500. [Google Scholar]
  38. Moradi, M.; Ghadiri, N. Quantifying the informativeness for biomedical literature summarization: An itemset mining method. Comput. Methods Programs Biomed. 2017, 146, 77–89. [Google Scholar] [CrossRef]
  39. Steinberger, J. Evaluation measures for text summarization. Comput. Inform. 2009, 28, 251–275. [Google Scholar]
  40. MacAvaney, S.; Sotudeh, S.; Cohan, A.; Goharian, N.; Talati, I.; Filice, R.W. Ontology-Aware Clinical Abstractive Summarization. In Proceedings of the 42nd International ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR’19), Paris, France, 21–25 July 2019; Association for Computing Machinery: New York, NY, USA, 2019; pp. 1013–1016. [Google Scholar] [CrossRef]
  41. Yongkiatpanich, C.; Wichadakul, D. Extractive Text Summarization Using Ontology and Graph-Based Method. In Proceedings of the 2019 IEEE 4th International Conference on Computer and Communication Systems (ICCCS 2019), Singapore, 23–25 February 2019; pp. 105–110. [Google Scholar]
  42. Gigioli, P.; Sagar, N.; Rao, A.; Voyles, J. Domain-Aware Abstractive Text Summarization for Medical Documents. In Proceedings of the 2018 IEEE International Conference on Bioinformatics and Biomedicine (BIBM), Madrid, Spain, 3–6 December 2018; pp. 2338–2343. [Google Scholar] [CrossRef]
  43. Manas, G.; Aribandi, V.; Kursuncu, U.; Alambo, A.; Shalin, V.L.; Thirunarayan, K.; Beich, J.; Narasimhan, M.; Sheth, A. Knowledge-Infused Abstractive Summarization of Clinical Diagnostic Interviews: Framework Development Study. JMIR Ment. Health 2021, 8, e20865. [Google Scholar] [CrossRef]
  44. Du, Y.; Li, Q.; Wang, L.; He, Y. Biomedical-domain pre-trained language model for extractive summarization. Knowl.-Based Syst. 2020, 199, 105964. [Google Scholar] [CrossRef]
  45. Moradi, M.; Dorffner, G.; Samwald, M. Deep contextualized embeddings for quantifying the informative content in biomedical text summarization. Comput. Methods Programs Biomed. 2020, 184, 105117. [Google Scholar] [CrossRef]
  46. Lee, E.K.; Uppal, K. CERC: An interactive content extraction, recognition, and construction tool for clinical and biomedical text. BMC Med. Inform. Decis. Mak. 2020, 20-S, 306. [Google Scholar] [CrossRef]
  47. Chen, Y.P.; Chen, Y.Y.; Lin, J.J.; Huang, C.H.; Lai, F. Modified Bidirectional Encoder Representations From Transformers Extractive Summarization Model for Hospital Information Systems Based on Character-Level Tokens (AlphaBERT): Development and Performance Evaluation. JMIR Med. Inform. 2020, 8, e17787. [Google Scholar] [CrossRef] [PubMed]
  48. Moradi, M.; Dashti, M.; Samwald, M. Summarization of biomedical articles using domain-specific word embeddings and graph ranking. J. Biomed. Inform. 2020, 107, 103452. [Google Scholar] [CrossRef] [PubMed]
  49. Davoodijam, E.; Ghadiri, N.; Shahreza, M.L.; Rinaldi, F. MultiGBS: A multi-layer graph approach to biomedical summarization. J. Biomed. Inform. 2021, 116, 103706. [Google Scholar] [CrossRef] [PubMed]
  50. Moradi, M. Frequent itemsets as meaningful events in graphs for summarizing biomedical texts. In Proceedings of the 2018 8th International Conference on Computer and Knowledge Engineering (ICCKE), Mashhad, Iran, 25–26 October 2018; pp. 135–140. [Google Scholar]
  51. Shah, D.J.; Yu, L.; Lei, T.; Barzilay, R. Nutri-bullets: Summarizing Health Studies by Composing Segments. In Proceedings of the Thirty-Fifth AAAI Conference on Artificial Intelligence, Thirty-Third Conference on Innovative Applications of Artificial Intelligence and the Eleventh Symposium on Educational Advances in Artificial Intelligence (Assoc Advancement Artificial Intelligence), Online, 22 February–1 March 2021; Volume 35, pp. 13780–13788. [Google Scholar]
  52. Xie, T.; Zhen, Y.; Li, T.; Li, C.; Ge, Y. Self-supervised extractive text summarization for biomedical literatures. In Proceedings of the 2021 IEEE 9th International Conference on Healthcare Informatics (ICHI), Victoria, BC, Canada, 9–12 August 2021; pp. 503–504. [Google Scholar] [CrossRef]
  53. S, D.; N, L.K.; S, S. Extractive Text Summarization for COVID-19 Medical Records. In Proceedings of the 2021 Innovations in Power and Advanced Computing Technologies (i-PACT), Vellore, India, 27–29 November 2021; pp. 1–5. [Google Scholar] [CrossRef]
  54. Vinod, P.; Safar, S.; Mathew, D.; Venugopal, P.; Joly, L.M.; George, J. Fine-tuning the BERTSUMEXT model for Clinical Report Summarization. In Proceedings of the 2020 International Conference for Emerging Technology (INCET), Belgaum, India, 5–7 June 2020; pp. 1–7. [Google Scholar] [CrossRef]
  55. Nguyen, E.; Theodorakopoulos, D.; Pathak, S.; Geerdink, J.; Vijlbrief, O.; van Keulen, M.; Seifert, C. A Hybrid Text Classification and Language Generation Model for Automated Summarization of Dutch Breast Cancer Radiology Reports. In Proceedings of the 2020 IEEE Second International Conference on Cognitive Machine Intelligence (CogMI), Atlanta, GA, USA, 28–31 October 2020; pp. 72–81. [Google Scholar] [CrossRef]
  56. Rai, A.; Sangwan, S.; Goel, T.; Verma, I.; Dey, L. Query Specific Focused Summarization of Biomedical Journal Articles. In Proceedings of the 2021 16th Conference on Computer Science and Intelligence Systems (FedCSIS), Online, 2–5 September 2021; pp. 91–100. [Google Scholar] [CrossRef]
  57. Purbawa, D.P.; Malikhah; Esti Anggraini, R.N.; Sarno, R. Automatic Text Summarization using Maximum Marginal Relevance for Health Ethics Protocol Document in Bahasa. In Proceedings of the 2021 13th International Conference on Information Communication Technology and System (ICTS), Surabaya, Indonesia, 20–21 October 2021; pp. 324–329. [Google Scholar] [CrossRef]
  58. Sibunruang, C.; Polpinij, J. Finding Clinical Knowledge from MEDLINE Abstracts by Text Summarization Technique. In Proceedings of the 2018 International Conference on Information Technology (InCIT), Khon Kaen, Thailand, 24–25 October 2018; pp. 1–6. [Google Scholar] [CrossRef]
  59. Rouane, O.; Belhadef, H.; Bouakkaz, M. Combine clustering and frequent itemsets mining to enhance biomedical text summarization. Expert Syst. Appl. 2019, 135, 362–373. [Google Scholar] [CrossRef]
  60. Allahyari, M.; Pouriyeh, S.; Assefi, M.; Safaei, S.; Trippe, E.D.; Gutierrez, J.B.; Kochut, K. Text summarization techniques: A brief survey. arXiv 2017, arXiv:1707.02268. [Google Scholar] [CrossRef]
  61. Wang, M.; Wang, M.; Yu, F.; Yang, Y.; Walker, J.; Mostafa, J. A systematic review of automatic text summarization for biomedical literature and EHRs. J. Am. Med. Inform. Assoc. 2021, 28, 2287–2297. [Google Scholar] [CrossRef]
  62. Li, W. Abstractive multi-document summarization with semantic information extraction. In Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing, Lisbon, Portugal, 17–21 September 2015; pp. 1908–1913. [Google Scholar]
  63. Chu, E.; Liu, P. Meansum: A neural model for unsupervised multi-document abstractive summarization. In Proceedings of the International Conference on Machine Learning, PMLR, Long Beach, CA, USA, 9–15 June 2019; pp. 1223–1232. [Google Scholar]
  64. Banerjee, S.; Mitra, P.; Sugiyama, K. Multi-document abstractive summarization using ilp based multi-sentence compression. In Proceedings of the Twenty-Fourth International Joint Conference on Artificial Intelligence, Buenos Aires, Argentina, 25–31 July 2015. [Google Scholar]
  65. Pasunuru, R.; Celikyilmaz, A.; Galley, M.; Xiong, C.; Zhang, Y.; Bansal, M.; Gao, J. Data augmentation for abstractive query-focused multi-document summarization. In Proceedings of the AAAI Conference on Artificial Intelligence (AAAI 2021), Online, 2–9 February 2021; pp. 13666–13674. [Google Scholar]
  66. Lin, C.Y. ROUGE: A Package for Automatic Evaluation of Summaries. In Text Summarization Branches Out; Association for Computational Linguistics: Barcelona, Spain, 2004; pp. 74–81. [Google Scholar]
  67. Amer, E.; Fouad, K.M. Keyphrase Extraction methodology from short abstracts of Medical Documents. In Proceedings of the 8th Cairo International Biomedical Engineering Conference (CIBEC), Cairo, Egypt, 15–17 December 2016; pp. 23–26. [Google Scholar]
  68. Olaronke, I.; Olaleke, J. A Systematic Review of Natural Language Processing in Healthcare. Int. J. Inf. Technol. Comput. Sci. 2015, 08, 44–50. [Google Scholar] [CrossRef]
  69. Deaton, J. Transformers and Pointer-Generator Networks for Abstractive Summarization. 2019. Available online: https://www.semanticscholar.org/paper/Transformers-and-Pointer-Generator-Networks-for-Deaton/46adc063c1c46e02f6457e45503cbb65495f6494 (accessed on 29 June 2022).
  70. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.; Polosukhin, I. Attention Is All You Need. In Proceedings of the Advances in Neural Information Processing Systems 30 (NIPS 2017), Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
  71. Gambhir, M.; Gupta, V. Recent Automatic Text Summarization Techniques: A Survey. Artif. Intell. Rev. 2017, 47, 1–66. [Google Scholar] [CrossRef]
  72. Zhang, T.; Kishore, V.; Wu, F.; Weinberger, K.Q.; Artzi, Y. BERTScore: Evaluating Text Generation with BERT. arXiv 2019, arXiv:1904.09675. [Google Scholar]
  73. Jurafsky, D.; Martin, J.H. Speech and Language Processing: An Introduction to Natural Language Processing, Computational Linguistics, and Speech Recognition; Pearson/Prentice Hall: Hoboken, NJ, USA, 2009. [Google Scholar]
  74. About PMC. Available online: https://www.ncbi.nlm.nih.gov/pmc/about/intro/ (accessed on 19 July 2022).
  75. Bada, M.; Eckert, M.; Evans, D.; Garcia, K.; Shipley, K.; Sitnikov, D.; Baumgartner, W., Jr.; Cohen, K.; Verspoor, K.; Blake, J.; et al. Concept annotation in the CRAFT corpus. BMC Bioinform. 2012, 13, 161. [Google Scholar] [CrossRef]
  76. Craft: The Colorado Richly Annotated Full Text Corpus. Available online: http://bionlp-corpora.sourceforge.net/CRAFT/ (accessed on 19 July 2022).
  77. Tsatsaronis, G.; Balikas, G.; Malakasiotis, P.; Partalas, I.; Zschunke, M.; Alvers, M.R.; Weissenborn, D.; Krithara, A.; Petridis, S.; Polychronopoulos, D.; et al. An overview of the BIOASQ large-scale biomedical semantic indexing and question answering competition. BMC Bioinform. 2015, 16, 138. [Google Scholar] [CrossRef] [PubMed]
  78. PubMed.gov. Available online: https://pubmed.ncbi.nlm.nih.gov/ (accessed on 19 July 2022).
  79. BioMed Central. Available online: https://www.biomedcentral.com/ (accessed on 19 July 2022).
  80. MEDLINE. Available online: https://www.nlm.nih.gov/medline/index.html (accessed on 19 July 2022).
  81. Download MEDLINE/PubMed Data. Available online: https://www.nlm.nih.gov/databases/download/pubmed_medline.html (accessed on 19 July 2022).
  82. Savery, M.; Ben Abacha, A.; Gayen, S.; Demner-Fushman, D. Question-Driven Summarization of Answers to Consumer Health Questions. Sci. Data 2020, 7, 1–9. [Google Scholar] [CrossRef] [PubMed]
  83. Wang, L.L.; Lo, K.; Chandrasekhar, Y.; Reas, R.; Yang, J.; Burdick, D.; Eide, D.; Funk, K.; Katsis, Y.; Kinney, R.M.; et al. CORD-19: The COVID-19 Open Research Dataset. In Proceedings of the 1st Workshop on NLP for COVID-19 at ACL 2020, Online; Association for Computational Linguistics: Stroudsburg, PA, USA, 2020. [Google Scholar]
  84. DAIC-WOZ Database & Extended DAIC Database. Available online: https://dcapswoz.ict.usc.edu/ (accessed on 19 July 2022).
  85. Gratch, J.; Artstein, R.; Lucas, G.; Stratou, G.; Scherer, S.; Nazarian, A.; Wood, R.; Boberg, J.; DeVault, D.; Marsella, S.; et al. The Distress Analysis Interview Corpus of Human and Computer Interviews; Technical Report; University of Southern California Los Angeles: Los Angeles, CA, USA, 2014. [Google Scholar]
  86. ClinicalTrials.gov. Available online: https://clinicaltrials.gov/ (accessed on 19 July 2022).
Figure 1. Flow chart of systematic review for text summarization in the biomedical domain.
Figure 1. Flow chart of systematic review for text summarization in the biomedical domain.
Information 13 00393 g001
Figure 2. Classification of summarization factors.
Figure 2. Classification of summarization factors.
Information 13 00393 g002
Figure 3. Number of studies following first stage of eligibility criteria. (a) Geographical distribution. (b) Year of publication. (c) Type of publication.
Figure 3. Number of studies following first stage of eligibility criteria. (a) Geographical distribution. (b) Year of publication. (c) Type of publication.
Information 13 00393 g003
Figure 4. Classification of studies following second stage of eligibility criteria according to summarization factors. (a) Input. (b) Purpose. (c) Output. (d) Method. (e) Evaluation Metric. (f) Human Evaluation.
Figure 4. Classification of studies following second stage of eligibility criteria according to summarization factors. (a) Input. (b) Purpose. (c) Output. (d) Method. (e) Evaluation Metric. (f) Human Evaluation.
Information 13 00393 g004
Table 1. Research questions.
Table 1. Research questions.
QuestionPurpose
Q1What are the most prevalent methods used for text summarization in the biomedical domain?To determine which techniques have been applied in text summarization in the biomedical domain.
Q2What data types are used in text summarization in the biomedical domain?To identify which types of text are most common, either single or multiple document. This will also allow us to assess the most frequently used application in biomedical literature or EHR.
Q3Which areas in the biomedical field have applied text summarization techniques?To find out which medical areas have implemented summarization methods.
Q4What are the most common evaluation metrics of text summarization in the biomedical field?To assess and identify suitable evaluation metrics to use when comparative studies are carried out on text summarization, mainly in the field of health care.
Table 2. Search strategy.
Table 2. Search strategy.
Medical KeywordsTechnical KeywordsSearch Strategy
Biomedical OR biomedicine OR medical OR medicine OR healthcare OR health OR “patient care” OR clinical OR disease OR diseases OR therapy OR therapies OR treatment OR treatment OR diagnosis OR diagnoses OR diagnostic OR etiology“Text summarization” OR “abstractive
summarization” OR “extractive
summarization” OR “abstractive text
summarization” OR “extractive text
summarization” OR “single document
summarization” OR “multi-document
summarization” OR “query-based
summarization” OR “generic
summarization” OR “hugging face”
All fields (Medical Keywords) AND All fields (Technical Keywords) AND (1 January 2014: 15 March 2022)
Table 3. Inclusion and exclusion criteria.
Table 3. Inclusion and exclusion criteria.
Criteria
InclusionExclusion
Complete records.
Studies published in journals or at conferences, where the words obtained from the search strategy appear in the title and abstract.
Studies that describe the evaluation component (metric) and method(s) used.
Studies not written in English, editorials, or opinion papers.
Studies based on summarization techniques in fields other than the biomedical domain.
Unavailable records.
Table 4. Number and frequency of studies categorized according to geographical distribution, year, and type of publication following first eligibility stage.
Table 4. Number and frequency of studies categorized according to geographical distribution, year, and type of publication following first eligibility stage.
ParametersCategoryFrequency
No. Studies%
LocationEastern Africa11.09%
Northern Africa33.26%
Africa44.35%
Eastern Asia2223.91%
Southern Asia3032.61%
Western Asia11.09%
Asia5357.61%
Northern Europe11.09%
Southern Europe44.35%
Western Europe99.78%
Europe1415.22%
North America1617.39%
South America33.26%
America1920.65%
Australia/Oceania22.17%
Year201433.26%
201566.52%
201655.43%
201755.43%
20181213.04%
20191516.30%
20202021.74%
20212223.91%
202244.35%
Type of publicationConference4953.26%
Journal4346.74%
Table 5. Number and frequency of studies categorized according to summarization factors following second eligibility stage.
Table 5. Number and frequency of studies categorized according to summarization factors following second eligibility stage.
ParametersCategoryFrequency
No. Studies%
InputSingle-document (SD)2589.29%
Multiple-document (MD)13.57%
Single-multiple-document (SMD)27.14%
Biomedical literature (BL)2071.43%
EHR (EHR)828.57%
PurposeQuery-based (QB)310.71%
Generic (Ge)2589.29%
OutputExtractive (Ex)2175.00%
Abstractive (Ab)621.43%
Extractive and abstractive (EA)13.57%
MethodMathematical/Statistical (M/S)828.57%
Machine Learning (ML)1657.14%
Hybrid (Hy)414.29%
Evaluation MetricRouge (Rg)2485.71%
Rouge and others (R/O)27.14%
Other (O)27.14%
Human EvaluationHuman evaluation (HE)725.00%
No human evaluation (NHE)2175.00%
Table 6. Studies included and their characteristics according to summarization factors.
Table 6. Studies included and their characteristics according to summarization factors.
TitleC/JLoc.YearInputPurposeOutMethod (Best)Metric (Best)H. Evaluation
Ontology-Aware Clinical Abstractive Summarization [40]CUSA2019SD, EHR: Radiology ReportsGeAbML: pointer–generator based on Seq2Seq modelRg 1:38.42 2:23.29 L:37.02HE: Radiologist (Readability, Accuracy, Completeness)
Extractive Text Summarization using Ontology and Graph-Based Method [41]CSingapore2019SD, BL: Review papersGeExM/S: Graph-based method (PageRank)Rg-P 1:25.46 L:23.61NHE
Domain-Aware Abstractive Text Summarization for Medical Documents [42]CSpain2019SD, BL: abstracts from PubMed datasetGeAbML: deep-reinforced pointer–generator networkR/O- 1:42.43 2:21.59 L:36.89 TFIDF UMLS MeSHNHE
Knowledge-Infused Abstractive Summarization of Clinical Diagnostic Interviews: Framework Development Study [43]JUSA2021SD, EHR: Diagnostic interviews by mental health professionalsQB: clinical diagnostic interviewsAbM/S: knowledge-infused abstractive summarization (KiAS)Rg-L R:24.46 F1:32.57HE: Mental Health professionals (GQCC, GQUC, meaningful responses)
Extractive summarization of clinical trial descriptions [22]JGermany2019SD, EHR: clinical trial descriptions from clinicaltrials.govGeExML: TextRankRg-L P:30.95 R:33.86 F1:30.03HE: Human reviewers (Contains all information, Helpfulness)
Biomedical-domain pretrained language model for extractive summarization [44]JChina2020SD, BL: titles and abstracts from PubMed dataset (Task 6a)GeExML: domain-aware bidirectional language model (BioBERTSum)Rg-F1 1:37.45 2:17.59 L:29.58NHE
Deep contextualized embeddings for quantifying the informative content in biomedical text summarization [45]JAustria2020SD, BL: articles from BioMed Central databaseGeExHy: deep bidirectional language model and clustering method (BERT-based, BERT-large)Rg 1:75.04 2:33.12NHE
CERC: an interactive content extraction, recognition, and construction tool for clinical and biomedical text [46]JEngland2020SD, BL: abstracts from MedlineGeExML: multistage algorithm (MINTS)Rg 1:41.4 2:13.6 SU4:17.1NHE
Evolutionary Algorithm based Ensemble Extractive Summarization for Developing Smart Medical System [6]JIndia2021SD, BL: PubMed and MEDLINE journal citationsGeExHy: Multiobjective Evolutionary Algorithm based on Decomposition (MOEAD)Rg-F1 1:70.7 2:65.5 SU:47.9NHE
Different approaches for identifying important concepts in probabilistic biomedical text summarization [7]JIran2018SD, BL: Biomedical articlesGeExM/S: Bayesian methodRg 1:78.86 2:35.29 SU4:41.04NHE
CIBS: A biomedical text summarizer using topic-based sentence clustering [31]JIran2018SMD, BL: abstracts from PubMed and BioMedGeExML: Clustering and Itemset mining (CIBs)Rg 2:34.75 SU4:39.78NHE
Modified Bidirectional Encoder Representations From Transformers Extractive Summarization Model for Hospital Information Systems Based on Character-Level Tokens (AlphaBERT): Development and Performance Evaluation [47]JTaiwan2020SD, EHR: diagnoses from National Taiwan University HospitalGeExML: BERT-based structure with a two-stage training method (AlphaBERT)Rg 1:76.9 2:61.0 L:75.1HE: Doctor feedback (Score)
Summarization of biomedical articles using domain-specific word embeddings and graph ranking [48]JAustria2020SD, BL: articles from PubMedGeExHy: domain-specific word embedding and graph-based modelRg 1:76.87 2:34.91NHE
MultiGBS: A multilayer graph approach to biomedical summarization [49]JIran2021SD, BL: articles from BioMed CentralGeExM/S: graph-based creation and sentence selection model (MultiGBS)Rg/O Rg-F1 1:16.4 2:05.2 L:14.6 SU4:07.5 Bertscore F1:80.6NHE
Quantifying the informativeness for biomedical literature summarization: An itemset mining method [38]JIran2017SD, BL: Scientific papersGeExM/S: Itemset miningRg 1:75.83 2:33.81 SU4:38.89NHE
Frequent itemsets as meaningful events in graphs for summarizing biomedical texts [50]CIran2018SD, BL: scientific articlesGeExM/S: Graph-based methodRg 2:34.03 SU4:38.51NHE
Nutri-bullets: Summarizing Health Studies by Composing Segments [51]CUSA2021MD, BL: scientific abstracts from PubMed and ScienceDirectGeAbML: reinforcement learning (Blank Language Model—BLM)O-Meteor Me:15.0HE: (Faithfulness, Relevance, Fluency)
Self-supervised extractive text summarization for biomedical literature [52]CUSA2021SD, BL: Radiation Therapy scientific articles from PubMedGeExML: BERTRg-R 1:71.00 2:59.00NHE
A Hybrid Multianswer Summarization Model for the Biomedical Question-Answering System [32]CVietnam2021SMD, BL: Medical Question-Answer Summarization dataset (MEDIQA-AnS)QB: Question-driven filtering phaseEAML: Denoising autoencoder and BART (Extractive Abstractive hybrid model - EAHS)Rg-F1 1:30.00 2:22.00 L:25.00NHE
Towards neural abstractive clinical trial text summarization with sequence to sequence models [23]CKenya2019SD, EHR: clinical trial descriptions from clinical trials.govGeAbML: Seq2Seq model with attentionRg-F1 1:40.4 2:15.0 L:33.8NHE
Extractive Text Summarization for COVID-19 Medical Records [53]CIndia2021SD, BL: COVID-19 research articles from PubMed, Microsoft Academic and WHO COVID-19GeExML: Generative Pre-Trained Transformer 2 (GPT-2)Rg-F1 1:78.22 2:71.17 L:78.22NHE
Fine-tuning the BERTSUMEXT model for Clinical Report Summarization [54]CIndia2020SD, EHR: clinical report summarization datasetGeExML: Fine-tuned BERTSUMTEXTRg-F1 1:50.07 2:39.85 L:49.59HE: Doctor’s opinion
A Hybrid Text Classification and Language Generation Model for Automated Summarization of Dutch Breast Cancer Radiology Reports [55]CNetherlands2020SD, EHR: Dutch breast cancer radiology reportsGeAbML: encoder–decoder attention model (EDA)Rg-F1 1:54.0 2:38.8 L:51.5HE: Radiologists (correctness, relevance, comprehensible)
Query Specific Focused Summarization of Biomedical Journal Articles [56]CIndia2021SD, BL: articles from COVID-19 Open Research Dataset (CORD-19)QB: User required informationExM/S: Optimization and contextual methodRg 1:47.61 2:19.62 L:44.74NHE
Exploring Multi-Feature Optimization for Summarizing Clinical Trial Descriptions [24]CIndia2020SD, EHR: Clinical Trial Descriptions from Mendeley datasetsGeExM/S: Multi Feature Optimization (MFO)Rg-R 1:70.0 2:39.0 L:50.0NHE
Automatic Text Summarization using Maximum Marginal Relevance for Health Ethics Protocol Document in Bahasa [57]CIndonesia2021SD, BL: Health research ethics protocolGeExM/S: Maximum Marginal Relevance (MMR)Rg-4 P:34.0 R:71.0 F1:46.0NHE
Finding Clinical Knowledge from MEDLINE Abstracts by Text Summarization Technique [58]CThailand2018SD, BL: cervical cancer in clinical trials from MEDLINE abstractsGeExML: BM25 term-weighting and text filtering techniquesO P:100.0 R:84.0 F1:91.0NHE
Combining clustering and frequent item set mining to enhance biomedical text summarization [59]JUSA2019SD, BL: articles from BioMed central databaseGeExHy: clustering and frequent itemset meaningRg 1:23.84 2:08.71 SU4:11.45NHE
Table 7. Most cited algorithms for each method.
Table 7. Most cited algorithms for each method.
MethodAlgorithmOutput
Mathematical-StatisticalBayesian [7]Extractive
Machine LearningPointer–generator network [40]Abstractive
HybridDeep bidirectional language model and clustering (BERT-based, BERT-large) [45]Extractive
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Chaves, A.; Kesiku, C.; Garcia-Zapirain, B. Automatic Text Summarization of Biomedical Text Data: A Systematic Review. Information 2022, 13, 393. https://doi.org/10.3390/info13080393

AMA Style

Chaves A, Kesiku C, Garcia-Zapirain B. Automatic Text Summarization of Biomedical Text Data: A Systematic Review. Information. 2022; 13(8):393. https://doi.org/10.3390/info13080393

Chicago/Turabian Style

Chaves, Andrea, Cyrille Kesiku, and Begonya Garcia-Zapirain. 2022. "Automatic Text Summarization of Biomedical Text Data: A Systematic Review" Information 13, no. 8: 393. https://doi.org/10.3390/info13080393

APA Style

Chaves, A., Kesiku, C., & Garcia-Zapirain, B. (2022). Automatic Text Summarization of Biomedical Text Data: A Systematic Review. Information, 13(8), 393. https://doi.org/10.3390/info13080393

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop