Improved Graph-Based Arabic Hotel Review Summarization Using Polarity Classification
Abstract
:1. Introduction
- -
- Can the Arabic automatic review summarization improve by considering the reviews’ sentiment factor?
2. Background
- The number of documents simultaneously processed—single- and multi-document [9].
- The type of retrieved sentences, extractive abstractive, and hybrid summarization [5]. Extractive summarization selectively includes in the summary some sentences from the original text without modifying the sentence structure, while abstractive summarization maintains the text meaning without adhering to its structure.
- General-based and query-based summary; the general-based summary retrieves the sentence regardless of any consideration for question or relations with the title, whereas the query-based summary is returned due to the summary and questions between the title and sentence [2].
- PageRank: PageRank is the backbone of the infamous search engine Google, created by Brin and Page in 1998. Google revolutionized web search by introducing PageRank, which finds web pages not only by considering keywords and indexing but also according to the page importance compared to other web pages. In this algorithm, the web pages are represented by nodes, and edges represent the links between pages. A page with a high rank is a page with a high in-degree value, i.e., many other pages are pointing to it. Additionally, this is done recursively, so if the edge comes from a significant page, it will have extra importance.
- TextRank: PageRank is the heart of TextRank. TextRank is a graph-based ranking model developed by Mihalcea and Tarau [15] for extracting important keywords and sentences from the text. TextRank is considered an extractive text summarization method. It generates a fully connected, undirected graph of sentences or words of a single document. In this graph, each sentence within the document is represented by a node. In contrast, the similarity between sentences measured as a function of joint concepts is defined by an edge. A weight value marks the importance of each edge. The sentences are ordered based on their scores, thus, creating a list of ranked sentences. For text summarization, usually, the top-ranked sentences are the most representative.
- LexRank: a graph-based extractive text summarization technique [14], similar to TextRank, where a node represents each sentence; however, the edge weights are defined by computing the cosine similarities. LexRank uses graph centrality measures such as degree and eigenvector centrality to find important sentences. During the summarization process, the relevance between the sentences is considered by calculating their importance for neighboring sentences. The positive contribution enhances the importance value of the neighbor sentence, and the negative contribution minimizes the importance value of neighbor sentences. Then, the sentences are ordered based on rank [16].
3. Related Work
4. Methodology
4.1. Data Collection
4.2. Data Preprocessing
Before Preprocessing | After Preprocessing | Translation |
---|---|---|
Fun and beautiful resort. All the staff did a great job in the service and dealt with politeness and respect. We will always get back to you from Riyadh to the resort. You get 10 out of 10 |
4.3. Polarity Classification
- Classifying is based on the original users’ rating, which includes ratings from 1 (lowest rating) to 5 (highest rating).
- Classifying based on users’ positive, negative, or neutral ratings. To perform this approach, the users’ ratings of 5 and 4 were considered positive, three as neutral, and 2 and 1 as negative reviews.
- Classifying based on sentiment analysis of all the reviews in the dataset, i.e., 409,312 reviews for the 1173 hotels. This classification was also based on the users’ original ratings; however, it considers the linguistic features and patterns modeled via the training of deep learning architecture. This procedure is illustrated in Algorithm 1 below. The deep learning architecture is discussed in the following subsection.
- This approach is similar to the previous approach and is based on sentiment analysis of all reviews in the dataset using deep learning modeling. However, the classification model classifies the reviews into positive, negative, and neutral using the same method described above for the second approach.
- Instead of using the rating score for sentiment in this experiment, we relabeled each review automatically using a pretrained model Mazajak [37], a deep learning model trained on Arabic Twitter data using CNN-LSTM models, more about Mazajak is presented below. This API labeled each review as positive, negative, or neutral.
Review | Translation | Rating Based on the Proposed Deep Learning Model | Original Rating |
---|---|---|---|
Worth the experience, great place, everything was great | 5 | 2 | |
Slow services | 3 | 4 | |
Wonderful, I highly recommend it, comfort, calm, and luxury | 5 | 4 | |
Very nice hotel, late breakfast | 4 | 5 | |
Unacceptable, the room was terrible, less than three stars | 2 | 3 | |
Good, breakfast was excellent | 4 | 3 |
Hotel | Total Reviews | Disagreements | Percentage | Rating Distribution |
---|---|---|---|---|
1 | 224 | 42 | 18.7% | |
2 | 224 | 12 | 5.35% |
Algorithm 1: The proposed polarity classification based on sentiment analysis. |
Procedure: Classifying reviews based on sentiment analysis |
Input: |
: a collection of labeled review |
Output: |
: a collection of updated labeled reviews , |
Construct the review polarity classification model: |
Add embedding layer |
Add Bidirectional layer |
Add dense layer |
Add dropout layer # to avoid overfitting |
Add softmax layer to include all possible class labels |
Use categorical_crossentropy loss function |
Train the the review polarity classification model |
validation_split = 0.1, |
number of epochs = 5, |
batch_size= 128 |
Plot Accuracy (training and validation) learning curves |
Finetune the model |
Evaluate and test the model |
Classify reviews and update |
4.3.1. Deep Learning for Polarity Classification
4.3.2. Mazajak API for Polarity Classification
4.4. TextRank Summarization
4.5. Evaluation
5. Experimental Results
6. Discussions
7. Conclusions and Future Work
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
References
- Nayar, R. Role of Web 3.0 in Service Innovation. In The Handbook of Service Innovation; Springer: Berlin/Heidelberg, Germany, 2015. [Google Scholar]
- Alami, N.; En-Nahnahi, N.; Ouatik, S.A.; Meknassi, M. Using Unsupervised Deep Learning for Automatic Summarization of Arabic Documents. Arab. J. Sci. Eng. 2018, 43, 7803–7815. [Google Scholar] [CrossRef]
- Chen, Y.; Chang, C.; Gan, J. A template approach for summarizing restaurant reviews. IEEE Access 2021, 9, 115548–115562. [Google Scholar] [CrossRef]
- Marzijarani, S.B.; Sajedi, H. Opinion mining with reviews summarization based on clustering. Int. J. Inf. Technol. 2020, 12, 1299–1310. [Google Scholar] [CrossRef]
- Elsaid, A.; Mohammed, A.; Ibrahim, L.F.; Sakre, M.M. A Comprehensive Review of Arabic Text Summarization. IEEE Access 2022, 10, 38012–38030. [Google Scholar] [CrossRef]
- Amoudi, G.; Albalawi, R.; Baothman, F.; Jamal, A.; Alghamdi, H.; Alhothali, A. Arabic rumor detection: A comparative study. Alex. Eng. J. 2022, 61, 12511–12523. [Google Scholar] [CrossRef]
- Elbarougy, R.; Behery, G.; El Khatib, A. Extractive Arabic Text Summarization Using Modified PageRank Algorithm. Egypt. Inform. J. 2020, 21, 73–81. [Google Scholar] [CrossRef]
- Suhara, Y.; Wang, X.; Angelidis, S.; Tan, W.-C. OpinionDigest: A Simple Framework for Opinion Summarization; Association for Computational Linguistics: Stroudsburg, PA, USA, 2020; pp. 5789–5798. [Google Scholar]
- Haque, M.M.; Pervin, S.; Begum, Z. Literature Review of Automatic Multiple Documents Text Summarization. Int. J. Innov. Appl. Stud. 2013, 3, 121–129. [Google Scholar]
- Etaiwi, W.; Awajan, A. Graph-based Arabic NLP Techniques: A Survey. Procedia Comput. Sci. 2018, 142, 328–333. [Google Scholar] [CrossRef]
- Alami, N.; Meknassi, M.; Ouatik, S.A.; Ennahnahi, N. Arabic text summarization based on graph theory. In Proceedings of the 2015 IEEE/ACS 12th International Conference of Computer Systems and Applications (AICCSA), Marrakech, Morocco, 17–20 November 2015; pp. 1–8. [Google Scholar]
- Ibrahim, M.N.; Maria, K.A.; Jaber, K.M. Summarization Systems (AMD-SS). In Proceedings of the 2017 8th International Conference on Information Technology (ICIT), Amman, Jordan, 17 May 2017; pp. 1013–1022. [Google Scholar]
- Varade, S.; Sayyed, E.; Nagtode, V.; Shinde, S. Text Summarization using Extractive and Abstractive Methods. ITM Web Conf. 2021, 40, 03023. [Google Scholar] [CrossRef]
- Erkan, G.; Radev, D.R. LexRank: Graph-based Lexical Centrality as Salience in Text Summarization. J. Artif. Intell. Res. 2004, 22, 457–479. [Google Scholar] [CrossRef] [Green Version]
- Brin, S.; Page, L. The anatomy of a large-scale hypertextual Web search engine BT—Computer Networks and ISDN Systems. Comput. Netw. ISDN Syst. 1998, 30, 107–117. [Google Scholar] [CrossRef]
- Mihalcea, R.; Tarau, P. TextRank: Bringing Order into Text. In Proceedings of the 2004 Conference on Empirical Methods in Natural Language Processing, Barcelona, Spain, 25–26 July 2004; pp. 404–411. [Google Scholar]
- Chouigui, A.; Khiroun, O.B.; Elayeb, B. An Arabic Multi-Source News Corpus: Experimenting on Single-Document Extractive Summarization. Arab. J. Sci. Eng. 2021, 46, 3925–3938. [Google Scholar] [CrossRef]
- Luhn, H.P. The Automatic Creation of Literature Abstracts. IBM J. Res. Dev. 1958, 2, 159–165. [Google Scholar] [CrossRef] [Green Version]
- Gunawan, D.; Harahap, S.H.; Rahmat, R.F. Multi-document Summarization by using TextRank and Maximal Marginal Relevance for Text in Bahasa Indonesia. In Proceedings of the 2019 International Conference on ICT for Smart Society (ICISS), Bandung, Indonesia, 19–20 November 2019; pp. 1–5. [Google Scholar]
- Gabriela, N.H.; Siautama, R.; Amadea, C.I.A.; Suhartono, D. Extractive Hotel Review Summarization based on TF/IDF and Adjective-Noun Pairing by Considering Annual Sentiment Trends. Procedia Comput. Sci. 2021, 179, 558–565. [Google Scholar]
- Al-Abdallah, R.Z.; Al-Taani, A.T. Arabic Single-Document Text Summarization Using Particle Swarm Optimization Algorithm. Procedia Comput. Sci. 2017, 117, 30–37. [Google Scholar] [CrossRef]
- Qaroush, A.; Farha, I.A.; Ghanem, W.; Washaha, M.; Maali, E. An efficient single document Arabic text summarization using a combination of statistical and semantic features. J. King Saud Univ. Comput. Inf. Sci. 2021, 33, 677–692. [Google Scholar] [CrossRef]
- El-Haj, M.; Kruschwitz, U.; Fox, C. Experimenting with Automatic Text Summarisation for Arabic. In Proceedings of the Language and Technology Conference, Poznan, Poland, 6–8 November 2011; pp. 490–499. [Google Scholar]
- Fejer, H.N.; Omar, N. Automatic Arabic text summarization using clustering and keyphrase extraction. In Proceedings of the 6th International Conference on Information Technology and Multimedia, Barcelona, Spain, 1–3 April 2014; pp. 293–298. [Google Scholar]
- Haboush, A.; Al-Zoubi, M.; Momani, A.; Tarazi, M. Arabic text summarization model using clustering techniques. World Comput. Sci. Inf. Technol. J. ISSN 2012, 2, 741–2221. [Google Scholar]
- Al Qassem, L.; Wang, D.; Barada, H.; Al-Rubaie, A.; Almoosa, N. Automatic Arabic Text Summarization Based on Fuzzy Logic. In Proceedings of the 3rd International Conference on Natural Language and Speech Processing, Trento, Italy, 11–12 September 2019; pp. 42–48. [Google Scholar]
- Elgamal, M.; Hamada, P.S.; Aboelezz, P.R.; Abou-kreisha, M. Better Results in Automatic Arabic Text Summarization System Using Deep Learning based RBM than by Using Clustering Algorithm based LSA. Int. J. Sci. Eng. Res. 2019, 10, 781–786. [Google Scholar]
- Zaki, A.M.; Khalil, M.I.; Abbas, H.M. Deep Architectures for Abstractive Text Summarization in Multiple Languages. In Proceedings of the 2019 14th International Conference on Computer Engineering and Systems (ICCES), Cairo, Egypt, 17–18 December 2019; pp. 22–27. [Google Scholar]
- Elmadani, K.N.; Elgezouli, M.; Showk, A. BERT Fine-tuning for Arabic Text Summarization. arXiv 2020, arXiv:2004.14135. [Google Scholar]
- Al-Maleh, M.; Desouki, S. Arabic text summarization using deep learning approach. J. Big Data 2020, 7, 109. [Google Scholar] [CrossRef]
- Etaiwi, W.; Awajan, A. SemG-TS: Abstractive Arabic Text Summarization Using Semantic Graph Embedding. Mathematics 2022, 10, 3225. [Google Scholar] [CrossRef]
- Wazery, Y.M.; Saleh, M.E.; Alharbi, A.; Ali, A.A. Abstractive Arabic Text Summarization Based on Deep Learning. Comput. Intell. Neurosci. 2022, 2022, 1566890. [Google Scholar] [CrossRef] [PubMed]
- Elayeb, B.; Chouigui, A.; Bounhas, M.; Khiroun, O.B. Automatic Arabic Text Summarization Using Analogical Proportions. Cognit. Comput. 2020, 12, 1043–1069. [Google Scholar] [CrossRef]
- El-Halees, A.M.; Salah, D. Feature-Based Opinion Summarization for Arabic Reviews. In Proceedings of the 2018 International Arab Conference on Information Technology (ACIT), Werdanye, Lebanon, 28–30 November 2018. [Google Scholar]
- El-Haj, M.; Kruschwitz, U.; Fox, C. Using Mechanical Turk to Create a Corpus of Arabic Summaries. In Proceedings of the Language Resources and Evaluation Conference (LREC), Valleta, Malta, 17–23 May 2010; pp. 36–39. [Google Scholar]
- Elnagar, A.; Khalifa, Y.S.; Einea, A. Hotel arabic-reviews dataset construction for sentiment analysis applications. Stud. Comput. Intell. 2018, 740, 35–52. [Google Scholar]
- Farha, I.A.; Magdy, W. Mazajak: An online arabic sentiment analyser. In Proceedings of the Fourth Arabic Natural Language Processing Workshop, Florence, Italy, 1 August 2019; pp. 192–198. [Google Scholar]
- Alwehaibi, A.; Bikdash, M.; Albogmi, M.; Roy, K. A study of the performance of embedding methods for Arabic short-text sentiment analysis using deep learning approaches. J. King Saud Univ. Comput. Inf. Sci. 2021, 34, 6140–6149. [Google Scholar] [CrossRef]
- Huan, J.L.; Sekh, A.A.; Quek, C.; Prasad, D.K. Emotionally charged text classification with deep learning and sentiment semantic. Neural Comput. Appl. 2022, 34, 2341–2351. [Google Scholar] [CrossRef]
- Almuzaini, H.A.; Azmi, A.M. Impact of Stemming and Word Embedding on Deep Learning-Based Arabic Text Categorization. IEEE Access 2020, 8, 127913–127928. [Google Scholar] [CrossRef]
- Heikal, M.; Torki, M.; El-Makky, N. Sentiment Analysis of Arabic Tweets using Deep Learning. Procedia Comput. Sci. 2018, 142, 114–122. [Google Scholar] [CrossRef]
- Papineni, K.; Roukos, S.; Ward, T.; Zhu, W.-J. BLEU: A Method for Automatic Evaluation of Machine Translation. In Proceedings of the 40th Annual Meeting on Association for Computational Linguistics—ACL ’02, Philadelphia, PA, USA, 7–12 July 2002; p. 311. [Google Scholar]
- Rehurek, R.; Sojka, P. Software Framework for Topic Modelling with Large Corpora. In Proceedings of the LREC 2010 Workshop on New Challenges for NLP Frameworks, Valleta, Malta, 22 May 2010; pp. 45–50. [Google Scholar]
- Lin, C.-Y. ROUGE: A Package for Automatic Evaluation of Summaries. In Text Summarization Branches Out; Association for Computational Linguistics: Stroudsburg, PA, USA, 2004; pp. 74–81. [Google Scholar]
Reference | Dataset | Language | Method | Evaluation | Results | Limitations |
---|---|---|---|---|---|---|
[11] | 25 Arabic articles gathered from different websites on different subjects | Arabic | Graph-based method, PageRank, and MMR to filter redundant sentences | Human-generated summaries | Precision: 0.79 Recall: 0.72 F-measure: 0.75 | Limited dataset, not using the standard BLEU and ROUGE metrics |
[2] | Wikipedia pages and articles from some popular Arabic newspapers. | Arabic | Deep learning, variational auto-encoder (VAE) model with graph-based and query-based approaches | Human-generated summaries | ROUGE-1: 0.561–0.660 on 45% summary size | Focused on the news domain with a relatively small dataset |
[34] | Hotel reviews from the Tripadvisor website | Arabic | Feature-base and sentiment-based summaries were generated using TF-IDF and cosine similarity | Subjective evaluation | Accuracy: 72.84% | Using subjective evaluation only |
[4] | Reviews of two hotels from the Tripadvisor website | English | Clustering, Gaussian Mixture Model (GMM) algorithm | Comparing different clustering methods using statistical analysis | GMM method showed a more significant “usefulness” grade than the other methods | English only, focused on comparing two approaches using statistical analysis only |
[7] | EASC dataset (153 Arabic articles) [35] | Arabic | PageRank with the number of nouns in the sentence as starting value for each node | Five human summaries for each document, generated using Mechanical Turk | Precision: 0. 6875 Recall: 0.7294 F-measure: 0.6799 | Single document, focusing on clean news articles, not replicable with noisy reviews or tweets |
[33] | EASC [35] and ANT [16] | Arabic | Analogical reasoning, an AI approach for learning by analogy | Human-generated summaries | ROUGE-1: 0.75 BLEU: 0.47 | Single document, focusing on news, depends on keywords |
[19] | Hotel reviews from the Tripadvisor website | English | Two methods: 1. Selecting relevant sentences for the summary base on (TF-IDF) score 2. Pairing adjective to the nearest noun and considering the polarity | Human-generated summaries | Method 1: ROUGE-1: 0.2101 BLEU: 0.7820 Method 2: ROUGE-1: 0.0670 BLEU: 0.03672 | The polarity-based method showed poor performance |
[3] | Restaurant reviews from the Tripadvisor website | English | Topic-based and sentiment-based summary using TextRank | Subjective evaluation using informativeness, clarity, helpfulness, and likes indicators | According to users’ evaluation, the summaries provided sufficient and precise information about the target restaurant | Using subjective evaluation only |
[31] | 8385 news articles from Aljazira.net news website | Arabic | Graph embedding abstractive summarization | ROUGE and subjective evaluation | F1 score of 0.047 | Low performance. Focused on well-written articles, unreproducible with Web 2.0 noisy nature |
[32] | The Arabic Headline Summary (AHS) [31] and The Arabic Mogalad_Ndeef (AMN) [28] datasets, both comprise news articles | Arabic | Neural networks, including GRU, LSTM, and BiLSTM | ROUGE and BLEU | ROUGE-1 score of 51.49 | Focused on news only data, requires high computing power |
Layer (Type) | Number of Parameters |
---|---|
embedding (Embedding) | 9,227,072 |
bidirectional (Bidirectional) | 12,672 |
dense (Dense) | 2080 |
dropout (Dropout) | 0 |
dense_1 (Dense) | 165 |
Total params: 9,241,989 | |
Trainable params: 9,241,989 | |
Non-trainable params: 0 |
Sentence 1 | Sentence 2 | Before Stemming | After Stemming | ||||||
---|---|---|---|---|---|---|---|---|---|
BLEU | ROUGE-1 | ROUGE-2 | ROUGE-L | BLEU | ROUGE-1 | ROUGE-2 | ROUGE-L | ||
1 | F: 0.9 | F: 0.9 | F: 0.9 | 1 | F: 0.9 | F: 0.9 | F: 0.9 | ||
P: 1.0 | P: 1.0 | P: 1.0 | P: 1.0 | P: 1.0 | P: 1.0 | ||||
R: 1.0 | R: 1.0 | R: 1.0 | R: 1.0 | R: 1.0 | R: 1.0 | ||||
0 | F: 0 | F: 0 | F: 0 | 0.6 | F: 0.66 | F: 0 | F: 0.66 | ||
P: 0 | P: 0 | P: 0 | P: 0.66 | P: 0 | P: 0.66 | ||||
R: 0 | R: 0 | R: 0 | R: 0.66 | R: 0 | R: 0.66 | ||||
0 | F: 0 | F: 0 | F: 0 | 0.6 | F: 0.66 | F: 0 | F: 0.66 | ||
P: 0 | P: 0 | P: 0 | P: 0.66 | P: 0 | P: 0.66 | ||||
R: 0 | R: 0 | R: 0 | R: 0.66 | R: 0 | R: 0.66 | ||||
0.3 | F: 0.39 | F: 0 | F: 0.39 | 0.6 | F: 0.79 | F: 0 | F: 0.79 | ||
P: 0.5 | P: 0 | P: 0.5 | P: 0.66 | P: 0 | P: 0.66 | ||||
R: 0.33 | R: 0 | R: 0.33 | R: 1 | R: 0 | R: 1 | ||||
0.5 | F: 0.49 | F: 0 | F: 0.49 | 1 | F: 0.9 | F: 0.9 | F: 0.9 | ||
P: 0.5 | P: 0 | P: 0.5 | P: 1.0 | P: 1.0 | P: 1.0 | ||||
R: 0.3 | R: 0 | R: 0.5 | R: 1.0 | R: 1.0 | R: 1.0 | ||||
0.8 | F: 0.59 | F: 0 | F: 0.39 | 1 | F: 0.9 | F: 0 | F: 0.39 | ||
P: 0.6 | P: 0 | P: 0.4 | P: 1.0 | P: 0 | P: 0.4 | ||||
R: 0.6 | R: 0 | R: 0.4 | R: 1.0 | R: 0 | R: 0.4 |
Summary Type | Summary |
---|---|
Combined reviews | |
Translation | Very well. The place is clean. The presence of a nightclub and not mentioned it on the booking website. Very weak. Tight parking. The smell of the hotel is nasty. The disco sound is so annoying that I’m on the sixth floor and can’t sleep. Car service is slow. The buffet is very bad. Excellent staff, cleanliness, free parking, easy access to it, and it is only 5 km from Naif market, and the breakfast is excellent and wonderful, which is an open buffet. I lived on the fifth floor, it was quiet, and there was no disturbance from the nightclub except for very light trembles, and the hotel price was suitable. There is no free internet. Worst hotel I’ve stayed in in 25 years The location is good, The staff is not friendly, the reception staff is unskilled, breakfast is bad and the time is not suitable, the breakfast menu is Indian only There are no good choices of food, the hotel is not suitable for children The noise is very loud from the disco from the neighboring rooms, there are insects in the bathroom, the facilities are dilapidated not worth a star, only Indian TV channels. Disappointing only the manager, WIFI is overpriced, and the hour is 10 dirhams. The reception is bad. Parking is very small and complex. |
Positive reviews | |
Translation | Excellent staff, cleanliness, free parking, easy access to it, and it is only 5 km from Naif market, and the breakfast is excellent and wonderful, which is an open buffet. I lived on the fifth floor, it was quiet, and there was no disturbance from the nightclub except for very light trembles, and the hotel price was suitable. There is no free internet. |
Negative reviews | |
Translation | Disappointing only the manager, WIFI is overpriced, and the hour is 10 dirhams. The reception is bad. Parking is very small and complex. |
Neutral reviews | |
Translation | Convenient, as I found its price cheap. The location and the presence of security in the parking lots to help stand up and treat the staff. The amount was withdrawn twice from the visa, and I asked for an increase for another day. They said that I must pay an additional 120 dirhams, so I asked for the withdrawn amount in cash, so they agreed and asked to compensate me for the currency difference and the bank fees because it is their fault, and they bear all the consequences. They were trying to negotiate in an unsuccessful way, the water on the bathroom floor was accumulating, the parking was cramped, the breakfasts choices were few and modest, the channels were few and mostly Indian. |
Experiment | Hotel 1 | Hotel 2 | |||||||
---|---|---|---|---|---|---|---|---|---|
BLEU | ROUGE-1 | ROUGE-2 | ROUGE-L | BLEU | ROUGE-1 | ROUGE-2 | ROUGE-L | ||
Human 1 and 2 agreement score | 0.4744 | R: 0.4586 | R: 0.1295 | R: 0.4173 | 0.2677 | R: 0.6590 | R: 0.263 | R: 0.6206 | |
P: 0.3881 | P: 0.1289 | P: 0.3531 | P: 0.3917 | P: 0.1373 | P: 0.3690 | ||||
F: 0.4204 | F: 0.1292 | F: 0.3825 | F: 0.4914 | F: 0.1806 | F: 0.4628 | ||||
H1 | TextRank | 0.2983 | R: 0.3462 | R: 0.1187 | R: 0.3026 | 0.4030 | R: 0.6791 | R: 0.4395 | R: 0.6604 |
P: 0.5000 | P: 0.2307 | P: 0.4370 | P: 0.4965 | P: 0.2950 | P: 0.4829 | ||||
F: 0.4091 | F: 0.1568 | F: 0.3576 | F: 0.5736 | F: 0.3530 | F: 0.5578 | ||||
Rating grouping | 0.4327 | R: 0.4241 | R: 0.1756 | R: 0.3620 | 0.4807 | R: 0.6666 | R: 0.3435 | R: 0.6430 | |
P: 0.4300 | P: 0.2058 | P: 0.3671 | P: 0.5148 | P: 0.2646 | P: 0.4965 | ||||
F: 0.4270 | F: 0.1895 | F: 0.3645 | F: 0.5809 | F: 0.2989 | F: 0.5604 | ||||
Sentiment grouping | 0.3213 | R: 0.3707 | R: 0.1385 | R: 0.3159 | 0.5469 | R: 0.6648 | R: 0.3644 | R: 0.6432 | |
P: 0.4965 | P: 0.2466 | P: 0.4230 | P: 0.5603 | P: 0.3119 | P: 0.5421 | ||||
F: 0.4245 | F: 0.1773 | F: 0.3617 | F: 0.6081 | F: 0.3361 | F: 0.5883 | ||||
API grouping | 0.3484 | R: 0.3881 | R: 0.1446 | R: 0.3450 | 0.5766 | R: 0.6414 | R: 0.3778 | R: 0.6212 | |
P: 0.5034 | P: 0.2398 | P: 0.4475 | P: 0.5785 | P: 0.3412 | P: 0.5603 | ||||
F: 0.4383 | F: 0.1804 | F: 0.3896 | F: 0.6083 | F: 0.3585 | F: 0.5892 | ||||
F: 0.3866H2 | TextRank | 0.3355 | R: 0.3535 | R: 0.1152 | R: 0.3171 | 0.4147 | R: 0.3676 | R: 0.1057 | R: 0.3271 |
P: 0.6033 | P: 0.2250 | P: 0.5413 | P: 0.4521 | P: 0.1360 | P: 0.4022 | ||||
F: 0.4458 | F: 0.1524 | F: 0.3999 | F: 0.4054 | F: 0.1189 | F: 0.3608 | ||||
Rating grouping | 0.3909 | R: 0.3448 | R: 0.0810 | R: 0.3034 | 0.3771 | R: 0.3716 | R: 0.1096 | R: 0.3421 | |
P: 0.4132 | P: 0.0954 | P: 0.3636 | P: 0.4827 | P: 0.1619 | P: 0.4444 | ||||
F: 0.3759 | F: 0.0876 | F: 0.3308 | F: 0.4199 | F: 0.1307 | F: 0.3866 | ||||
Sentiment grouping | 0.3283 | R: 0.3368 | R: 0.0902 | R: 0.2976 | 0.3504 | R: 0.3540 | R: 0.0986 | R: 0.3297 | |
P: 0.5330 | P: 0.1613 | P: 0.4710 | P: 0.5019 | P: 0.1619 | P: 0.4674 | ||||
F: 0.4127 | F: 0.1157 | F: 0.3647 | F: 0.4152 | F: 0.1226 | |||||
API grouping | 0.3647 | R: 0.3557 | R: 0.1105 | R: 0.3288 | 0.3613 | R: 0.3535 | R: 0.1097 | R: 0.3131 | |
P: 0.5454 | P: 0.1840 | P: 0.5041 | P: 0.5363 | P: 0.1900 | P: 0.4750 | ||||
F: 0.4306 | F: 0.1381 | F: 0.3980 | F: 0.4261 | F: 0.1391 | F: 0.3774 |
Experiment | Hotel 1 | Hotel 2 | |||||||
---|---|---|---|---|---|---|---|---|---|
BLEU | ROUGE-1 | ROUGE-2 | ROUGE-L | BLEU | ROUGE-1 | ROUGE-2 | ROUGE-L | ||
Human 1 and 2 agreement score | 0.4744 | R: 0.4586 | R: 0.1295 | R: 0.4173 | 0.2677 | R: 0.6590 | R: 0.263 | R: 0.6206 | |
P: 0.3881 | P: 0.1289 | P: 0.3531 | P: 0.3917 | P: 0.1373 | P: 0.3690 | ||||
F: 0.4204 | F: 0.1292 | F: 0.3825 | F: 0.4914 | F: 0.1806 | F: 0.4628 | ||||
H1 | TextRank | 0.2983 | R: 0.3462 | R: 0.1187 | R: 0.3026 | 0.4030 | R: 0.6791 | R: 0.4395 | R: 0.6604 |
P: 0.5000 | P: 0.2307 | P: 0.4370 | P: 0.4965 | P: 0.2950 | P: 0.4829 | ||||
F: 0.4091 | F: 0.1568 | F: 0.3576 | F: 0.5736 | F: 0.3530 | F: 0.5578 | ||||
Rating grouping | 0.3223 | R: 0.3726 | R: 0.1476 | R: 0.3324 | 0.5743 | R: 0.6744 | R: 0.4218 | R: 0.6536 | |
P: 0.4860 | P: 0.2511 | P: 0.4335 | P: 0.5899 | P: 0.3738 | P: 0.5717 | ||||
F: 0.4218 | F: 0.1859 | F: 0.3763 | F: 0.6294 | F: 0.3964 | F: 0.6099 | ||||
Sentiment grouping | 0.3303 | R: 0.3692 | R: 0.1310 | R: 0.3180 | 0.5730 | R: 0.6666 | R: 0.4189 | R: 0.6484 | |
P: 0.4790 | P: 0.2149 | P: 0.4125 | P: 0.5831 | P: 0.3693 | P: 0.5671 | ||||
F: 0.4170 | F: 0.1628 | F: 0.3592 | F: 0.6221 | F: 0.3925 | F: 0.6051 | ||||
H2 | TextRank | 0.3355 | R: 0.3535 | R: 0.1152 | R: 0.3171 | 0.4147 | R: 0.3676 | R: 0.1057 | R: 0.3271 |
P: 0.6033 | P: 0.2250 | P: 0.5413 | P: 0.4521 | P: 0.1360 | P: 0.4022 | ||||
F: 0.4458 | F: 0.1524 | F: 0.3999 | F: 0.4054 | F: 0.1189 | F: 0.3608 | ||||
Rating grouping | 0.3248 | R: 0.3190 | R: 0.0784 | R: 0.2895 | 0.3699 | R: 0.3567 | R: 0.1067 | R: 0.3255 | |
P: 0.4917 | P: 0.1340 | P: 0.4462 | P: 0.5249 | P: 0.1814 | P: 0.4789 | ||||
F: 0.3869 | F: 0.0989 | F: 0.3512 | F: 0.4248 | F: 0.1343 | F: 0.3875 | ||||
Sentiment grouping | 0.3342 | R: 0.3234 | R: 0.0772 | R: 0.2938 | 0.3643 | R: 0.3541 | R: 0.1008 | R: 0.3203 | |
P: 0.4958 | P: 0.1272 | P: 0.4504 | P: 0.5210 | P: 0.1706 | P: 0.4712 | ||||
F: 0.3915 | F: 0.0961 | F: 0.3556 | F: 0.4217 | F: 0.1268 | F: 0.3813 |
Method | Hotel 1 | Method | Hotel 2 | Average | |
---|---|---|---|---|---|
H1 | TextRank | 0.2983 | TextRank | 0.4030 | 0.3507 |
Three-class grouping: rating grouping | 0.4327 | Three-class grouping: API grouping | 0.5766 | 0.5047 | |
Five-class grouping: rating grouping | 0.3223 | Five-class grouping: rating grouping | 0.5743 | 0.4483 | |
H2 | TextRank | 0.3355 | TextRank | 0.4147 | 0.3751 |
Three-class grouping: rating grouping | 0.3909 | Three-class grouping: rating grouping | 0.3771 | 0.3840 | |
Five-class grouping: Sentiment grouping | 0.3342 | Five-class grouping: rating grouping | 0.3699 | 0.3521 |
Method | Hotel 1 | Method | Hotel 2 | Average | |
---|---|---|---|---|---|
H1 | TextRank | 0.4091 | TextRank | 0.5736 | 0.4914 |
Three-class grouping: rating grouping | 0.4270 | Three-class grouping: API Grouping | 0.6083 | 0.5177 | |
Five-class grouping: rating grouping | 0.4218 | Five-class grouping: rating grouping | 0.6294 | 0.5256 | |
H2 | TextRank | 0.4458 | TextRank | 0.4054 | 0.4256 |
Three-class grouping: API grouping | 0.4306 | Three-class grouping: API Grouping | 0.4261 | 0.4284 | |
Five-class grouping: sentiment grouping | 0.3915 | Five-class grouping: rating grouping | 0.4248 | 0.4082 |
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations. |
© 2022 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Amoudi, G.; Almansour, A.; Alghamdi, H.S. Improved Graph-Based Arabic Hotel Review Summarization Using Polarity Classification. Appl. Sci. 2022, 12, 10980. https://doi.org/10.3390/app122110980
Amoudi G, Almansour A, Alghamdi HS. Improved Graph-Based Arabic Hotel Review Summarization Using Polarity Classification. Applied Sciences. 2022; 12(21):10980. https://doi.org/10.3390/app122110980
Chicago/Turabian StyleAmoudi, Ghada, Amal Almansour, and Hanan Saleh Alghamdi. 2022. "Improved Graph-Based Arabic Hotel Review Summarization Using Polarity Classification" Applied Sciences 12, no. 21: 10980. https://doi.org/10.3390/app122110980
APA StyleAmoudi, G., Almansour, A., & Alghamdi, H. S. (2022). Improved Graph-Based Arabic Hotel Review Summarization Using Polarity Classification. Applied Sciences, 12(21), 10980. https://doi.org/10.3390/app122110980