1. Introduction
Generative artificial intelligence (hereafter termed GenAI) is a rapidly developing technology which has been employed in the development of ChatGPT by OpenAI (OpenAI:
https://openai.com/ (accessed on 10 April 2024)). In a broad and diverse range of applications, GenAI plays a significant role in disruptive innovation (DI), where merging technologies can support smart applications [
1]. In addition, GenAI has many societal, ethical, technological, and practical risks, as expressed in
Section 2. GenAI models can accommodate multiple domains and the development of GenAI applications can be found in financial systems, computing systems, analysis, technological, and human resources [
2,
3,
4].
In the realm of AI, while there are multiple GenAI systems (both open source and proprietary systems), a significant focus has been on ChatGPT, a domain stemming from natural language processing (NLP) [
5,
6,
7]. The development trajectory of ChatGPT was primarily fueled by the objective to engineer an AI language model of high sophistication and versatility. This model is tailored for a spectrum of tasks encompassing text generation, language translation, and analysis of data. At the heart of ChatGPT’s foundational technology is the Transformer architecture, a pivotal evolution in AI language processing initially introduced in Ref. [
8]. This architecture was designed as a solution to the limitations inherent in previous NLP models, specifically recurrent neural networks (RNNs) and convolutional neural networks (CNNs). Many applications using large language models (LLMs) consider reasoning mechanisms in LLMs combined with ChatGPT for responses [
9,
10]. An integration of GenAI and LLMs can enable personalized service provision and decision making using engaging technologies in dynamic virtual environments which adapt and respond to users’ actions.
A goal of GenAI is to enhance interactions between a chatbot and an LLM(s) in a multiplicity of domains and systems to enable the creation of content including media, images, video, text, and audio. It supports innovative automated interactions in GenAI, NLP, image processing, and computer vision [
11]. GenAI provides novel approaches for creating content by filling gaps in the development of the ’metaverse’. Furthermore, LLM(s) and ChatGPT can enhance their responses as they relate to knowledge experience and information generation.
However, a recognized limitation lies in the difficulty in dealing with hidden rules in large datasets and the resulting responses by using a chatbot. In real-world applications, extracting information from large datasets using GenAI systems results in high computational cost and significant hardware and staff resources, as noted above; while large organizations have the resources to implement GenAI, SMEs generally lack the required resources.
In this paper, we present a novel model (hereafter termed GenAI-Algebra) which utilizes a combination of hedge algebra approaches and LLM(s) to find hidden rules in large datasets by incorporating the GenAI of ChatGPT. The GenAI-Algebra:
Extracts natural language knowledge from large datasets by leveraging fuzzy rules quantified by hedge algebra.
Has been designed to extract hidden rules in large datasets with automated question–response interactions in a broad and diverse range of domains and systems.
Has been developed for resource-limited SME(s).
In a case study in the medical domain predicated on the human heart (based on the UCI datasets to evaluate the effectiveness of the proposed model), the reported experimental results validate the effectiveness of the proposed model.
Our contributions may be summarized as follows:
Our GenAI-Algebra method can adapt to a multiplicity of domains in both Vietnamese and English. In the case study, GenAI-Algebra generates a comprehensive list of potential heart disease diagnoses based on a patient’s reported symptoms and medical history by analyzing the patient’s information using rules drawn from medical knowledge.
The customization and fine-tuning of ChatGPT integrated with knowledge bases allows the identification of hidden fuzzy rules quantified by hedge algebra in large datasets.
Our GenAI-Algebra method provides an effective basis upon which the simulation of real-time/real-world interactions [in both English and Vietnamese] can be realised.
The GenAI-Algebra method contributes to symptom analysis, supports differential diagnosis, collects real-time data, and enhances decision-support for clinicians.
Furthermore, the proposed GenAI-Algebra method and ChatGPT can play a valuable role in early detection by extracting relevant historical patient data and prognoses from large datasets; this can ultimately lead to improved patient policy outcomes.
The GenAI-Algebra model is trained by using ‘low-rank adaptation’ (LoRA) together with ‘DeepSpeed’ and mass datasets, which results in low computational overhead with reductions in inference time and cost that can lead to enhanced data protection and safety.
This research aims to address the problem by creating a GenAI model for a chatbot complete with an LLM [
12,
13] in both the Vietnamese and English languages.
In experimental testing, the proposed GenAI-Algebra model achieves a significant performance improvement. In the case study, the proposed model is compared to existing chatbot models, achieving a 92% performance based on the English benchmark.
The remainder of this paper is structured as follows: The state of the art and related research are considered in
Section 2 with the proposed
GenAI-Algebra model introduced in
Section 4. The experimental testing is introduced in
Section 6. The results with an analysis are set out in
Section 7.
Section 8 presents a discussion along with open research questions and directions for future research. The paper closes with concluding observations in
Section 9.
4. The Proposed GenAI-Algebra Model
In this section, we introduce our
GenAI-Algebra model, consisting of proprietary data and user questions as inputs, outputs as answers, the vector database, and the submodel. The proposed model aims to create a multilingual chatbot with its GenAI for instant responses. An overview of the proposed system architecture is shown in the conceptual model in
Figure 4.
The proposed GenAI-Algebra model can be applied to advance the diagnosis of heart disease and extract datasets by analyzing patient data to support doctors, leveraging its LLMs for responses in real time.
Proprietary data: Datasets are preprocessed and parameters are adjusted to process these data based on rules in the submodel hedge algebra hidden rule-based model, including heart datasets in the mass datasets.
User questions: Users can give questions and make requests from the proposed system, as well as interactive prompts, contexts, and original questions.
Hedge algebra hidden rule-based model: The submodel is to execute hidden rules considered from fuzzy rules with hedge algebra into the vector database. These rules are also updated to the vector database, which responds to LLMs.
Vector database: Prompts from questions and contexts of a domain can be requested from the database, which responds to LLMs.
LLMs: Stanford University has provided an approach which utilizes a publicly accessible backbone called
LLaMA [
56] and fine-tunes it using BLOOM on their public website. The adaptability of BLOOM [
57] to both English and Vietnamese allows the development of a multilingual chatbot that is capable of generating contextually relevant responses in both the English and Vietnamese languages.
To optimize hardware resources for model training, reducing the training time and costs, the proposed method allows organizations (including SMEs) to implement a chatbot adapted for both English and Vietnamese; the aim is the development of a multilingual chatbot capable of generating contextually relevant responses in both languages. The approach uses BLOOM [
57] with optimization for the training process and efficiently utilizes GPU memory; the LoRA [
58] with the DeepSpeed ZeRO-Offload [
59] method are used to optimize parameters to enable hardware performance.
In the proposed model, the input to the model consists of instruction prompts which can be in the form of inputs for the chatbot to respond to, as given by Equation (
2):
where dataset
C contains
N samples, for example,
i, and
N is the number of
instruction–output pairs.
is the
nth instruction, and
is the output for the
nth instruction.
To input texts of length
L, the attention scores for the
ith query
, (
) in each head, given the first
i keys
, where
d presents a head dimension, are given by Equation (
3):
4.2. Proposed LSmd Algorithm
This section introduces the proposed LSmd algorithm in order to generate LS sentences of the form “Q F y is/have S”, and the truth value T of each LS sentence, which is quantified from hidden rules in large datasets. These LS sentences will be updated to a vector database for LLMs of GenAI application.
Step 1: Select parameters for the HA architecture corresponding to
Let c−, c+ be the negative and positive generating elements, respectively, is the basic level frame of cognition, “0” is the label with the smallest semantic value, “W” is the label with the average semantic value, “1” is the one with the greatest semantic value, H is the set of labels, is the measure of fuzziness of the average label, is the fuzzy calculation range of the label x, and m is the calculation level.
Step 2: Generate a frame of cognition
Corresponding to the trapezoid of the fuzzy set representing the label x, we denote as the semantic core of x, as the left vertex ordinate of the big bottom, is the ordinate of the top right of the big bottom, is the ordinate of the top left of the small bottom, is the ordinate to the right of the small bottom, is the ordinate interval between the two small bottom peaks. , are the labels immediately before and after x in the ordinal set under consideration, respectively.
Call the m-level frame of cognition of fk , for each label , the fuzzy set of labels x is denoted as . To determine , the four vertices of the trapezoid need to be determined: .
Step 3: Calculate the average value of corresponding to each label as described in Algorithm 1
Let
be the average value of label x in the frame of cognition calculated over all records that satisfy the filter condition.
Algorithm 1: Calculating average value of term |
|
Step 4: Calculate the truth value of the conclusion corresponding to each quantifier
Let LSs be the set of conclusion sentences, T() is the truth value of the result sentence , Q is the frame of cognition of the quantifier, q is a label in Q, and is the membership function of fuzzy set q.
Step 5: Indicate all results of the sentences are updated to the vector database, which interacts with prompts through LLMs
8. Discussion
This study addresses the creation of a chatbot utilizing GenAI and LLM(s). The novel feature in our proposedGenAI-Algebra model is the identification of hidden rules in large datasets with appropriate question–response interactions. Moreover, the proposed GenAI-Algebra method has the capability to reduce the resource requirements, thus providing an effective basis upon which an SME can implement a multilingual chatbot.
The model training using ‘low-rank adaptation’ contributed to a reduction in training time and computational cost. In addition, we posit that our proposed model will be used for other languages. A reinforcement learning from human feedback (RLHF) method can be designed to improve the quality and safety of chatbot responses to questions and the quality of the extracting rules.
When reviewing large datasets of projects [for example in the medical domain used in the case study], the GenAI-Algebra outlines a framework describing the five levels of GenAI solutions through seven different levels of complexity. By using the GenAI-Algebra model, organizations can clearly understand their current position in the proposed model. This understanding will help them plan specific strategies to achieve their business goals.
To align internal skills and capabilities with desired business outcomes, enterprises can realistically assess their current position according to the GenAI-Algebra model. They should then consider the business outcomes they aim to achieve and evaluate what needs to be achieved to reach that future maturity state. This involves technical aspects and allows for practical adjustments in initiatives, skill development, support, and build-or-buy decisions. Understanding their maturity level will assist them in transforming to realize the desired business outcomes.
GenAI-Algebra enhances data strategy, processes, sharing, and more, alongside predictive AI in deploying end-to-end applications. In the preparation of datasets, it focuses on creating, managing, and preparing data—the essential raw material for GenAI models. This involves collecting large datasets, cleaning them, and ensuring their quality and relevance for training purposes. All of the LS sentences have truth values T of each LS sentence, which are quantified from hidden rules in large datasets. These LS sentences are updated to a vector database for prompts in LLMs of the GenAI application. In multiple domains, we can set up multiple models such as GenAI-Algebra and GenAI models of chatbots.
Author Contributions
Conceptualization, methodology, H.V.P. and P.M.; software, H.V.P.; validation, H.V.P. and P.M.; formal analysis, H.V.P.; investigation, H.V.P. and P.M.; data, H.V.P. and P.M.; writing—original draft preparation, H.V.P.; writing—review and editing, H.V.P. and P.M.; project administration, H.V.P. All authors have read and agreed to the published version of the manuscript.
Funding
This research was funded by Ministry of Science and Technology of Viet Nam under Program KC4.0, No. KC-4.0-38/19-25.
Institutional Review Board Statement
Not applicable.
Informed Consent Statement
Not applicable.
Data Availability Statement
Acknowledgments
The authors thank Nguyen Ha Thanh Dat and technical Engineering, Hanoi University of Science and Technology, doctor from Traditional Medical Hospital, and language experts participating in the experiments in ChatGPT for both English and Vietnamese languages. This work has been supported by Ministry of Scicence and Technology of VietNam under Program KC4.0, No. KC-4.0-38/19-25.
Conflicts of Interest
The authors declare no conflicts of interest. This research does not involve any human or animal participation. All authors have checked and agreed with the submission.
References
- Christensen, C.M.; McDonald, R.; Altman, E.J.; Palmer, J.E. Disruptive Innovation: An Intellectual History and Directions for Future Research. J. Manag. Stud. 2018, 55, 1043–1078. [Google Scholar] [CrossRef]
- Van Pham, H.; Thai, K.P.; Nguyen, Q.H.; Le, D.D.; Le, T.T.; Nguyen, T.X.D.; Phan, T.T.K.; Thao, N.X. Proposed Distance and Entropy Measures of Picture Fuzzy Sets in Decision Support Systems. Int. J. Fuzzy Syst. 2023, 44, 6775–6791. [Google Scholar] [CrossRef]
- Pham, H.V.; Duong, P.V.; Tran, D.T.; Lee, J.H. A Novel Approach of Voterank-Based Knowledge Graph for Improvement of Multi-Attributes Influence Nodes on Social Networks. J. Artif. Intell. Soft Comput. Res. 2023, 13, 165. [Google Scholar] [CrossRef]
- VPham, V.H.; Nguyen, Q.H.; Truong, V.P.; Tran, L.P.T. The Proposed Context Matching Algorithm and Its Application for User Preferences of Tourism in COVID-19 Pandemic. In International Conference on Innovative Computing and Communications; Springer Nature: Singapore, 2023; Volume 471. [Google Scholar] [CrossRef]
- Eysenbach, G. The role of ChatGPT, generative language models, and artificial intelligence in medical education: A conversation with ChatGPT and a call for papers. JMIR Med Educ. 2023, 9, e46885. [Google Scholar] [CrossRef]
- Michail, A.; Konstantinou, S.; Clematide, S. UZH_CLyp at SemEval-2023 Task 9: Head-First Fine-Tuning and ChatGPT Data Generation for Cross-Lingual Learning in Tweet Intimacy Prediction. arXiv 2023, arXiv:2303.01194. [Google Scholar]
- Haleem, A.; Javaid, M.; Singh, R.P. An era of ChatGPT as a significant futuristic support tool: A study on features, abilities, and challenges. BenchCouncil Trans. Benchmarks Stand. Eval. 2022, 2, 100089. [Google Scholar] [CrossRef]
- Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. Adv. Neural Inf. Process. Syst. 2017, 30, 6000–6010. Available online: https://dl.acm.org/doi/10.5555/3295222.3295349 (accessed on 8 June 2024).
- Hagendorff, T.; Fabi, S.; Kosinski, M. Human-like intuitive behavior and reasoning biases emerged in large language models but disappeared in ChatGPT. Nat. Comput. Sci. 2023, 3, 833–838. [Google Scholar] [CrossRef]
- Anam Nazir, Z.W. A comprehensive survey of ChatGPT: Advancements, applications, prospects, and challenges. Meta-Radiology 2023, 1, 100022. [Google Scholar] [CrossRef]
- Chiarello, F.; Giordano, V.; Spada, I.; Barandoni, S.; Fantoni, G. Future applications of generative large language models: A data-driven case study on ChatGPT. Technovation 2024, 133, 103002. [Google Scholar] [CrossRef]
- Crosthwaite, P.; Baisa, V. Generative AI and the end of corpus-assisted data-driven learning? Not so fast! Appl. Corpus Linguist. 2023, 3, 100066. [Google Scholar] [CrossRef]
- Tuan, N.T.; Moore, P.; Thanh, D.H.V.; Pham, H.V. A Generative Artificial Intelligence Using Multilingual Large Language Models for ChatGPT Applications. Appl. Sci. 2024, 14, 3036. [Google Scholar] [CrossRef]
- Khosla, M.; Anand, A.; Setty, V. A comprehensive comparison of unsupervised network representation learning methods. arXiv 2019, arXiv:1903.07902. [Google Scholar]
- Sun, Q.; Zhao, C.; Tang, Y.; Qian, F. A survey on unsupervised domain adaptation in computer vision tasks. Sci. Sin. Technol. 2022, 52, 26–54. [Google Scholar] [CrossRef]
- Zhang, Y.; Yang, Q. A survey on multi-task learning. IEEE Trans. Knowl. Data Eng. 2021, 34, 5586–5609. [Google Scholar] [CrossRef]
- Wang, Y.; Yao, Q.; Kwok, J.T.; Ni, L.M. Generalizing from a few examples: A survey on few-shot learning. ACM Comput. Surv. (CSUR) 2020, 53, 63. [Google Scholar] [CrossRef]
- Beck, J.; Vuorio, R.; Liu, E.Z.; Xiong, Z.; Zintgraf, L.; Finn, C.; Whiteson, S. A survey of meta-reinforcement learning. arXiv 2023, arXiv:2301.08028. [Google Scholar]
- Dong, Q.; Li, L.; Dai, D.; Zheng, C.; Wu, Z.; Chang, B.; Sun, X.; Xu, J.; Sui, Z. A survey for in-context learning. arXiv 2022, arXiv:2301.00234. [Google Scholar]
- Wu, T.; He, S.; Liu, J.; Sun, S.; Liu, K.; Han, Q.L.; Tang, Y. A brief overview of ChatGPT: The history, status quo and potential future development. IEEE/CAA J. Autom. Sin. 2023, 10, 1122–1136. [Google Scholar] [CrossRef]
- Cao, Y.; Li, S.; Liu, Y.; Yan, Z.; Dai, Y.; Yu, P.S.; Sun, L. A comprehensive survey of ai-generated content (aigc): A history of generative ai from gan to chatgpt. arXiv 2023, arXiv:2303.04226. [Google Scholar]
- Borji, A. A categorical archive of chatgpt failures. arXiv 2023, arXiv:2302.03494. [Google Scholar]
- Alkaissi, H.; McFarlane, S.I. Artificial hallucinations in ChatGPT: Implications in scientific writing. Cureus 2023, 15, e35179. [Google Scholar] [CrossRef]
- Cotton, D.R.; Cotton, P.A.; Shipway, J.R. Chatting and cheating: Ensuring academic integrity in the era of ChatGPT. Innov. Educ. Teach. Int. 2023, 61, 1–12. [Google Scholar]
- Howard, A.; Hope, W.; Gerada, A. ChatGPT and antimicrobial advice: The end of the consulting infection doctor? Lancet Infect. Dis. 2023, 23, 405–406. [Google Scholar] [CrossRef]
- Dwivedi, Y.K.; Kshetri, N.; Hughes, L.; Slade, E.L.; Jeyaraj, A.; Kar, A.K.; Baabdullah, A.M.; Koohang, A.; Raghavan, V.; Ahuja, M.; et al. Opinion Paper: “So what if ChatGPT wrote it?” Multidisciplinary perspectives on opportunities, challenges and implications of generative conversational AI for research, practice and policy. Int. J. Inf. Manag. 2023, 71, 102642. [Google Scholar] [CrossRef]
- Biswas, S.S. Role of chat gpt in public health. Ann. Biomed. Eng. 2023, 51, 868–869. [Google Scholar] [CrossRef] [PubMed]
- McGee, R.W. Is Chat GPT Biased against Conservatives? An Empirical Study (15 February 2023). Available online: https://ssrn.com/abstract=4359405 (accessed on 5 May 2024).
- Ali, M.J.; Djalilian, A. Readership awareness series—Paper 4: Chatbots and chatgpt-ethical considerations in scientific publications. In Seminars in Ophthalmology; Taylor & Francis: Abingdon, UK, 2023; pp. 1–2. [Google Scholar]
- Naumova, E.N. A mistake-find exercise: A teacher’s tool to engage with information innovations, ChatGPT, and their analogs. J. Public Health Policy 2023, 44, 173–178. [Google Scholar] [CrossRef]
- King, M.R.; ChatGPT. A conversation on artificial intelligence, chatbots, and plagiarism in higher education. Cell. Mol. Bioeng. 2023, 16, 1–2. [Google Scholar] [CrossRef]
- Thorp, H.H. ChatGPT Is Fun, but Not an Author, 2023. Available online: https://www.science.org/doi/full/10.1126/science.adg7879 (accessed on 5 May 2024).
- Wu, C.; Yin, S.; Qi, W.; Wang, X.; Tang, Z.; Duan, N. Visual chatgpt: Talking, drawing and editing with visual foundation models. arXiv 2023, arXiv:2303.04671. [Google Scholar]
- Li, M.; Wang, R. Chatbots in e-commerce: The effect of chatbot language style on customers’ continuance usage intention and attitude toward brand. J. Retail. Consum. Serv. 2023, 71, 103209. [Google Scholar] [CrossRef]
- Chen, Z.; Jiang, F.; Chen, J.; Wang, T.; Yu, F.; Chen, G.; Zhang, H.; Liang, J.; Zhang, C.; Zhang, Z.; et al. Phoenix: Democratizing ChatGPT across Languages. arXiv 2023, arXiv:2304.10453. [Google Scholar]
- Mackenzie, D. Surprising Advances in Generative Artificial Intelligence Prompt Amazement—and Worries. Engineering 2023, 25, 9–11. [Google Scholar] [CrossRef]
- Evans, O.; Wale-Awe, O.; Osuji, E.; Ayoola, O.; Alenoghena, R.; Adeniji, S. ChatGPT impacts on access-efficiency, employment, education and ethics: The socio-economics of an AI language model. BizEcons Q. 2023, 16, 1–17. [Google Scholar]
- Baidoo-Anu, D.; Owusu Ansah, L. Education in the era of generative artificial intelligence (AI): Understanding the potential benefits of ChatGPT in promoting teaching and learning. J. AI 2023, 7, 52–62. [Google Scholar] [CrossRef]
- Kohnke, L.; Moorhouse, B.L.; Zou, D. Exploring generative artificial intelligence preparedness among university language instructors: A case study. Comput. Educ. Artif. Intell. 2023, 5, 100156. [Google Scholar] [CrossRef]
- Martínez-Plumed, F.; Gómez, E.; Hernández-Orallo, J. Futures of artificial intelligence through technology readiness levels. Telemat. Inform. 2021, 58, 101525. [Google Scholar] [CrossRef]
- Sætra, H.S. Generative AI: Here to stay, but for good? Technol. Soc. 2023, 75, 102372. [Google Scholar] [CrossRef]
- Brown, T.; Mann, B.; Ryder, N.; Subbiah, M.; Kaplan, J.D.; Dhariwal, P.; Neelakantan, A.; Shyam, P.; Sastry, G.; Askell, A.; et al. Language models are few-shot learners. Adv. Neural Inf. Process. Syst. 2020, 33, 1877–1901. [Google Scholar]
- Ramesh, A.; Dhariwal, P.; Nichol, A.; Chu, C.; Chen, M. Hierarchical text-conditional image generation with clip latents. arXiv 2022, arXiv:2204.06125. [Google Scholar]
- Elhage, N.; Nanda, N.; Olsson, C.; Henighan, T.; Joseph, N.; Mann, B.; Askell, A.; Bai, Y.; Chen, A.; Conerly, T.; et al. A mathematical framework for transformer circuits. Transform. Circuits Thread 2021, 1, 12. [Google Scholar]
- Qiu, X.; Sun, T.; Xu, Y.; Shao, Y.; Dai, N.; Huang, X. Pre-trained models for natural language processing: A survey. Sci. China Technol. Sci. 2020, 63, 1872–1897. [Google Scholar] [CrossRef]
- Devlin, J.; Chang, M.W.; Lee, K.; Toutanova, K. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv 2018, arXiv:1810.04805. [Google Scholar]
- Liu, Z.; Lin, W.; Shi, Y.; Zhao, J. A robustly optimized BERT pre-training approach with post-training. In China National Conference on Chinese Computational Linguistics; Springer: Berlin/Heidelberg, Germany, 2021; pp. 471–484. [Google Scholar]
- Yang, Z.; Dai, Z.; Yang, Y.; Carbonell, J.; Salakhutdinov, R.R.; Le, Q.V. Xlnet: Generalized autoregressive pretraining for language understanding. Adv. Neural Inf. Process. Syst. 2019, 32, 5753–5763. Available online: https://dl.acm.org/doi/10.5555/3454287.3454804 (accessed on 8 June 2024).
- Banarescu, L.; Bonial, C.; Cai, S.; Georgescu, M.; Griffitt, K.; Hermjakob, U.; Knight, K.; Koehn, P.; Palmer, M.; Schneider, N. Abstract meaning representation for sembanking. In Proceedings of the 7th Linguistic Annotation Workshop and Interoperability with Discourse, Sofia, Bulgaria, 8–9 August 2013; pp. 178–186. [Google Scholar]
- Huang, S.; Dong, L.; Wang, W.; Hao, Y.; Singhal, S.; Ma, S.; Lv, T.; Cui, L.; Mohammed, O.K.; Liu, Q.; et al. Language is not all you need: Aligning perception with language models. arXiv 2023, arXiv:2302.14045. [Google Scholar]
- Chen, S.; Liu, J.; Wang, H.; Xu, Y.; Augusto, J.C. A linguistic multi-criteria decision making approach based on logical reasoning. Inf. Sci. 2014, 258, 266–276. [Google Scholar] [CrossRef]
- Nguyen, C.H.; Tran, T.S.; Pham, D.P. Modeling of a semantics core of linguistic terms based on an extension of hedge algebra semantics and its application. Knowl.-Based Syst. 2014, 67, 244–262. [Google Scholar] [CrossRef]
- Klir, G.K.; Yuan, B. Fuzzy Sets and Fuzzy Logic: Theory and Applications; Prentice Hall: Hoboken, NJ, USA, 1995. [Google Scholar]
- Berkan, R.C.; Trubatch, S.L. Fuzzy Systems Design Principles: Building Fuzzy IFTHEN Rule Bases; IEEE Press: Piscataway, NJ, USA, 1997. [Google Scholar]
- Zadeh, L.A. Fuzzy sets. Inf. Control 1965, 8, 338–353. [Google Scholar] [CrossRef]
- Touvron, H.; Lavril, T.; Izacard, G.; Martinet, X.; Lachaux, M.A.; Lacroix, T.; Rozière, B.; Goyal, N.; Hambro, E.; Azhar, F.; et al. LLaMA: Open and Efficient Foundation Language Models. arXiv 2023, arXiv:2302.13971. [Google Scholar]
- Scao, T.L.; Fan, A.; Akiki, C.; Pavlick, E.; Ilić, S.; Hesslow, D.; Castagné, R.; Luccioni, A.S.; Yvon, F.; Gallé, M.; et al. BLOOM A 176B-Parameter Open-Access Multilingual Language Model. arXiv 2022, arXiv:2211.05100. [Google Scholar]
- Sun, X.; Ji, Y.; Ma, B.; Li, X. A Comparative Study between Full-Parameter and LoRA-based Fine-Tuning on Chinese Instruction Data for Instruction Following Large Language Model. arXiv 2023, arXiv:2304.08109. [Google Scholar]
- Ren, J.; Rajbhandari, S.; Aminabadi, R.Y.; Ruwase, O.; Yang, S.; Zhang, M.; Li, D.; He, Y. ZeRO-Offload: Democratizing Billion-Scale Model Training. arXiv 2021, arXiv:2101.06840. [Google Scholar]
- Lapp, D. Heart Disease Dataset|Kaggle—kaggle.com. 2019. Available online: https://www.kaggle.com/datasets/johnsmith88/heart-disease-dataset (accessed on 2 March 2024).
- Laurençon, H.; Saulnier, L.; Wang, T.; Akiki, C.; Villanova del Moral, A.; Le Scao, T.; Von Werra, L.; Mou, C.; González Ponferrada, E.; Nguyen, H.; et al. The BigScience ROOTS Corpus: A 1.6 TB Composite Multilingual Dataset. In Advances in Neural Information Processing Systems; Koyejo, S., Mohamed, S., Agarwal, A., Belgrave, D., Cho, K., Oh, A., Eds.; Curran Associates, Inc.: Red Hook, NY, USA, 2022; Volume 35, pp. 31809–31826. [Google Scholar]
- Zheng, L.; Chiang, W.L.; Sheng, Y.; Zhuang, S.; Wu, Z.; Zhuang, Y.; Lin, Z.; Li, Z.; Li, D.; Xing, E.P.; et al. Judging LLM-as-a-judge with MT-Bench and Chatbot Arena. arXiv 2023, arXiv:2306.05685. [Google Scholar] [CrossRef]
- Hu, E.; Shen, Y.; Wallis, P.; Allen-Zhu, Z.; Li, Y.; Wang, S.; Wang, L.; Chen, W. LoRA: Low-Rank Adaptation of Large Language Models. arXiv 2022, arXiv:2106.09685. [Google Scholar]
- Checkland, P.; Holwell, S. Information, Systems and Information Systems: Making Sense of the Field; John Wiley and Sons: Oxford, UK, 1997. [Google Scholar]
- Murawsky, S. The struggle with transnormativity: Non-binary identity work, embodiment desires, and experience with gender dysphoria. Soc. Sci. Med. 2023, 327, 115953. [Google Scholar] [CrossRef]
Figure 1.
The taxonomy of pre-trained language models.
Figure 2.
An example of fuzzy set mapping of sub figures (a) and (b) for a numerical reference domain.
Figure 3.
An example of five fuzzy sets semantically representing the linguistic values of the variable age in the reference domain [0, 100] (unit: age).
Figure 4.
System architecture overview with data processing pipeline, model architecture, training process, and deployment.
Figure 5.
System architecture to extract information from heart disease database.
Figure 6.
Fuzzy sets of terms in the frame of cognition “age”.
Figure 7.
A frame of cognition of Q quantifiers.
Figure 8.
A list of records in the database.
Figure 9.
List of records in the database.
Figure 20.
The operational mechanism of LoRA is delineated through the flow depicted in the image.
Figure 21.
Prompt is applied to testing for GenAI-Algebra in heath care questions.
Figure 22.
Prompt is applied to LS sentence of the GenAI-Algebra in heart questions.
Table 1.
Ages of 10 patients.
No. | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 | 10 |
Age | 52 | 53 | 70 | 61 | 62 | 58 | 58 | 55 | 46 | 54 |
Table 2.
Coordinates of 4 vertices of trapezoidal fuzzy set of terms.
Word Class | 0 | c− | W | c+ | 1 |
---|
| 0 | 45 | 55 | 65 | 75 |
| 40 | 50 | 60 | 70 | 100 |
| 0 | 40 | 50 | 60 | 70 |
| 45 | 55 | 65 | 75 | 100 |
Table 3.
The membership of each age attribute to each term.
No. | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 | 10 |
---|
Age | 52 | 53 | 70 | 61 | 62 | 58 | 58 | 55 | 46 | 54 |
| 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
| 0.6 | 0.4 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0.2 |
| 0.4 | 0.6 | 0 | 0.8 | 0.6 | 1 | 1 | 1 | 0 | 0.8 |
| 0 | 0 | 1 | 0.2 | 0.4 | 0 | 0 | 0 | 0 | 0 |
| 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
Table 4.
Dependence with quantifiers.
Value | | | | | |
---|
| 1 | 0 | 0 | 0.4 | 1 |
| 0 | 1 | 0 | 0.6 | 0 |
| 0 | 0 | 0.8 | 0 | 0 |
| 0 | 0 | 0.2 | 0 | 0 |
| 0 | 0 | 0 | 0 | 0 |
Table 9.
Comparison of methods.
| Time/Epoch | Batch Size | Memory |
---|
Proposed model (BLOOM) | 54.5 h | 1 | 3.59 GB |
Proposed model (BLOOM) + LoRA | 4 h | 1 | 39.5 GB |
Proposed model (BLOOM) + LoRA + DeepSpeed | 4 h | 1 | 36.5 GB |
Proposed model (BLOOM) + LoRA + DeepSpeed | 3 h | 2 | 39.5 GB |
Table 10.
Details of the number of wins for each model over the categories in both English and Vietnamese. The bold numbers indicate the model that won in each category.
| English | | Vietnamese |
---|
Category | Phoenix | GenAI-Algebra | Total | Phoenix | GenAI-Algebra |
Heart common | 2 | 5 | 7 | 3 | 4 |
Health sense | 3 | 6 | 10 | 4 | 6 |
Health care | 4 | 6 | 10 | 5 | 5 |
Consultant | 4 | 6 | 10 | 6 | 4 |
Generic | 3 | 7 | 10 | 6 | 4 |
Knowledge | 3 | 7 | 10 | 3 | 7 |
Math | 6 | 4 | 10 | 6 | 4 |
Heart dialog | 4 | 6 | 10 | 4 | 6 |
Common sense | 6 | 4 | 10 | 5 | 5 |
Total wins | 12 | 43 | 87 | 18 | 21 |
Table 11.
Performance ratio (%) of GenAI-Algebra compared to Phoenix in the comparison on the English benchmark.
Performance Ratio | English | English in Specification |
---|
Phoenix | 97.89 | 95.72 |
GenAI-Algebra | 97.50 | 96.70 |
| Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).