1. Introduction
During the last two years, the huge breakthrough of Generative Pretrained Transformers (GPT) on Artificial Intelligence (AI) led to a boost and democratization of the use of Generative AI (GenAI) solutions, in the form of chatbots. Several companies have launched their solutions to the public in order to train, test, and acquire customers for their non-free versions, with Open AI and its ChatGPT, or Google and its Gemini being the most prevalent items. Students around the world are changing their habits, and now these tools are their favorite way to search for information [
1]. As a reference, it could be stated that 27% of the students were using AI writing tools in spring 2023, and by that year’s fall season, the percentage had boosted to 49% [
2].
Although this is a powerful new tool that can be used for student learning, concerns arise about the possible misuse of received information, such as cheating [
3] or taking the received answer without further mental processing [
4]. Some actions that can prevent it are being taken into account, such as creating new approaches to academic tasks and assessment [
5]. Other efforts are being made to verify the validity of these tools for serving as study partners, based on their performance on exams or practical assignments [
6]. Although the role of Generative AI in learning is a controversial topic, due to the risk of undermining student learning skills, there are studies that turn this last point of view upside down, even in memorization skills [
7].
Facing this new threat, for some people, opportunity for others, several international higher-education institutions have raised recommendations, following the first studies on the impact, possibilities, and dangers of GenAI use in this environment. One of the most used is the work of Gimpel et al. at the University of Hohenheim [
8], where they stressed the importance for students to know and respect regulations, focus on learning objectives rather than on the tool, and appealed to their use of critical thinking whenever they interact with these GenAIs. Meanwhile, for teachers, it is recommended to provide rules from the beginning of the season and to use GenAI to provide personalized training experiences to students, always oriented to improve their critical thinking. Another interesting study is [
9], where the authors recommended up to seven possible roles for GenAI when it comes to higher-education classrooms: as a mentor to provide feedback, as a tutor to provide direct instruction, as a coach to give deep reflection on the study and habits, as a team member in teamwork tasks to provide new approaches to the group, as a simulated student to force the real student to explain any item to it and by that means reinforce their learning, as a simulator to act as a practical trainer, and as a tool to improve students (and professor) efficiency. Although there is a full set of example prompts to implement them all, each one of the roles has a counterpart: the narrowing of critical thinking, biases, mismatch with the student, conflicts, hallucinations (unreal answers), low fidelity, or writing-without-thinking danger, respectively.
From then on, the aforementioned higher-education institutions have published recommendations following these lines in several ways. All of them stress the importance of changing some traditional assessments, such as typical written assignments, where GenAI is definitely going to excel, turning them into challenging situations for students in order to appeal their critical thinking [
10]. In fact, tracking this student’s thinking process could become a powerful new way of assessment, even with the help of GenAI [
11]. Policies for the use of GenAI at all levels are recommended in documents such as [
12] in the United States, where a human-in-the-loop strategy is stressed. This is increasingly common in the latest publications and is even regulated by the European Union as mandatory in any kind of assessment that may affect future employability, taking this issue as a red line or reaching a “high-risk” level in the use cases of AI [
13]. Other recommendations in [
12] are directed towards GenAI developers, to align them with educational purposes from scratch, to implement information and formation policies for faculty members who develop specific guidelines and guardrails, and to enhance trust and safety in the use of GenAI. The responsibility of educational personnel to keep track of these tools and learn their use is depicted in [
14], added to department or institution collaborative opportunities to apply them, and promoting cooperation in actions between institutions. All of these works show the importance of counting on students voices and perspectives in every process, as well as reviewing academic programs. Many of the current academic programs will be impacted by the use of these tools since professional activities will include the use of GenAI once students leave universities to enter the labor market [
15]. Indeed, this latter study shows not only the risks about the threat that GenAI entails to current jobs that may disappear, or at least be deeply transformed but also the opportunities that GenAI open to more creative, empowering and human-enhancing skilled jobs. Another very interesting recent study shows that students claim that a higher-education transformation is needed for future-ready employment in a GenAI-powered society, with the need for new learning skills and more interdisciplinary and hands-on topics [
16]. The trends that students point out are these competences for future workforce, AI literacy and new literacies to come, interdisciplinary teaching, innovative pedagogies, and new assessment formats.
Another major concern is the ethical use of GenAI, including copyright issues, responsible and trustworthy use, privacy and data governance (and not only where the uploaded data end but with what use: further training, publication, information collection, etc.), transparency, accountability, societal well-being, and knowledge of carbon footprint impact of the tools; all this is studied and recommended in the European Union in [
17]. There is a real possibility of reopening or deepening the digital gap, which is a priority to ensure equitable access to AI tools on campuses [
18]. Even more, these tools can be taken as an opportunity to increase inclusion in higher-education areas by creating specific adapted content or providing GPT assistants for students with disabilities, or removing biases affecting minorities, races, gender, religion, etc. [
19].
In addition to these general published recommendations, there has been an increasing number of published papers in 2024 providing insights on implications of Generative AI for education and research, also covering bias and inclusion, ethics and regulation, offering usage approaches as a support system or intelligent tutoring one, or in assessment [
20]. For example, ref. [
21] is a deep reflection on how universities must creatively revisit their teaching by developing AI-powered learning assistants or adaptive learning systems, and even faculty assistants. It is stated that learning improvements may be added to alleviate the pressure on stressed staff.
More specifically for the present work, when it comes to computer science, these tools are likely going to impact jobs in this area: it is a real concern because not only can GenAI perform programming tasks in a very good way but also many job sectors are increasingly using it [
22]. Indeed, AI is reshaping the syllabus of machine learning or data science. Therefore, both the syllabus and the way of teaching computer science are affected and should tend to adapt [
23]. The challenge for faculty members is also to stay updated on rapidly advancing technologies, maybe even more critically in this area than in other ones. A balanced approach is needed, integrating AI to enrich learning while carefully managing its implications for educational practices and student development. Thus, misuse can lead to the cheating or narrowing of programming skills of students, but good use can really improve their critical thinking and high-level computer science skills [
24].
Recent evidence shows the ability of these tools to solve computer science problems at the higher-education level, as well as the aim of students to adapt them to their study, such as the one performed in the Netherlands in [
25]. A good attempt to incorporate GenAI into a renowned online course at Harvard University—the CS50 course—can be found in [
26]. Interestingly, they show how the integration of AI can enhance the learning experience, providing students with continuous and personalized support, leaving educators more time to address higher-level pedagogical issues. Additionally, AI tools demonstrate their usefulness in explaining code snippets or improving code style, as well as answering any kind of administrative or curricular queries.
On the other hand, GenAI is a tool that can promote student programming self-efficacy and motivation, provided that teachers take specific actions to teach students prompt engineering skills [
27]. It has been shown that students not only need this guide but also demand it when facing the study with the help of these tools [
28], showing it to be more necessary, as the programming task to be addressed is more complex and higher level [
29].
Diving deeper into the inherent GenAI skills, in [
23], there is a description of an experiment using GenAI to solve a computer science complex problem, showing that it is not a good approach, and that it is rather better to divide it into small problems, a skill that students should gain, together with designing programs rather than just programming. At the same time, it shows a reflection about how coding skills can be quickly overcome in the teaching process to focus on understanding processes and the topic foundations. These two conclusions can be reinforced in recent works: Ref. [
22] looked for the limitations of ChatGPT in solving software development problems, and it was shown that it is good only at the easy- and medium-level ones. As a matter of fact, a survey shows that programmers are using AI tools to create better-performing, shorter, and more readable code (i.e., to become more productive, not to solve big tasks). On the other hand, Ref. [
30] drew similar conclusions about the positive role of GenAI in clearing the learning toward the fundamentals, as well as in solving the calculator dilemma. Relevant for the work presented in this paper, it is applied to the topic of database systems.
Regarding GenAI solutions for creating teaching aids, a very recent paper [
31] explored the possibility of using a RAG (Retrieved Augmented Generation) chatbot in a computer science classroom. A total of 46 students participated in this study. The results showed that the chatbot performed with a high accuracy rate (88% of the answers were considered right by teaching assistants based on the course materials), and the students felt more engaged with the more personalized training. However, they also claimed an additional need for human interaction with teachers, who, on their side, certainly reduced their administrative burden. Notwithstanding, once again, critical thinking and the teacher’s guide are mandatory, as the detection of inaccurate statements by a bot-generated tool is not efficient enough. This study not only evaluates the precision of bot responses, but also analyzes student acceptance, congruence with the course material, the comprehensiveness of the support offered, and associated costs, filling some of the gaps identified in the literature on LLM-driven educational chatbots.
Other higher-education institutions are developing some solutions for AI-assisted learning [
32]. The European School of Management and Business (EUDE) developed EDU, a virtual co-tutor powered by Generative AI, using IBM’s suite of AI solutions including Watson Assistant and Watson Discovery. This system supports administrative, academic, and logistic queries in real-time, which could be applied to database administration courses. Loyola University Chicago implemented LUie, an AI-powered digital assistant built on the Oracle Digital Assistant platform. Although not specifically for database administration, this system demonstrates how AI can be integrated into university systems to provide continuous student support and personalized information.
Focusing on data management, it has been reported that the GenAI scope can fill the gap when automation fails to understand the semantics of the data. Thus, difficult database problems can take a new approach, boosting data discovery, query synthesis, schema matching, and the like [
33]. By their nature, these GenAI tools are predictive models; therefore, they can perform well when it comes to combining them with information retrieval. MIT researchers found that SQL does not allow one to incorporate probabilistic AI models, and, at the same time, solutions with probabilistic models making inferences do not support complex database queries. To fill this gap, they created GenSQL [
34], a Generative AI system for databases that allows users to perform complex statistical analyses on tabular data without extensive technical knowledge. This tool could be particularly useful in teaching database administration by allowing students to make predictions, detect anomalies, guess missing values, fix errors, or generate synthetic data.
Ref. [
35] explored the use of ChatGPT as an assistant to automatically correct SQL exercises in a university database course. ChatGPT effectively identifies syntactic and semantic errors, provides explanations, and assigns grades with a strong correlation to human teacher ratings, although, interestingly, it tends to be stricter. Its accuracy decreases with more complex exercises due to subjectivity in error assessment. Although ChatGPT excels at detecting syntax errors (88.9%), it often confuses them with semantic ones, and its suggestions are only valuable in about half of the cases. Comparisons of Few-Shot and Zero-Shot Learning reveal that the latter produces clearer explanations, while the former relies heavily on provided examples. The study concludes that ChatGPT shows potential as a teaching assistant, but further research is needed to refine its application in education.
In ref. [
36], a chatbot is used to support students in completing SQL queries in a step-by-step process. Students who used it performed better on the final exam (4 out of 10 passed the exam, a number halved among students who did not use the tool), while teachers received insightful performance metrics on student progress, aiding them in instructional decisions. Anderson et al. research also highlighted the critical role of prompt engineering in obtaining successful results with GenAI for SQL practice [
37], highlighting that students benefited significantly from teacher guidance in developing effective prompts.
Together, these studies underline Generative AIs potential as an SQL learning aid, while also highlighting the importance of instructor support for effective prompt formulation and student guide for GenAI usage. They also show the need for further research on the utility of these tools in database administration courses, as there are still some open questions about (i) the combined use of traditional and new tools, (ii) the ability of these tools to solve complex problems at the university level, and (iii) the possible improvement in learning database administration using them.
In consideration of the aforementioned findings and previous efforts, we have undertaken an empirical study which addresses the following research questions (RQs):
RQ1. How widespread is the use of ChatGPT among students to learn database administration, and what is the perceived reliability and usefulness of this educational resource compared to other traditional resources?
RQ2. How does prior knowledge about database administration affect the use of ChatGPT to solve database administration problems?
RQ3. Is there a relationship between the usage of ChatGPT and the success in solving practical database administration problems?
This study aims to advance the experience of using GenAIs in teaching computer science, and more specifically database administration, following the aim of the recommendations described above. The rest of the article is structured as follows. The materials and methods are described in
Section 2. The results are presented and discussed, respectively, in
Section 3 and
Section 4. Lastly,
Section 5 presents the conclusions, limitations, and future work drawn from this study.
5. Conclusions
This article has presented an empirical study to explore the usage of ChatGPT to learn about database administration in computer science education. We found that ChatGPT was moderately used, but it was more used than traditional internet learning resources, and a trend of increasing use could be observed compared to the previous year. We also found uneven usage by students and that the students who end up earning better grades were the ones who used it more and had better perceptions about ChatGPT. In fact, positive correlations were found between the use of this tool and the success of solving database administration problems during a practical exam. Lastly, we found that one of the elements that could encourage students to use this tool intensively and productively was the theoretical background they had about database administration.
Therefore, it can be concluded that ChatGPT can be an excellent educational instrument in the context of database administration and that to use it properly, it is necessary for students to have a solid theoretical basis. However, as mentioned in the Introduction, it should be noted that GenAI tools such as ChatGPT have a number of limitations to consider: the possibility of hallucinating, the lack of critical thinking that could be generated in the GenAI users, and even prohibitions on the use of these tools in the process of software development by contractors.
In addition, during the execution of this case study, we have found that prompting skills are also important, and teachers can use a guided practice strategy, where prompting and guided step-by-step practice are employed to train these skills as well as to increase the student’s theoretical background. Training students in the use of GenAI tools like ChatGPT is key to prevent the appearance of new digital trenches between students with solid knowledge and prompting skills that can further increase their performance through the proper use of GenAI tools and students with a lower degree of the fundamentals and worse prompting skills, who lag behind and will obtain even worse results.
The presented conclusions are strongly connected to computer science education since they are based on a case study in that context, but we deem that they are fully transferable to any engineering or science discipline in which the practical mastery of technologies and problem solving play a key role.
Lastly, some limitations and future work are presented. First, the sample size is only 40 students, so the conclusions reported here should be taken with caution. Furthermore, there are many difficulties associated with the generalization of single case studies like the presented [
49], so more studies like this should be conducted to consolidate our findings. Once this occurs and the presented framework has been applied several times, we could provide a more reliable and generalizable mathematical model. Second, the gender distribution is unbalanced, which is common in computer science studies, and unfortunately we do not have a way to solve this limitation in the study. Third, the study design precludes the assertion that ChatGPT caused higher grades. The conduction of a randomized controlled trial in which there was an experimental group with access to ChatGPT and a control group without access to ChatGPT would allow one to isolate the ChatGPT variable and assert cause-and-effect relationships. However, although this is very desirable from a research point of view, from a teaching perspective it is complicated, as it violates the principle of student equity. Fourth, the conclusions reported are based on various data, including those obtained through the questionnaire. This instrument provides self-reported data that can influence responses due to social desirability bias and inaccurate self-reporting. In detail, the prompting skills of the students are measured using this self-perception questionnaire, rather than more objective tests. In the future, the prompting skills of the students could be measured by analyzing their prompts. Lastly, conducting studies in the future using qualitative data or assessing long-term knowledge retention would provide a more complete picture of ChatGPT’s impact on learning.