Next Article in Journal
A Bibliometric Analysis and Systematic Review on E-Marketplaces, Open Innovation, and Sustainability
Next Article in Special Issue
The Effects of the Crisis Management Skills and Distance Education Practices of Universities on Student Satisfaction and Organizational Image
Previous Article in Journal
CO2 Emission Reduction Potential of Road Transport to Achieve Carbon Neutrality in China
Previous Article in Special Issue
Creating IoT-Enriched Learner-Centered Environments in Sports Science Higher Education during the Pandemic
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Using Digital Technologies for Testing Online Teaching Skills and Competencies during the COVID-19 Pandemic

by
Mohammed Abdullatif Almulla
Department of Curriculum and Instruction, Faculty of Education, King Faisal University, Al Ahsa 31982, Saudi Arabia
Sustainability 2022, 14(9), 5455; https://doi.org/10.3390/su14095455
Submission received: 7 April 2022 / Revised: 28 April 2022 / Accepted: 28 April 2022 / Published: 1 May 2022

Abstract

:
This study examines the online teaching abilities and competencies needed to teach online courses in Saudi Arabia’s higher education system. As a result, the goal of this study was to examine and expand the technology acceptance model (TAM) to assess online teaching abilities and competences utilizing digital technologies in higher education during the COVID-19 epidemic. Therefore, it aimed to develop a new model to measure and explore critical factors that influence online teaching skills, competencies, and actual use of digital tools in higher education. The participants in the study were 350 lecturers at King Faisal University. The research data were analyzed using structural equation modeling (AMOS-SEM). The findings revealed that: (a) perceived ease of use and perceived usefulness on using digital tools during the COVID-19 pandemic has a direct positive impact on perceived teaching self-efficacy, perceived enjoyment, online teaching skills, and digital tools access; (b) perceived ease of use and perceived usefulness of using digital tools have a direct positive impact on lecturers’ attitude toward use and lecturers’ behavioral intention to use digital tools during the COVID-19 pandemic; and (c) perceived ease of use. As a consequence of the findings, a validated instrument was designed to assess and investigate crucial elements that impact lecturers’ real usage of digital technologies for teaching and learning in Saudi Arabia’s higher education.

1. Introduction

The coronavirus 2019 (COVID-19) pandemic has become a global threat, according to UNESCO, affecting several areas, including education [1]. On 30 January 2020, the World Health Organization (WHO) designated the COVID-19 virus as a public health concern; then, on 11 March 2020, it was recognized as a worldwide pandemic [2]. COVID-19 has generated serious challenges in the field of education for students, teachers, and educational institutions all across the world [3]. Due to the epidemic, Saudi Arabia’s Ministry of Education (MOE) has ordered that all academic activities be totally transformed to online formats [4]. Many higher education courses, particularly those that require physical interaction, are experiencing difficulties. This epidemic has affected over 87 percent of the world’s student population in more than 160 nations, according to [1]. As a result, conventional learning is being replaced by digital learning. Furthermore, there is no consensus in the research on how to define traditional (sometimes conventional) learning precisely. According to [5], a model of contextual facilitators for learning activities involving technology for both on-site and distance learning environments in higher education on digital teaching and learning is proposed, as well as a roadmap for future research to understand the complex dynamic factors that lead to successful digital teaching and learning in higher education through learning activities. In this paradigm, the teacher is the active party, transmitting information to students through lectures and printed materials, as well as a sort of interaction between students and both the instructor and the subject. Traditional learning, according to [6], is “delivering learning material face-to-face with no usage of the Internet for teaching and learning”. Traditional learning, according to [7], is “a method of learning that takes place under the supervision of a teacher in a physical learning environment while employing physical learning instruments and direct synchronous communication”. Only those courses that do not use any internet technologies are classified as conventional by [8]. Traditional learning, according to these definitions, is a learning process, a teaching style, and a medium for imparting education. Because the instructor and learner are physically and simultaneously present in the same room, they communicate and speak in real time without the need of the Internet or online technologies. Surprisingly, the effectiveness of a model or teaching approach is frequently discussed in educational literature as a comparison to traditional learning or teaching. According to [9], faculty learning approach goals were also positively connected with considering the move to online teaching as a good challenge and as beneficial to their personal competence development. In a large body of literature, traditional learning has been used as an antonym for new modes of learning, such as e-learning, as well as virtual, cyber, hybrid, and online learning and teaching [10]. Others have used the term “traditional learning” to refer to particular teaching methods, such as cooperative learning [11,12], problem-based learning [13,14], project-based learning [15,16], and game-based learning [15,16,17,18]. This raises the question of what constitutes a standard education. Allen and Seaman [8] categorize courses depending on how much Internet technology is utilized in their delivery compared to traditional courses. Web-facilitated courses are face-to-face courses that use the Internet to offer less than 30% of the content by distributing learning materials and assignments on an LMS. A hybrid or blended course is one in which the Internet is used for less than 80% of the content and activities. The majority of online courses are given and taught exclusively via the internet. As a result, it might be an opportune time to push for the adoption of a new learning paradigm. It may be helpful to consider web-facilitated or blended learning as the next standard learning approach [19]. Since its debut, a variety of terminology and concepts have been used to characterize online learning. Some of the terminology used to characterize online learning includes e-teaching, virtual learning, cyber learning, Internet learning, distributed learning, web-facilitated learning, web-based learning, remote learning, computer-based learning, resource-based learning, and technology-based learning [20,21,22].
The phrase “online teaching and learning” will be used throughout this paper. Education in the country has changed dramatically as a result of the increased demand for online teaching adoption. According to [23], institutions that expect their teaching staff to use e-learning systems for teaching and learning should assess their behavioral intention to use e-learning systems for teaching and learning. The technology acceptance model (TAM) is the most commonly utilized model in research of consumer technology adoption [24,25]. The model’s major goal is to explain consumers’ attitudes toward technology adoption [26,27]. As a result, using the technology acceptance model theory, this study explores teachers’ intentions to utilize digital tools during the COVID-19 epidemic.

2. Research Model and Hypotheses

Online teaching skills and competencies must be determined in order to help assess online teaching. Teaching online necessitates a shift in educational thinking. Whereas traditional teaching focuses on the instructor attempting to impart their information to the students, online teaching focuses on the interaction between the teacher and the student as well as the student and knowledge. An online course’s efficacy necessitates more from a teacher than just technical competence; it also necessitates a thoughtful and inventive mentality that prioritizes the student-centered learning process.
Therefore, this research develops a model that is unique in extending the technology acceptance model (TAM) to measure online teaching skills and competencies in using digital tools during the COVID-19 pandemic. It does so by exploring critical factors that influence lecturers’ actual use of digital tools for teaching and learning in Saudi Arabia’s higher education through perceived teaching self-efficacy, perceived enjoyment, online teaching skills, digital tool access, perceived ease of use, perceived usefulness, attitude toward using digital tools, behavioral intention to use digital tools, and actual use of digital tools for teaching and learning, see Figure 1.

2.1. Perceived Teaching Self-Efficacy (PTSE)

The success of online teaching and learning is influenced by teachers’ perceived teaching self-efficacy and attitude toward adopting digital technology for teaching and learning [28]. According to [29], self-efficacy beliefs influenced changes in behavior and how people regulate and impact their environment, and the notable conclusion from these studies is that teachers with high self-efficacy beliefs perceived teaching efficacy to have positive correlations with desired outcomes that benefit student learning [29,30,31]. As a result of these advantages, teachers will be more eager to experiment with new ideas [32] and demonstrate a greater commitment to the teaching profession. According to [33], there is a strong link between instructors’ comfort and competency with technology and the degree to which they use it. Additional hurdles, such as bandwidth, access, and technical skill, may also affect high technology acceptance. Teachers’ impressions of how classroom technology interacts with teaching can be influenced by their self-efficacy. As a result, the following hypotheses emerged from this study:
Hypothesis 1 (H1).
There is a strong relationship between PTSE & PEU.
Hypothesis 2 (H2).
There is a strong relationship between PTSE & PU.
Hypothesis 3 (H3).
There is a strong relationship between PTSE & PE.

2.2. Perceived Enjoyment (PE)

Perceived enjoyment, according to Davis et al. [24], is similar to intrinsic motivation, which drives the execution of an activity that is unrelated to anything other than the process of executing the activity. Venkatesh [34] also observed that the influence of enjoyment on perceived ease of use was stronger as users gained more direct experience with the system. The degree to which users believe using the system to be enjoyable has an impact on perceived ease of use, according to these studies. Davis et al. [25] observed that perceived usefulness was influenced by ease of use, whereas [34] discovered that perceived usefulness was influenced by ease of use. Perceived usefulness is a statistic that measures how people believe technology may help them be more productive and effective. Teo, Lim, and Lai [35] investigated the effects of perceived ease of use, perceived utility, and perceived enjoyment on Internet usage, finding that perceived usefulness and perceived enjoyment had an impact on respondents’ enjoyment of the Internet. They also observed that the amount of perceived satisfaction had a significant influence on the frequency with which people used the product. As a result, the following hypotheses emerged from this study:
Hypothesis 4 (H4).
There is a strong relationship between PE & PEU.
Hypothesis 5 (H5).
There is a strong relationship between PE & PU.

2.3. Online Teaching Skills (OTS)

Online teaching abilities must be identified and prioritized based on the jobs that educators will actually perform [36]. Institutions may provide technical assistance for adopting LMSs and other technologies, which may lessen the workload of online educators. It is, nevertheless, critical to achieve sufficient understanding of how to diagnose and resolve issues [37]. Teachers’ skills and competencies can aid in the design and creation of online educator professional development activities. As mature, self-regulated, and self-determined learners, online teachers may utilize these to self-evaluate their abilities and subsequently detect their own learning and training requirements [38]. Competencies may also be used as a process to guarantee that teachers are qualified and ready to teach in online learning environments [39]. The duties, talents, and skills that are necessary to be a competent and successful online teacher should be identified and emphasized by educational institutions, online learning organizations and authorities, and online learning theorists. As a result, online instructors need a framework and rules to help them stay afloat, improve their skills, and offer relevant training programs [40]. Bawane and Spector [36] also provide a general framework for developing and implementing teacher professional development programs. Furthermore, if online instructional staff members are exposed to examples of exceptional online instructors and view them as role models, they may obtain a better understanding of their function as online professors [38]. As a result, the following hypotheses emerged from this study:
Hypothesis 6 (H6).
There is a strong relationship between OTS & PEU.
Hypothesis 7 (H7).
There is a strong relationship between OTS & PU.

2.4. Digital Tools Access (DTA)

Technology ushers in fundamental structural changes which can be critical in driving significant productivity improvements. To enhance both teaching and learning, technology enriches classrooms with digital learning aids, including computers and handheld devices. It promotes student involvement and motivation while also expanding course offerings, experiences, and learning resources [41,42,43,44]. According to a meta-analysis on the integration of technology into education [45], the most common obstacle to technology integration is the availability of technological resources. Teachers have explained how digital technology may help diversify learning and “increase grades, keep students’ interest, and even enhance the confidence of many kids,” [46]. However, more recent research has demonstrated that growing technology accessibility does not always imply increased or improved technology usage quality [47,48]. The findings, which are consistent with earlier studies, emphasize the factors that impact instructors’ usage of technology [49,50]. As a result, the following hypotheses emerged from this study:
Hypothesis 8 (H8).
There is a strong relationship between DTA & PEU.
Hypothesis 9 (H9).
There is a strong relationship between DTA & PU.

2.5. Perceived Ease of Use (PEU)

The extent to which a user feels that using digital tools for teaching and learning will be free of effort is described as perceived ease of use in the context of utilizing digital technologies for teaching and learning [51]. It influences perceived usefulness as well as attitudes toward adopting digital resources for teaching and learning [25,52,53]. Furthermore, a number of studies has confirmed that adopting digital tools for teaching and learning is one of the most important indicators of attitude toward technology adoption [54,55]. We investigated instructors’ perceptions of how easy it is to use digital technologies for teaching and learning in higher education in this study. As a result, the following hypotheses emerged from this study:
Hypothesis 10 (H10).
There is a strong relationship between PEU & PU.
Hypothesis 11 (H11).
There is a strong relationship between PEU & ATUD.
Hypothesis 12 (H12).
There is a strong relationship between PEU & BIUD.

2.6. Perceived Usefulness (PU)

Perceived usefulness of utilizing digital tools for teaching and learning is defined by Lin et al. [51] as the degree to which a user feels that using digital tools for teaching and learning may assist them in achieving learning objectives. According to previous research, perceived utility is one of the most important elements influencing consumers’ attitudes toward technology [53,55,56]. Furthermore, perceived utility influences behavioral intention both directly and indirectly [57,58]. As a result, the following hypotheses emerged from this study:
Hypothesis 13 (H13).
There is a strong relationship between PU & ATUD.
Hypothesis 14 (H14).
There is a strong relationship between PU & BIUD.

2.7. Attitude toward Using Digital Tools for Teaching and Learning (ATUD)

Kaplan [59] defined attitude as a predisposition to react positively or negatively to an occurrence. Much research on the acceptability of digital technologies in education has found that attitude is a strong predictor of behavioral intention to use digital tools in education [60,61]. The link between user attitude and behavioral intention meant that users are more likely to engage in specific actions if they have a good attitude toward them [62]. Furthermore, behavioral intention to utilize digital tools for education is totally mediated by attitudes about technology. As a result, the following hypotheses emerged from this study:
Hypothesis 15 (H15).
There is a strong relationship between ATUD & BIUD.
Hypothesis 16 (H16).
There is a strong relationship between ATUD & AUDT.

2.8. Behavioral Intention to Use Digital Tools for Teaching and Learning (BIUD)

The TAM approach produces two outcome variables: behavioral intention to use and actual usage. The behavioral intention to utilize digital tools is described as a behavioral inclination to continue utilizing technology in the future; as a result, it influences technological adoption [63]. Previous research has found that having the desire to utilize digital resources for education has a beneficial impact on actually using them for teaching and learning. Furthermore, previous research has found that perceived utility [64], perceived ease of use [52,64], and attitude toward utilizing digital tools all impact behavioral intention to use digital tools [65,66]. As previously stated, behavioral intention to utilize digital tools is a dependent variable in this study. As a result, the following hypothesis emerged from this study:
Hypothesis 17 (H17).
There is a strong relationship between BIUD & AUDT.

2.9. Actual Use of Digital Tools for Teaching and Learning (AUDT)

Distance learning has becoming more popular in higher education institutions across the world [67]. Students and instructors use technology to aid the learning process in distant learning. Teaching with technology does not apply to everyone. Its execution will be determined by the technology resources available and students’ educational demands [68]. According to research conducted at Hong Kong University, students’ computer skills, attitudes toward technology, learning styles, and instructors’ and peers’ support are all factors that influence technology use for learning [69]. “The use of electronic technology to deliver, assist, and enhance both learning and teaching, and involves communication between learners and teachers utilizing online information,” according to a definition of distance learning using digital technologies [70]. Distance learning may help students shift from a teacher-centered approach, in which they are a passive observer following the teacher’s instructions, to a student-centered approach, in which they are active participants in their own education as well as other students’. Individuals may benefit from adopting digital tools in distance learning in coping with a wide range of learning environments and in providing self-paced learning and content [71]. The eight main types of digital educational tools include learning management systems (LMS), Massive online open courses (MOOCs), mind mapping, interactive tools, web conferencing, infographics, research visibility tools, and web-based content management systems. Only one third of nursing students favored distant learning using digital technologies according to a recent survey [72,73]. Furthermore, digital platforms for education and learning were linked to students’ greater propensity to finish research work successfully [74]. Lesson study and learning have benefited from digital technology because it has permitted rigorous cooperation, synchronous observations, improved data management, and a deeper grasp of material [75].

3. Research Methodology

With a cross-sectional questionnaire, we used a quantitative method [76]. A quantitative technique may produce trustworthy, valid, objective, and generalizable results, and questionnaires can be sent to a large number of people. Furthermore, a quantitative technique [76], allows for population-wide generalizations. Quantitative research also relies on hypothesis testing, which may be performed with explicit instructions and objectives [77]. During the COVID-19 pandemic, hypotheses were explored to evaluate online teaching skills and competencies utilizing digital technologies.

3.1. Participants in the Study

The participants in this study were university lecturers from Saudi Arabia’s King Faisal University who taught using an online learning system. Participants completed the online training over the course of a semester. The disciplines of instruction were educational technology, and instruction via an online learning system was introduced in 2021 and is now available at all Saudi Arabian universities. The research was carried out in one of the allocated areas at King Faisal University, which will provide training for 240 instructors in 2021.

3.2. Data Gathering and Analyze

An online questionnaire was used in this study since it was simple to administer and available from a variety of devices [76]. The bulk of those who took part were reached by WhatsApp and email. We used Google Forms to run the online survey, emailing participants a link and keeping the survey open for five weeks. In order to test the online teaching abilities and competences utilizing digital technologies during the COVID-19 pandemic, structural equation modelling (SEM), AMOS-SEM, and SPSS were used. As a result, AMOS-SEM was used to perform confirmatory factor analysis (CFA) and test the model’s reliability, validity, and internal consistency. The hypotheses were proven, and a structural model was created [78]. Therefore, composite reliability and validity can be calculated as follows:
C R = λ i 2 λ i 2 + ε i
The standardized factor loading is represented by (λ) with (i) denoting the item index, whereas error variance is represented by (ε). Error variance is defined as the following:
ε i = 1 λ i 2
Meanwhile, as given in the following, r-squared represents items variances percentage justified through latent constructs. R-squared is defined as follows:
r 2 = λ 1 2 = 1 ε i

3.3. Instruments Model

As stated in, a survey instrument was used to meet the study goals through an in-depth analysis. There were ten factors with thirty-one indicators. Perceived teaching self-efficacy was adapted to three items as recommended by [30], perceived enjoyment was adapted to three items as recommended by [24,25], perceived ease of use and perceived usefulness were adapted to three items for each factor as recommended by [24], online teaching skills was adapted to three items as recommended by [36], digital tools access was adapted to three items as recommended by [46], attitude toward using digital tools was adapted to three items as recommended by [61], behavioral intention to use digital tools was adapted to three items as recommended by [63]. Finally, actual use of digital tools for teaching and learning was adapted to five items as recommended by [68].

4. Data Analysis and Results

The demographic data is presented in Table 1. Among 345 useable ques-tionnaires, 200 (57.1%) were from male respondents while 150 were from female respondents (42.9%). Additionally, 106 (30.3%) were 28–31 years old, 194 (55.4%) were 32–38 years old, 25 (7.1%) were 39–45 years old, 17 (4.9%) were 46–50 years old, and 8 (2.3%) were more than 51 years old. With regards to academic level, 273 (78.0%) were lecturers and 77 (22.0%) were senior lecturers. Finally, regarding the respondents’ faculties, 101 (28.9%) were from the faculty of education, 89 (25.4%) were from the faculty of art, 62 (17.7%) were from the faculty of law, and 98 (28.0%) were from the faculty of management, see Table 1.

4.1. Structured Equation Modeling

In order to test the online teaching abilities and competencies utilizing digital technologies during the COVID-19 pandemic, structural equation modelling (SEM), AMOS-SEM, and SPSS were used. As a result, SPSS 23 was used to evaluate quantitative data. In this study, descriptive statistics were used to describe the characteristics of the participants, and AMOS-SEM was used to investigate hypotheses and significant predictors in the extended TAM model, which included nine variables for actual use of digital tools for teaching and learning as well as outcome expectancy variables. Additionally, to ensure that the scales were meaningful, all factors were examined by Cronbach’s alpha as shown by [78], see Table 2.

4.2. Model Fit Mesuerment

The CMN/DF ratio in Table 3 is 2.265, which is lower than the necessary threshold (5.00). The RMR value is below the threshold of 0.33 (0.05), AGFI (0.914) is a valid value, GFI (0.931) is a valid value, CFI (0.959) is a valid value, TLI (0.952) is a valid value, IFI (0.959) is a valid value, RFI (0.934) is a valid value, and NFI (0.943) is a valid value as suggested by [78]. Figure 2 shows all items and factor values. This shows that the measurement model was acceptable and well-suited to the structural model. See Table 3 and Figure 2.

4.3. Reliability, Validity, and Measurement Model

The SEM-AMOS measurement model for each idea has its own set of characteristics, such as reliability and validity. The structural model was used to investigate the intensity of the link direction using confirmatory factor analysis (CFA) and model fit. The survey’s psychometric qualities are evaluated in order to determine its reliability, validity, and collinearity. The variables’ reliability was investigated using item loadings, Cronbach’s alpha (CA), and composite reliability (CR), with acceptance values of (>0.7). Table 3 shows that all of the constructs’ Cronbach’s alpha (CA) values varied from 0.877 to 0921, the CR values ranged from (0.889 to 0.928), and the loadings of the indicators’ items ranged from (0.711–0.885), showing strong internal consistency [78]. Loadings of indicators, average variance extracted (AVE) values, and the square roots of AVE values were used to assess convergent and discriminant validity (0.591–0.672). Table 4 shows the results.

4.4. Measurement Discriminant and Convergent Validity

Table 5 displays the Fornell–Larcker criteria for determining discriminant validity, where the square roots of AVE for all items were larger than their correlations with other components [79]. The AVEs’ square roots are on the diagonal. Table 5 further reveals that at the 0.80 threshold, when applying the heterotrait-monotrait criteria, which imposes a stricter examination, all constructs are clearly different [80].

4.5. Evaluation of the Structural Model

This study’s structural model included nine components with two degrees of linkages. As a consequence, the model hypotheses depicted in Figure 3 is evaluated using the AMOS approach, which uses bootstrap resampling to examine each hypothesis path’s significance levels. Table 5 illustrates the hypothesis testing findings, with a two-tail test used to establish the statistical significance of each hypothesis (0.000). Furthermore, the path coefficient value estimate (β), which measures the association between variables based on their degree of significance (p value), is significant when p = 0.05, see Figure 3.

4.6. Hypotheses Testing Results

Based on the results shown in Figure 3 and Table 6, the relationship between perceived teaching self-efficacy and perceived ease of use (β = 0.253; C.R = 11.045, p < 0.000), the relationship between perceived teaching self-efficacy and perceived usefulness (β = 0.054; C.R = 2.034, p < 0.000), and the relationship between perceived teaching self-efficacy and perceived enjoyment (β = 0.585; C.R = 26.657, p < 0.000) were accepted. Similarly, the relationship between perceived enjoyment and perceived ease of use (β = 0.216; C.R = 8.945, p < 0.000) and the relationship between perceived enjoyment and perceived usefulness (β = 0.109; C.R = 4.529, p < 0.000) were accepted. Additionally, the relationship between online teaching skills and perceived ease of use (β = 0.235; C.R = 10.246, p < 0.000) and the relationship between online teaching skills and perceived usefulness (β = 0.255; C.R = 11.068, p < 0.000) were accepted. Moreover, the relationship between digital tools access and perceived ease of use (β = 0.143; C.R = 5.817, p < 0.000) and the relationship between digital tools access and perceived usefulness (β = 0.093; C.R = 3.885, p < 0.000) were accepted. Furthermore, the relationship between perceived ease of use and perceived usefulness (β = 0.540; C.R = 20.167, p < 0.000), the relationship between perceived ease of use and attitude toward using digital tools (β = 0.324; C.R = 9.712, p < 0.000), and the relationship between perceived ease of use and behavioral intention to use digital tools (β = 0.273; C.R = 9.818, p < 0.000) were accepted. Additionally, the relationship between perceived usefulness and attitude toward using digital tools (β = 0.306; C.R = 10.405, p < 0.000) was accepted) and the relationship between perceived usefulness and behavioral intention to use digital tools (β = 0.216; C.R = 8.773, p < 0.000) were accepted. Also, the relationship between attitude toward using digital tools and behavioral intention to use digital tools (β = 0.322; C.R = 14.459, p < 0.000) and the relationship between attitude toward using digital tools and actual use of digital tools for teaching and learning (β = 0.382; C.R = 17.234, p < 0.000) were accepted. Finally, the relationship between behavioral intention to use digital tools and actual use of digital tools for teaching and learning (β = 0.482; C.R = 20.941, p < 0.000) was accepted.

5. Factors Described and Analyzed

The standard deviation (SD) and mean are two statistics that describe how measurements in a population deviate from the average (mean) or expected value. The majority of the data points are close to the mean when the standard deviation is low. The data is more distributed if the standard deviation is high. As a consequence, as shown in Figure 4, all values were accepted and the majority was agree and strongly agree, meaning that the critical factors that influence online teaching skills and competencies by using digital tools during the COVID-19 pandemic through exploring critical factors that influence lecturers’ actual use of digital tools for teaching and learning in Saudi Arabia’s higher education were perceived teaching self-efficacy, perceived enjoyment, online teaching skills, digital tools access, perceived ease of use, perceived usefulness, attitude toward using digital tools, behavioral intention to use digital tools, and actual use of digital tools for teaching and learning; see Figure 4.

5.1. Discussion and Implications

This study aims to measure the role of lecturers’ attitudes toward using digital tools, behavioral intentions to use digital tools, and actual usage of digital technologies for online teaching and learning as an external variable to the TAM and assess the impact of perceived teaching self-efficacy, perceived enjoyment, online teaching skills, and digital tools access as an external variable to the TAM. In this research, all of the hypotheses were found to be acceptable. This is in line with prior studies, which found that attitude toward technology usage was a strong predictor of desire to utilize technology, particularly under obligatory settings [24,25,81]. Therefore, this research differs from others in that it was performed during the COVID-19 epidemic and looked into lecturers who used digital technology to improve their online teaching skills. As a result, lecturers’ attitudes about utilizing digital tools, behavioral intentions to use digital tools, and actual usage of digital technologies for online teaching and learning must all be examined. We intended to improve our understanding of the characteristics that influence lecturers’ usage of online teaching, especially for professors who have prior experience with utilizing technology to build online teaching abilities. Therefore, to achieve the research goal, the TAM model [25] was used together with an external component of instructor experience in using technology for the development of online teaching skills. The hypothesis of a directed relationship between TAM scales and external influences was investigated. Thus, the COVID-19 pandemic was a significant driving force for this study. Furthermore, lecturers’ attempts to move from traditional teaching techniques to a more tech-oriented approach cause problems [82]. A number of lecturers discussed the nature of the instructional application as well as external and internal issues. As a result, a new category known as developing difficulties has emerged, which is tied to the limitations that teachers face while using Google Classroom. Furthermore, because this style of teaching demands quick and stable connectivity to permit interaction between the student and the educator, the efficiency of remote learning is significantly hampered by unreliable internet access [83]. According to a recent systematic analysis, the new educational standard has had an impact on the United Nations’ Sustainable Development Goals, with an increased risk of sustainability in tertiary education [84]. When they required it, however, students received insufficient social assistance and security protection from other students and their professors [85]. Furthermore, universities face challenges in maintaining course content consistency and relevance, communicating clearly with the academic community, and acquiring and recruiting students [86]. The concept of supporting evidence-based practices to promote scholarly teaching practices is also a challenge [87]. As a result, higher education institutions must guarantee that education is inclusive, egalitarian, and of high quality in order to bridge the digital gap and promote sustainable activities [88]. This involves utilizing digital tools to provide learners with focused and timely feedback as well as analyzing data on learner performance to influence future teaching tactics. Because of their student status during COVID-19 pandemic, the lecturers in this study may not have considered their use of technology to be required, but they use digital technologies to test online teaching skills and competencies during the COVID-19 pandemic. Despite the fact that lecturers were expected to take courses on various aspects of technology usage and were encouraged to use technology for their learning and assignments, their interactions with technology as lecturers, attitude toward using digital tools, behavioral intention to use digital tools, and actual use of digital tools for teaching and learning were examined in three ways.
(1)
First, lecturers have more time to experiment with technology throughout training, and they have more self-efficacy in instructing.
(2)
Second, lecturers utilize technology primarily to improve their online teaching abilities and prepare for their future profession as a university professor.
(3)
Third, lecturers have additional assistance for using technology as part of their instruction as well as access to digital tools.
The results of this study show that perceived teaching self-efficacy, perceived enjoyment, online teaching skills, and digital tool access have a significant impact on the TAM factors’ core constructs; perceived ease of use, perceived usefulness, lecturers’ attitude toward using digital tools, and lecturers’ behavioral intention to use digital tools, in turn, affect actual use of digital tools for teaching and learning (see Figure 3 and Table 6). As a consequence, this research shows that adopting digital tools during the COVID-19 epidemic improved online teaching abilities and competences. Furthermore, our research has resulted in the creation of a validated instrument to assess lecturers’ attitudes about utilizing digital tools, behavioral intentions to use digital tools, and actual usage of digital tools in higher education teaching and learning.
Finally, the scientific contributions as follows:
  • Regarding the independent factor hypotheses on the actual use of digital tools for teaching and learning by lecturers at universities; perceived teaching self-efficacy, perceived enjoyment, online teaching skills, and digital tools access were found to affect perceived usefulness and perceived ease of use digital tools for teaching.
  • Regarding the mediator factor hypotheses on the on the actual use of digital tools for teaching and learning by lecturers at universities; perceived usefulness, and perceived ease of use digital tools for teaching were found to affect lecturers’ attitude toward using digital tools and lecturers’ behavioral intention to use digital tools for teaching.
  • Regarding the mediator factor hypotheses on the on the actual use of digital tools for teaching and learning by lecturers at universities; lecturers’ attitudes toward using digital tools were found to affect lecturers’ behavioral intention to use digital tools for teaching.
  • Regarding the dependent factors hypotheses on the actual use of digital tools for teaching and learning by lecturers at universities; lecturers’ behavioral intention to use digital tools was found to affects actual use of digital tools for teaching and learning.

5.2. Conclusions and Future Work

In conclusion, this study discovered three main challenges that impacted the use of digital technologies for online teaching skills and competencies during the COVID-19 pandemic: (a) lecturers voiced their unbiased concerns about digital tools access (Internet and technology); (b) lecturers discussed the challenges of equity and the concern they had for social interaction and peer connection during the pandemic; nonetheless, they viewed this attempt as a challenge, and they were eager to promote perceived teaching self-efficacy; (c) the lecturers were concerned that students’ motivation and engagement would be lower in the online learning environment. This research has some limitations as well. Because this study only considered one public university in Saudi Arabia, the sample size was very small. As a consequence, future studies might include more public and private higher education institutions, as well as student and lecturers’ perspectives on the usage of digital tools in the classroom. Qualitative data, such as interviews, must be incorporated in order to properly analyze and explain findings. Future research on the use of digital technology to improve online teaching abilities should concentrate on lecturers’ interactions and engagement as well as how pedagogy and course designs impact their willingness to employ digital technologies in the classroom.

Funding

This work was supported through the Annual Funding track by the Deanship of Scientific Research, Vice Presidency for Graduate Studies and Scientific Research, King Faisal University, Saudi Arabia [Project No. AN00076].

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The author declares no conflict of interest.

References

  1. UNESCO. COVID-19 Educational Disruption and Response. 2021. Available online: https://en.unesco.org/themes/education-emergencies/coronavirus-school-closures (accessed on 13 December 2021).
  2. Cucinotta, D.; Vanelli, M. WHO declares COVID-19 a pandemic. Acta Bio Med. Atenei Parm. 2020, 91, 157. [Google Scholar]
  3. Mailizar, M.; Almanthari, A.; Maulina, S.; Bruce, S. Secondary School Mathematics Teachers’ Views on E-learning Implementation Barriers during the COVID-19 Pandemic: The Case of Indonesia. Eurasia J. Math. Sci. Technol. Educ. 2020, 16, em1860. [Google Scholar] [CrossRef]
  4. Ministry of Education in Saudi Arabia (MOE). 2020. Available online: https://moe.gov.sa/en/LifeEvents/Pages/default.aspx (accessed on 12 February 2022).
  5. Sailer, M.; Schultz-Pernice, F.; Fischer, F. Contextual facilitators for learning activities involving technology in higher education: The Cβ-model. Comput. Hum. Behav. 2021, 121, 106794. [Google Scholar] [CrossRef]
  6. Lee, S.W.-Y.; Tsai, C.-C. Students’ perceptions of collaboration, self-regulated learning, and information seeking in the context of Internet-based learning and traditional learning. Comput. Hum. Behav. 2011, 27, 905–914. [Google Scholar] [CrossRef]
  7. Targamadzė, A.; Petrauskienė, R. Impact of information technologies on modern learning. Inf. Technol. Control. 2010, 39, 169–175. [Google Scholar]
  8. Allen, I.E.; Seaman, J. Grade Change: Tracking Online Education in the United States; Babson Survey Research Group and Quahog Research Group: New York, NY, USA, 2014. [Google Scholar]
  9. Daumiller, M.; Rinas, R.; Hein, J.; Janke, S.; Dickhäuser, O.; Dresel, M. Shifting from face-to-face to online teaching during COVID-19: The role of university faculty achievement goals for attitudes towards this sudden change, and their relevance for burnout/engagement and student evaluations of teaching quality. Comput. Hum. Behav. 2021, 118, 106677. [Google Scholar] [CrossRef]
  10. Moazami, F.; Bahrampour, E.; Azar, M.R.; Jahedi, F.; Moattari, M. Comparing two methods of education (virtual versus traditional) on learning of Iranian dental students: A post-test only design study. BMC Med Educ. 2014, 14, 45. [Google Scholar] [CrossRef] [Green Version]
  11. Al-Rahmi, W.; Yahaya, N.; Alamri, M.; Aljarboa, N.A.; Kamin, Y.; Moafa, F.A. A Model of Factors Affecting Cyber Bullying Behaviors among University Students. IEEE Access 2018, 7, 2978–2985. [Google Scholar] [CrossRef]
  12. Basak, T.; Yildiz, D. Comparison of the effects of cooperative learning and traditional learning methods on the im-provement of drug-dose calculation skills of nursing students undergoing internships. Health Educ. J. 2014, 73, 341–350. [Google Scholar] [CrossRef]
  13. Almulla, M.A. The Efficacy of Employing Problem-Based Learning (PBL) Approach as a Method of Facilitating Students’ Achievement. IEEE Access 2019, 7, 146480–146494. [Google Scholar] [CrossRef]
  14. Deo, S.K. Human Resources and Logistic Requirements in Problem Based Learning Compared to Traditional Learning. Nepal Orthop. Assoc. J. 2014, 3, 46–47. [Google Scholar] [CrossRef]
  15. Almulla, M.A. The Effectiveness of the Project-Based Learning (PBL) Approach as a Way to Engage Students in Learning. SAGE Open 2020, 10, 3. [Google Scholar] [CrossRef]
  16. Ibik, A.S.; Yalçin, N. The effect of project based learning supported with analogies method on success and under-standing level for electric current concept. J. Turk. Sci. Educ. TUSED 2013, 10, 123–136. [Google Scholar]
  17. Liao, Y.K.C. Game-based learning verse traditional instruction on student affective outcomes in Taiwan: A meta-analysis. J. Inf. Technol. Appl. 2011, 5, 28–36. [Google Scholar]
  18. Ronimus, M.; Kujala, J.; Tolvanen, A.; Lyytinen, H. Children’s engagement during digital game-based learning of reading: The effects of time, rewards, and challenge. Comput. Educ. 2014, 71, 237–246. [Google Scholar] [CrossRef]
  19. Palloff, R.M.; Pratt, K. Lessons from the Virtual Classroom: The Realities of Online Teaching; John Wiley & Sons: Oxford, UK, 2013. [Google Scholar]
  20. Moore, J.L.; Dickson-Deane, C.; Galyen, K. e-Learning, online learning, and distance learning environments: Are they the same? Internet High. Educ. 2011, 14, 129–135. [Google Scholar] [CrossRef]
  21. Moafa, F.A.; Ahmad, K.; Al-Rahmi, W.; Yahaya, N.; Kamin, Y.; Alamri, M.M. Develop a Model to Measure the Ethical Effects of Students Through Social Media Use. IEEE Access 2018, 6, 56685–56699. [Google Scholar] [CrossRef]
  22. Rudestam, K.E.; Schoenholtz-Read, J. Handbook of Online Learning; SAGE Publications: Thousand Oaks, CA, USA, 2010. [Google Scholar]
  23. Cigdem, H.; Topcu, A. Predictors of instructors’ behavioral intention to use learning management system: A Turkish vocational college example. Comput. Hum. Behav. 2015, 52, 22–28. [Google Scholar] [CrossRef]
  24. Davis, F.D. User acceptance of information technology: System characteristics, user perceptions and behavioral impacts. Int. J. Man-Machine Stud. 1993, 38, 475–487. [Google Scholar] [CrossRef] [Green Version]
  25. Davis, F.D.; Bagozzi, R.P.; Warshaw, P.R. User acceptance of computer technology: A comparison of two theoretical models. Manag. Sci. 1989, 35, 982–1003. [Google Scholar] [CrossRef] [Green Version]
  26. Chang, C.-T.; Hajiyev, J.; Su, C.-R. Examining the students’ behavioral intention to use e-learning in Azerbaijan? The General Extended Technology Acceptance Model for E-learning approach. Comput. Educ. 2017, 111, 128–143. [Google Scholar] [CrossRef]
  27. Al-Rahmi, A.M.; Al-Rahmi, W.M.; Alturki, U.; Aldraiweesh, A.; Almutairy, S.; Al-Adwan, A.S. Exploring the factors affecting mobile learning for sustainability in higher education. Sustainability 2021, 13, 7893. [Google Scholar] [CrossRef]
  28. Cardullo, V.; Wang, C.-H.; Burton, M.; Dong, J. K-12 teachers’ remote teaching self-efficacy during the pandemic. J. Res. Innov. Teach. Learn. 2021, 14, 32–45. [Google Scholar] [CrossRef]
  29. Bandura, A.; Barbaranelli, C.; Caprara, G.V.; Pastorelli, C. Multifaceted impact of self-efficacy beliefs on academic functioning. Child Dev. 1996, 67, 1206–1222. [Google Scholar] [CrossRef] [PubMed]
  30. Humphries, C.A.; Hebert, E.; Daigle, K.; Martin, J. Development of a Physical Education Teaching Efficacy Scale. Meas. Phys. Educ. Exerc. Sci. 2012, 16, 284–299. [Google Scholar] [CrossRef]
  31. Klassen, R.M.; Tze, V.M.C.; Betts, S.M.; Gordon, K.A. Teacher Efficacy Research 1998–2009: Signs of Progress or Unfulfilled Promise? Educ. Psychol. Rev. 2019, 23, 21–43. [Google Scholar] [CrossRef]
  32. Gibson, S.; Dembo, M.H. Teacher efficacy: A construct validation. J. Educ. Psychol. 1984, 76, 569. [Google Scholar] [CrossRef]
  33. Schechter, A. Political and Technology Efficacy among Millennials. Ph.D. Thesis, University of Delaware, Newark, DE, USA, 2013. [Google Scholar]
  34. Venkatesh, V. Determinants of Perceived Ease of Use: Integrating Control, Intrinsic Motivation, and Emotion into the Technology Acceptance Model. Inf. Syst. Res. 2000, 11, 342–365. [Google Scholar] [CrossRef] [Green Version]
  35. Teo, T.S.; Lim, V.K.; Lai, R.Y. Intrinsic and extrinsic motivation in Internet usage. Omega 1999, 27, 25–37. [Google Scholar] [CrossRef]
  36. Bawane, J.; Spector, J.M. Prioritization of online instructor roles: Implications for competency-based teacher educa-tion programs. Distance Educ. 2009, 30, 383–397. [Google Scholar] [CrossRef]
  37. Alman, S.W.; Tomer, C. Designing Online Learning: A Primer for Librarians; ABC-CLIO: Santa Barbara, CA, USA, 2012. [Google Scholar]
  38. Baran, E.; Correia, A.-P.; Thompson, A.D. Tracing Successful Online Teaching in Higher Education: Voices of Exemplary Online Teachers. Teach. Coll. Rec. Voice Sch. Educ. 2013, 115, 1–41. [Google Scholar] [CrossRef]
  39. Baran, E.; Correia, A.-P. A professional development framework for online teaching. TechTrends 2014, 58, 95–101. [Google Scholar] [CrossRef]
  40. Carril, P.C.M.; Sanmamed, M.G.; Hernández-Sellés, N. Pedagogical roles and competencies of university teachers practicing in the e-learning environment. Int. Rev. Res. Open Distrib. Learn. 2013, 14, 462–487. [Google Scholar] [CrossRef] [Green Version]
  41. ElSaheli-Elhage, R. Access to Students and Parents and Levels of Preparedness of Educators during the COVID-19 Emergency Transition to e-Learning. Int. J. Stud. Educ. 2021, 3, 61–69. [Google Scholar] [CrossRef]
  42. Montiel, I.; Delgado-Ceballos, J.; Ortiz-De-Mandojana, N.; Antolin-Lopez, R. New Ways of Teaching: Using Technology and Mobile Apps to Educate on Societal Grand Challenges. J. Bus. Ethic. 2020, 161, 243–251. [Google Scholar] [CrossRef]
  43. Ferdig, R.E.; Baumgartner, E.; Hartshorne, R.; Kaplan-Rakowski, R.; Mouza, C. (Eds.) Teaching, Technology, and Teacher Education during the COVID-19 Pandemic: Stories from the Field; Association for the Advancement of Computing in Education: Waynesvilla, NC, USA, 2020. [Google Scholar]
  44. Moreira-Fontán, E.; García-Señorán, M.; Conde-Rodríguez, A.; González, A. Teachers’ ICT-related self-efficacy, job resources, and positive emotions: Their structural relations with autonomous motivation and work engagement. Comput. Educ. 2019, 134, 63–77. [Google Scholar] [CrossRef]
  45. Hew, K.F.; Brush, T. Integrating technology into K-12 teaching and learning: Current knowledge gaps and rec-ommendations for future research. Educ. Technol. Res. Dev. 2007, 55, 223–252. [Google Scholar] [CrossRef]
  46. Williams, C. Enter the Classroom of 2018 [WWW Document]. App Store. Available online: https://itunes.apple.com/gb/story/id1427202128 (accessed on 12 October 2021).
  47. Ertmer, P.A.; Ottenbreit-Leftwich, A.T. Teacher technology change: How knowledge, confidence, beliefs, and culture intersect. J. Res. Technol. Educ. 2010, 42, 255–284. [Google Scholar] [CrossRef]
  48. Vongkulluksn, V.W.; Xie, K.; Bowman, M.A. The role of value on teachers’ internalization of external barriers and externalization of personal beliefs for classroom technology integration. Comput. Educ. 2018, 118, 70–81. [Google Scholar] [CrossRef]
  49. Ertmer, P.A. Responsive instructional design: Scaffolding the adoption and change process. Educ. Technol. 2001, 41, 33–38. [Google Scholar]
  50. Sadaf, A.; Newby, T.J.; Ertmer, P.A. An investigation of the factors that influence preservice teachers’ intentions and integration of Web 2.0 tools. Educ. Technol. Res. Dev. 2015, 64, 37–64. [Google Scholar] [CrossRef]
  51. Lin, K.-M.; Chen, N.-S.; Fang, K. Understanding e-learning continuance intention: A negative critical incidents perspective. Behav. Inf. Technol. 2011, 30, 77–89. [Google Scholar] [CrossRef]
  52. Wu, B.; Zhang, C. Empirical study on continuance intentions towards E-Learning 2.0 systems. Behav. Inf. Technol. 2014, 33, 1027–1038. [Google Scholar] [CrossRef]
  53. Chang, C.-C.; Yan, C.-F.; Tseng, J.-S. Perceived convenience in an extended technology acceptance model: Mobile technology and English learning for college students. Australas. J. Educ. Technol. 2012, 28. [Google Scholar] [CrossRef] [Green Version]
  54. Briz-Ponce, L.; García-Peñalvo, F.J. An Empirical Assessment of a Technology Acceptance Model for Apps in Medical Education. J. Med. Syst. 2015, 39, 176. [Google Scholar] [CrossRef] [PubMed]
  55. Hamid, A.A.; Razak, F.Z.A.; Abu Bakar, A.; Abdullah, W.S.W. The Effects of Perceived Usefulness and Perceived Ease of Use on Continuance Intention to Use E-Government. Procedia Econ. Financ. 2016, 35, 644–649. [Google Scholar] [CrossRef] [Green Version]
  56. Mou, J.; Shin, D.-H.; Cohen, J. Understanding trust and perceived usefulness in the consumer acceptance of an e-service: A longitudinal investigation. Behav. Inf. Technol. 2016, 36, 125–139. [Google Scholar] [CrossRef]
  57. Teo, T. A path analysis of pre-service teachers’ attitudes to computer use: Applying and extending the technology ac-ceptance model in an educational context. Interact. Learn. Environ. 2010, 18, 65–79. [Google Scholar] [CrossRef]
  58. Wong, G.K.W. Understanding technology acceptance in pre-service teachers of primary mathematics in Hong Kong. Australas. J. Educ. Technol. 2015, 31, 6. [Google Scholar] [CrossRef]
  59. Kaplan, K.J. On the ambivalence-indifference problem in attitude theory and measurement: A suggested modification of the semantic differential technique. Psychol. Bull. 1972, 77, 361–372. [Google Scholar] [CrossRef]
  60. Cheung, R.; Vogel, D. Predicting user acceptance of collaborative technologies: An extension of the technology ac-ceptance model for e-learning. Comput. Educ. 2013, 63, 160–175. [Google Scholar] [CrossRef]
  61. Tosuntas, S.B.; Karadağ, E.; Orhan, S. The factors affecting acceptance and use of interactive whiteboard within the scope of FATIH project: A structural equation model based on the Unified Theory of acceptance and use of technology. Comput. Educ. 2015, 81, 169–178. [Google Scholar] [CrossRef]
  62. Keong, Y.C.; Albadry, O.; Raad, W. Behavioral Intention of EFL Teachers to Apply E-Learning. J. Appl. Sci. 2014, 14, 2561–2569. [Google Scholar] [CrossRef] [Green Version]
  63. Alharbi, S.; Drew, S. Using the technology acceptance model in understanding academics’ behavioural intention to use learning management systems. Int. J. Adv. Comput. Sci. Appl. 2014, 5, 143–155. [Google Scholar] [CrossRef]
  64. Tarhini, A.; Elyas, T.; Akour, M.A.; Al-Salti, Z. Technology, Demographic Characteristics and E-Learning Acceptance: A Conceptual Model Based on Extended Technology Acceptance Model. High. Educ. Stud. 2016, 6, 72. [Google Scholar] [CrossRef] [Green Version]
  65. Hussein, Z. Leading to Intention: The Role of Attitude in Relation to Technology Acceptance Model in E-Learning. Procedia Comput. Sci. 2017, 105, 159–164. [Google Scholar] [CrossRef]
  66. Taat, M.S.; Francis, A. Factors Influencing the Students’ Acceptance of E-Learning at Teacher Education Institute: An Exploratory Study in Malaysia. Int. J. High. Educ. 2019, 9, 133. [Google Scholar] [CrossRef] [Green Version]
  67. Poon, J. Blended learning: An institutional approach for enhancing students’ learning experiences. J. Online Learn. Teach. 2013, 9, 271–288. [Google Scholar]
  68. Orlando, J.; Attard, C. Digital natives come of age: The reality of today’s early career teachers using mobile devices to teach mathematics. Math. Educ. Res. J. 2015, 28, 107–121. [Google Scholar] [CrossRef]
  69. Lee, C.; Yeung, A.S.; Ip, T. Use of computer technology for English language learning: Do learning styles, gender, and age matter? Comput. Assist. Lang. Learn. 2015, 29, 1035–1051. [Google Scholar] [CrossRef]
  70. Durmaz, A.; Dicle, A.; Cakan, E.; Cakir, S. Effect of screen-based computer simulation on knowledge and skill in nursing students’ learning of preoperative and postoperative care management: A randomized controlled study. CIN Computer Inform. Nurs. 2012, 30, 196–203. [Google Scholar] [CrossRef] [PubMed]
  71. Biggs, J.; Tang, C. Teaching for Quality Learning at University; McGraw-Hill Education: London, UK, 2011. [Google Scholar]
  72. Al-Rahmi, A.M.; Shamsuddin, A.; Alturki, U.; Aldraiweesh, A.; Yusof, F.M.; Al-Rahmi, W.M.; Aljeraiwi, A.A. The Influence of Information System Success and Technology Acceptance Model on Social Media Factors in Education. Sustainability 2021, 13, 7770. [Google Scholar] [CrossRef]
  73. Al-Rahmi, W.M.; Alkhalaf, S. An empirical investigation of adoption Big Data in higher education sustainability. Entrep. Sustain. Issues 2021, 9, 108–122. [Google Scholar] [CrossRef]
  74. Likhachev, S.; Likhacheva, T.; Silchenkova, L.; Krivorotova, E.; Plesnik, L. Controlling Research Activity of Students by Digital Tools. SHS Web Conf. 2020, 79, 01010. [Google Scholar] [CrossRef]
  75. Weaver, J.C.; Matney, G.; Goedde, A.M.; Nadler, J.R.; Patterson, N. Digital tools to promote remote lesson study. Int. J. Lesson Learn. Stud. 2021, 10, 187–201. [Google Scholar] [CrossRef]
  76. Fraenkel, J.R.; Wallen, N.E.; Hyun, H.H. How to Design Research in Education and Evaluate Research in Education [Internet]. Kiefer, S., Ed. 2011. Available online: https://saochhengpheng.files.wordpress.com/2017/03/jack_fraenkel_norman_wallen_helen_hyun-how_to_design_and_evaluate_research_in_education_8th_edition_-mcgraw-hill_humanities_social_sciences_languages2011.pdf (accessed on 2 February 2022).
  77. Shank, G.; Brown, L. Exploring Educational Research Literacy; Routledge: London, UK, 2013. [Google Scholar] [CrossRef]
  78. Hair, J.F.; Ringle, C.M.; Sarstedt, M. Partial Least Squares: The Better Approach to Structural Equation Modeling? Long Range Plan. 2012, 45, 312–319. [Google Scholar] [CrossRef]
  79. Fornell, C.; Larcker, D.F. Evaluating structural equation models with unobservable variables and measurement error. J. Mark. Res. 1981, 48, 39–50. [Google Scholar] [CrossRef]
  80. Henseler, J.; Ringle, C.M.; Sarstedt, M. A new criterion for assessing discriminant validity in variance-based structural equation modeling. J. Acad. Mark. Sci. 2015, 43, 115–135. [Google Scholar] [CrossRef] [Green Version]
  81. Venkatesh, V.; Morris, M.G.; Davis, G.B.; Davis, F.D. User acceptance of information technology: Toward a unified view. MIS Q. 2003, 27, 425–478. [Google Scholar] [CrossRef] [Green Version]
  82. Al-Azawei, A.; Parslow, P.; Lundqvist, K. The Effect of Universal Design for Learning (UDL) Application on E-learning Acceptance: A Structural Equation Model. Int. Rev. Res. Open Distrib. Learn. 2017, 18, 54–87. [Google Scholar] [CrossRef]
  83. Hussein, M.H.; Ow, S.H.; Ibrahim, I.; Mahmoud, M.A. Measuring instructors continued intention to reuse Google Classroom in Iraq: A mixed-method study during COVID-19. Interact. Technol. Smart Educ. 2020, 18, 380–402. [Google Scholar] [CrossRef]
  84. Crawford, J.; Cifuentes-Faura, J. Sustainability in Higher Education during the COVID-19 Pandemic: A Systematic Review. Sustainability 2022, 14, 1879. [Google Scholar] [CrossRef]
  85. Cifuentes-Faura, J.; Obor, D.O.; To, L.; Al-Naabi, I. Cross-cultural impacts of COVID-19 on higher education learning and teaching practices in Spain, Oman, Nigeria and Cambodia: A cross-cultural study. J. Univ. Teach. Learn. Pr. 2021, 18, 135–151. [Google Scholar] [CrossRef]
  86. Marinoni, G.; Van’t Land, H.; Jensen, T. The impact of COVID-19 on higher education around the world. IAU Glob. Surv. Rep. 2020, 23, 1–50. [Google Scholar]
  87. García-Morales, V.J.; Garrido-Moreno, A.; Martín-Rojas, R. The Transformation of Higher Education after the COVID Disruption: Emerging Challenges in an Online Learning Scenario. Front. Psychol. 2021, 12, 616059. [Google Scholar] [CrossRef] [PubMed]
  88. Faura-Martínez, U.; Lafuente-Lechuga, M.; Cifuentes-Faura, J. Sustainability of the Spanish university system during the pandemic caused by COVID-19. Educ. Rev. 2021, 1–19. [Google Scholar] [CrossRef]
Figure 1. Research model and hypotheses.
Figure 1. Research model and hypotheses.
Sustainability 14 05455 g001
Figure 2. Measurement model.
Figure 2. Measurement model.
Sustainability 14 05455 g002
Figure 3. Path and T-values results.
Figure 3. Path and T-values results.
Sustainability 14 05455 g003
Figure 4. Factors described and analyzed.
Figure 4. Factors described and analyzed.
Sustainability 14 05455 g004
Table 1. Demographic Data.
Table 1. Demographic Data.
GenderFactorsFrequencyPercentLevel of studyFactorsFrequencyPercent
Male 20057.1Lecturer27378.0
Female 15042.9Senior lecturer7722.0
Total350100.0Total350100.0
Age28–3110630.3FacultyEducation10128.9
32–3819455.4Art8925.4
39–45257.1Law6217.7
46–50174.9Management9828.0
>5182.3Total350100.0
Total350100.0
UniversityKFU20358.0
KSU14742.0
Total350100.0
Table 2. The Reliability Coefficient for All Variables.
Table 2. The Reliability Coefficient for All Variables.
FactorsCodePilot TestFinal Test
Perceived Teaching Self-EfficacyPTSE0.7130.896
Perceived EnjoymentPE0.7540.907
Online Teaching SkillsOTS0.7290.921
Digital Tools AccessDTA0.8040.899
Perceived UsefulnessPU0.7990.877
Perceived Ease of UsePEU0.7920.900
Behavioral Intention to Use Digital ToolsBIUD0.8170.912
Attitude Toward Using Digital ToolsATUD0.7220.891
Actual Use of Digital Tools for Teaching and LearningAUDT0.7290.918
Table 3. Model fit evaluation.
Table 3. Model fit evaluation.
Model FitNFIRFIIFITLICFIGFIAGFIRMR
Default model0.9430.9340.9590.9520.9590.9310.9140.033
Saturated model1.000 1.000 1.0001.000 0.000
Independence model0.0000.0000.0000.0000.0000.1570.1010.331
Table 4. Reliability, validity, and measurement model.
Table 4. Reliability, validity, and measurement model.
No EstimateCRCAAVE
1PTES1<---Perceived
Teaching
Self-Efficacy
0.7300.8890.8960.591
2PTES2<---0.779
3PTES3<---0.758
4PE1<---Perceived
Enjoyment
0.7920.9240.9070.611
5PE2<---0.865
6PE3<---0.802
7PE4<---0.720
8OTK1<---Online
Teaching
Skills
0.7580.9280.9210.632
9OTK2<---0.754
10OTK3<---0.731
11DTA1<---Digital
Tools
Access
0.7530.9020.8990.672
12DTA2<---0.820
13DTA3<---0.813
14PU1<---Perceived
Usefulness
0.7830.8910.8770.642
15PU2<---0.885
16PU3<---0.855
17PEU1<---Perceived
Ease of Use
0.7540.9080.9000.598
18PEU2<---0.772
19PEU3<---0.786
20BIUD1<---Behavioral
Intention
to Use
Digital Tools
0.7110.9230.9120.672
21BIUD2<---0.778
22BIUD3<---0.860
23BIUD4<---0.850
24ATUD1<---Attitude Toward
Using Digital
Tools
0.7760.9070.8910.643
25ATUD2<---0.864
26ATUD3<---0.852
27AUDT1<---Actual Use of
Digital Tools
for Teaching
and Learning
0.7520.9050.9180.668
28AUDT2<---0.824
29AUDT3<---0.857
30AUDT4<---0.814
31AUDT5<---0.838
Table 5. Discriminant and convergent validity.
Table 5. Discriminant and convergent validity.
FactorsPTSEPEOTSDTAPUPEUBIUDATUDAUDT
Perceived Teaching Self-Efficacy0.855
Perceived Enjoyment0.2700.888
Online Teaching Skills0.3690.2890.837
Digital Tools Access0.2910.3440.3130.871
Perceived Usefulness0.2970.3290.3430.3250.816
Perceived Ease of Use0.3510.3420.4300.3610.4450.863
Behavioral Intention to Use Digital Tools0.2820.3330.3030.3810.3030.3470.822
Attitude Toward Using Digital Tools0.2820.3300.3280.3480.3350.3770.3260.883
Actual Use of Digital Tools for Teaching and Learning0.3080.3670.3400.3390.3550.3940.3570.3580.903
Table 6. Hypothesis testing.
Table 6. Hypothesis testing.
NoRelationships Estimate (β)S.E.C.R.pResults
H1PEU<---PTSE0.2530.02311.0450.000Accepted
H2PU<---PTSE0.0540.0232.0340.002Accepted
H3PE<---PTSE0.5850.02226.6570.000Accepted
H4PEU<---PE0.2160.0248.9450.000Accepted
H5PU<---PE0.1090.0244.5290.000Accepted
H6PEU<---OTS0.2350.02310.2460.000Accepted
H7PU<---OTS0.2550.02311.0680.000Accepted
H8PEU<---DTA0.1430.0255.8170.000Accepted
H9PU<---DTA0.0930.0243.8850.000Accepted
H10PU<---PEU0.5400.02720.1670.000Accepted
H11AT<---PEU0.3240.0339.7120.000Accepted
H12BIU<---PEU0.2730.0289.8180.000Accepted
H13AT<---PU0.3060.02910.4050.000Accepted
H14BIU<---PU0.2160.0258.7730.000Accepted
H15BIU<---AT0.3220.02214.4590.000Accepted
H16AUD<---AT0.3820.02217.2340.000Accepted
H17AUD<---BIU0.4820.02320.9410.000Accepted
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Almulla, M.A. Using Digital Technologies for Testing Online Teaching Skills and Competencies during the COVID-19 Pandemic. Sustainability 2022, 14, 5455. https://doi.org/10.3390/su14095455

AMA Style

Almulla MA. Using Digital Technologies for Testing Online Teaching Skills and Competencies during the COVID-19 Pandemic. Sustainability. 2022; 14(9):5455. https://doi.org/10.3390/su14095455

Chicago/Turabian Style

Almulla, Mohammed Abdullatif. 2022. "Using Digital Technologies for Testing Online Teaching Skills and Competencies during the COVID-19 Pandemic" Sustainability 14, no. 9: 5455. https://doi.org/10.3390/su14095455

APA Style

Almulla, M. A. (2022). Using Digital Technologies for Testing Online Teaching Skills and Competencies during the COVID-19 Pandemic. Sustainability, 14(9), 5455. https://doi.org/10.3390/su14095455

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop