Next Article in Journal
Empirical Study on Bikesharing Brand Selection in China in the Post-Sharing Era
Next Article in Special Issue
Sustainability Teaching Tools in the Digital Age
Previous Article in Journal
Analysis of Fatal Accidents and Their Causes in the Korean Construction Industry
Previous Article in Special Issue
Teacher Training in Lifelong Learning—The Importance of Digital Competence in the Encouragement of Teaching Innovation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Evaluation of an Intervention Programme in Teacher Training for Geography and History: A Reliability and Validity Analysis

by
Jairo Rodríguez-Medina
1,
Cosme J. Gómez-Carrasco
2,
Pedro Miralles-Martínez
2 and
Inmaculada Aznar-Díaz
3,*
1
Department of Pedagogy, Faculty of Education, University of Valladolid, 47011 Valladolid, Spain
2
Department of Mathematics and Social Sciences Teaching, Faculty of Education at University of Murcia, 30100 Murcia, Spain
3
Department of Didactics and School Organization, Faculty of Education Sciences at University of Granada, 18071 Granada, Spain
*
Author to whom correspondence should be addressed.
Sustainability 2020, 12(8), 3124; https://doi.org/10.3390/su12083124
Submission received: 11 February 2020 / Revised: 20 March 2020 / Accepted: 10 April 2020 / Published: 13 April 2020
(This article belongs to the Special Issue Teacher Training in Active Methodologies for Ecosystem Learning)

Abstract

:
We evaluated a teacher training intervention programme aimed at improving the teaching and learning process relating to history in the secondary classroom. This was carried out via the implementation of several teaching units during the period of teaching practice of trainee teachers specialising in geography and history. The design of the teaching units was based on historical thinking competencies and on the introduction of active learning strategies. The programme was evaluated via a quasi-experimental A-B type methodological approach employing a pretest and a post-test. Both tools were designed on the basis of four dimensions (methodology, motivation, satisfaction and perception). The content of the tools was validated using the interjudge process via a discussion group in the first round and with a Likert scale questionnaire (1–4) with seven experts in the second round. The reliability of the tools has been estimated via three indices (Cronbach’s alpha, composite reliability and omega), and the validity of the construct via an exploratory (EFA) and confirmatory factor analysis (CFA) with the structural equation model. The results regarding reliability and validity have been adequate. Furthermore, the descriptive results show an improvement in all of the dimensions following the implementation of the teaching units, particularly with regard to group work, the use of digital resources and work with primary sources.

1. Introduction

1.1. Historical Competencies and Active Learning Methods

The introduction into school curriculums of education competencies has significantly affected the way in which history is approached in the classroom. Since the turn of the century, history as a school subject has been agitated by at least two factors: on the one hand, the profound social changes brought about by the impact of new technologies and the ever more visible effects of globalization; and, on the other hand, the introduction into educational curriculums of competency-based education. The effect of the introduction of competencies can be observed in two fundamental issues: first of all, the dichotomy of competency-based education compared with rote learning and the conceptual model of teaching and assessment of history, which is so common in our context; and, secondly, the difficulty of fitting transversal or general competencies into subjects which have no direct relationship with them and the danger of diffusing the competencies of the subject itself [1].
The epistemological, pedagogical and cognitive bases of each subject must be taken as a point of reference for the application of the education competencies to teaching and learning processes. Pellegrino, Chudowski and Glaser [2] point out that all assessment, without taking into account its purpose, must be upheld by three pillars: a theoretical model on the way in which students represent their knowledge and develop the skills of the subject; the tasks or situations which allow for the observation of those skills; and a method for interpreting these tasks [3,4]. It is therefore necessary, first of all, to define the cognitive model of the learning of history in order to adapt teaching processes and the assessment of competencies to the subject.
Several decades ago, studies on the teaching of history took a cognitive turn [5]—one which was not taken at the same time or at the same pace in all countries. The origin of this change can be traced to the 1970s, at the time when Bruner’s theories and Bloom and Krathwohl’s taxonomies of educational objectives began to have a decisive influence on proposals regarding history teaching. One turning point in achieving this change in history teaching and learning took place in the United Kingdom in 1972 with the educational project History Project 13-16, which later came to be known as the School History Project (SHP). This project had the aim of enabling the learner to “make” history and not just to memorise past events. In other words, the learner was encouraged to develop historical thinking. It had huge repercussions on the teaching of history and on the official curriculum of the United Kingdom. Indeed, it was the origin of some extremely interesting projects in the 1990s, such as Concepts of History and Teaching Approaches [6,7]. Studies on the definition of historical thinking, and the concepts and skills of which it consists, have become widespread throughout the world: in Canada [8,9], the USA [4,10], Australia [11], Spain [12,13], Portugal [14] and the Netherlands [15,16], as well as in Latin America [17,18].
Historical knowledge weds first-order contents, relating to concepts, dates and events, to second-order concepts, such as the handling of historical sources, empathy and historical perspective [9]. It is the second, more complex, type of skills which facilitates the comprehension of history via the simulation of the work of the historian. In VanSledright’s [4] opinion, history is a construct and it must be taught as such. In this sense, history teachers must have a solid theoretical understanding of the formation of historical thinking and understanding in their students, of the way the subject is learned and of the search for markers of cognitive progression.
Furthermore, the development of historical thinking requires a methodological change which favours the active participation of the student in the process of the construction of historical knowledge. In the study by Miralles, Gómez and Rodríguez [19], some of the strategies which can be used during the teacher training process are shown. Among them are case studies (which make it possible to apply historical knowledge, and which help to understand and analyse present-day society), debates and simulations. These strategies are valid for mobilising the three types of knowledge and for work based on the resolution of problems, which enables the establishment of synthetic discourse to facilitate the ordering and structuring of historical information.
Teacher training is an essential element for overcoming the problems of history teaching in compulsory education. The need to train up highly qualified teachers in order to improve teaching and learning processes is a much-debated issue worldwide [20,21]. Despite the fact that there is a broad bibliography on teacher training, certain authors mention the need for greater comparative research, stating that empirical studies in higher education need to be linked systematically with previous results [22,23].
Among the issues dealt with in international studies, the analysis of trainee teachers’ knowledge has become an area of considerable interest and a way of focusing intervention in initial teacher training programmes [24]. The empirical findings of recent years have provided detailed information on how the learning opportunities presented in training programmes have a clear correlation with the knowledge and the competencies of teachers at the end of their training [25,26,27].
Some previous studies have highlighted the importance of analysing the teaching methods and strategies used in history classes, mainly in relation to the use of digital resources [28,29]. Indicators of motivation, satisfaction and perceived learning have often been used to assess potential changes resulting from the inclusion of such resources [30]. Motivation is a crucial factor for academic success and several studies have shown that active learning can improve it [31,32]. The importance of student satisfaction with learning is well documented and is highly related to motivation and commitment [33]. Furthermore, understanding the level of student satisfaction with a course or activity is basic to its design. Finally, perceived learning has been defined as the student’s perceptions of their own levels of ability and knowledge [34]. Therefore, understanding the factors that affect the perception of learning could help future teachers to improve both design and assessment, in order to enhance students’ learning experience. There is therefore a need for studies to identify the factors that affect students’ motivation, satisfaction and perception of learning in history class.

1.2. Research Objectives

The objective of this paper is to analyse the implementation in teaching practice classes of the teaching units designed within a teacher training programme in a master’s degree in secondary education. These teaching units were based on methodological changes (active strategies, the use of debates, group work and digital resources), and on the development of historical competencies (research via primary sources, historical empathy, etc.). In order to achieve this general objective, the following specific objectives were proposed:
SO1: to estimate the reliability of the data collection tools;
SO2: to analyse the validity of the construct of the data collection tools; and
SO3: to evaluate the results obtained in the implementation of the teaching units on the methodology employed, motivation, satisfaction and the perception of learning and social knowledge transfer.

2. Methodology

2.1. Research Approach

A type A-B (pretest–post-test) quasi-experimental approach was chosen. The quasi-experimental design has the aim of evaluating the impact of the treatments in situations in which the subjects are not assigned to the groups in accordance with a random criterion. In this case, the election of the groups was linked to the assignation of the teaching practice centers. A large proportion of education and social research employs this type of approach [35].

2.2. Participants

The intervention was implemented in 18 classes in the autonomous region of Murcia, with the participation of 14 schools (13 state-run and 1 private). The previously designed teaching units were implemented in the four years of compulsory secondary education (12–16 years of age) and in the two years of baccalaureate (16–18 years of age). The sample comprised schools from nine different local councils in the autonomous community: Mazarrón, Cieza, Cartagena, La Unión, Murcia, Molina de Segura, Águilas, San Javier and Alcantarilla. There were 473 secondary pupils who took part in the project (Table 1). Six pupils were eliminated from the sample for not completing more than a third of the items. The selection of the sample was related to the assignation of teaching practice centers of the trainee teachers who would implement the teaching units (Table 2).

2.3. Design of the Intervention Programme

An intervention programme was designed for the speciality of geography and history in the master’s degree in teacher training, in order to improve the competencies of the future teachers in the design of activities and teaching units. This programme combined epistemological (affecting historical thinking competencies) and methodological (active teaching strategies, research methods, digital resources, etc.) elements. The proposed aim was for the trainee teacher to modify their epistemological concepts (what to teach and why history must be taught) and methodological ideas (how to teach history).
This programme was implemented in the subject entitled “Methods and resources for the teaching of geography, history and the history of art”. The training programme consisted of eight four-hour sessions. The first three were devoted to working with active learning methods: project method, case studies, problem-based learning, simulations, gamification and flipped classroom. The following two sessions were given over to working with primary sources and digital resources. The last three sessions were dedicated to the construction of teaching units, applying the prior theoretical work to the specific teaching unit which would be implemented in the secondary classroom during the students’ teaching practice. The students were required to design teaching units which combined work with historical competencies (working with sources, empathy, causes and consequences, etc.) via active methods and digital resources. Of the trainee teachers, 18 decided to evaluate the implementation of these units in the schools assigned to them for their teaching practice. Specifically, the descriptions of the sessions were as follows:
Session 1: Why is a change in teaching model necessary for Geography and History classes? Analysis of diagnostic and comparative researches with England and Canada. Incidence in epistemological aspects (six competences of historical thinking proposed by Seixas) for the change of didactic model.
Session 2: Inquiry strategies (I). Incidence in the work of inquiry with students (search, selection and analysis of information). Work with cooperative techniques (Aronson puzzle, etc.), use of classroom discussions, communication strategies, Project-Based Learning/Problems, case studies and Service-Learning. Debate on concrete practical examples.
Session 3: Inquiry strategies (II). Simulation strategies (dramatizations, historical perspective, etc.), flipped-classroom, gamification and fieldwork through didactic experimentation itineraries. Debate on concrete practical examples.
Session 4: Primary sources and heritage. Analysis of the usefulness of the use of primary sources in the classroom, work with studies on heritage education, examples of typology of heritage assets, guided work on where to look for primary sources on the web and how to insert them in the teaching units. Specific work on the website of PARES and Carmesí Project of the CARM.
Session 5: Digital resources. Digital competence, web resources, general applications (WebQuest, Wikis, Blog...) and specific applications useful for Geography and History (Google Earth, National Geographic MapMaker Interactive, virtual museums, virtual recreations, etc.). Work with Kahoot and Socrative to introduce gamification.
Session 6: Curriculum frame of the didactic unit. Objectives, contents and competencies.
Session 7: Methodology and activities. Methodological explanation and sequence of activities: initiation-motivation, introduction of knowledge, synthesis and application.
Session 8: Assessment. Procedure, techniques and evaluation instruments. What, who, how and when to assess?

2.4. Data Collection Tool

In order to evaluate the implementation of the teaching units, two tools were designed: one pretest and one post-test (Supplementary document). These tools were designed to address four categories (methodology, motivation, satisfaction and learning), in accordance with other studies which have evaluated training programmes based on active methods such as gamification [36,37,38,39]. The pretest and post-test items were the same. While the pretest evaluates the history classes received by the pupils up to that moment, the post-test evaluates the implementation of the teaching unit designed by the trainee teachers. The first of the subscales, Section 1 (methodology), is composed of 13 items relating to methodology, teaching strategies and resources used by the teacher.
Two items representative of this scale are “The most frequently used resource is the textbook” and “Historical documents are used in the classroom to learn history”. Section 2, concerning student motivation, was composed of 8 items grouped together regarding intrinsic motivation (“The classes motivate me to know more about history”) and extrinsic motivation (“The history classes motivate me because we work in groups”). The third section dealt with student satisfaction and contained 6 items. Learner satisfaction is generally measured by self-reporting on their satisfaction with the learning environment. Sample items of the student satisfaction are: “I am satisfied with the role I have as a learner” or “I am satisfied with the way in which the teacher approaches the topics”. Finally, the fourth section consisted of 13 items relating to perceived learning of historical knowledge and knowledge transfer. This dimension was evaluated through items grouped together regarding the learning of historical knowledge (“In the history classes I learn about the main historical events”) and items grouped together regarding knowledge transfer (“Thanks to the history classes, I am more respectful towards people of other cultures and with opinions which differ from my own”). The questionnaire also included information about background characteristics such as age, gender and teaching grade. Respondents were asked to rate each statement on a five-point Likert scale, anchored between (1) strongly disagree and (5) strongly agree.
The validation of the content was carried out via the interjudge procedure based on the categories of relevance and the clarity of the items of the tool. In the first round, the option was taken to form a discussion group of seven experts (two lecturers in the teaching of the social sciences; two secondary geography and history teachers; two primary social sciences teachers; and one lecturer from the Department of Research Methods and Diagnostics in Education, an expert in research methodology) to validate the content. After the necessary modifications, a second round was carried out with the experts in order to give a definitive validation of the two tools. In this second round we used a validation guide through a Likert 1-4 scale questionnaire (Supplementary material) with the same seven experts. On the first page of the validation instrument, the objective of the pretest and post-test was explained, as well as their function within the objectives of the research project. The validation instrument has three parts: in the first part, the instructions and identification data of the students are evaluated. In the second part, the experts must assess the design and formal aspects of the questionnaire. Finally, the experts must assess the contents of the questionnaire and its relationship to achieving the objectives of the research project. For this, the dimensions of clarity and relevance of the four pretest/post-test subscales were used: methodology, motivation, satisfaction, and learning and transfer. All the items obtained a mean higher than 3. In addition, we calculated the concordance between judges in the dimensions of clarity and relevance. We obtained good concordance results using the Bandigwala’s weighted agreement coefficient (BWN). Specifically, we obtained 0.86 for clarity of the items, and 0.91 for relevance of the items.

2.5. Research Procedure and Data Analysis

Both the procedure designed for the research and the data collection tools were positively evaluated by the ethics committee of the University of Murcia. An informed consent protocol (Supplementary document) was designed for the students and the families of the participants. In order to ensure the reliability of the implementation, a protocol was established with the trainee teachers and their tutors with all the steps to be followed, both in the teaching of the units and the collection of the data via the pretest and post-test.
The data were collected in two separate files (one for each tool), with each teaching unit differentiated from the others with an identification number. The pupils were identified, both in the pretest and the post-test, via a list number which was the same for both tools. In this way, it was possible to carry out an individualised study without gathering any personal information. Once the data had been collected, the R package lavaan was used to carry out the analysis [40].
For this paper, reliability analyses (Cronbach’s alpha, composite reliability and omega) were carried out, along with construct validity analyses (exploratory (EFA) and confirmatory factor analysis (CFA) via structural equation modelling) and descriptive statistics to detail the results of the post-test compared to the pretest in each of the dimensions established (methodology, motivation, satisfaction, perception of learning and social knowledge transfer).
Reliability can be defined as the degree of precision offered by a measurement. In order to be reliable, a scale must have the capacity for exhibiting consistent results in successive measurements of the same phenomenon. It has the objective of determining, in terms of probability, the degree of variation attributable to random or causal errors which are not linked to the construction of the tool. It guarantees the consistency expressed in the determination of the degree of error contained in the application of the scale and, therefore, in the measurement of the phenomenon. We studied the reliability of the two scales used with three indices: Cronbach’s alpha, composite reliability and McDonald’s omega.
With the validity analysis, we hope to analyse a construct in order to visualise the different dimensions which make up a concept via the identification of latent properties and variables (factors). Each factor is represented by the indicators which reach greater correlations.
In order to examine the validation of the construct in more depth, we carried out structural equation modelling (SEM) to confirm the existence of a series of constructs in the questionnaire. With a structural equation model, what is really being done is a logistical regression, in which the dependent variable would be the construct (in this case, each section of the questionnaire), and the items would be the independent variables. By calculating the models, the values of those coefficients can be estimated.
We compared the covariance matrix derived from the variables observed and the covariance matrix reproduced by the model. In this way, it was possible to contrast the hypothesis that the difference between the matrix from the data collected from the questionnaire and the theoretical matrix defined in the conceptual model was not statistically significant. As it was a Likert scale, the assumption of normality was not fulfilled. Therefore, we decided to make a robust estimation of the statistic χ2 via the DWLS (Diagonal Weighted Least Squares) estimator [41].
In order to establish the adjustment indices of the model, we used the Tucker–Lewis index (TLI) values, also known as the Non-Normed Fit Index (NNFI) and Comparative Fit Index (CFI). These take values of between 0 and 1, in which values closer to 1 indicate a good fit [42]. We also considered the RMSEA (Root Mean Square Error of Approximation) value, which measures the absolute difference between the proposed theoretical model and the data observed, taking into account the number of estimators and the sample size [43]. It takes values of between 0 and 1, with values closer to 0 indicating a good fit. Before showing the different adjustment indices, it must be mentioned that there is a certain degree of controversy relating to them, in the sense that there is no established agreement in the scientific community regarding their use. Some authors believe that only the Chi-square should be interpreted. Other authors [44] advocate the cautious use of adjustment measures, due to the fact that their limits can be deceitful if they are used badly.

3. Results

3.1. Reliability Analysis

First of all, we employed the internal consistency method based on Cronbach’s alpha, which makes it possible to estimate the reliability of a measurement tool made up of a set of items, for example a 5-point Likert scale, which we hope will measure the same theoretical dimension (the same construct). In this way, the items are summable in a single score which measures a feature, which is important in the theoretical construction of the tool. The reliability of the scale must always be obtained with the data of each sample to guarantee the reliable measurement of the construct in the specific research sample. We obtained an alpha value of 0.92 in the pretest and 0.92 in the post-test. Both values were considered excellent.
It is also considered important in scales of matrix correlations which are ordinal in nature to offer composite reliability data for each of the critical dimensions, as it analyses the relations between the responses to the items and the latent variable measured, as well as the variance extracted for studying the validity of the scale. The composite reliability coefficient (Composite Reliability) is considered to be more suitable than Cronbach’s alpha because it does not depend on the number of attributes associated with each concept. It is considered that the minimum value should be 0.70 [45].
We obtained an overall composite reliability index of 0.91 in the pretest and 0.92 in the post-test. Both values were considered excellent.
Other authors propose the omega coefficient, also known as Jöreskog’s rho, as it is not affected by the number of items, by the number of alternative responses or by the proportion of variance of the test [46]. The omega coefficient is based on factor loading, which is the weighted sum of the standardised variables. In the pretest, we obtained an omega of 0.721, which is considered acceptable. However, in the post-test, we obtained a lower value (.49), which is considered to be questionable.
Overall, the reliability results for both tools can be considered to be appropriate, although there are some unbalanced elements (Table 3). There are excellent overall results for Cronbach’s Alpha and the composite reliability index (higher than 0.90). Section 2 (motivation) and Section 4 (perception of learning) obtained results of between 0.80 and 0.90 in all of the indices, whereas Section 3 (satisfaction) scored close to 0.80 in all the indices. Section 1 (methodology) was more heterogeneous, with results oscillating between acceptable and questionable.

3.2. Construct Validity: EFA and CFA

3.2.1. Pretest

Before carrying out the analysis, it is recommendable to examine the correlation matrix to search for variables which do not correlate well with any other (with correlation coefficients less than 0.3), and variables which correlate too well with others (variables with some correlation coefficients higher than 0.9). The former should be eliminated from the analysis while the latter can be maintained, albeit taking into account that they may cause problems of multicollinearity. In our case, there were no problems of this kind.
We also carried out Bartlett’s test of sphericity in order to check that it was significant, in other words, that our matrix was not similar to an identity matrix. Indeed, we obtained a p-value of p < 0.05, indicating that the matrix was factorizable.
We also estimated the Keiser–Meyer–Olkin (KMO) coefficient. For the factor analysis (FA) carried out on Sections 1, 2, 3 and 4, all of the KMO coefficients were above 0.7, with values of 0.72, 0.87, 0.80 and 0.90 respectively. It should be remembered that the KMO coefficient is better when it is closer to 1, which indicates that the application here of an FA was correct and that Section 1 was the least stable. When the FA was applied to the questionnaire as a whole, a KMO of 0.91 was obtained.
The overall EFA of the questionnaire demonstrates a distribution in five dimensions, explaining 43% of the total variance (Figure 1).
Table 4 shows that Dimension 1 groups together the majority of the items of Sections 3 and 4 (satisfaction and learning of historical contents) and that the remaining items are distributed between the rest of the dimensions. These groupings explain 43% of the variance of the questionnaire.
In the following section, we shall perform a more in-depth examination with a confirmatory factor analysis of each section via structural equation modelling.
Section 1
By applying the hypothesis contrast, it could be observed that the DWLS estimator had a statistic of 350.2525565 (robust estimation 386.6210834), with 65 degrees of freedom and a significant p-value (p < 0.05). All of the p-values were significant with the exceptions of items 1 and 4. Therefore, with the exception of these two items, all of the variables were different to zero. That is to say, to a greater or lesser degree, they contributed to the model.
The model with all of the items from the questionnaire did not fit correctly (TLI = 0.82; CFI =.85; RMSEA 1.01). Given that the model did not fit well with the data from Section 1, we proceeded to eliminate the variables which contributed least to the model, which was associated with the internal error of each variable. We eliminated the variables with an internal error greater than 0.85, leaving us with the variables from 1.5 to 1.11 (items relating to innovation methodology), and checked the model again. The model now fitted correctly (TLI = 0.95; CFI = 0.96; RMSEA = 0.08) (Table 5).
Figure 2 shows the definition of the structural equation model, in which the two-way arrows represent the covariances between the latent variables (ellipses) and the one-way arrows symbolise the influence of each latent variable (constructs) on their respective observed variables (items). Lastly, the two-way arrows over the squares (items) show the error associated to each observed variable. The three variables which contribute most to the model are 1.9 (use of the internet), 1.11 (use of research in history classes) and 1.8 (use of audio-visual resources).
Section 2
By applying the hypothesis contrast, it could be observed that the DWLS estimator had a statistic of 82.3929415 (robust estimation 136.8868406), with 20 degrees of freedom and a significant p-value (p < 0.05) (Figure 3). It can be seen that, except for item 2.18 (“History classes only motivate me to pass exams”), all of the p-values were significant, and all of the variables were different to zero; to a greater or lesser extent they contributed to the model. When item 2.18 was eliminated, it was observed that the TLI and CFI were greater than 0.99 (Table 6). Therefore, the model fit well. In this case, there was an RMSEA value of 0.0858781 and a non-significant p-value, which meant that the model did indeed fit well with the data.
Section 3
By applying the hypothesis contrast, it could be observed that the DWLS estimator had a statistic of 82.3929415 (robust estimation 136.8868406), with 20 degrees of freedom and a significant p-value (p < 0.05) (Figure 4). All of the p-values were significant, and all of the variables were different to zero; to a greater or lesser extent they contributed to the model. The model with all of the items of the questionnaire did not fit well (TLI = 0.96; CFI = 0.98; RMSEA = 0.1). Given that the model did not fit the data of Section 3, we proceeded to eliminate variable 3.24 (“I am satisfied with the work of my classmates when we work in groups”), which is that which contributed least to the model. In this case, the model fit correctly (TLI = 0.99; CFI = 0.99; RMSEA = 0.05) (Table 7).
Section 4
By applying the hypothesis contrast, it could be observed that the DWLS estimator had a statistic of 82.3929415 (robust estimation 136.8868406), with 20 degrees of freedom and a significant p-value (p < 0.05) (Figure 5). All of the p-values were significant and all of the variables were different to zero; to a greater or lesser extent they contributed to the model. It is clear that variables 4.28 and 4.29 (items relating to the knowledge of historical events and figures) are those which contributed most to the model, whereas 4.36 (“I learn to work in groups with my classmates”) contributed almost nothing. The model with all of the items of the questionnaire did not fit correctly (TLI = 0.97; CFI = 0.97; RMSEA = 0.11). Given that the model does not fit well with the data of Section 3, we proceeded to eliminate the variables which contributed least to the model (4.36 and 4.38). In this case, the model fitted better, although the RMSEA was questionable (TLI = 0.98; CFI = 0.98; RMSEA = 0.10) (Table 8).
In conclusion, for Section 1, the SEM eliminated half of the questions, just as the exploratory factor analysis distributed the items of this section into different constructs. For Section 2, the SEM retained all of the items, although it indicated that 2.18 did not contribute to the model. For Section 3, the SEM retained all of the items, with the exception of 3.24, which did not contribute to the model, coinciding completely with the FA, which kept the items together and placed 3.24 in another dimension. Finally, for Section 4, the SEM was not able to fit the model. It indicated that 4.36 and 4.38 did not contribute to the model. Again, this coincided with the FA, which kept the items together and placed those questions in other dimensions, along with 4.39, which is the one which contributed least according to the SEM. In general, it could be observed that the results obtained here were in agreement with those obtained in the general FA, and proposed a division of Section 1, whereas the rest of the sections were better adjusted.

3.2.2. Post-test

In the analysis of the correlation matrix, there were no variables which did not correlate well or with a correlation coefficient greater than 0.9. Likewise, in the Bartlett sphericity analysis, we obtained a p-value of p < 0.05, indicating that the matrix was not similar to the identity matrix. In the FA carried out on Sections 1, 2, 3 and 4, all of the KMO coefficients were close to or above 0.7, with values of 0.68, 0.86, 0.77 and 0.91 respectively. This indicated that the application here of an FA was well considered and that Section 1 was the least stable.
When the EFA was applied to the whole questionnaire, we obtained a distribution in 4 dimensions, explaining 41% of the total variance, with a KMO of 0.92 (Figure 6).
Table 9 shows that the first dimension groups together many of the items relating to traditional methodology (assessment via different techniques, use of the internet and audio-visual resources, critical work with sources, etc.), with the majority of the items from Sections 2, 3 and 4. The variance explained by a single factor per process is 16%, 43%, 37% and 37%, with Section 1, again, being the most heterogeneous.
In the following section, we shall look in more detail at the validation of the construct for each of the sections of the questionnaire.
Section 1
By applying the hypothesis contrast, it could be observed that the DWLS estimator had a statistic of 760.9003772 (robust estimation 693.4870295), with 65 degrees of freedom and a significant p-value (p < 0.05) (Figure 7). In our case, variables 1.13, 1.12 and 1.6 were those which contributed most to the model.
The model with all of the items from Section 1 did not fit correctly (TLI = 0.57; CFI = 0.64; RMSEA = 0.16). Given that the model did not fit well with the data of Section 1, we proceeded to eliminate the variables which contributed least to the model. We eliminated the variables with an internal error greater than 0.80 and checked the model again. We eliminated 7 variables: 1.1, 1.2, 1.3, 1.8, 1.9, 1.10 and 1.11. In this way, it was possible to achieve a better fit of the model (Table 10), although the RMSEA was questionable (TLI = 0.92, CFI = 0.95; RMSEA = 0.1). In this model, the items relating to innovative methodology had a negative load (1.5, 1.6, 1.7, 1.12 and 1.13), whereas item 1.4 (“In order to pass, I learn the contents by rote”) had a positive load.
Section 2
By applying the hypothesis contrast, it could be observed that the DWLS estimator had a statistic of 96.2166894 (robust estimation 161.1879894), with 20 degrees of freedom and a significant p-value (p < 0.05) (Figure 8). With the exception of item 2.18, all of the p-values were significant, and all of the variables were different to zero and, to a greater or lesser extent, contributed to the model. It could be seen that variables 2.14, 2.15 and 2.16 (“The classes motivate me to learn history, to make an effort and to understand social reality”) were those which contributed most to the model, whereas item 2.18 (“History classes only motivate me to pass the exams”) contributed nothing. The model with all of the items of the section fit correctly (TLI = 0.98; CFI = 0.98; RMSEA = 0.09) (Table 11).
Section 3
By applying the hypothesis contrast, it could be observed that the DWLS estimator had a statistic of 96.2166894 (robust estimation 161.1879894), with 20 degrees of freedom and a significant p-value (p < 0.05) (Figure 9). All of the p-values were significant, all of the variables were different to zero and, to a greater or lesser extent, they contributed to the model. Variables 3.22, 3.23 and 3.25 (“I am satisfied with my role in the classroom and with the working atmosphere in the classroom”) were those which contributed most to the model. The model with all of the items from Section 3 fit correctly, albeit with a questionable RMSEA (TLI = 0.96; CFI = 0.97; RMSEA = 1) (Table 12).
Section 4
By applying the hypothesis contrast, it could be observed that the DWLS estimator had a statistic of 96.2166894 (robust estimation 161.1879894), with 20 degrees of freedom and a significant p-value (p < 0.05) (Figure 10). All of the p-values were significant, all of the variables re different to zero and, to a greater or lesser extent, they contributed to the model. Variables 4.28, 4.32 and 4.40 (“I have learnt about the main historical events, changes and continuities and to debate issues relating to current affairs”) were those which contributed most to the model, whereas item 4.36 (“I have learnt to carry out group work”) hardly contributed anything. When item 4.36 was eliminated, the model fit correctly (TLI = 0.97; CFI = 0.98; RMSEA = 0.08) (Table 13).
In conclusion, for Section 1, the SEM eliminated half of the questions, leaving 6 of the initial 13, just as the factor analysis distributed the items from this section into different constructs. For Section 2, the SEM retained all of the items, although it indicated that 2.18 did not contribute to the model. The FA kept the items together, although it moved 2.18 to one dimension and 2.20 to another, with the latter being the second least-explained item in the SEM. For Section 3, the SEM retained all of the items, coinciding completely with the FA, which kept all of the items together, with the exception of 3.24, which contributed little to the SEM model and was placed in another dimension. Finally, for Section 4, the SEM retained all of the items, with the exception of 4.36. Again, it coincided with the FA in this aspect, although the FA also removed 4.37 and 4.38. In general, it can be observed that the results obtained here were in agreement with those obtained in the general FA and proposed a division of Section 1.

3.3. Descriptive Results

As can be appreciated in Table 14, items relating to traditional methodology scored less well in the post-test, unlike items grouped with innovative methodology. A particular difference can be noted for textbooks and the use of exams, with more than a point of difference. The pupils evaluated positively the use of information technology, group work and the introduction of strategies related to historical methodology (the use of sources, research, simulations, critical evaluation, etc.) in the teaching units. There is more than a point of difference in the items relating to group work, carrying out research, the use of the internet and the use of simulations and dramatizations.
Table 15 shows a more positive evaluation of motivation (particularly intrinsic motivation) in the teaching units implemented. Specifically, the pupils valued motivation via group work, being able to contribute their own point of view and their own knowledge and the motivation brought about by the use of digital resources. The three items with the biggest difference in scores were motivation due to group work, having used resources other than the textbook and being able to give their point of view.

4. Results Regarding the Perception of Satisfaction

Table 16 shows a more positive evaluation of satisfaction in the teaching units implemented compared with the history classes the pupils had received beforehand. They expressed particular satisfaction with the way in which the teacher approached the topics in the classroom, with group work and with the positive atmosphere in the classroom.

5. Results Regarding Perception of Learning of Historical Knowledge and its Social Transfer

Table 17 shows a more positive evaluation of historical knowledge and the transfer thereof in the teaching units implemented compared with the history classes which the pupils had received beforehand. The pupils particularly valued the transfer of knowledge thanks to the different ways of using IT, the transfer of learnt knowledge to be more respectful towards other cultures and opinions and the transfer relating to debating and understanding current affairs. The improvement in the perception of the learning of historical knowledge is more moderate (approximately 0.3 improvement), with the exception of the use of documents and primary sources, which represented an improvement of 0.7. The improvement in the items grouped together regarding the transfer of knowledge was higher (0.7).

6. Discussion and Conclusions

The results show an improvement in the evaluation of history classes by secondary school pupils following the implementation of the teaching units based on historical thinking competencies and on a methodological change. The items relating to the carrying out of group work, research, the development of simulations in the classroom and the use of documents and historical sources are those which received a higher score. On the other hand, the pre-eminent use of textbooks and assessment based on written exams and rote learning fell.
As far as the category relating to motivation is concerned, once again, extremely positive results were obtained. The pupils expressed the view that the classes motivated them to learn and make more of an effort, not only to achieve better marks or to pass the exams, but also to know more about history. Also worthy of note is the idea that the pupils considered that they were able to contribute their own opinions and, above all, to carry out projects in groups. These results are in line with other work that has concluded that support for autonomy, class structure, and active participation contribute to improved student motivation [47,48].
As far as the category relating to satisfaction is concerned, all of the items in the questionnaire demonstrated an increase in this aspect when active methods are employed in the classroom and the learners see another way of teaching and learning about history. In direct relation to this aspect, the way in which the topics are approached in class stands out. These results are consistent with studies by Burgess, Senior and Moores [49] and Langan and Harris [50] who found organization, classroom management and teaching quality impacts on and is related to student satisfaction.
Lastly, regarding the category of perception of learning and knowledge transfer, the pupils gave positive evaluations in all of the items. Again, working in groups was the aspect which received the best score, although the different uses of information technologies and the use of documents and historical sources in the classroom also stand out [19,40]. These results show that learner-content interaction and learner-instructor interaction are critical factors in student satisfaction and perceived learning [33]. On the other hand, the use of digital resources and active learning have shown an increase in perceived learning in previous works and have also been related to increased motivation [51].
The use of active learning methods in conjunction with the theory of historical thinking reflects positively on the effects of this programme in classroom methodology. This is clearly reflected in the pupils’ perception, with a decrease being evident in aspects relating to traditional methodology and an increase in innovative methodology. It is extremely revealing that the programme had better results in intrinsic, rather than extrinsic, motivation. This marks a path which can be followed for the ongoing improvement of history classes. The pupils proved to be more motivated in their classes due to the mere fact of wanting to learn about history, rather than passing their exams or increasing their marks.
Equally relevant was the effect on the pupils’ level of satisfaction. The perception shown is that the training programme also increased the appreciation of the pupils with regard to new ways of working in the classroom. Last of all, the effects of the programme again showed a positive response on the part of the learners as far as the perception of learning of historical knowledge and its social transfer are concerned. The latter aspect is of particular importance as the items related to the use and application of historical knowledge in the pupils’ daily lives acquired a higher score than the items related to the learning of historical knowledge, although this also received a positive evaluation.

Supplementary Materials

The following are available online at https://www.mdpi.com/2071-1050/12/8/3124/s1, Figure S1: title, Table S1: title, Video S1: title.

Author Contributions

Conceptualization, C.J.G.-C., P.M.-M. and J.R.-M.; methodology, C.J.G.-C., P.M.-M. and J.R.-M.; software, C.J.G.-C. and J.R.-M.; validation, J.R.-M., C.J.G.-C. and P.M.-M.; formal analysis, J.R.-M., C.J.G.-C. and I.A.-D.; investigation, C.J.G.-C. and P.M.-M.; data curation, J.R.-M. and C.J.G.-C.; writing—original draft preparation, J.R.M, C.J.G.-C. and I.A.-D.; writing—review and editing, I.A.-D. and P.M.-M.; project administration, P.M.-M. and C.J.G.-C.; funding acquisition, C.J.G.-C. and P.M.-M. All authors have read and agreed to the published version of the manuscript.

Funding

This article has been possible thanks to the research project “Methodological concepts and active learning methods for the improvement of teaching competencies” (PGC2018-094491-B-C33), funded by Ministry of Science, University and Innovation, co-funded by FEDER, and project “Teacher competencies and active learning methods. An evaluative research with trainee teachers of social sciences” (20638/JLI/18), funded by Seneca Foundation. Agency of Science and Technology from Region of Murcia.

Acknowledgments

We want to thank teachers, school managers and the institutions that participated for their collaboration.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Gómez-Carrasco, C.J.; Ortuño-Molina, J.; Miralles-Martínez, P. Enseñar Ciencias Sociales con Métodos Activos de Aprendizaje, Reflexiones y Propuestas a Través de la Indagación; Octaedro: Barcelona, Spain, 2018. [Google Scholar]
  2. Pellegrino, J.W.; Chudowski, N.; Glaser, R. Knowing What Students Know: The Science and Design of Educational Assessment; National Academies Press: Washington, DC, USA, 2001. [Google Scholar]
  3. Gómez, C.J.; Miralles, P. Historical Skills in Compulsory Education: Assessment, Inquiry Based Strategies and Argumentation. New Approaches Educ. Res. 2016, 5, 139–146. [Google Scholar]
  4. VanSledright, B.A. Assessing Historical Thinking and Understanding, Innovation Design for New Standards; Routledge: New York, NY, USA, 2014. [Google Scholar]
  5. Carretero, M.; López, C. Estudios cognitivos sobre el conocimiento histórico: Aportaciones para la enseñanza y alfabetización histórica. Enseñanza de las Ciencias Sociales 2009, 8, 79–93. [Google Scholar]
  6. Lee, P. Putting Principles into Practice: Understanding History. In How Students Learn: History in the Classroom; Donovan, M.S., Bransford, J.D., Eds.; National Academies Press: Washington, DC, USA, 2005. [Google Scholar]
  7. Lee, P.; Ashby, R. Progression in Historical Understanding Among Students Ages 7–14. In Knowing, Teaching, and Learning History: National and International Perspectives; Stearns, P.N., Seixas, P., Wineburg, S., Eds.; University Press: New York, NY, USA, 2000. [Google Scholar]
  8. Clark, P. New Possibilities for the Past, Shaping History Education in Canada; UBC Press: Vancouver, BC, Canada, 2011. [Google Scholar]
  9. Seixas, P.; Morton, T. The Big Six Historical Thinking Concepts; Nelson College Indigenous: Toronto, ON, Canada, 2013. [Google Scholar]
  10. Wineburg, S. Historical Thinking and other Unnatural Acts, Charting the Future of Teaching the Past; Temple University Press: Philadelphia, PA, USA, 2001. [Google Scholar]
  11. Parkes, R.J.; Sharp, H. Nietzschean perspectives on representations of national history in Australian school textbooks: What should we do with Gallipoli? Ensayos Revista de la Facultad de Educación de Albacete 2014, 29, 159–181. [Google Scholar]
  12. López-Facal, R. La LOMCE y la competencia histórica. Ayer 2014, 94, 273–285. [Google Scholar]
  13. Martínez-Hita, M.; Gómez, C.J. Nivel cognitivo y competencias de pensamiento histórico en los libros de texto de Historia de España e Inglaterra. Un estudio comparativo. Revista de Educación 2018, 379, 145–169. [Google Scholar]
  14. Barca, I. Till New Facts Are Discovered: Students’ Ideas About Objectivity in History. In International Review of History Education, Vol. 4, Understanding History: Recent Research in History Education; Ashby, R., Gordon, P., Lee, P., Eds.; Routledge Falmer: New York, NY, USA, 2005. [Google Scholar]
  15. Van Drie, J.; Van Boxtel, C. Historical reasoning: Towards a framework for analyzing student’s reasoning about the past. Educ. Psychol. Rev. 2008, 20, 87–110. [Google Scholar] [CrossRef] [Green Version]
  16. Van Boxtel, C.; Grever, M.; Klein, S. Heritage as a Resource for Enhancing and Assessing Historical Thinking: Reflections from the Netherlands. In New Directions in Assessing Historical Thinking; Ercikan, K., Seixas, P., Eds.; Routledge: New York, NY, USA, 2015. [Google Scholar]
  17. Mora, G.; Ortiz, R. El modelo de educación histórica. Experiencias de innovación en la Educación Básica de México. Ensayos Revista de la Facultad de Educación de Albacete 2012, 11, 87–98. [Google Scholar]
  18. Schmidt, M.A. Jóvenes brasileños y europeos: Identidad, cultura y enseñanza de la historia (1998–2000). Enseñanza de las Ciencias Sociales 2005, 4, 53–64. [Google Scholar]
  19. Miralles, P.; Gómez, C.J.; Rodríguez, R. Patrimonio, competencias históricas y metodologías activas de aprendizaje. Un análisis de las opiniones de los docentes en formación en España e Inglaterra. Estudios Pedagógicos 2017, 43, 161–184. [Google Scholar] [CrossRef] [Green Version]
  20. Cochran-Smith, M.; Zeichner, K.M. Studying Teacher Education: The Report of the AERA Panel on Research and Teacher Education; Routledge: New York, NY, USA, 2005. [Google Scholar]
  21. Darling-Hammond, L.; Bransford, J.D. Preparing Teachers for a Changing World: What Teachers Should Learn and Be Able to Do; Jossey-Bass: San Francisco, CA, USA, 2005. [Google Scholar]
  22. König, J.; Ligtvoet, R.; Klemenz, S.; Rothland, M. Effects of opportunities to learn in teacher preparation on future teachers’ general pedagogical knowledge: Analyzing program characteristics and outcomes. Stud. Educ. Eval. 2017, 53, 122–133. [Google Scholar] [CrossRef]
  23. Schmidt, W.H.; Blömeke, S.; Tatto, M.T. Teacher Education Matters, A Study of the Mathematics Teacher Preparation from Six Countries; Teacher College Press: New York, NY, USA, 2011. [Google Scholar]
  24. Darling-Hammond, L. Assessing teacher education. The usefulness of multiple measures for assessing program outcomes. J. Teach. Educ. 2006, 57, 120–138. [Google Scholar] [CrossRef]
  25. Konig, J.; Blomeke, S. Future Teachers’ General Pedagogical Knowledge from Comparative Perspective. Does School Experience Matter? ZDM Int. J. Math. Educ. 2012, 44, 341–354. [Google Scholar] [CrossRef]
  26. Blomeke, S.; Suhl, U.; Kaiser, G.; Dohrmann, M. Family background, entry selectivity and opportunities to learn: What matters in primary teacher education? An international comparison of fifteen countries. Teach. Teacher Educ. 2012, 28, 44–55. [Google Scholar] [CrossRef]
  27. Schimdt, W.H.; Cogan, L.; Houang, R. The role of opportunity to learn in teacher preparation: An international context. J. Teach. Educ. 2011, 62, 138–153. [Google Scholar] [CrossRef]
  28. Ibáñez, A.; Vicent, N.; Asensio, M. Aprendizaje informal, patrimonio y dispositivos móviles. Evaluación de una experiencia en educación secundaria. Didáctica de las Ciencias Experimentales y Sociales 2012, 26. [Google Scholar] [CrossRef] [Green Version]
  29. Miralles, P.; Gómez, C.J.; Arias, V.B.; Fontal, O. Digital resources and didactic methodology in the initial training of history teachers. Comunicar 2019, 61, 45–56. [Google Scholar] [CrossRef]
  30. Maloy, J.; Fries, L.; Laski, F.; Ramirez, G. Seductive details in the flipped classroom: The impact of interesting but educationally irrelevant information on student learning and motivation. CBE Life Sci. Educ. 2019, 18, 3. [Google Scholar] [CrossRef] [Green Version]
  31. Armbruster, P.; Patel, M.; Johnson, E.; Weiss, M. Active learning and student-centered pedagogy improve student attitudes and performance in introductory biology. CBE Life Sci. Educ. 2009, 8, 203–213. [Google Scholar] [CrossRef]
  32. Hospel, V.; Galand, B. Are both classroom autonomy support and structure equally important for students’ engagement? A multilevel analysis. Learn. Instruct. 2016, 41, 1–10. [Google Scholar] [CrossRef]
  33. Alqurashi, E. Predicting student satisfaction and perceived learning within online learning environments. Distance Educ. 2019, 40, 133–148. [Google Scholar] [CrossRef]
  34. Alavi, M.; Marakas, G.M.; Youngjin, Y. A Comparative Study of Distributed Learning Environments on Learning Outcomes. Inf. Syst. Res. 2002, 4, 404–415. [Google Scholar] [CrossRef]
  35. Fernández, P.; Vallejo, G.; Livacic, P.; Tuero, E. Validez Estructurada para una investigación cuasi-experimental de calidad. Se cumplen 50 años de la presentación en sociedad de los diseños cuasi-experimentales. Anales de Psicología 2014, 30, 756–771. [Google Scholar]
  36. Da Rocha, L.; Sandro, A.; De Melho, I. Effectiveness of gamification in the engagement of students. J. Comput. Hum. Behav. 2016, 58, 48–63. [Google Scholar] [CrossRef]
  37. De Marcos, L.; García-López, E.; García-Cabot, A. On the effectiveness of game-like and social approaches in learning: Comparing educational gaming, gamification & social networking. Comput. Educ. 2016, 95, 99–113. [Google Scholar]
  38. Han-Huey, C.; Kofinas, A.; Luo, J. Enhancing student learning experience with technology-mediated gamification: An empirical study. Comput. Educ. 2018, 121, 1–17. [Google Scholar]
  39. Hanus, M.D.; Fox, J. Assessing the effects of gamification in the classroom: A longitudinal study on intrinsic motivation, social comparison, satisfaction, effort and academic performance. Comput. Educ. 2015, 80, 152–161. [Google Scholar] [CrossRef]
  40. Rosseel, Y. lavaan: An R package for structural equation modeling. J. Stat. Softw. 2012, 48, 1–36. [Google Scholar] [CrossRef] [Green Version]
  41. Beaujean, A.A. Latent Variable Modeling Using R: A Step-By-Step Guide; Routledge: New York, NY, USA, 2014. [Google Scholar]
  42. Kenny, D.A. Structural Equation Modeling; SAGE: Thousand Oaks, CA, USA, 2012. [Google Scholar]
  43. Steiger, J.H. Structural model evaluation and modification: An interval estimation approach. Multivar. Behav. Res. 1990, 25, 173–180. [Google Scholar] [CrossRef] [Green Version]
  44. Hayduk, L.; Cummings, G.; Boadu, K.; Pazderka-Robinson, H.; Boulianne, S. Testing! Testing! One, two, three—Testing the theory in structural equation models! Personal. Individ. Differ. 2007, 42, 841–850. [Google Scholar] [CrossRef]
  45. Hair, J.F., Jr.; Black, W.C.; Babin, B.J.; Anderson, R.E. Multivariate Data Analysis, 7th ed.; Pearson Prentice Hall: Upper Saddle River, NJ, USA, 2009. [Google Scholar]
  46. Ventura-León, J.L.; Caycho-Rodríguez, T. El coeficiente omega: Un método alternativo para la estimación de la confiabilidad. Revista Latinoamericana de Ciencias Sociales Niñez y Juventud 2017, 15, 625–627. [Google Scholar]
  47. Ahn, I.; Patrick, H.; Chiu, M.M.; Levesque-Bristol, C. Measuring teacher practices that support student motivation: Examining the factor structure of the teacher as social context questionnaire using multilevel factor analyses. J. Psychoeduc. Assess. 2019, 37, 743–756. [Google Scholar] [CrossRef]
  48. Owens, D.C.; Sadler, T.D.; Barlow, A.T.; Smith-Walters, C. Student motivation from and resistance to active learning rooted in essential science practices. Res. Sci. Educ. 2020, 50, 253–277. [Google Scholar] [CrossRef]
  49. Burgess, A.; Senior, C.; Moores, E. A 10-year case study on the changing determinants of university student satisfaction in the UK. PLoS ONE 2018, 13, 1–15. [Google Scholar] [CrossRef] [PubMed]
  50. Langan, A.M.; Harris, W.E. National student survey metrics: Where is the room for improvement? High. Educ. 2019, 78, 1075–1089. [Google Scholar] [CrossRef] [Green Version]
  51. Makransky, G.; Lilleholt, L. A structural equation modeling investigation of the emotional value of immersive virtual reality in education. Educ. Technol. Res. Dev. 2018, 66, 1141–1164. [Google Scholar] [CrossRef]
Figure 1. Numbers of components.
Figure 1. Numbers of components.
Sustainability 12 03124 g001
Figure 2. Structural equation model, Section 1.
Figure 2. Structural equation model, Section 1.
Sustainability 12 03124 g002
Figure 3. Structural equation model, Section 2.
Figure 3. Structural equation model, Section 2.
Sustainability 12 03124 g003
Figure 4. Structural equation model, Section 3.
Figure 4. Structural equation model, Section 3.
Sustainability 12 03124 g004
Figure 5. Structural equation model, Section 4.
Figure 5. Structural equation model, Section 4.
Sustainability 12 03124 g005
Figure 6. Numbers of components of post-test.
Figure 6. Numbers of components of post-test.
Sustainability 12 03124 g006
Figure 7. Structural equation model, Section 1.
Figure 7. Structural equation model, Section 1.
Sustainability 12 03124 g007
Figure 8. Structural equation model, Section 2.
Figure 8. Structural equation model, Section 2.
Sustainability 12 03124 g008
Figure 9. Structural equation model, Section 3.
Figure 9. Structural equation model, Section 3.
Sustainability 12 03124 g009
Figure 10. Structural equation model, Section 4.
Figure 10. Structural equation model, Section 4.
Sustainability 12 03124 g010
Table 1. Participants in the research according to sex.
Table 1. Participants in the research according to sex.
Sexn%
Male22247.54
Female24552.46
Total467100.00
Table 2. Participants in the research according to the academic year in which the teaching units were implemented.
Table 2. Participants in the research according to the academic year in which the teaching units were implemented.
Yearn%
1 Secondary15032.120
2 Secondary8117.345
3 Secondary194.069
4 Secondary12526.767
1 Baccalaureate7616.274
2 Baccalaureate163.426
Total467100.000
Table 3. Summary of the scores obtained in the reliability analysis according to sections and overall.
Table 3. Summary of the scores obtained in the reliability analysis according to sections and overall.
Cronbach’s Alpha PretestComposite Reliability PretestOmega PretestCronbach’s Alpha PosttestComposite Reliability PosttestOmega Posttest
Section 10.690.560.700.660.430.67
Section 20.810.820.830.810.830.84
Section 30.770.780.780.760.770.77
Section 40.860.870.880.870.880.84
Overall0.920.910.720.920.920.48
Table 4. Components of matrix.
Table 4. Components of matrix.
Comp. 1Comp. 2Comp.3Comp.4Comp.5
2.161.51.61.11.2
3.221.101.72.141.3
3.231.111.82.151.4
3.252.201.92.172.18
3.263.241.122.192.21
3.274.361.134.39
4.28 4.31
4.29 4.38
4.30
4.32
4.33
4.34
4.35
4.37
4.40
Table 5. Adjustment indices of the model of Section 1.
Table 5. Adjustment indices of the model of Section 1.
Tucker–Lewis Index (TLI)Comparative Fit Index (CFI)Root Mean Square Error of Approximation (RMSEA)
Model of Section 1 with variables 1.5-1.110.950.960.08
Table 6. Adjustment indices of the model of Section 2.
Table 6. Adjustment indices of the model of Section 2.
TLICFIRMSEA
Model of Section 2 with all the items0.990.990.08
Table 7. Adjustment indices of the model of Section 3.
Table 7. Adjustment indices of the model of Section 3.
TLICFIRMSEA
Model of Section 3 without variable 3.240.990.990.05
Table 8. Adjustment indices of the model of Section 4.
Table 8. Adjustment indices of the model of Section 4.
TLICFIRMSEA
Model of Section 4 without variables 4.36 and 4.380.980.980.10
Table 9. Components of post-test matrix.
Table 9. Components of post-test matrix.
Comp. 1Comp. 2Comp.3Comp.4
1.51.101.11.7
1.62.201.21.11
1.83.241.31.12
1.94.361.42.18
1.134.38 4.37
2.14
2.15
2.16
2.17
2.19
2.21
3.22
3.23
3.25
3.26
3.27
4.28
4.29
4.30
4.31
4.32
4.33
4.34
4.35
4.39
4.40
Table 10. Adjustment indices of the model of Section 1.
Table 10. Adjustment indices of the model of Section 1.
TLICFIRMSEA
Model of Section 1 without the 7 variables with internal error greater than 0.800.920.950.10
Table 11. Adjustment indices of the model of Section 2.
Table 11. Adjustment indices of the model of Section 2.
TLICFIRMSEA
Model of Section 2 with all of the items0.980.980.09
Table 12. Adjustment indices of the model of Section 3.
Table 12. Adjustment indices of the model of Section 3.
TLICFIRMSEA
Model of Section 3 with all of the items0.960.970.1
Table 13. Adjustment indices of the model of Section 4.
Table 13. Adjustment indices of the model of Section 4.
TLICFIRMSEA
Model of Section 4 with all of the items0.970.980.08
Table 14. Differences of averages between the pretest and post-test in the evaluation of the methodology.
Table 14. Differences of averages between the pretest and post-test in the evaluation of the methodology.
ItemsAverage PretestAverage Post-TestDifference
History classes are based on the teacher’s explanation3.953.63−0.32
The most frequently used resource is the textbook3.532.10−1.43
The most important part of the assessment are the written exams4.093.00−1.09
In order to pass the exams, I learn the contents by heart 3.963.02−0.94
Different tools are used for assessment (notebooks, written work, rubrics, portfolio, etc.)3.624.150.53
Historical documents are used in the classroom to learn history2.693.680.99
In history classes we visit museums and other places of historical and heritage interest1.771.70−0.07
In history classes we use audio-visual resources (presentations, films, documentaries, etc.)3.394.350.96
In history classes we use the internet to study the topics 3.164.251.09
In history classes we carry out group work2.664.021.36
In history classes we carry out research2.513.641.13
We put ourselves in the shoes of a historical figure (theatre, simulations, etc.) in order to understand his/her actions1.682.731.05
We critique historical events and processes 2.653.490.84
Items of traditional methodology grouped together3.882.94−0.95
Items of innovative methodology grouped together2.683.560.88
Table 15. Differences of averages between the pretest and post-test in the evaluation of motivation.
Table 15. Differences of averages between the pretest and post-test in the evaluation of motivation.
ItemsAverage PretestAverage PosttestDifference
The classes motivate me to know more about history3.113.920.81
The history classes improve my motivation to learn and to make more of an effort 2.983.810.83
The history classes motivate me because I gain a better understanding of the social and cultural reality with which I am in contact 3.293.850.56
The history classes motivate me to achieve better marks3.213.780.57
I am only motivated in the history classes in order to pass the exams2.633.410.78
I am motivated in the history classes because I can contribute my point of view and my own knowledge 2.823.680.86
The history classes motivate me because we work in groups2.383.561.18
I am motivated in the history classes because we use resources other than the textbook (Internet, audio-visual resources, historical documents, etc.)3.214.120.91
Items grouped together regarding intrinsic motivation2.963.830.87
Items grouped together regarding extrinsic motivation2.923.600.68
Table 16. Differences of averages between the pretest and post-test in the evaluation of satisfaction.
Table 16. Differences of averages between the pretest and post-test in the evaluation of satisfaction.
ItemsAverage PretestAverage PosttestDifference
I am satisfied with the role I have as a learner3.514.050.54
I am satisfied with the working atmosphere in the class3.444.060.62
I am satisfied with my classmates’ work when we work in small groups3.253.910.66
I am satisfied with what I learn3.604.200.6
I am satisfied with the way in which the teacher approaches the topics3.554.320.77
Until now I am satisfied with my marks3.413.930.52
Table 17. Differences of averages between the pretest and post-test in the evaluation of learning and social transfer of knowledge.
Table 17. Differences of averages between the pretest and post-test in the evaluation of learning and social transfer of knowledge.
ItemsAverage PretestAverage PosttestDifference
In the history classes I learn about the main historical events3.824.150.33
In the history classes I learn about the main historical figures3.743.920.18
In the history classes I learn to use chronology3.313.640.33
In the history classes I learn to handle documents and historical sources3.053.750.7
In the history classes I learn about the changes and continuities of history3.633.970.34
In the history classes I learn that all historical figures and events are equally important3.403.730.33
In the history classes I learn about the causes and continuities of historical events3.654.000.35
In the history classes I learn about the reasons which led people in the past to act in a particular way and to critically evaluate their actions3.603.940.34
In the history classes I learn to carry out group work with my classmates2.813.891.08
In the history classes I learn to value more the heritage of our surrounding area3.303.780.48
In the history classes I learn about different ways of using IT for the teaching of the social sciences 2.693.670.98
Thanks to the history classes, I am more respectful towards people of other cultures and with opinions which differ from my own3.123.750.63
The history classes help me to understand and debate current affairs3.504.150.65
Items grouped together regarding the learning of historical knowledge3.523.890.37
Items grouped together regarding knowledge transfer3.093.840.75

Share and Cite

MDPI and ACS Style

Rodríguez-Medina, J.; Gómez-Carrasco, C.J.; Miralles-Martínez, P.; Aznar-Díaz, I. An Evaluation of an Intervention Programme in Teacher Training for Geography and History: A Reliability and Validity Analysis. Sustainability 2020, 12, 3124. https://doi.org/10.3390/su12083124

AMA Style

Rodríguez-Medina J, Gómez-Carrasco CJ, Miralles-Martínez P, Aznar-Díaz I. An Evaluation of an Intervention Programme in Teacher Training for Geography and History: A Reliability and Validity Analysis. Sustainability. 2020; 12(8):3124. https://doi.org/10.3390/su12083124

Chicago/Turabian Style

Rodríguez-Medina, Jairo, Cosme J. Gómez-Carrasco, Pedro Miralles-Martínez, and Inmaculada Aznar-Díaz. 2020. "An Evaluation of an Intervention Programme in Teacher Training for Geography and History: A Reliability and Validity Analysis" Sustainability 12, no. 8: 3124. https://doi.org/10.3390/su12083124

APA Style

Rodríguez-Medina, J., Gómez-Carrasco, C. J., Miralles-Martínez, P., & Aznar-Díaz, I. (2020). An Evaluation of an Intervention Programme in Teacher Training for Geography and History: A Reliability and Validity Analysis. Sustainability, 12(8), 3124. https://doi.org/10.3390/su12083124

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop