1. Introduction
The assessment literacy of students, in general, as well as that of prospective teachers, in particular, is all the more addressed in literature, thus responding to the emerging needs in society (
Smith et al. 2013;
Massey et al. 2020;
Noble 2021). The trend is explained by an increasing interest in equipping novice teachers with expertise in performing assessments as part of their job. The influx of assessment-literate teachers in the profession will allow for the redirection of administrative resources from in-service training in the
stable knowledge base acquisition (
Brown and Bailey 2008) to more demanding topics of assessment, such as online classroom assessment, alternative assessment, etc.
Unlike practicing teachers, pre-service trainees in LTA do not have hands-on experience in teaching/assessment to reflect and rely on. As
Kvasova (
2020) states, undergraduates do not approach LTA from a teacher perspective, but associate themselves primarily with assesses, drawing mostly on their actual experience as students being assessed. This experience, together with beliefs, values, and attitudes, shapes student teachers’ pre-existing conceptions of assessment, which can be positive/negative and dynamic/persistent. In any case, the conceptions take a much longer time to develop than the duration of an academic course in LTA (
Deneen and Brown 2016). The development process is affected by multiple factors, such as practicum (
Xu and He 2019), context, and career stage (
Coombe et al. 2020), as well as personal development and enhanced educational and professional levels. We may assume that the popular saying ‘teachers teach as they were taught’ may relate to the pre-existing conceptions in novice teachers’ consciousness, which guide them in their practice, despite the newly received formal training. As
Smith et al. (
2014) found, the teachers’ assessment beliefs were formed by their past experiences, rather than by the content they were taught.
Although cognitive psychology and education research widely recognize the influence of pre-existing knowledge and experiences in shaping the cognition and behavior of prospective instructors (e.g., see
Oleson and Hora 2013), the above quoted claim has not been supported empirically as of yet. In fact, Cox provided evidence that “there is a significant difference between how teachers teach and how they were taught during their own educational experience” (
Cox 2014, p. ii), which needs further investigation. Regarding LAL, Grainger and Adie argue that “learning to be an assessor through being assessed can be problematic, particularly when working between the different levels of education” (
Grainger and Adie 2014, p. 90). This seems to apply mostly to the faculty, who have not received formal training in pedagogical methods (
Oleson and Hora 2013) and, therefore, are unable to overcome the intuitive and imitative ‘folkways of teaching’ (
Lortie 1975).
The current study focuses on the impact of student teachers’ pre-existing conceptions of assessment on the more conscious formal training in LAL. More particularly, the pre-existing conceptions of assessment developed tacitly in a General English (GE) course classroom are examined and how these influence LAL development within a course in LTA.
2. Literature Review
Assessment literacy consists of knowledge and attitudes, or conceptions (
Deneen and Brown 2011), similar to any other competence. Following
Deneen and Brown (
2016), the term conception is, in this paper, inclusive of attitudes, perceptions, dispositions, and other terms that suggest belief about a phenomenon. Conceptions of assessment (CoA) are key elements of assessment competence (
Hill and Eyers 2016) and constitute an important part of a teacher’s assessment identity (
Looney et al. 2018). They guide teacher assessment practices (
Barnes et al. 2014;
Brown 2008) and play a pivotal role in achieving assessment literacy (AL) (
Deneen and Brown 2016).
Deneen and Brown (
2016) argue that CoA have significant implications for teacher education. Their research shows that the theoretical knowledge acquired, and the practical assessment skills developed are not sufficient for the overall development of AL. The authors reiterate that shaping conceptions requires a lot of time and students’ gains in knowledge and skills across a brief training course may not, regrettably, be accompanied by enhanced CoA, which eventually impedes students’ AL achievement.
Smith et al. (
2013) define student AL as “students’ understanding of the rules surrounding assessment in their course context, their use of assessment tasks to monitor, or further, their learning, and their ability to work with the guidelines on standards in their context to produce work of a predictable standard” (p. 46). As
Noble (
2021) recaps this definition, AL is “students’ ability to understand the purpose and processes of assessment, and accurately judge their own work” (p. 1). It presupposes (1) understanding key assessment terminology and different assessment methods and procedures, in the context of university, (2) being familiar with performance standards and criteria used to judge student work, and (3) being able to judge or evaluate own/peers’ performances on assessment tasks marked by criteria and standards (
Noble 2021).
In a similar vein to
Smith’s et al. (
2013) treatment of student AL,
O’Donovan et al. (
2004) focus on the development of student understanding of assessment standards and approaches to assessment knowledge transfer. This development may proceed in either an explicit or tacit way, with varied student engagement—passive or active. The authors claim that the future lies in the active student engagement, facilitated by the social constructivist approach (through active student engagement in formal processes devised to communicate tacit knowledge of standards) or the ‘Cultivated’ community of practice approach (wherein tacit standards communicated through participation in informal knowledge exchange networks ‘are seeded’ by specific activities). The authors advocate the ‘social constructivist process model’ of assessment, where “students are actively engaged with every stage of the assessment process in order that they truly understand the complex and contextual requirements of assessment praxis, particularly the criteria and standards being applied, and how they may subsequently produce better work” (
O’Donovan et al. 2004, p. 208).
The scholars argue that developing tacit dimensions of knowledge is more effective than the transfer of explicit information, since learners obviously need to be engaged in meaningful activities to be able to construct meaning for themselves.
O’Donovan et al. (
2008) contend that tacit knowledge is built up through socialisation and practice and can be internalized by the student to shape their understanding of standards and expectations. The implication from the above is that, in order to come to grips with assessment standards, students need not only to gain the knowledge transferred explicitly, but also to take advantage of tacit acquisition of knowledge, whenever this is possible, including the situations created by teachers and embedded in the instruction.
Finally, O’Donovan et al. propose facilitating learners’ ‘pedagogical intelligence’ by inducting them, not only into the discourse of their chosen discipline, but also into the discourse of learning and teaching. The authors concur with
Hutchings (
2005), who proposes that “pedagogical intelligence would require students not only to be able to critically evaluate their own learning, but also their learning environment including the teaching” (quoted in
O’Donovan et al. 2008, p. 214). The idea to engage students, especially student teachers, with concepts, principles, and technicalities of assessment in a tacit paradigm, within any discipline of their study, appears quite constructive. It resonates with Tenet 5 (integrating assessment literacy into course design), as proposed by the
Higher Education Academy Project Report (
2012). The tenet mentions that ”Active engagement with assessment standards needs to be an integral and seamless part of course design and the learning process in order to allow students to develop their own, internalized conceptions of standards and to monitor and supervise their own learning” (p. 21).
Given that significant time and effort are required for serious conceptual shift to happen, it may seem quite logical for student teacher assessment conceptions to start shaping prior to the beginning of student teachers’ formal training in LTA. There is evidence that teachers’ CoA arise from their experiences of being assessed as learners (
Hill and Eyers 2016) and prior personal experiences of being assessed before and during teacher education play a significant role in structuring their assessment conceptions (
Crossman 2007;
Smith et al. 2014). It is worth mentioning the experiences of apprenticeship of observation (
Lortie 1975) that involve subconscious learning about teaching through observing their own teachers’ teaching. These significantly impact pre-service teachers’ initial conceptions and account for their teaching practice, as they offer insights into what to do as teachers (
Knapp 2012;
Boyd et al. 2013; quoted in
Xu and He 2019).
Emphasizing the role of CoA within AL, Deneen and Brown argue that “courses in assessment, perhaps more than any other aspect of teacher education, must address preexisting conceptions and beliefs and their causes” (
Deneen and Brown 2016, p. 14). Following this line of thought, the authors call for a sustained program-level engagement with CoA that will lead to innovating the curricula content and structure and modernizing the concept of AL itself.
As far as LAL is concerned, after being identified by
Davies (
2008) as knowledge, skills, and principles that stakeholders are required to master, in order to perform assessment tasks, the concept has evolved significantly. More extended definitions of LAL are proposed by
Fulcher (
2012),
Inbar-Lourie (
2013),
Vogt and Tsagari (
2014),
Giraldo (
2018), and others. One of the most recent generic definitions of LAL is offered by
Coombe et al. (
2020), who, referring to
Inbar-Lourie (
2008),
Pill and Harding (
2013), and
Stiggins (
1999), interpreted LAL as “a repertoire of competences, knowledge of using assessment methods, and applying suitable tools in an appropriate time that enables an individual to understand, assess, construct language tests, and analyze test data” (p. 2).
While the “knowledge-skills-principles” paradigm is frequently referred to in the definitions of the above-mentioned and other authors,
Scarino (
2013) contends that there is one more integral part of teacher LAL—their self-awareness as assessors. Drawing on the impact of sociocultural theories of learning on assessment practices in second language education, she argues for the need to enable teachers to explore and evaluate their own preconceptions, beliefs, personal theories, and experience, so as to enhance critical awareness of their own assessment practices and students’ second language learning.
The assessor’s personal beliefs and attitudes occupy a specific area within differential profiles of LAL, as proposed by
Taylor (
2013). In the profile for professional language testers, these aspects are in an equally high place with all other aspects, such as knowledge, principles, etc. In the profile for language teachers, personal beliefs/attitudes are ranked second, together with technical skills, sociocultural values, and local practices, following the top-ranked language pedagogy. In the second position, there are also the personal beliefs/attitudes in the version of Taylor’s description, revised by
Bøhn and Tsagari (
2021), based on their empirical study of teacher educator conceptions of LAL. Does the inclusion of personal beliefs/attitudes in LAL profiles mean that they should be included in the LAL construct and taken into thorough consideration? In any case, we cannot deny that teachers’ contexts for language assessment should contribute to the meaning of LAL, given that one’s own preconceptions, understandings, and opinions may “inform [one’s] conceptualizations, interpretations, judgments and decisions in assessment” (
Scarino 2013, p. 109).
Before voices were raised for the necessity of building LAL in various stakeholders (
Fulcher 2012;
Pill and Harding 2013;
Taylor 2013;
Kremmel and Harding 2019), LAL was mostly associated with the LAL of two major stakeholder groups—language testers and teachers (
Volante and Fazio 2007;
Brookhart 2011;
Fulcher 2012;
Vogt and Tsagari 2014;
DeLuca et al. 2016;
Xu and Brown 2016). According to
Fulcher (
2012), assessment literacy for language testers comprises the knowledge, skills, and abilities required to design, develop, maintain, or evaluate large-scale standardized and/or classroom-based tests, familiarity with test processes, and awareness of principles and concepts that guide and underpin practice, including ethics and codes of practice. As for teacher assessment literacy, it has been considered mostly from the perspective of in-service teacher training in LTA, based on the needs that had been empirically explored (
Hasselgreen et al. 2004;
Fulcher 2012;
Vogt and Tsagari 2014;
Kvasova and Kavytska 2014;
Massey et al. 2020;
Bøhn and Tsagari 2021). Teachers’ beliefs and perceptions concerning the relevance of language assessment and the challenges it poses have also been studied (
Wach 2012;
Tsagari and Vogt 2017;
Giraldo 2019;
Vogt et al. 2020). Moreover, some works report on planning and implementing in-service teacher training courses in LTA (e.g.,
Tsagari et al. 2018).
Malone (
2017) claims that it is time to expand the understanding of LAL by including students within the stakeholders and that, so far, “only limited research … has explored students’ understanding of assessment, assessment results and their own performances” (p. 9). In fact,
Malone (
2017) warns against the erroneous transfer of practicing teachers’ understanding of major assessment concepts onto students during training, since a lack of relevant research reflects that “language assessment literacy efforts for students represent still more complexity for the field” (p. 83). The past few years, research into student assessment literacy seems to be gaining momentum through studies that discuss the organization of training courses for prospective/pre-service teachers (
Beziat and Coleman 2015;
Lam 2015;
Yetkin 2018;
Ukrayinska 2018;
Hildén and Fröjdendahl 2018;
Fröjdendahl 2018;
Giraldo and Murcia 2019;
Kvasova 2020). The common element in these studies is that their authors proceed from identifying the gaps in in-service teacher LAL, in order to define the training needs aimed at bridging these gaps and, finally, building up curricula for pre-service training courses. Proceeding, in this case, from teacher training needs, as a starting point, is viewed as absolutely justified. It is also true, however, that the differences between the objectives of courses aimed at two cohorts of trainees are either ignored or minimized; therefore, the curriculum for practicing teachers is considered easily applicable to prospective teachers’ training (e.g.,
Anam and Putri 2021).
Tsagari (
2017) argues that the acquisition and implementation of LAL is context-specific, depending on social and educational values, as well as the beliefs shared within a particular assessment culture. Besides, LAL is not static, but dynamic, since it is affected by the teachers’ individual perceptions and knowledge about learning and assessment. Acknowledging the importance of contextualization of LAL, we agree that LAL of two major stakeholder groups—teachers and students—is specific. It is situated in particular contexts and determined by teachers and/or students’ social roles, education level, status, age, maturity, etc. Thus, specificity of LAL presupposes “identification of assessment priorities and the development of assessment training strategies that are contextually situated within effective modes of training” (
Tsagari 2017, p. 83).
After looking into research in this field, what seems to be lacking is a comprehensive definition of student teacher LAL. This paper attempts to contribute in this respect.
4. Data Collection and Analysis
4.1. Exploring GE Course Delivery
The curriculum of the multi-skill GE course for Term 1, which comprises two modules, addresses mainly general interest topics, such as student life (studies and pastime), appearance, and personal characteristics. The teacher’s book, similar to most teacher’s books, does not contain any assessment materials, e.g., quizzes, tests, and rating scales, to assess speaking and writing. Course activities include a variety of classroom activities such as, questioning, written home assignment, whole-group discussion, paired/small team discussion, role plays, quizzes, revision translation, individual long turn, written tests, and writing tasks. The assessment phase of the activities is planned by the teacher, with respect to the objectives of the activities and their formats. Therefore, revision translation, individual long turn (oral test), written test, and a writing task are viewed as summative assessments, whereas the rest of the activities are assessed continuously and formatively. Thus, across the term, the students are exposed to a variety of formative and summative assessment techniques, with the exception of peer- and self-assessment, which, according to the curriculum, are offered in Term 2. The course is completed with administering Questionnaire 1.
Since the instructor’s research and professional interests lay within the field of LTA, she conducted assessments ‘by the assessment book’, drawing from the works of
Coombe et al. (
2007),
Green (
2014), and
Tsagari et al. (
2018). In particular, she made a point of explaining the tasks to be performed/assessed in a meticulous fashion, indicating such necessary parameters as timing, score, example, answer grid (for the test format of tasks), and the assessment criteria (for productive skills). She would also strive for the test papers to be appropriately laid-out and have positive face validity, thus seeking to develop the learners’ diligence while using and explaining the tests. Regarding the assessment of speaking and writing, she made a point of introducing criteria, and even rating scales, to students and used and explained the necessary metalanguage, e.g., ‘coherence and cohesion’. Students’ performances were commented on throughout, with comments frequently containing suggestions as to what and how it should be improved.
As the classes during the academic year were conducted mostly online, due to the COVID-19 pandemic, the teacher made it a rule to assess students’ written assignments via Google classroom, providing corrections and suggestions wherever possible. When assessing written compositions electronically, the teacher would indicate relevant grids in the rating scale to make the scores—per criterion and total—more explicit and comprehensive. Additionally, the final score for written works was normally delayed, providing the students with an option to edit their work, thus making the most of the teacher’s written feedback.
It is worth noting that, although quite concentrated, the efforts the teacher put into instilling the norms of conducting assessments properly were far from being an intervention, as they did not stand out purposely and disrupt the natural flow of classroom activities. This was confirmed in conversations with the instructor, who was teaching the same curriculum in the other group of second year students. Additionally, the teaching and assessment process was monitored by the education managers, who did not notice any deviations from the normal instructional procedure. However, there was another purpose of conducting assessment by the book, beyond ensuring fair assessment within the course. The teacher meant for the students to observe assessment procedures repeatedly, systemically, and in compliance with testing principles or ‘cornerstones’ (termed so by
Coombe et al. 2007). As was hypothesized, the students could develop certain preconceptions of good assessment practice in a tacit way that would facilitate their further formal acquisition of knowledge about the principles of LTA. Last but not least, the teacher also considered the possibility of facilitating ‘pedagogical intelligence’ (
O’Donovan et al. 2008) in the GE class by inducting prospective teachers into the discourse of their future profession. Such early professionalization of learning had all the necessary prerequisites, since the students had already received training in pedagogy, pedagogical psychology, and the methods of teaching foreign languages.
4.2. Exploring LTA Course Delivery
The LTA course, under examination, addresses undergraduate students and has been conducted by the author of this article for four years. When the author first designed the LTA course, she ensured that the content is consistent with the students’ future professional needs, learning motivation, and mental maturity. It is also consistent with the language assessment culture adopted in the community, especially in respect of the compatibility of assessment practices with institutional or individual ideologies, social expectations, and attitudes and values, as
Inbar-Lourie (
2008) states that such a course should be. Moreover, the course contributes to shaping critical awareness of the students’ prospective assessment practices, following
Scarino’s views (
2013).
The course content comprises the
stable knowledge base (
Brown and Bailey 2008), which includes: purposes of tests, types of tests, parameters of test usefulness, principles of test development, scoring, and feedback, as well as theory and practice of item/task writing for testing receptive skills. The course is implemented as a series of workshops that are built on the generic principles of workshop design, i.e., with the objectives for constructing new knowledge set at the beginning, technique of heuristic talk employed throughout the class, theoretical input combined with hands-on experience of doing a variety of tasks by learners in teams, and eventual statement of the outcomes of the workshop activities. It is worth mentioning that the workshop format is still considered innovative, as it is not practiced broadly in the local higher education.
The course is delivered in English, although it is substantially tailored to trainees‘ English proficiency by using simpler structures and more frequent lexis, resorting to explanation of terminology, making links to the LTA course delivered in L1, delivering at a slower pace, and translating new concepts into L1. Additionally, the course is supported by multiple and diverse handouts, short glossaries, and presentation slides, available before the class (on students’ request). The formative assessment implemented includes teacher observation, questioning, quizzes, and items/tasks developed by teams. The summative assessment is comprised of the exit test and grades for the test tasks to test grammar, vocabulary, and reading, as developed by students individually. Profiling the content according to the tenets of task-based learning allows for covering the selected course content at a good pace within the allotted 14 contact hours. The course is completed with administering the exit test and Questionnaire 2.
Due to the pandemic of 2020–2021, the course was conducted remotely, which, to a certain extent, optimized its delivery. Sharing workshop slides enabled trainees to follow the oral input closely and in a focused, synchronous manner. It proved meaningful during theoretical input, as well as when performing practical tasks. Additionally, interaction in breakout rooms allowed the trainees to trial items/tasks collaboratively, thus recreating the collegiality of real-life test developers’ work.
Although some alterations did occur since the course was first described by
Kvasova (
2020), they pertained primarily to the teaching techniques (adaptations of the scenario), but not to the content of teaching. This fact allows for comparison of the learning outcomes achieved by the students of the current study (Set 1 of participants) with those achieved by the students in the previous research (
Kvasova 2020) (Set 2 of participants).
4.3. Analysis
The descriptive data analysis method was employed to interpret the results of the questionnaires.
6. Discussion
The data of the study seem to vividly demonstrate that the cohort of students that had previously been taught the GE course by the assessment-minded instructor have coped quite well with the content of the LTA course. Comparing the results of their exit test with the results of Questionnaire 2, it seems that the students, through their perception of their mastery of the course content, confirmed the quantitative data obtained. When asked about the most positive impact of their prior knowledge and experiences, they largely mentioned the GE course. This fact confirms that the active, although tacit, engagement of students with assessment concepts and principles, through being taught and assessed during this course, appears effective. The assumption about the feasibility of conceptions of “understanding of feedback” and “knowledge of assessment construct and criteria” seems positive.
It is reasonable to now present the author’s vision of the two differential types of LAL addressed in this study. The first is student initial AL, developed within the learners of the GE course. It may be defined as the understanding of assessment procedures, existing in a particular context, i.e., the ability to understand, interpret, and use feedback provided on one’s performance, as well as the conscious knowledge of assessment criteria/standards enabling the control and evaluation of own performance. This LAL is developed in an active, though tacit, way via a systemic exposure to assessment procedures properly organised.
The emerging ‘pedagogical intelligence’, or rather interaction of the instructor with the future colleagues, suggests the idea of certain apprenticeship. Unlike ‘folkways of learning’ in ‘apprenticeship of observation’ by
Lortie (
1975), this relationship stems from the community of interests, mentor’s willingness to share their teaching/assessing expertise, and mentee’s aspiration to acquire the profession. The lessons and insights gleaned from these learning experiences may further affect the prospective teachers’ identity and toolkit. However, special research is required to look into this issue if we want to discriminate between the mentor’s influence and the influence of course of life, circumstances, and other factors that affect becoming a professional.
The other type of student LAL is the one developed via the course in LTA. This is the prospective teacher LAL, which, following
Malone (
2017), should not be confused with the LAL of a practicing teacher. In line with
Scarino (
2013) and
Tsagari (
2017), the LAL of prospective teachers is viewed as fully context-dependent. We may describe it in a plethora of ways building on the well-known definitions of LAL, though with a significant amendment—the one regarding its relevance to, and consistence with, the curriculum of the course taught. The construct of this LAL will be collated with the curriculum and learner’s background and maturity, as well as the purpose of their studying LTA.
For instance,
Ukrayinska (
2018), delineating the construct of classroom LAL of the English majors in their 5th year of study, provides a ramified system of skills and subskills, divided further into lower- and higher-level subskills. In the case of the LTA course aimed at second year Ukrainian majors, as was discussed throughout the article, the course content is concentrated and organized in a constructivist framework, with the language of the course delivery adapted to the learners’ proficiency in English and techniques made as interactive as possible. For instance, when trained in a writing item to test language skills and reading, the students are engaged in peer-trialing and further modifications of tasks. Additionally, they are familiarized with assessing listening by doing test tasks of different formats themselves and further reporting on their experience and knowledge acquired. When it comes to the assessment of productive skills, the students are shown how to use rating scales and have an opportunity to assess original scripts/oral performances.
Both types of student LAL cannot be viewed outside the course content, institutional context, and, moreover, irrespective of individual perceptions and beliefs, as well as the predisposition towards learning and assessment, as parts of pedagogical work.