Next Article in Journal
Multilingual, Multimodal, and Multidisciplinary: Deaf Students and Translanguaging in Content Area Classes
Next Article in Special Issue
Teacher Beliefs and Practices of Language Assessment in the Context of English as a Lingua Franca (ELF): Insights from a CPD Course
Previous Article in Journal
Examining the Role of Phoneme Frequency in First Language Perceptual Attrition
Previous Article in Special Issue
A Bibliometric Analysis of Peer Assessment in Online Language Courses
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

The Relationship between Language Educators’ Perceptions and Assessment Practices during the COVID-19 Crisis

1
Higher Institute of Applied Studies in Humanities, University of Tunis, 92, Boulevard 9 Avril, Tunis 1938-1007, Tunisia
2
Department of Primary and Secondary Teacher Education, Oslo Metropolitan University, 0130 Oslo, Norway
*
Author to whom correspondence should be addressed.
Languages 2023, 8(1), 54; https://doi.org/10.3390/languages8010054
Submission received: 6 June 2022 / Revised: 28 January 2023 / Accepted: 30 January 2023 / Published: 12 February 2023
(This article belongs to the Special Issue Recent Developments in Language Testing and Assessment)

Abstract

:
This study investigated language educators’ readiness in coping with language assessment during the global outbreak of the COVID-19 pandemic and the enforced move to Emergency Remote Teaching (ERT). This pandemic prompted debates on language assessors’ roles in adjusting to a New Normal. While a good body of research has investigated the role of teacher assessment perceptions versus assessment behavior for many decades, little is known about the factors that may have impacted language educators’ assessment perceptions and practices during the recent crisis. To address this issue, an online survey was administered to 256 language educators. Pearson correlations and simple linear regression were utilized to determine if the language educators’ perceptions of (1) the official assessment measures, (2) purposes, and (3) their assessment self-efficacy were predictors of their assessment practices during this crisis. The results revealed a total absence of any correlations between these variables. The findings suggest that the assessment accommodations adopted by the teachers were not determined by their assessment perceptions. Other factors such as assessment policy and the assessment culture may have shaped their practices during this crisis.

1. Introduction

The global outbreak of the COVID-19 pandemic has dramatically changed the educational landscape around the world, affecting the way language educators deal with assessment. Educational stakeholders at all educational levels have encountered numerous challenges in ensuring pedagogical continuity in their educational systems, including the way they plan and implement assessment (Ghanbari and Nowroozi 2021). Among the challenges encountered were the difficult planning and administration of end-of-year assessments during this long-lasting crisis (UNESCO 2020a; Clark et al. 2020; Nădrag 2020). These unprecedented conditions prompted debates in various educational disciplines on ways of coping with the “New Normal” (d’Orville 2020; Pacheco 2020) for teachers and other education stakeholders worldwide. The New Normal “forces education further into technologization, a development already well underway” (Pacheco 2020, p. 5), leading to a rethinking of language educators’ preparedness (Gouëdard et al. 2020; Hodges et al. 2020; Hodges and Barbour 2021; Paramour 2020) in terms of their perceptions and their possible impact on their language assessment practices when coping with ERT (Emergency Remote Teaching). The latter refers to the sudden move to remote education because of the distortion of onsite schooling. This has forced many educators worldwide to organize and implement assessments differently (Ockey 2021; Rahim 2020; Reimers and Schleicher 2020; Zhang et al. 2020) while maintaining quality and professionalism (Muhammad and Ockey 2021).

2. Background

Today’s global educational landscape under the COVID-19 crisis calls upon language teachers to use their skills as assessors perhaps more than ever. Research has emphasized teachers’ roles in implementing assessment policy while being at “the frontline as designers and users of language assessments” (Kremmel and Harding 2020, p. 101). The turmoil caused by school closure during the pandemic has initiated a debate on teachers’ role in planning and implementing education (Anderson 2020; Lightfoot 2020; Rahim 2020; UNESCO 2020a, 2020b, 2020c) and language assessment (British Council 2020; Gao and Zhang 2020) worldwide. A study conducted by the British Council (2020) with language teachers and teacher educators in different countries has shown that the forced and unplanned shift to online teaching may be particularly challenging for teachers as they need to develop their skills in effective online assessment. Unlike previous educational emergency research (Salerni and Vaccarelli 2019; MacNeil and Topping 2007) concerned with the question of resilience and “psychological, social, and community dimensions’’ (Vaccarelli 2018, p. 29), the main challenge during the COVID-19 crisis has been to take urgent measures regarding both teaching and assessment during the lockdown (Anderson 2020). The current crisis has had implications mainly on high-accountability systems in the USA (US Department of Education 2020), on high-stakes exams such as the GCSE and the A-level in the UK (Lightfoot 2020; Miller 2020; UK Government 2020), and on end-of-year assessments in the countries affected by the pandemic. The main challenge decision-makers faced was related to the management of scheduled assessments, including high-stakes ones (d’Orville 2020; UNESCO 2020c, 2020d, 2020e, 2020f).
The post-COVID research has raised issues about teachers’ “pedagogical readiness” (Rapanta et al. 2020; Tuah and Naing 2021) in coping with assessment challenges under ERT (Hodges et al. 2020). Limited education (Rahim 2020; Reimers and Schleicher 2020) but almost no language assessment research has pointed to the need for proactive measures to prepare teachers for more principled and effective assessment measures. The unprepared shift from onsite to remote education during ERT was the only possible measure to reduce the destabilizing effect of the COVID-19 crisis on schooling. Educators had to adapt to an online mode for their teaching and assessment alike (Rahim 2020; Reimers and Schleicher 2020). Nevertheless, the lack of theory-based and planned solutions regarding assessment (Iglesias-Pradas et al. 2021) has compelled language educators to implement responsive measures (British Council 2020; Dill et al. 2020; US Department of Education 2020; Paesani 2020; Radcliffe et al. 2020; UNESCO 2020a; Whittle et al. 2020; Woolcock 2020) and assessment accommodations (Marinoni et al. 2020; Reimers and Schleicher 2020; UNESCO 2020c) as an adaptation to this novel situation. Traditionally, assessment accommodations (Abedi 2013) concern changes to meet learners’ special needs (Abedi et al. 2000; Pennock-Roman and Rivera 2011; Taylor 2012; Tsagari and Sperling 2017). The concept of assessment accommodations is used in this study to refer to a strategy, change, or adjustment to the new situation by canceling, postponing, or rescheduling planned exams, moving to online assessment, or opting for alternative modes of assessments (UNESCO 2020a).
Of relevance to the question of teacher preparedness (Anderson 2020; Gouëdard et al. 2020) is teachers’ assessment perceptions (Mäkipää et al. 2021; Rapanta et al. 2020). The ERT context requires the consideration of assessors’ perceptions as a possible determining factor of their behavior from a “teacher as assessor” perspective (Looney et al. 2017). While a plethora of studies has attended to the role of teacher perceptions in determining assessment behavior for many decades (Brown 2004; Scarino 2013), limited post-COVID exploratory research has emerged to bring to light language teachers’ perceptions of online assessment, the assessment methods utilized (Abduh 2021), and the difficulties encountered during the lockdown (Abduh 2021; Ghanbari and Nowroozi 2021). Prior to the current crisis, Zulaiha et al.’s (2020) study revealed a mismatch between English as a foreign language teachers’ assessment perceptions and practices. In the context of ERT, Rapanta et al. (2020) examined teachers’ perceptions of the effective monitoring of students’ learning online and its implications for assessment. Their findings confirm the teachers’ need for more adaptation time and effort for more effective and stress-free assessment practices. In their report, Reimers and Schleicher (2020) reveal compelling findings regarding teachers’ views of the official decisions during the COVID-19 crisis. This study aims at examining language educators’ perceptions of the responsive measures taken and the assessment purposes during the crisis. Such purposes are based on an adapted taxonomy of Brown (2004), including the two main purposes of (a) improving learning through feedback and (b) school accountability.
Among the key aspects of educator perceptions is self-efficacy as a developmental construct (Bandura 1994), which could be a key predictor of educators’ pedagogical practice (e.g., Bandura 1997; Myyry et al. 2021; Poulou et al. 2019). Despite its recognition as a personal construct in educational assessment (Brown 2004; Xu and Brown 2016), limited research has addressed it in the language assessment literature (Levy-Vered and Nasser-Abu Alhija 2015). According to social-cognitive theory, four main factors can shape people’s perceptions of their own efficacy. “Mastery experiences”, including certification, professional development, and years of experience, are the most effective way of developing a strong sense of efficacy (Bandura 1994). Assessment self-efficacy (Bandura 1994; Kruse et al. 2020; Levy-Vered and Nasser-Abu Alhija 2015) is operationally defined for this study as ‘assessors’ personal beliefs about their confidence in planning and performing language assessment tasks and assessment accommodations during the crisis’. Research has shown that assessment self-efficacy can be shaped by assessment literacy and experience (Levy-Vered and Nasser-Abu Alhija 2015). In the context of a crisis such as the current one, teachers’ reported self-efficacy could be associated with the efforts they would deploy “on an activity, how long they will persevere when confronting obstacles, and how resilient they will prove in the face of adverse situations” (Pajares 1996, p. 544). Exceptionally, studies employing quantitative analyses including correlation (Mueller et al. 2008) sometimes failed to reveal any positive relationship between teacher self-efficacy and technology use as an example of classroom practice.
Despite the overall consensus on the significance of teacher factors in determining their practice, little research has focused on such issues during the crisis. Studies accounting for teachers’ roles (Cahapay and Anoba 2021; Ghanbari and Nowroozi 2021), cognition (Gao and Zhang 2020), or teachers’ formative assessment engagement in English as a foreign language settings (Zou et al. 2021) under the crisis have been rather scant. A study using a duoethnographic approach has explored three English language teachers’ assessment literacy (i.e., assessment knowledge and skills) before and after COVID-19 (Tian et al. 2021). The teachers who sought solutions to the post-COVID assessment issues reported an expanded sense of self-questioning, leading to a felt need to develop their assessment literacy. Studies by Ghanbari and Nowroozi (2021) and Watermeyer et al. (2021) have also explored language teachers’ ability and preparedness to deal with the demands of ERT. Jelińska and Paradowski’s (2021) study examined teachers’ engagement during the transition to ERT. Watermeyer et al.’s (2021) survey of teachers’ reactions to ERT provided evidence of these educators’ strong confidence in their ability as both teachers and assessors online. Post-COVID research on language teachers’ experience during ERT has shown that teachers with no prior online teaching experience exerted personal efforts to be capable of teaching remotely (Hodges et al. 2020; Juárez-Díaz and Perales 2021). Juárez-Díaz and Perales (2021) alluded to the teachers’ need to adjust to the assessment conditions of ERT. Even though language teachers from 150 countries demonstrated moderately high confidence in their ability to teach online, some teachers overtly reported that they were rather overwhelmed by the sudden transition to ERT (British Council 2020).
On the basis of the above discussions, this paper will then try to test the following hypotheses:
Hypothesis 1 (H1).
There is a significant relationship between the language educators’ perceptions of the assessment measures mandated by the authorities of their countries and their assessment practices during the crisis.
Hypothesis 2 (H2).
There is a significant relationship between the language educators’ perceptions of assessment purposes during the crisis and their own assessment practices.
Hypothesis 3 (H3).
There is a significant relationship between the language educators’ self-efficacy and their reported post-COVID assessment practices.

3. Materials and Methods

3.1. Research Site and Participants

The study data were collected based on a combination of both purposeful and snowball sampling techniques, with the sample not necessarily representing the larger population of language educators worldwide. A cross-sectional survey questionnaire was developed, piloted, and repiloted to assess the instrument’s internal reliability and validity among the teacher network of the current authors. To reduce the participants’ reluctance and aversion to respond to the survey, a Google Form of the questionnaire was created and disseminated by email or uploaded on the authors’ professional and personal networks during the lockdown from May to July 2020. The researchers managed to recruit a total of 256 participants through a non-probability sampling technique.
The participants’ selection can be justified by their shared “specialized knowledge” (Ruel et al. 2016). Consequently, no generalizable claims can be made. These participants were from four main regions. They were all language educators, 88.7% of whom were language teachers from secondary and higher education, and 11.3% were language educators including teacher trainers and test developers. While 38.4% were from the Middle East and North Africa (MENA) region, 35.1% were from Europe; the remaining participants were from America and other countries around the globe. It is also worth mentioning that 41% had never taught online before the crisis. According to Figure 1, the participants also reported that 40.63% of them “never” and 34.77% “sometimes” taught online prior to the crisis. Question 4 in Section B also revealed that the majority (76.6%) of the respondents used online assessment during ERT (see Figure 1 below).

3.2. Data Collection

To understand language educators’ assessment self-efficacy, it was also necessary to examine their assessment background. The results in Table 1 below are based on the analysis of the responses to Question 7 of our survey (see Appendix A). They indicate the relative frequency values of the participants’ assessment literacy source. It is clear from the results that the majority (0.863) of the participants developed their assessment literacy mainly thanks to pre-service training, while in-service training occupied the second most important professional assessment development activity (0.773). This was followed by professional development events such as conferences, workshops, and seminars (0.667). The fourth source of assessment literacy development for these language educators was experience on the job (0.617), followed by online resources such as websites, courses, and webinars (0.61). Of equal importance to the development of the participants’ assessment literacy were doing research in assessment (0.49) and in-service training (0.52). Their assessment literacy also developed thanks to collaborating with colleagues and the institution (0.57) in which the participants work.

3.3. The Survey

This study was motivated by COVID-19. It examined three types of associations between the three independent variables of (1) language educators’ perceptions of the assessment measures, (2) their perceptions of assessment purposes, (3) their assessment self-efficacy, and one dependent variable, (4) the variable of language assessment practices during the crisis. The language educators’ questionnaire consisted of four sections (see Appendix A). The items were designed based on the hypotheses guiding this study. Section A was devoted to reporting on the measures taken in the participants’ contexts regarding language assessment under COVID-19. It included one main question allowing the respondents to choose as many options as needed. Section B elicited information on the educators’ assessment practices during the COVID-19 crisis. Question 2, in particular, concerned the assessment accommodations they adopted during the crisis using multiple-choice items, while Question 3 focused on the assessment tasks they used. Question 4 aimed to understand whether these educators conducted online assessments during the crisis using a three-point scale of “yes”, “no”, and “not yet”. Section C, using an agreement Likert scale of five points, elicited information regarding the participants’ perceptions of the language assessment measures taken in their contexts (Question 5), their perceptions of assessment purposes (Question 6), and their perceived assessment self-efficacy (Question 7). The last section collected data on their profiles, including their assessment background (Questions 8 to 17).

3.4. Data Analysis

The data analysis was mainly quantitative. The internal reliability of the survey items comprising Likert scales was examined using Cronbach’s alpha. The findings indicated that the alpha for the total scale was equal to 0.7 and hence deemed acceptable. All statistical analyses were carried out using the Statistical Package for the Social Sciences (SPSS) version 23. In addition to percentages to describe the educational contexts in terms of the assessment measures taken and the participants’ profiles, Pearson correlations and simple linear regression were employed to determine if the participants’ factors of perceptions predicted their assessment practices during this crisis.
A principal components analysis was conducted on the 10 items of the perceptions of language assessment measures during the COVID-19 crisis to determine whether they represent a single construct. An exploratory factor analysis was performed employing a principal component analysis and varimax rotation. In factor extraction, three main criteria were used: (1) a minimum of three items in one factor with an eigenvalue of one or greater; (2) factor loadings less than 0.4 were deleted and not counted in any factor; and (3) items with double loadings were removed. This analysis yielded one factor with an eigenvalue of 2.879, accounting for 47.981% of the total variance for this factor. All items loaded higher than 0.60 on the factor. Factor analysis results indicated that six survey questions clustered together around teachers’ perceptions of the measures taken about high-stakes exams with a Cronbach alpha of 0.767. This indicates that the construct is valid for further analysis. The KMO measure (Kaiser–Meyer–Olkin; Kaiser 1970) of sampling adequacy is 0.756.
For Question 5 regarding these language educators’ perceptions of assessment purposes during the crisis, a principal components analysis was performed on the six items to determine whether they represent a single construct. The analysis yielded one factor with an eigenvalue of 2.153, accounting for 53.833% of the total variance with a Cronbach alpha of 0.708, indicating that the construct is valid for further analysis. The KMO measure of sampling adequacy is 0.683.
Regarding the participants’ perceived assessment self-efficacy, a principal components analysis was equally conducted on the nine items to determine whether they represent a single construct. This analysis yielded two factors, with an eigenvalue of 4.802 for Factor 1, accounting for 53.36% of the total variance and an eigenvalue of 1.129, and a total variance of 12.545 for Factor 2 and a Cronbach alpha of 0.881. The factor analysis results indicated that four survey questions clustered together around the language educators’ self-efficacy. The KMO measure of sampling adequacy is 0.888.
For the last variable of the post-COVID assessment practices, the 27 items received a principal components analysis and resulted in seven factors with eigenvalues of 6.497, 3.031, 2.350, 2.194, 1.942, 1.741, and 1.439 for Factors 1, 2, 3, 4, 5, 6, and 7, respectively. The KMO measure of sampling adequacy is 0.730, and has Cronbach alpha of 0.692. The seven factors were calculated for further analysis.

4. Results

4.1. The Assessment Context in the Different Countries

The statistical analysis below serves first to describe the assessment landscape immediately after the outbreak of the COVID-19 crisis. As displayed in Table 2 (see below), the official responsive measures due to school closure indicate that there was a partial shift to organizing exams online with a relative frequency of 0.515 in the participants’ countries. In their answers to the first survey question, the educators reported that the crisis led to postponing exams for a later date (0.363), replacing exams with alternative assessments such as projects and essays (0.351), reducing the subject content (0.285), offering exam(s) as before (0.257), using written take-home exams (0.234), offering part of the exams by opting for either an oral or written form (0.222), using a plagiarism detector for online exams (0.195), reducing the number of subjects tested (0.164), canceling exam(s) altogether in only a few contexts (0.101), and ‘Using an online proctoring service’ (=online surveillance, e.g., Proctorio, ProctorU, Examity, etc.) and ‘Using test results from previous semesters/years’ (0.093). However, ‘Using teacher assessments from previous semesters/years’ represented only 0.082.

4.2. Perceptions of Assessment Measures during the Pandemic

The results in Table 3 below summarize these language educators’ perceptions of the responsive assessment measures under COVID-19. When asked if their countries were managing the assessment situation effectively, the participants were either neutral (29.3%) or disagreed with the statement (28%). In addition, about half of them (45.3%) refused the decision of exam cancellation despite the near-total closures of educational institutions. About half of these educators also believe that exams should be maintained because of their importance to learners (54.5%) and society (42.2%). However, 72% disagree with the decision to maintain only exams. It is important to highlight that 67.5% are against exam cancellation because of the importance they represent to them as language teaching and testing stakeholders.
Table 4 displays the participants’ opinions concerning assessment purposes during the crisis. While the majority of opinions (95.8%) reflected an awareness of the importance of assessment during a crisis to create opportunities for feedback, 89.8% agree with the need for assessment because it is part of the teaching and learning cycle. A total of 79.6% reported that they would use assessment during a crisis because it is mandated in learning curricula, and 75.8% think that assessment is needed for decision making, including student progress to the next level or for exit requirements.

4.3. The Language Educators’ Assessment Self-Efficacy

One of the study goals was to understand the role assessment self-efficacy might have played in the study of participants’ assessment accommodations. Thus, the participants were invited to describe their self-efficacy as assessors in Question 5. As displayed in Table 5, the results show that the educators generally reported having moderate to high self-confidence in their roles as assessors. Around half of them were moderately confident (59.8%) in their ability to design test questions and scoring criteria (46.9%) and give feedback (66%). Their assessment self-efficacy was somewhat higher for scoring written papers, with 37.1% who were moderately confident and 54.3% who were very confident. However, they were equally moderately (41.4%) confident and highly confident (40.6%) in their ability to assess their students via self- or peer assessment.

4.4. The Impact of the Perceptions of the Measures on Language Assessment Practices

This section aims to report whether there is a significant relationship between the language educators’ perceptions of the assessment measures taken by their countries and their assessment practices during the crisis. First, correlation analyses were used to examine this relationship and yielded a correlation of −0.118. Next, simple linear regression was used to test if language educators’ perception of the assessment measures significantly predicted their assessment practices. The fitted regression model was as follows: assessment practices = 81.586 + 1.148 (perception of the assessment measures). The overall regression was statistically insignificant (R2 = 0.014, F(1, 254) = 3.575, p < 0.060), indicating that the model explained only 1.4% of the variance in the data that can be explained by the language educators’ perception of the assessment measures. As displayed in Table 6, the language educators’ perceptions of the assessment measures did not significantly predict assessment practices (β = −0.014, p = 0.060).

4.5. Perceptions of Language Assessment Purposes as a Predictor of Assessment Practices

Concerning the correlation between perceptions of assessment purposes during the crisis and assessment practices = 0.069. Simple linear regression was also used to test if the language educators’ perceptions of the assessment purposes during the crisis significantly predicted their assessment practices during this period.
The fitted regression model was as follows: assessment practices = 82.625 + 0.109 (language educators’ perception of assessment purposes during the crisis).
According to Table 7, the overall regression was statistically insignificant (R2 = 0.001, F(1, 254) = 0.337, p < 0.562) and indicated that the model explained only 1% of the variance in the data.
It was found that such perceptions of assessment purposes did not significantly predict the educators’ assessment practices during the crisis (β = −0.006, p = 0.562).

4.6. The Correlation between Self-Efficacy and Assessment Practices

To test the hypothesis of whether there is a significant relationship between the language educators’ assessment self-efficacy and their reported post-COVID assessment practices, simple linear regression was used. Table 8 provides a summary of the analysis, showing that assessment practices were the dependent variable and assessment self-efficacy was the predictor variable.
The fitted regression model was as follows: assessment practices = 82.336 + 0.399 (the educators’ perceptions of their assessment self-efficacy). The overall regression was statistically insignificant (R2 = 0.005, F(1, 254) = 1.231, p < 0.268), indicating that the model explained only 5% of the variance in the data. It was found that their perceived assessment self-efficacy during the crisis did not significantly predict their assessment practices (β = −0.014, p = 0.268).

5. Discussion

This study was specifically designed to gather evidence concerning the hypotheses about language educators’ perceptions predicting their assessment practices during the crisis. First, the results provided a description of the assessment accommodation characteristics in the various educational contexts investigated. One of the first findings describing the assessment contexts revealed that the assessment measures in their respective countries concerned mainly moving assessment online despite the limited experience these educators had prior to the crisis. This confirms that this New Normal was quite challenging in terms of the teachers’ adaptation to the online assessment mode, as put forth by the majority of researchers during the crisis (Gao and Zhang 2020). Additionally, practical issues were probably at stake not only in the administration of face-to-face assessment but also in online assessment during the crisis, especially because of the teachers’ lack of preparedness in terms of the online assessment.
The findings also revealed that despite the overwhelming nature of the crisis and the absence of clear assessment guidelines, these language educators’ perceptions reflected contradictory perceptions about the assessment measures. Despite the near-total closures of educational institutions during the crisis, two patterns of perceptions appeared in the findings. While about half of these participants believe that exams should be maintained for different professional and accountability reasons, the other half demonstrated more flexibility as to the adoption of alternative assessment methods. The pattern found among these language educators’ perceptions is consistent with previous research findings reflecting beliefs about assessment of learning and assessment for learning as two major assessment paradigms (Brown 2004). Thus, the move to ERT did not systematically compel them to abandon their previous views of the need for formal and summative assessment, most probably in exam-driven assessment cultures. This shows that the crisis did not necessarily lead to a total shift in the language educators’ assessment perceptions as a whole.
It was also found that these educators’ assessment self-efficacy was moderate to high, reflecting their general confidence in their roles as language assessors despite the destabilizing effect of the crisis. This may show that the crisis did not affect their self-perceptions and ability to carry out both formative (i.e., continuous) and summative assessments. These findings are in line with Watermeyer et al.’s (2021) results. However, according to the regression analysis results, assessment self-efficacy did not determine these language educators’ assessment practices despite the forced move to ERT and the importance of self-efficacy in preparing these education stakeholders to face the new assessment demands.
Regarding H1, the findings show that a change in these educators’ assessment practices is not predicted by a change in their assessment perceptions. This finding may confirm that assessors’ perceptions were not necessarily a determining factor of their behavior in the ERT context (Looney et al. 2017). However, this confirms previous research findings (Zulaiha et al. 2020), revealing a possible gap between teachers’ assessment perceptions and practices. Concerning H2, no significant relationship was found between the language educators’ perceptions of assessment purposes and their own assessment practices during the crisis. Likewise, H3 was rejected as no significant relationship between the language educators’ self-efficacy and their reported post-COVID assessment practices was revealed. Unlike previous research findings on the role of self-efficacy (Pajares 1996; Poulou et al. 2019), assessment self-efficacy did not shape these educators’ pedagogical practice and efforts on assessment during the first period of the crisis.
Likewise, these language educators’ perceptions of the assessment measures examined in this study were not found to be predictors of their assessment accommodations and adjustment to the New Normal. Interestingly, the crisis did not generally lead to a conceptual shift in the way these language educators perceived language assessment purposes, despite their confrontation with various assessment challenges imposed by school closure. This may raise more issues about the language teachers’ and teacher trainers’ flexibility in accepting some necessary changes during the crisis. In other words, despite their high assessment self-efficacy, their beliefs that formal assessment through exams had to be maintained may be indicative that they were not fully ready for this crisis (Tuah and Naing 2021). The complexity of the situation might have been intensified by these language educators’ relative openness to adaptations due to the crisis on the one hand and the difficulties of embracing any changes on the other, mainly because of their limited experience in online assessment. Institutional factors such as the subject taught, the learners’ level, the available resources, and the local assessment culture might have also played a role in shaping both their beliefs and practices. The results seem to confirm initial concerns about educational crisis management issues in different contexts worldwide (UNESCO 2020a, 2020b, 2020c).
The absence of any correlations between the assessment perceptions and practices of the participating language educators during the COVID-19 crisis may generally suggest that the accommodations were not dependent on teacher factors such as their assessment self-efficacy as much as other contextual factors. The educational context, prevailing assessment culture, the educators’ language assessment literacy, and technological infrastructure may have played a role in the adoption of the measures and the implementation of the assessment accommodation strategies by the study participants. All these factors may have contributed to the complexity of managing the situation effectively from a language assessment perspective.

6. Conclusions

The findings of the present study have interestingly shown an absence of any relationship between the language educators’ assessment perceptions and practices during the COVID-19 crisis. Researching the questions related to the responsive measures vs. perceptions and practices regarding assessment during crises has been found to be fraught with some challenges. The difficulty in addressing these issues was partly due to the large scale of the crisis and its impact on assessment at various educational levels. In addition to this, the use of a single data collection instrument and quantitative analyses may also need to be further confirmed in more specific educational settings while accounting for contextual factors such as the available technology infrastructure, learner computer literacy, and attitudes to online assessment. This study focused on language educators’ perceptions and practices during their expected adjustment to a demanding New Normal. Future research on assessment during crises should probably deal with the impact of these assessment accommodations on students’ perceptions and performance. This would lead to a better understanding of learners’ needs during the pandemic. Future research should develop an assessment model and/or guidelines to help language educators cope with the crisis more effectively. This could contribute to the adoption and implementation of more informed decisions regarding ERT in general and assessment in particular. From a pedagogical perspective, language educators need to gain more knowledge and skills related to managing assessment for learning sustainability rather than just pedagogical continuity and student resilience in times of crisis. From this perspective, given the long-term nature of the crisis and the need for ERT as a strategic step for learning continuity, teaching and assessment should move beyond mere “psychosocial support” (Anderson 2020) to incorporate more planned research-based assessment strategies at national and international levels.

Author Contributions

Conceptualization, A.M., D.T. and H.D.; methodology, A.M., D.T. and H.D.; software, A.M., D.T. and H.D.; validation, D.T.; formal analysis, A.M., D.T. and H.D.; investigation, A.M., D.T. and H.D.; resources, A.M., D.T. and H.D.; data curation, A.M., D.T. and H.D.; writing—original draft preparation, A.M., D.T. and H.D.; writing—review and editing, A.M., D.T. and H.D.; visualization, A.M., D.T. and H.D.; supervision, A.M., D.T. and H.D.; project administration, A.M., D.T. and H.D. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Ethical review and approval were waived for this study, after consultation with the Norwegian Centre for Research Data due to anonymous data collection which did not include sensitive or identifying personal data.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

Data supporting the reported results is securely stored.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

Assessment in Education during COVID-19 Survey
Dear colleagues,
We would like to kindly invite you to take part in this study that examines the way language teachers in different educational sectors have been responding to assessment challenges in the COVID-19 pandemic.
Should you wish to participate, please note:
-
The survey contains 17 questions and will take about 15 minutes to complete.
-
Your participation will remain completely anonymous.
-
The data collected will be analyzed by the researchers mentioned below and used only for research purposes related to educational assessment during the COVID-19 pandemic. You can withdraw your data at any time by contacting [Details omitted for double-blind reviewing] at [Details omitted for double-blind reviewing]
Please feel free to share this questionnaire with your colleagues!
We really appreciate your contributions and time!
Thank you!
I have read the guidelines and agree to take part in the survey
No
Yes
Section A. Language Assessment under COVID-19
In this section, we would like to know about language assessment during the COVID-19 crisis.
1. Select all the measures taken in your context regarding assessment under COVID-19 (select as many as apply). *
-
Offering exam(s) as before
-
Cancelling exam(s) altogether
-
Postponing the dates of the exam(s) for a later date
-
Offering part of the exam(s), e.g., only oral or only written
-
Reducing the number of subjects tested
-
Reducing the subject content
-
Conditional admission to a university, complemented by remedial courses upon school reopening
-
Organising exam(s) online
-
Using written take-home exams
-
Using an online proctoring service (=online surveillance, e.g., Proctorio, ProctorU, Examity, etc.)
-
Using plagiarism-detection services (e.g., Turnitin)
-
Replacing exams with alternative assessments (e.g., projects, essays, etc.)
-
Using test results from previous semesters/years
-
Using teacher assessments from previous semesters/years
-
Other:
Section B. Your Assessment Practices during the Crisis
2. The assessment measures I have taken under COVID-19 concerned changing (choose as many as you use) *
The assessment form e.g., replaced exams/tests with projects, presentations, quizzes, assignments…
The scope of assessment e.g., assigned less or more content
The focus of assessment, e.g., the types of knowledge or skills
The assessment medium, e.g., using digital assessment
The assessment time, e.g., postponed until later in the semester or next academic year
The assessment duration e.g., allowed more or less time
The assessment weight e.g., allocated more or less weight/grades on the basis of exam results
Other:
3. I have used the following assessment tasks: *
1. before COVID-19; 2. during COVID-19; 3. not yet; 4. not allowed to use it
1234
Essays
Multiple-choice questions
Short-answer questions
Oral exam
Fill in the gaps/cloze questions
Oral exam on the basis of a project or set task
Multiple-matching questions
Self-assessment tasks
Peer-assessment tasks
E-portfolios
Project-based tasks
Open-book questions
 
4. Have you conducted any type of online assessment during COVID-19? *
Yes
No
Not yet
Section C. Assessment Perceptions
In this section, we would like to know about your opinions regarding the assessment.
5. I would assess students’ learning during a pandemic. *
1. Strongly agree; 2. Agree; 3. Neutral; 4. Disagree; 5. Strongly disagree
12345
To maintain ‘normality’/routine?
To create opportunities to provide feedback
To meet societal expectations
Because assessment is part of the teaching/learning cycle
Because assessment is mandated in learning curricula
Because of decision-making, e.g., progress to the next level, exit requirement, etc.
6. Please indicate your opinion concerning the measures taken regarding language assessment in your context as a result of COVID-19. *
1. Strongly agree; 2. Agree; 3. Neutral; 4. Disagree; 5. Strongly disagree
I believe that 12345
Cancelling exams because of the pandemic is unacceptable
Exams are so important to learners that they should not be cancelled
Exams are so important for society in general that they should not be cancelled
Exams are important to me as an educator/expert
My country is managing the assessment situation in an effective way
Assessment can take place online under COVID-19 in my context
Assessment can take other forms, e.g., projects, presentations, assignments, etc.
Only exams should be maintained as a way of assessment
Institutional/regional exams can be replaced by teacher assessments that took place before COVID-19
National exams can be replaced by teacher assessments that took place before COVID-19
 
7. How confident in your ability are you to perform the following? *
1. Very confident; 2. Moderately Confident; 3. Somewhat confident; 4. Not too confident; 5. Not at all confident
12345
Designing test questions
Designing scoring criteria
Scoring written papers
Scoring oral performance
Giving students feedback based on assessment results
Using self- or peer-assessment
Using portfolio assessment
Using other types of assessment (e.g., presentations, projects…)
Using online tools for assessment
Designing test questions
Designing scoring criteria
Scoring written papers
Scoring oral performance
Giving students feedback based on assessment results
Using self- or peer-assessment
Using portfolio assessment
Using other types of assessment (e.g., presentations, projects...)
Using online tools for assessment
 
Section D. Demographics and assessment background
In this section, we would like to know about you and your teaching background.
8. What is your gender? *
-
Male
-
Female
-
Prefer not to say
9. Which country are you working in now? *
10. What is your main job? *
-
Teacher
-
Teacher educator/trainer
-
University lecturer
-
Other:
11. Where do you mainly work? *
-
Primary School
-
Secondary School
-
Higher/Further Education
-
Language School/Institution
-
Examination Committee/Agency
-
I work freelance
-
Other:
12. Which sector of education do you work for? *
Public  Private
13. What is your highest qualification/certification? *
-
Bachelor or equivalent
-
MA/MSc
-
Doctorate/Ph.D.
-
Other
14. How long have you been working on your post above? *
-
Less than two years
-
2–5 years
-
6–10
-
11–15
-
16–20
-
More than 20 years
15. What language do you mainly teach/train in?
-
English
-
French
-
Spanish
-
German
-
Chinese
-
Italian
-
Arabic
-
Greek
-
Other
16. Have you ever taught online before COVID-19?
-
Always
-
Very often
-
Often
-
Sometimes
-
Never
17. I learned about assessment through … (you may choose more than one answer): *
-
Pre-service training
-
In-service training
-
Professional conferences, workshops, seminars, etc.
-
Experience on the job
-
Μy colleagues/institution
-
Online resources such as websites, courses, webinars, etc.
-
Doing research in this area
-
Other:
Follow-up
Thank you for taking the time to complete the questionnaire. If you have any questions or comments, please contact [Details omitted for double-blind reviewing]
 
We will be doing follow-up interviews to find out more about your assessment experiences and challenges under the COVID-19 pandemic. If you would like to participate in the follow-up study, please provide your email below.
 
If you do, we would like to assure you that your responses will be de-identified for any publication or dissemination purposes.
 
Your email: ----------------------------------------------------------------------------------------------------

References

  1. Abduh, Mariam Yousef Mohammed. 2021. Full-time Online Assessment during COVID-19 Lockdown: EFL Teachers’ Perceptions. Asian EFL Journal Research Articles 28: 11–21. Available online: https://www.researchgate.net/publication/349103573_Full-time_Online_Assessment_during_COVID_-19_Lockdown_EFL_Teachers%27_Perceptions (accessed on 5 November 2022).
  2. Abedi, Jamal. 2013. Accommodations in the assessment of English language learners. The Companion to Language Assessment 3: 1115–29. [Google Scholar] [CrossRef]
  3. Abedi, Jamal, Carol Lord, Carolyn Hofstetter, and Eva Baker. 2000. Impact of accommodation strategies on English language learners’ test performance. Educational Measurement: Issues and Practice 19: 16–26. [Google Scholar] [CrossRef]
  4. Anderson, Allison. 2020. COVID-19 Outbreak Highlights Critical Gaps in School Emergency Preparedness. Available online: https://www.brookings.edu/blog/education-plus-development/2020/03/11/covid-19-outbreak-highlights-critical-gaps-in-school-emergency-preparedness/ (accessed on 10 November 2021).
  5. Bandura, Albert. 1994. Self-Efficacy. Encyclopedia of Human Behavior. Edited by Vilayanur S. Ramachaudran. New York: Academic Press, pp. 71–81. [Google Scholar]
  6. Bandura, Albert. 1997. Self-Efficacy: The Exercise of Control. New York: Freeman. [Google Scholar]
  7. British Council. 2020. English Language Teaching and COVID-19 A Survey of Teacher and Teacher Educator Needs during the COVID-19 Pandemic April–May 2020. London: British Council. Available online: https://www.teachingenglish.org.uk/sites/teacheng/files/covid19-teacher-teacher-educator-survey.pdf (accessed on 15 September 2020).
  8. Brown, Galvin. 2004. Teachers’ conceptions of assessment: Implications for policy and professional development. Assessment in Education: Principles, Policy & Practice 11: 301–18. [Google Scholar] [CrossRef]
  9. Cahapay, Michael B., and Jeorge Louie D. Anoba. 2021. Technological pedagogical knowledge self-efficacy and continuance intention of Philippine teachers in remote education amid COVID-19 crisis. Journal of Pedagogical Research 5: 68–79. [Google Scholar] [CrossRef]
  10. Clark, Ted M., Christopher S. Callam, Noel M. Paul, Matthew W. Stoltzfus, and Daniel Turner. 2020. Testing in the time of COVID-19: A sudden transition to unproctored online exams. Journal of Chemical Education 97: 3413–17. [Google Scholar] [CrossRef]
  11. d’Orville, Hans. 2020. COVID-19 causes unprecedented educational disruption: Is there a road towards a new normal? Prospects 49: 11–15. [Google Scholar] [CrossRef]
  12. Dill, Emma, Karin Fischer, Beth Mcmurtrie, and Beckie Supiano. 2020. As coronavirus spreads, moving classes online is the first step. What’s next. The Chronicle of Higher Education, 4–7. Available online: https://my.aup.edu/sites/default/files/imce/file/TLC/As%20Coronavirus%20Spreads%2C%20Moving%20Classes%20Online%20is%20the%20First%20Step.%20What%27s%20next%3F.pdf (accessed on 10 January 2021).
  13. Gao, Lori Xingzhen, and Lawrence Jun Zhang. 2020. Teacher learning in difficult times: Examining foreign language teachers’ cognitions about online teaching to tide over COVID-19. Frontiers in Psychology 11: 2396. [Google Scholar] [CrossRef]
  14. Ghanbari, Nasim, and Sima Nowroozi. 2021. The practice of online assessment in an EFL context amidst COVID-19 pandemic: Views from teachers. Language Testing in Asia 11: 1–18. [Google Scholar] [CrossRef]
  15. Gouëdard, Pierre B., Beatriz Pont, and Romane Viennet. 2020. Education Responses to COVID-19: Implementing a Way Forward. OECD Education Working Papers no. 224. Paris: OECD Publishing. [Google Scholar] [CrossRef]
  16. Hodges, Charles B., and Michael K. Barbour. 2021. Assessing learning during emergency remote education. Italian Journal of Educational Technolog 29: 85–98. [Google Scholar] [CrossRef]
  17. Hodges, Charles B., Stephanie Moore, Barbara B. Lockee, Torrey Trust, and M. Aaron Bond. 2020. The difference between emergency remote teaching and online learning. Educause Review 27. Available online: ttps://er.educause.edu/articles/2020/3/the-difference-between-emergency-remote-teaching-and-online-learning (accessed on 15 October 2021).
  18. Iglesias-Pradas, Santiago, Ángel Hernández-García, Julián Chaparro-Peláez, and José Luis Prieto. 2021. Emergency remote teaching and students’ academic performance in higher education during the COVID-19 pandemic: A case study. Computers in Human Behavior 119: 106713. [Google Scholar] [CrossRef]
  19. Jelińska, Magdalena, and Michał B. Paradowski. 2021. Teachers’ perception of student coping with Emergency Remote Instruction during the COVID-19 pandemic: The relative impact of educator demographics and professional adaptation and adjustment. Frontiers in Psychology 12: 648443. [Google Scholar] [CrossRef]
  20. Juárez-Díaz, Catalina, and Moisés Perales. 2021. Language teachers’ emergency remote teaching experiences during the COVID-19 confinement. Profile Issues in TeachersProfessional Development 23: 121–35. [Google Scholar] [CrossRef]
  21. Kaiser, Henry F. 1970. A second generation little jiffy. Psychometrika 35: 401–15. [Google Scholar] [CrossRef]
  22. Kremmel, Benjamin, and Luke Harding. 2020. Towards a comprehensive, empirical model of language assessment literacy across stakeholder groups: Developing the language assessment literacy survey. Language Assessment Quarterly 17: 100–20. [Google Scholar] [CrossRef]
  23. Kruse, Lance, Whitney Impellizeri, Claire E. Witherel, and Toni A. Sondergeld. 2020. Evaluating the impact of an assessment course on preservice teachers’ classroom assessment literacy and self-efficacy. Mid-Western Educational Researcher 32: 107–32. [Google Scholar]
  24. Levy-Vered, Adi, and Fadia Nasser-Abu Alhija. 2015. Modelling beginning teachers’ assessment literacy: The contribution of training, self-efficacy, and conceptions of assessment. Educational Research and Evaluation 21: 378–406. [Google Scholar] [CrossRef]
  25. Lightfoot, Liz. 2020. Will GCSE and A-Level Students Get a Fair Deal When Coronavirus Has Cancelled Exams? Available online: https://www.theguardian.com/education/2020/apr/14/will-gcse-and-a-level-students-get-a-fair-deal-when-coronavirus-has-cancelled-exams (accessed on 15 October 2020).
  26. Looney, Anne, Joy Cumming, Fabienne van Der Kleij, and Karen Harris. 2017. Reconceptualising the role of teachers as assessors: Teacher assessment identity. Assessment in Education: Principles, Policy & Practice 25: 442–67. [Google Scholar]
  27. MacNeil, Wilson, and Keith Topping. 2007. Crisis management in schools: Evidence based. The Journal of Educational Enquiry 7: 64–94. Available online: https://ojs.unisa.edu.au/index.php/EDEQ/article/view/495 (accessed on 14 March 2020).
  28. Mäkipää, Toni, Kaisa Hahl, and Milla Luodonpää-Manni. 2021. Teachers’ perceptions of assessment and feedback practices in Finland’s foreign language classes during the COVID-19 pandemic. CEPS Journal 11: 219–40. [Google Scholar] [CrossRef]
  29. Marinoni, Giorgio, Hilligje Van’t Land, and Trine Jensen. 2020. The Impact of COVID-19 on Higher Education around the World. IAU Global Survey Report. Paris: International Association of Universities. Available online: Iau_covid19_and_he_survey_report_final_may_2020.pdf (accessed on 20 September 2020).
  30. Miller, Michelle D. 2020. Going Online in a Hurry: What to do and Where to Start. The Chronicle of Higher Education. Available online: https://www.chronicle.com/article/going-online-in-a-hurry-what-to-do-and-where-to-start/ (accessed on 18 September 2020).
  31. Mueller, Julie, Eileen Wood, Teena Willoughby, Craig Ross, and Jacqueline Specht. 2008. Identifying discriminating variables between teachers who fully integrate computers and teachers with limited integration. Computers and Education 51: 1523–37. [Google Scholar] [CrossRef]
  32. Muhammad, Ananda Astrini, and Gary J. Ockey. 2021. Upholding Language Assessment Quality during the COVID-19 pandemic: Some final thoughts and questions. Language Assessment Quarterly 18: 51–55. [Google Scholar] [CrossRef]
  33. Myyry, Liisa, Terhi Karaharju-Suvanto, Anna-Maija K. Virtala, Marja R. Raekallio, Outi Salminen, Marjo Vesalainen, and Anne Nevgi. 2021. How self-efficacy beliefs are related to assessment practices: A study of experienced university teachers. Assessment & Evaluation in Higher Education 47: 155–68. [Google Scholar] [CrossRef]
  34. Nădrag, Lavinia. 2020. Traditional and online assessment tools for learning English. Ovidius University Annals, Series Economic Sciences 20: 412–17. [Google Scholar]
  35. Ockey, Gary J. 2021. An Overview of COVID-19′s Impact on English Language University Admissions and Placement Tests. Language Assessment Quarterly 18: 1–5. [Google Scholar] [CrossRef]
  36. Pacheco, José Augusto. 2020. The “new normal” in education. Prospects 51: 3–14. [Google Scholar] [CrossRef]
  37. Paesani, Kate. 2020. Teacher professional development and online instruction: Cultivating coherence and sustainability. Foreign Language Annals 53: 292–97. [Google Scholar] [CrossRef]
  38. Pajares, Frank. 1996. Self-Efficacy Beliefs in Academic Settings. Review of Educational Research 66: 543–78. [Google Scholar] [CrossRef]
  39. Paramour, Alexandra. 2020. Assessing Learning Online. Oxford: Oxford University Press, pp. 1–5. Available online: https://roycross.files.wordpress.com/2021/04/oup-focus-learning-online.pdf (accessed on 18 November 2020).
  40. Pennock-Roman, Maria, and Charlene Rivera. 2011. Mean effects of test accommodations for ELLs and non-ELLs: A meta-analysis of experimental studies. Educational Measurement: Issues and Practice 30: 10–28. [Google Scholar] [CrossRef]
  41. Poulou, Maria S., Linda A. Reddy, and Christopher M. Dudek. 2019. Relation of teacher self-efficacy and classroom practices: A preliminary investigation. School Psychology International 40: 25–48. [Google Scholar] [CrossRef]
  42. Radcliffe, John Scott, Debra K. Aaron, Jodi Sterle, Marina A. G. von Keyserlingk, Nancy Irlbeck, Martin Maquivar, Meghan Wulster-Radcliffe, and Cassandra Jones. 2020. Moving online: Roadmap and long-term forecast. Animal Frontiers 10: 36–45. [Google Scholar] [CrossRef]
  43. Rahim, Ahmad Fuad Abdul. 2020. Guidelines for online assessment in Emergency Remote Teaching during the COVID-19 pandemic. Education in Medicine Journal 12: 3. [Google Scholar] [CrossRef]
  44. Rapanta, Chrysi, Luca Botturi, Peter Goodyear, Lourdes Guàrdia, and Marguerite Koole. 2020. Online university teaching during and after the Covid-19 crisis: Refocusing teacher presence and learning activity. Postdigital Science and Education 2: 923–45. [Google Scholar] [CrossRef]
  45. Reimers, Fernando M., and Andreas Schleicher. 2020. A Framework to Guide an Education Response to the COVID-19 Pandemic of 2020. Paris: OECD. Available online: https://read.oecd-ilibrary.org/view/?ref=126_126988-t63lxosohs&title=A-framework-to-guide-an-education-response-to-the-Covid-19-Pandemic-of-2020 (accessed on 20 November 2021).
  46. Ruel, Erin, William Edward Wagner III, and Brian Joseph Gillespie. 2016. The Practice of Survey Research: Theory and Applications. New York: Sage Publications. [Google Scholar]
  47. Salerni, Anna, and Alessandro Vaccarelli. 2019. Supporting school resilience: A study on a sample of teachers after the 2016/2017 seismic events in Central Italy. Journal of Educational, Cultural and Psychological Studies (ECPS Journal) 19: 83–103. [Google Scholar] [CrossRef] [Green Version]
  48. Scarino, Angela. 2013. Language assessment literacy as self-awareness: Understanding the role of interpretation in assessment and in teacher learning. Language Testing 30: 309–27. [Google Scholar] [CrossRef]
  49. Taylor, Lynda. 2012. Accommodation in Language Testing. Cambridge Guide to Second Language Assessment. Edited by Christine Coombes, Peter Davison, Barry O’Sullivan and Stephen Stoynoff. Cambridge: Cambridge University Press, pp. 307–16. [Google Scholar]
  50. Tian, Wenwen, Stephen Louw, and Muhammad Kamal Khan. 2021. COVID-19 as a critical incident: Reflection on language assessment literacy and the need for radical changes. System 103: 102682. [Google Scholar] [CrossRef]
  51. Tsagari, Dina, and Ines Sperling. 2017. Assessing SLLs with SpLDs: Challenges and opportunities for equity in education. In At the Crossroads: Challenges of Foreign Language Learning. Edited by Ewa Piechurska-Kuciel, Elżbieta Szymańska-Czaplak and Magdalena Szyszka. Cham: Springer, pp. 175–88. [Google Scholar] [CrossRef]
  52. Tuah, Nik Ani Afiqah, and Lin Naing. 2021. Is online assessment in higher education institutions during COVID-19 pandemic reliable? Siriraj Medical Journal 73: 61–68. [Google Scholar] [CrossRef]
  53. UK Government. 2020. Coronavirus (COVID-19): Cancellation of GCSEs, AS and A Levels in 2020. Available online: https://www.gov.uk/government/publications/coronavirus-covid-19-cancellation-of-gcses-as-and-a-levels-in-2020/coronavirus-covid-19-cancellation-of-gcses-as-and-a-levels-in-2020 (accessed on 10 March 2021).
  54. UNESCO. 2020a. COVID-19—A Glance of National Coping Strategies on High-Stakes Examinations and Assessments. Available online: https://en.unesco.org/news/report-glance-national-coping-strategies-high-stakes-examinations-and-assessments (accessed on 10 March 2021).
  55. UNESCO. 2020b. Could Coronavirus Shape the Way Assessments Work Forever? Available online: https://gemreportunesco.wordpress.com/2020/03/20/could-coronavirus-shape-the-way-assessments-work-forever/ (accessed on 10 March 2021).
  56. UNESCO. 2020c. COVID-19—An Overview of National Coping Strategies on High-Stakes Examinations and Assessments. Working Document, Paris. Available online: https://en.unesco.org/sites/default/files/unesco-covid-19-ed-webinar-4-working-document-en.pdf (accessed on 10 March 2021).
  57. UNESCO. 2020d. National Education Responses to COVID-19: Summary Report of UNESCO’s Online Survey. Available online: https://unesdoc.unesco.org/ark:/48223/pf0000373322 (accessed on 10 November 2020).
  58. UNESCO. 2020e. Managing High-Stakes Assessments and Exams during Crisis. Issue Note N°4.3. UNESCO Covid-19 Education Response. Available online: https://unesdoc.unesco.org/ark:/48223/pf0000373387?fbclid=IwAR12R9mQPZoc-k3QdR2E9qHFpScDKxgbTmSR-4dfMl-tq6WP_uxxXO6ATx4 (accessed on 15 November 2020).
  59. UNESCO. 2020f. Summary Report of UNESCO’s Online Survey. Available online: https://unesdoc.unesco.org/ark:/48223/pf0000373322 (accessed on 10 November 2020).
  60. US Department of Education. 2020. Fact Sheet: Impact of COVID-19 on Assessments and Accountability under the Elementary and Secondary Education Act. Available online: https://oese.ed.gov/files/2020/03/COVID-19-OESE-FINAL-3.12.20.pdf (accessed on 10 November 2020).
  61. Vaccarelli, Alessandro. 2018. Pedagogy of emergency and teacher training after a catastrophe. Autobiographical narrative and resilience for supporting teacher professionalism. Ricerche di Pedagogia e Didattica. Journal of Theories and Research in Education 13: 29–47. [Google Scholar] [CrossRef]
  62. Watermeyer, Richard, Tom Crick, Cathryn Knight, and Janet Goodall. 2021. COVID-19 and digital disruption in UK universities: Afflictions and affordances of emergency online migration. High Education 81: 623–41. [Google Scholar] [CrossRef] [PubMed]
  63. Whittle, Clayton, Sonia Tiwari, Shulong Yan, and Jeff Williams. 2020. Emergency remote teaching environment: A conceptual framework for responsive online teaching in crises. Information and Learning Sciences 121: 311–19. [Google Scholar] [CrossRef]
  64. Woolcock, Nicola. 2020. Coronavirus: A-Level Students not Set Work by Teachers. Available online: https://www.thetimes.co.uk/article/coronavirus-a-level-students-not-set-work-by-teachers-jc006thr2 (accessed on 10 November 2021).
  65. Xu, Yueting, and Gavin T. L. Brown. 2016. Teacher assessment literacy in practice: A reconceptualization. Teaching and Teacher Education 58: 149–62. [Google Scholar] [CrossRef]
  66. Zhang, Haoyuen, Yan Yan, and Susie L. Gronseth. 2020. Adding flexibility to curriculum: A practical guide for student-directed assessment. In Teaching, Technology, and Teacher Education during the COVID-19 Pandemic: Stories from the Field. Edited by Richard E. Ferdig, Emily Baumgartner, Richard Hartshorne, Regina Kaplan-Rakowski and Chrystalla Mouza. AACE-Association for the Advancement of Computing in Education: pp. 113–18. Available online: https://www.learntechlib.org/d/216903 (accessed on 10 November 2020).
  67. Zou, Min, Delin Kong, and Icy Lee. 2021. Teacher engagement with online formative assessment in EFL writing during COVID-19 pandemic: The case of China. The Asia-Pacific Education Researcher 30: 487–98. [Google Scholar] [CrossRef]
  68. Zulaiha, Siti, Herri Mulyono, and Lies Ambarsari. 2020. An Investigation into EFL Teachers’ assessment literacy: Indonesian teachers’ perceptions and classroom practice. European Journal of Contemporary Education 9: 189–201. Available online: https://eric.ed.gov/?id=EJ1249376 (accessed on 10 November 2020).
Figure 1. Online teaching prior experience.
Figure 1. Online teaching prior experience.
Languages 08 00054 g001
Table 1. The educators’ assessment background.
Table 1. The educators’ assessment background.
FrequencyRelative Frequency
Pre-service training1130.863
In-service training1470.773
Professional conferences, workshops, seminars, etc.1980.667
Experience on the job2210.617
Μy colleagues/institution1580.574
Online resources such as websites, courses, webinars, etc.1710.535
Doing research in this area1370.441
Other200.078
Table 2. Relative frequency of the responsive assessment measures during the crisis.
Table 2. Relative frequency of the responsive assessment measures during the crisis.
Assessment MeasuresFrequencyRelative Frequency
Organizing exam(s) online1320.515
Postponing the dates of the exam(s) for a later date930.363
Replacing exams with alternative assessments (e.g., projects, essays, etc.)900.351
Reducing the subject content730.285
Offering exam(s) as before660.257
Using written take-home exams600.234
Offering part of the exam(s), e.g., only oral, or only written570.222
Using plagiarism-detection services (e.g., Turnitin)500.195
Reducing the number of subjects tested420.164
Cancelling exam(s) altogether260.101
Using an online proctoring service (=online surveillance, e.g., Proctorio, ProctorU, Examity, etc.)240.093
Using test results from previous semesters/years240.093
Using teacher assessments from previous semesters/years210.082
Conditional admission to the university, complemented by remedial courses upon school reopening50.019
Other00
Table 3. Teachers’ perceptions of assessment measures under COVID-19.
Table 3. Teachers’ perceptions of assessment measures under COVID-19.
SD/D *NeutralSA/A **
Canceling exams is unacceptable34%20.7%45.3%
Exams are important to learners28.5%17%54.5%
Exams are important for society35.5%22.3%42.2%
Only exams should be maintained for assessment72%12%16%
Exams are important to me as an educator/expert19.2%13.3%67.5%
My country is managing the assessment situation in an effective way28%29.3%41.8%
* SD/D: Strongly Disagree/Disagree, ** SA/A: Strongly Agree/Agree.
Table 4. Teachers’ perceptions of assessment purposes during the crisis.
Table 4. Teachers’ perceptions of assessment purposes during the crisis.
SD/DNeutralSA/A
To create opportunities to provide feedback2.8%7.495.8%
Because assessment is part of the teaching/learning cycle4.8%15.6%89.8%
Because assessment is mandated in learning curricula5.8%15.6%79.6%
Because of decision making, e.g., progress to the next level, exit requirement, etc.8.2%16%75.8%
Table 5. The educators’ self-efficacy about their roles as assessors.
Table 5. The educators’ self-efficacy about their roles as assessors.
Designing Test QuestionsDesigning Scoring CriteriaScoring Written PapersScoring Oral PerformanceGiving Students Feedback Based on Assessment ResultsUsing Self- or Peer Assessment
Not at all confident2.0%4.3%0.4%0.4%2.0%1.6%
Not too confident8.2%9.8%1.6%2.7%4.3%5.9%
Somewhat confident30.1%39.1%6.6%9.8%27.7%10.5%
Moderately Confident59.8%46.9%37.1%31.6%66.0%41.4%
Very confident2.0%4.3%54.3%55.5%2.0%40.6%
Table 6. Coefficients *.
Table 6. Coefficients *.
ModelUnstandardized CoefficientsStandardized Coefficients
BStd. ErrorBetatSig.
1(Constant)1.6860.141 11.9960
Perceptions of the assessment measures−0.0140.007−0.118−1.8910.06
* Dependent variable: assessment practices.
Table 7. Coefficients *.
Table 7. Coefficients *.
ModelUnstandardized CoefficientsStandardized Coefficients
BStd. ErrorBetatSig.
1(Constant)1.2730.270 4.7180.000
Perceptions of assessment purposes during the crisis0.0060.0110.0360.5800.562
* Dependent variable: assessment practices.
Table 8. Coefficients *.
Table 8. Coefficients *.
ModelUnstandardized CoefficientsStandardized Coefficients
BStd. ErrorBetatSig.
1(Constant)1.1530.251 4.6040.000
Assessment self-efficacy0.0140.0130.0691.1090.268
* Dependent variable: AP.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Maaoui, A.; Tsagari, D.; Dammak, H. The Relationship between Language Educators’ Perceptions and Assessment Practices during the COVID-19 Crisis. Languages 2023, 8, 54. https://doi.org/10.3390/languages8010054

AMA Style

Maaoui A, Tsagari D, Dammak H. The Relationship between Language Educators’ Perceptions and Assessment Practices during the COVID-19 Crisis. Languages. 2023; 8(1):54. https://doi.org/10.3390/languages8010054

Chicago/Turabian Style

Maaoui, Asma, Dina Tsagari, and Hanen Dammak. 2023. "The Relationship between Language Educators’ Perceptions and Assessment Practices during the COVID-19 Crisis" Languages 8, no. 1: 54. https://doi.org/10.3390/languages8010054

Article Metrics

Back to TopTop