Teacher-Versus AI-Generated (Poe Application) Corrective Feedback and Language Learners’ Writing Anxiety, Complexity, Fluency, and Accuracy

This study examines the effects of corrective feedback (CF) on language learners’ writing anxiety, writing complexity, fluency, and accuracy, and compares the effectiveness of feedback from human teachers with an AI-driven application called Poe. The study included three intact classes, each with 25 language learners. Using a quasi-experimental design with pretest and posttest measures, one class received feedback from the teacher, one from the Poe application, and the third received no response to their writing. Data were generated though tests and a writing anxiety scale developed for the study. Data analysis, conducted using one-way ANOVA tests, revealed significant effects of teacher and AI-generated feedback on learners’ writing anxiety, accuracy, and fluency. Interestingly, the group that received AI-generated feedback performed better than the group that received teacher feedback or no AI support. Additionally, learners in the AI-generated feedback group experienced a more significant reduction in writing anxiety than their peers. These results highlight the remarkable impact of AI-generated CF on improving writing outcomes and alleviating anxiety in undergraduate language learners at East China University of Political Science and Law . The study demonstrates the benefits of integrating AI applications into language learning contexts, particularly by promoting a supportive environment for students to develop writing skills. Educators, researchers, and developers can use these findings to inform pedagogical practices and technological interventions to optimize the language learning experience in primary school settings. This research highlights the effectiveness of AI-driven applications in language teaching. It highlights the importance of considering learners’ psychological well-being, particularly anxiety levels, when developing effective language learning interventions.


Introduction
Providing written feedback to correct errors made by language learners has long been a fundamental practice in teaching writing, capturing considerable attention in writing research in Second Language (L2).
Given the multifaceted and complex nature of writing, feedback encompasses a broad array of responses, offering insights into the accuracy, successful communication, and content of learners' expressions or discourse (Li & Vuono, 2019;Thi & Nikolov, 2021).Pedagogically, feedback is a crucial link between assessment and teaching, providing appropriate information about the language learners' correct performance and guidance to achieve the target learning goals.Consequently, considerable focus has been directed toward understanding the significant contribution of CF to the language learners' writing performance.Written Corrective Feedback (WCF) plays a pivotal role in enhancing learners' writing performance, accuracy, fluency, organizational skills, and task achievement (Karim & Nassaji, 2018, 2019;Leeman, 2010;Lim & Renandya, 2020;Liu & Brown, 2015;Liu & Huang, 2020;Luo & Liu, 2017).
Numerous studies have investigated the impact of written CF in the writing ability of foreign language learners, comparing various types of feedback (Han & Hyland, 2015;Truscott, 2010;Truscott & Hsu, 2008).
The findings from these studies have been synthesized through quantitative and qualitative systematic reviews, commonly known as meta-analyses.Earlier meta-analyses, including the work of Russell and Spada (2006), underscored the significant contribution of CF to the development of grammatical knowledge in language learners.
In subsequent analyses conducted by Hyland and Hyland (2006), the emphasis was on the diversity evident in student populations, writing genres, feedback practices, and research designs.Notably, Truscott (2010) posited a negative impact of error correction on students' ability to accurately report information, drawing this conclusion from an examination of twelve published studies.In later meta-analyses conducted by researchers (e.g., Biber et al., 2011;Kang & Han 2015;Liu & Brown, 2015;Sia & Cheung, 2017), valuable perspectives were offered regarding the efficacy of written CF.These analyses considered individual differences among learners and addressed methodological limitations in the existing literature.
Recent meta-analyses by Lim and Renandya (2020) presented evidence supporting the potential of written WCF to enhance L2 writing skills of English as a Foreign Language (EFL) learners, with a specific focus on improving grammatical accuracy.However, ongoing debates persist, addressing questions about the extent of the benefits derived from WCF and the sustained efficacy of various feedback treatments, particularly in comparing implicit and explicit approaches.The existing body of literature on CF in language learning predominantly revolves around traditional teacher-generated feedback, with limited exploration into the consequences of feedback generated by AI.Notably, there is a restricted investigation into the impact of AIgenerated feedback, primarily through applications, highlighting a gap in research in this area.While studies have examined the effectiveness of feedback on aspects such as accuracy, complexity, and anxiety, there is a noticeable gap in the research regarding a direct comparison between teacher-generated and AIgenerated CF.
One specific instance of an AI-powered application is the Personalized Online Experience (Poe).This application integrates AI technologies to tailor online interactions based on individual user behavior, preferences, and historical data.The Poe application optimizes content recommendations, user interfaces, and overall interaction design by continuously learning from user engagement.Through machine-learning algorithms, Poe evolves to anticipate user needs, delivering a more personalized and efficient online experience.This enhances user satisfaction and exemplifies the potential of AI-powered applications to revolutionize how we interact with digital platforms, creating a more intelligent and user-centric digital landscape.
Understanding these two feedback sources' potential differences and implications is essential for informing pedagogical practices and optimizing language learning experiences.The rationale for this study stems from the increasing integration of AI technologies in language education and the need to evaluate their efficacy compared to traditional teaching methods.With the emergence of AI applications such as Poe, which claim to provide nuanced and personalized corrective feedback, it is crucial to assess their impact on language learners.This study investigates how learners respond to feedback from AI applications compared to feedback from human teachers, specifically regarding writing anxiety, writing complexity, and accuracy.
The rationale behind this comparative analysis lies in the potential benefits and drawbacks of AI-generated feedback, which may differ from the interpersonal and contextual aspects associated with teachergenerated feedback.By addressing this gap, the research seeks to contribute valuable insights into the evolving landscape of language education.More specifically, this study attempts to answer the following questions: 1. Do AI-generated (through Poe application) and teacher corrective feedbacks equally reduce the EFL learners' writing anxiety?

Literature Review
Empirical considerations do not solely drive the research on WCF; it is also underpinned by theoretical frameworks that highlight its potential contributions to L2 development (Polio, 2012).Skill-acquisition theories, such as DeKeyser's (2007), underscore the importance of practice and explicit instruction in developing accuracy, a concept aligned with WCF's role in aiding learners to store and retrieve declarative knowledge.The theoretical foundations of the noticing hypothesis (Schmidt, 2012) and the interaction hypothesis (Long, 1980) further support WCF by helping learners identify gaps in interlanguage by providing the needed evidence.
In L2 writing, the distinction between corrective and non-corrective feedback, focusing on form and content, is emphasized (Luo & Liu, 2017;Zhang, 2021).Corrective feedback targets language learning by providing negative evidence to enhance accuracy, while non-corrective feedback addresses broader aspects such as content, organization, and linguistic performance.The role of WCF in L2 writing goes beyond traditional written commentary feedback, with feedback strategies ranging from direct to indirect and metalinguistic (Ellis, 2009a(Ellis, , 2009b(Ellis, , 2009c)).The choice between comprehensive and focused feedback treatments has been explored, with recent studies highlighting the benefits of focused feedback.However, some still argue for a comprehensive approach (Benson & DeKeyser, 2018).
Considering the significance of writing tasks in L2 writing, factors such as task types and complexity play a vital role (Liu & Huang, 2020).Both unfocused and focused writing tasks are used for evaluating the language learners' writing proficiency.The impact of different writing task genres on language use, each with distinct communicative and functional requirements, contributes to a deeper understanding of how diverse writing tasks influence linguistic performance, encompassing aspects such as accuracy and complexity (Polio & Yoon, 2018).
Research into WCF is empirically driven and rooted in theoretical frameworks exploring its potential contributions to L2 development (Zhang, 2021).Skill-acquisition theories, like DeKeyser's (2007), posit that accuracy results from practice, explicit instruction, and extensive practice, prerequisites for transforming declarative knowledge into procedural knowledge.The concept of written CF is in harmony with these concepts, with the objective of aiding learners in storing and retrieving declarative knowledge, specifically explicit knowledge related to the target language.Stressing the importance of both practice and feedback, Evans et al. (2014) and Hartshorn and Evans (2015) highlighted automatization's crucial role within the skill acquisition theory framework.
In L2 writing, scholars highlight the role of WCF in fostering students' writing skills and abilities.The crucial distinction between corrective and non-corrective feedback, which focuses on form and content, is essential (Long, 1980;Luo & Liu, 2017).Corrective feedback aims at negative evidence to promote learning the target language, specifically addressing accuracy.Conversely, non-corrective feedback provides commentary on broader aspects, including organization, linguistic performance, and format.Exploring different types of WCF beyond traditional written commentary feedback is gaining interest.
Empirical investigations into WCF examine its facilitative role through the comparison of feedback strategies against no-feedback conditions (Ellis, 2009a(Ellis, , 2009b(Ellis, , 2009c;;Kurzer, 2018) and assess the relative effectiveness of various feedback strategies (Riazantseva, 2012).Feedback interventions delineate between comprehensive and focused approaches, determining the extent of WCF provided to students.
While earlier studies leaned towards comprehensive error correction, recent research underscores the advantages of focused feedback.Scholars (e.g., Benson & DeKeyser, 2018;Stefanou & Révész, 2015) propose that correcting the errors in a focused way yields more significant benefits than addressing all errors indiscriminately.However, certain studies advocate for comprehensive feedback that addresses a range of errors rather than concentrating on a specific sort of error (Bonilla López et al., 2018).Ellis (2009c) and Robinson (2011) advocated for focusing on meaning in tasks, classifying them as unfocused or focused, based on general language use or specific linguistic features in L2 writing, in which both unfocused tasks and focused writing tasks serve as assessments of learners' proficiency.They also believe that recognizing how task demands influence L2 writing is paramount because tasks establish contexts which provide opportunities to uptake CF.While the impact of cognitive demands imposed by tasks on learners' accuracy has received limited attention in WCF research, empirical investigations into various genres of writing tasks reveal distinct communicative and functional requirements.

Studies on AI
AI constitutes a domain with a rich historical and philosophical background (Bozkurt et al., 2023;Cao, 2023).Its evolution has raised fundamental inquiries about machine cognition and the capacity for independent creativity beyond programmed instructions (Kurzweil, 2014;Winterson, 2022, pp. 9-32).
Despite these strides, AI has become so profoundly integrated into daily life that there are expectations of an era where human and artificial intelligence converge (Kurzweil, 2014).The ubiquitous influence of AI extends to communication and advisory roles across various professions, including media, accounting, and copywriting (Bozkurt et al., 2021).Since the inception of computerized AI, educators have expressed concerns about the potential obsolescence of their roles (Goksel & Bozkurt, 2019;Selwyn et al., 2023).More recently, apprehensions have emerged regarding the possibility of students completing assignments or responding to questions with undetectable AI assistance, raising concerns about academic integrity and the authenticity of students' work (Diebold, 2023;Luan et al., 2020;Ouyang et al., 2022).
The systematic exploration of AI technologies in educational settings predominantly revolves around their application for forecasting learner outcomes and behaviors which establish adaptive learning environments, improve academic performance, and enhance overall learning achievements and experiences (Chu et al., 2022;Zawacki-Richter et al., 2019).Recent examinations of literature in K-12 education reveal a broadening range of applications for artificial intelligence in education (AIED), including collaborative learning, modeling approaches, and visualization.This signifies a shift beyond conventional pedagogical methods, as noted by Humble and Mozelius (2022) and Zawacki-Richter et al. (2019).
However, the multifaceted adoption of AI in education necessitates a comprehensive understanding of its potential implications within the broader social, cultural, pedagogical, and organizational contexts.Despite the potential advantages of integrating AI into education, numerous persistent challenges and ethical considerations warrant attention.These challenges encompass attitudes toward AI, educators' proficiency in effectively using technological tools, ethical concerns, and various technological hurdles (Sharma et al., 2019).
Ethical considerations form a cornerstone in the discourse surrounding AI in education.Scholars argue for a comprehensive examination of ethical concerns, including privacy issues and data ownership, before the widespread adoption of AI technologies (Humble & Mozelius, 2022).The potential influence of major edtech organizations over educational institutions raises additional ethical questions, particularly regarding privacy and corporate control, as these organizations may have access to student and staff data for corporate gains (Bozkurt et al., 2021).
Moreover, the existing literature on AI in education often consists of descriptive studies, indicating a need for a robust theoretical foundation to propel the field forward (Chen et al., 2020).Establishing a theoretical framework is essential for advancing our understanding of the implications of AI in educational settings and guiding future research and implementation strategies.Therefore, a concerted effort to address these challenges and ethical considerations is imperative to harness responsibly AI's full potential in education.
In a focused examination of the AI-powered chatbot ChatGPT from an educational standpoint, Tlili et al. (2023) supported the use of ChatGPT in education.They advocated for a new teaching philosophy to effectively integrate AI-powered technologies into education, emphasizing the importance of responsible, humanized chatbots and the development of digital literacy competencies (Ng et al., 2021).Concerns about academic integrity have prompted discussions on AI tools' ethical and responsible use in education (Cox, 2021), with calls for updated policies and strategies.Researchers, instructors, and policymakers are cautioned to proactively address potential disruptions caused by integrating AI technologies (Tang et al., 2021).Additionally, it is acknowledged that novel assessment formats emphasizing creativity and critical thinking, areas where AI cannot entirely replace human judgment, may be essential (Dogan et al., 2023).

Methodology Sample and Procedure
The study involved three intact classes of language learners enrolled in online courses at the School of Foreign Studies, East China University of Political Science and Law, China.Each class consisted of 25 members.As both researcher and instructor, I recruited participants from these classes, all taking a writing course as part of their language curriculum.In order to homogenize the language learners based on writing accuracy, fluency, and complexity, a writing test comprising three tasks was administered to the entire pool of participants.Following the initial evaluation, participants were kept on for the treatment; however, the final analysis only included those whose writing test results were within the range of +/-1 standard deviation (SD) from the mean.The purpose of implementing this criterion was to guarantee a study group that was relatively homogeneous.The analysis comprised 75 language learners' pretest and posttest results, representing those who satisfied the predetermined requirements.All participants were native speakers of Mandarin, with English being their second language.The study focused on this specific population to explore the impact of the proposed treatment on the writing skills of Chinese English language learners at the University of X.In the initial phase of the study, three intact classes underwent a preassessment involving the administration of the Writing Anxiety Scale and a Writing Test.The classes were randomly assigned to three groups for the subsequent intervention.The first group received corrective feedback from the teacher, focusing on addressing issues in their written work.In the second group, students were trained to use the Poe application.This involved submitting their writings to Poe and requesting revisions, edits, and paraphrasing, specifically emphasizing grammar, accuracy, coherence, and complexity.The third group received no corrective written feedback during this intervention period.
Weekly assignments that matched the course material were given to participants, with the understanding that their submissions would be evaluated for correctness, organization, and content.Punctuation, spelling, and recognizable grammatical errors that might obstruct clear communication were all evaluated.
Assignments from the start, middle, and end of the semester were chosen for comparison in order to assess any changes in the intricacy, accuracy, and fluency of participants' work over the course of the semester.For this study, three assignments were selected from each participant, for a total of 225 assignments analyzed.All groups engaged in their specific interventions over the course of 14 sessions.
The same Writing Anxiety Scale and Writing Test were used for the post-assessment in all three intact classes after this intervention.After the data was gathered, it was analyzed to look for variations in the groups' performance and anxiety levels when writing.

Data Analysis
A T-unit is characterized as one main clause along with any subordinate clauses that are attached to or embedded within it.The classification of clauses involves the distinction between dependent and independent clauses, with an independent clause being self-sufficient.In contrast, a dependent clause, which includes adverbial, nominal, and adjectival clauses, comprises a finite verb and a subject (Wolfe-Quintero,1998).Following the frameworks proposed by Storch (2009), the evaluation of complexity was conducted through the examination of the ratios of clauses per T-unit (C/T) and dependent clauses per Tunit (DC/T).
The assessment of accuracy took into account the proportion of error-free T-units (EFT/T), the proportion of error-free clauses (EFC/C), and the total number of errors per total number of words (E/W).Errors were categorized into syntactic errors (e.g., word order, incomplete sentences), morphological errors (e.g., tense, agreement, use of articles), and errors in word choice.Notably, spelling and mechanical errors such as punctuation were excluded from consideration.Fluency metrics included the total number of words (W), the count of T-units, and the length of T-units measured in words per T-unit (W/T).
An additional coder was engaged to ensure coding reliability.The inter-coder reliability achieved high scores of .92 for T-unit identification and .97 for clause identification.Regarding the identification of errorfree clauses and T-units, the reliability scores were .91 and .93,respectively.The data analysis involved employing a one-way analysis of variance (ANOVA) to compare the scores of the three groups on writing accuracy, fluency, complexity, and writing anxiety tests.This method was chosen to determine whether there were any statistically significant differences among the means of the three independent groups.Before delving into ANOVA, key assumptions, including normality and homogeneity of variances, were thoroughly examined to ensure the reliability of the subsequent results.The null hypothesis assumed no significant differences existed between the group means, while the alternative hypothesis posited that at least one group mean differed.
Post hoc tests, Bonferroni, were then employed to pinpoint specific group differences if the ANOVA results were significant.This ensured that effect size measures, such as eta-squared or omega-squared, could be computed to offer insights into the practical significance of the observed differences.The same one-way ANOVA procedure was applied to posttest scores, allowing for an examination of changes or improvements within each group over the intervention period.

Research Question 1
The first research question delved into how different corrective written feedbacks-teacher, AI generated, and no correction-affect language learners' anxiety levels in writing.The data analysis yielded noteworthy results, presented in tables 1 and 2.

Research Question 2
The second research question centered on a comparison of the writing fluency of students in the three intact classes.Results are presented in tables 3 and 4. Concerning fluency, a one-way analysis of variance (ANOVA) revealed noteworthy distinctions in the proportion of error-free T-units (F (2, 72) = 3.56, p < .05,η 2 = 0.71).Upon a more detailed investigation into the proportion of error-free clauses, significant differences emerged among distinct groups (F (2, 72) = 4.12, p < .05,η 2 = 0.61).The results presented in Table 4 from pairwise comparisons underscored that both AI-generated and teacher-generated CF exhibited significantly higher proportions of error-free clauses compared to the no-correction group.Furthermore, the total number of errors per total number of words exhibited notable differences between groups (F (2, 72) = 6.25, p < .05,η 2 = .51).Subsequent pairwise comparisons elucidated that the error rate within the AI-generated feedback group was lower than that within the teacher-generated feedback group.Additionally, the error rate in the teacher-generated group was lower than in the no-correction group.As a result, these findings suggest an improvement in syntactic accuracy across the various feedback groups.

Research Question 3
The third research question concerned the writing complexity of students in the three intact classes.Results are presented in tables 5 and 6.A more in-depth examination, facilitated by a post hoc analysis (as outlined in Table 6), provides additional evidence affirming the superior performance of AI-generated CF over teacher-generated CF.Furthermore, the teacher-generated feedback, in comparison, demonstrates higher effectiveness when contrasted with the no-correction group.These findings shed light on the nuanced impact of different feedback approaches on the syntactic structure, offering valuable insights into the intricate dynamics of language learning and correction methods.As seen in Table 6, the mean differences in scores were computed for each combination of correction methods, revealing significant differences across all comparisons (p < .001).In the CT, AI-generated correction yielded the highest mean difference of 0.50 compared to teacher-generated with no correction (0.40).In the DCT, AI-generated correction also resulted in the highest mean difference of 0.60, followed by teacher-generated correction at 0.20 and no correction at 0.40.These findings suggest that AI-generated correction is more effective than teacher-generated or no correction in improving learners' performance in writing complexity.

Research Question 4
The fourth research question concerned comparing the writing fluency of students in the three intact classes.Results are presented in tables 7 and 8.The findings of the study reveal significant differences across the three types of corrections (AI generated, teacher generated, and no correction) regarding various linguistic variables.The total number of T-units produced by participants under the AI-generated correction condition was significantly higher (M = 31) compared to the teacher correction condition (M = 28) and the no-correction condition (M = 24).This difference was statistically significant (F = 80.149, p < .001),indicating a substantial impact of the correction method on the overall syntactic structure.Similarly, the total number of words in the AIgenerated correction condition (M = 750) surpassed those in the teacher correction condition (M = 640) and the no-correction condition (M = 590), with a significant overall difference (F = 95.38,p < .001).This suggests that AI-generated corrections influenced the participants to produce more words in their writing.
The words per T-unit (W/T) ratio significantly differed among the three conditions.Participants in the AIgenerated correction condition had a higher W/T ratio (M = 19.50)compared to the teacher correction condition (M = 17.30) and the no-correction condition (M = 15.10).This difference was statistically significant (F = 51.64,p < .001),indicating that AI-generated corrections influenced the number of words and the distribution of words within T-units.Results of the post hoc test (Table 8) also verified that the differences between the AI-generated and teacher-generated CF on three aspects of writing fluency were statistically significant (p = .001),favoring the AI-generated feedback group.The writing fluency of the nocorrection group was significantly lower than the writing fluency of the teacher-generated corrective feedback (p = .001).

Discussion
Incorporating AI into educational settings has become a focal point of scholarly inquiry, with researchers delving into its potential to augment learning outcomes.This discourse consolidates insights derived from a quasi-experimental study examining the influence of AI-generated CF on writing anxiety, fluency, accuracy, and complexity among EFL learners.The study systematically compares the efficacy of AIgenerated feedback against feedback provided by teachers.The results are contextualized within the broader literature on AI in education, the digital transformation in higher education, and the overarching domain of second language acquisition.The results indicate that both AI-generated and teacher-provided feedback significantly affected the language learners' writing accuracy, fluency, and complexity.
Interestingly, AI-generated feedback proves to be more effective than teacher-generated feedback.This aligns with the broader discourse on the efficacy of AI in education, as discussed by Bozkurt et al. (2021) and Chu et al. (2022).These studies emphasized the transformative potential of AI in enhancing educational practices and suggested that AI could provide personalized and timely feedback, addressing individual learning needs.
A noteworthy outcome of the study is the reduction in learners' writing anxiety facilitated by teacher and AI-generated feedback.This finding resonates with the work of Ellis (2009a), who highlighted the importance of feedback in creating a supportive learning environment and reducing learners' anxiety.The study contributes to the growing body of research that recognizes the emotional aspects of language learning and emphasizes the role of technology, including AI, in fostering a positive learning experience (Han & Hyland, 2015).
The comparison between AI-and teacher-generated feedback draws attention to the unique advantages of AI, as evidenced by the study's results.The AI system used in the research, the Poe application, outperformed human teachers in enhancing writing skills and reducing anxiety.This aligns with the findings of Bonilla López et al. ( 2018) who investigated the differential effects of feedback forms in secondlanguage writing.The discussion here underscores the potential of AI to provide consistent and objective feedback, addressing some limitations associated with human feedback, such as variability and subjectivity.
The theoretical underpinnings of the study draw support from skill acquisition theory (DeKeyser, 2007) and the task complexity framework (Robinson, 2011).Skill acquisition theory underscores the importance of practice and feedback in language learning, aligning with the study's focus on corrective feedback's impact on writing skills.Insights from the task complexity framework shed light on how cognitive demands embedded within writing tasks influence learners' language development, providing a valuable perspective for interpreting outcomes related to complexity.
Beyond its theoretical contributions, this study enriches the ongoing discourse on AI in education.In line with trends discussed by Bozkurt et al. (2023) and Chen et al. (2020), it emphasizes the need for a nuanced understanding of AI's role in education, considering both practical applications and theoretical implications.Bozkurt et al.'s (2023) systematic review and exploration of speculative futures for ChatGPT contributed to the broader dialogue on responsibly integrating generative AI into education.Therefore, this study not only provides insights into language learning dynamics but also aligns with and extends the broader conversation on the integration of AI into educational contexts.

Implications and Conclusions
The findings of this study hold practical implications for language educators and policymakers.Integrating AI-generated feedback systems, such as the Poe application, into language classrooms could enhance the quality and efficiency of feedback provision.However, as discussed by Kurzweil (2014) and Tlili et al. (2023), ethical considerations should guide the responsible implementation of AI in education.Teachers may need to adapt their roles to incorporate AI as a supportive tool rather than a replacement.In conclusion, the quasi-experimental study on the impact of AI-generated corrective feedback on EFL learners' writing skills contributes valuable insights to the evolving landscape of AI in education.The effectiveness of AI in fostering writing accuracy, fluency, and complexity, and reducing anxiety positions it as a promising tool for language learning.However, the responsible integration of AI into educational practices requires a thoughtful and ethical approach.This discussion bridges the study's findings with existing literature, providing a comprehensive understanding of the implications for language education in the digital transformation era.
2. Do AI-generated (through Poe application) and teacher corrective feedbacks equally foster EFL learners' writing fluency? 3. Do AI-generated (through Poe application) and teacher corrective feedbacks equally foster the EFL learners' writing accuracy?4. Do AI-generated (through Poe application) and teacher corrective feedbacks equally foster the EFL learners' writing complexity?

Table 1
ANOVA for Groups' Scores on Writing Anxiety Table1shows the mean scores for writing anxiety differed significantly across the three groups.The AIgenerated group (Poe group) exhibited the lowest anxiety levels, with a mean score of 1.60, followed by the teacher group at 2.10 and the no-correction group at 2.70.The effect size (η 2 ) of 0.70 suggests a significant impact.The findings in Table2reveal a significant reduction in writing anxiety among learners who received AI-generated CF compared to those receiving teacher-generated feedback or no correction (p = .001).

Table 2
Bonferroni for Comparisons Between the Groups' Writing Anxiety

Table 3
ANOVA for Groups' Scores on Writing Fluency EFC/C= proportion of error-free clauses, FFT/T= error-free T-units, E/W= the total number of errors per total number of words.

Table 4
Bonferroni for Comparisons Between the Groups' Writing FluencyNote.EFC/C= proportion of error-free clauses, FFT/T= error-free T-units, E/W= the total number of errors per total number of words p = .001

Table 5
ANOVA for Groups' Scores on Writing Complexity

Table 6
Bonferroni for Comparisons Between the Groups' Writing Complexity

Table 7
ANOVA for Groups' Scores on Writing Fluency

Table 8
Bonferroni for Comparisons Between the Groups' Writing Fluency