Questions? +1 (202) 335-3939 Login
Trusted News Since 1995
A service for global professionals · Thursday, April 17, 2025 · 804,304,694 Articles · 3+ Million Readers

New Study Reveals Why Students Reject ChatGPT Feedback in Academic Writing

A second-language student evaluating ChatGPT feedback on their academic writing, highlighting the factors leading to the acceptance or rejection of AI-generative content Image Credit: Jernej Furman from Wikimedia Commons

Research highlights several challenges in ChatGPT-generated feedback acceptance among second-language learners.

SHANGHAI, SHANGHAI, CHINA, April 9, 2025 /EINPresswire.com/ -- Generative AI tools like ChatGPT present extraordinary potential in promoting better language learning and teaching practice, such as timely feedback, ease of use, personalized support, simulations of the learning environment, and individualized recommendations on learning materials. However, a team of researchers in Hong Kong and Macao found that 45.9% of ChatGPT-generated feedback was rejected by second-language students, higher rejection rates for content-focused feedback (58.7%) than form-focused (41.3%).

To investigate why second-language learners reject feedback generated by ChatGPT. A team of researchers in Hong Kong and Macao, led by Associate Professor Wei Wei, from Macao Polytechnic University analyzed 45 Computer Science undergraduates tasked with revising argumentative essays using ChatGPT feedback. This study was published online on January 7, 2025, in the ECNU Review of Education.

Through mixed-method analysis of self-reflection and actual revision behaviors, the team identified four key factors driving rejection:
1. Mismatched expectations: Feedback often misinterpreted student intent or lacked clarity.
2. High workload: Overwhelming or vague suggestions discouraged engagement.
3. Conflicts with external references: Discrepancies between AI advice and teacher/peer feedback caused distrust.
4. Impeding conditions: Limited emotional support and non-personalized responses hindered usability.

Notably, content-focused feedback (e.g., argument structure, evidence quality) faced higher rejection due to its subjective nature, while form-focused feedback (e.g., grammar, vocabulary) was more readily accepted but still challenged by effort-related barriers.

“Our findings reveal that students aren’t dismissing AI feedback outright - they’re struggling with its relevance and practicality,” explains Prof. Wei. “For content-related suggestions, misalignment with their goals or external standards often leads to rejection. Meanwhile, even straightforward grammar tips can feel burdensome if not contextualized.”

The study also highlights a critical paradox: while ChatGPT’s accessibility is praised, its inability to provide emotionally supportive or proficiency-tailored feedback limits its effectiveness. “Students need more than raw suggestions; they require scaffolding to interpret and apply AI-generated insights,” Prof. Wei adds.

***

Reference

Titles of original papers: Unpacking the Rejection of L2 Students Toward ChatGPT-Generated Feedback: An Explanatory Research

Journal: ECNU Review of Education

DOI: 10.1177/20965311241305140

Melody Zhang
ECNU Review of Education
+86 21 6222 4545
roe@ecnu.edu.cn
Visit us on social media:
X

Powered by EIN Presswire

Distribution channels: Culture, Society & Lifestyle, Education, Science, Technology, World & Regional

Legal Disclaimer:

EIN Presswire provides this news content "as is" without warranty of any kind. We do not accept any responsibility or liability for the accuracy, content, images, videos, licenses, completeness, legality, or reliability of the information contained in this article. If you have any complaints or copyright issues related to this article, kindly contact the author above.

Submit your press release