Chrome Extension
WeChat Mini Program
Use on ChatGLM

Exploring Methods for Generating Feedback Comments for Writing Learning

Shizen gengo shori(2022)

Cited 14|Views23
No score
Abstract
The task of generating explanatory notes for language learners is known as feedback comment generation.Although various generation techniques are available, little is known about which methods are appropriate for this task.Nagata (2019) demonstrates the effectiveness of neural-retrieval-based methods in generating feedback comments for preposition use.Retrieval-based methods have limitations in that they can only output feedback comments existing in a given training data.Furthermore, feedback comments can be made on other grammatical and writing items than preposition use, which is still unaddressed.To shed light on these points, we investigate a wider range of methods for generating many feedback comments in this study.Our close analysis of the type of task leads us to investigate three different architectures for comment generation: (i) a neural-retrieval-based method as a baseline, (ii) a pointer-generator-based generation method as a neural seq2seq method, (iii) a retrieve-and-edit method, a hybrid of (i) and (ii).Intuitively, the pointer-generator should outperform neural-retrieval, and retrieve-andedit should perform best.However, in our experiments, this expectation is completely overturned.We closely analyze the results to reveal the major causes of these counter-intuitive results and report on our findings from the experiments. 1
More
Translated text
Key words
Language Modeling
AI Read Science
Must-Reading Tree
Example
Generate MRT to find the research sequence of this paper
Chat Paper
Summary is being generated by the instructions you defined