Abstract

AbstractIn this survey, we review recent progress on surface realization in natural language generation (NLG), highlighting how machine learning models have moved beyond n‐grams to successfully incorporate linguistic insights into increasingly rich models. We also advance the view that NLG still has much to gain by taking up insights from psycholinguistic studies – not only of human production but also of comprehension. We highlight how realization ranking models can be improved by modeling the role of memory in human language comprehension and discuss how surface realizers might transition to using grammars developed for incremental parsing in computational psycholinguistics, thereby making them more suitable for integration into real‐time incremental dialog systems. From a production standpoint, we suggest that the principle of uniform information density has the potential to enhance the theoretical basis for choice making in NLG and discuss two initial steps in this direction. Finally, we conclude our survey with a discussion of prospects for community‐based evaluation of surface realization systems.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call