Abstract

Paraphrase is an expression of a text with alternative words and orders to achieve a better clarity. Paraphrases have been found vital for augmenting training dataset, which aid to enhance performance of machine learning models that intended for various natural language processing (NLP) tasks. Thus, recently, automatic paraphrase generation has received increasing attention. However, evaluating quality of generated paraphrases is technically challenging. In the literature, the importance of generated paraphrases is tended to be determined by their impact on the performance of other NLP tasks. This kind of evaluation is referred as extrinsic evaluation, which requires high computational resources to train and test the models. So far, very little attention has been paid to the role of intrinsic evaluation in which quality of generated paraphrase is judged against predefined ground truth (reference paraphrases). In fact, it is also very challenging to find ideal and complete reference paraphrases. Therefore, in this study, we propose semantic or meaning oriented automatic evaluation metric that helps to evaluate quality of generated paraphrases against the original text, which is an intrinsic evaluation approach. Further, we evaluate quality of the paraphrases by assessing their impact on other NLP tasks, which is an extrinsic evaluation method. The goal is to explore the relationship between intrinsic and extrinsic evaluation methods. To ensure the effectiveness of proposed evaluation methods, extensive experiments are done on different publicly available datasets. The experimental results demonstrate that our proposed intrinsic and extrinsic evaluation strategies are promising. The results further reveal that there is a significant correlation between intrinsic and extrinsic evaluation approaches.

Highlights

  • Paraphrase is an expression that delivers the same information of the original text by using different words and order [1]

  • Paraphrases have been found vital for augmenting training dataset, which aid to enhance performance of machine learning models that intended for various natural language processing (NLP) tasks

  • Very little attention has been paid to the role of intrinsic evaluation in which quality of generated paraphrase is judged against predefined ground truth

Read more

Summary

Introduction

Paraphrase is an expression that delivers the same information of the original text by using different words and order [1]. It has been proved that paraphrases play a vital role for augmenting and bringing diversity to existing training dataset, which significantly improves performance of machine learning models that intended for NLP tasks. Machine learning models inherently require large training dataset. If the training sample is big enough, the models can capture all the discrepancies and learn relevant patterns. Failing to find enough dataset obscures application of machine learning techniques to natural language processing tasks, especially for low-resource and morphologically rich languages [2]. Hiring people to collect large dataset is not practical. Large volume of synthetic data can be used to train complex machine learning models. Barse et al [4] used generated artificial data to train fraud detection system

Methods
Results
Conclusion
Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.