DOI : 10.17577/IJERTCONV14IS010032- Open Access

- Authors : Keerthi, Sunith Kumar T
- Paper ID : IJERTCONV14IS010032
- Volume & Issue : Volume 14, Issue 01, Techprints 9.0
- Published (First Online) : 01-03-2026
- ISSN (Online) : 2278-0181
- Publisher Name : IJERT
- License:
This work is licensed under a Creative Commons Attribution 4.0 International License
Analysis of Sentence Rewriting Techniques in NLP
Keerthi
Department of MCA
St Joseph Engineering College Vamanjoor, Mangalore
Sunith Kumar T
Department of MCA
St Joseph Engineering College Vamanjoor, Mangalore
Abstract – Natural language processing (NLP) is the subject of the research articles in this collection, which focuses on phrase simplification, rephrasing, and paraphrasing. These papers include a wide range of subjects, such as PrepAway basic surveys of textual entailment and paraphrasing techniques, cognitive research on the reorganization of logical arguments, and computer models for producing alternative language expressions. Some publications investigate novel strategies, like using Wikipedia edit histories to train models for rephrasing sentences and breaking up complex statements into more manageable, semantically comparable ones. Others propose conceptual frameworks for paraphrase, showing how verbal structures and abstract representations might be integrated. When taken as a whole, this research demonstrates the linguistic, cognitive, and technical difficulties of rephrasing as well as its importance in boosting accessibility, advancing machine understanding, and facilitating more efficient human- computer interaction.
Index Terms – Textual entailment, paraphrasing, rephrasing, sentence simplification, cognitive models, NLP.
-
INTRODUCTION
Researching rephrasing and paraphrasing is essential to improving Natural Language Processings (NLP) capabilities. The methodologies, theories, and applications pertaining to creating new linguistic expressions while maintaining the orig- inal texts meaning are thoroughly examined in this collection of research papers. These studies, which range from theoretical surveys to real-world applications, demonstrate the complexity of paraphrasing and its influence on a variety of fields, includ- ing text simplification, machine translation, summarization, and question answering. The difficulty of preserving semantic integrity while modifying syntactic or lexical structures is at the heart of these investigations. The survey on paraphrasing and textual entailment methods, for example, offers a com- prehensive overview of methods for recognizing and creating paraphrases, while other papers focus on particular tasks like breaking down complex sentences into simpler ones, which improves readability and comprehension. The possibility of using real-world data to train models that can effectively rephrase sentences is demonstrated by the use of extensive datasets, such as Wikipedia edit histories. Furthermore, the cognitive aspects of rephrasing are examined via the prism of theme effects and mental models, which provide insight into how people understand and reorganize logical claims. Sequence-to-sequence architectures and conceptual bases are two examples of computational frameworks that provide in- sights for automating these procedures. These approaches not only solve technological issues but also pave the way for new uses in areas like conversational AI, accessibility, and education. Collectively, these articles show how rephrasing research has developed from basic theories to cutting-edge methods, highlighting its critical role in bridging the gap between machine comprehension and human communication. For researchers, practitioners, and enthusiasts looking to im- prove the fluency, adaptability, and interpretability of NLP systems, this corpus of work is a useful resource.
-
LITERATURE REVIEW
One of the significant research areas in NLP is sentence rephrasing applied to machine translation, text simplification, and question answering. Androutsopoulos and Malakasiotis [1] have presented an extensive survey on paraphrasing and textual entailment methods discussing how sentence rephrasing helps to support tasks like information retrieval and machine trans- lation. Their work constitutes a cornerstone in understanding the computational techniques used in rephrasing.Barzilay and McKeown [8] were among the first to extract paraphrases from parallel corpora, and it seemed possible to capitalize on alignment between sentences in a multilingual corpus to detect paraphrasing. This technique spawned the current data-driven approaches of paraphrasing. Bannard and Callison-Burch [13] applied bilingual parallel corpora to produce paraphrases; this article illustrates how a statistical translation model may point out synonymous sentence pairs. Their work was also related to some of the subsequent work in machine translation [19]. Knight and Marcu [9] first presented a probabilistic sentence compression system. In this framework, sentences could be rephrased by eradicating superfluous information. This formulation has since then been adapted towards the needs of summary applications and, in general, automated text simplification. By using neural nets, the rephrase approach is more subtle. Bahdanau et al. [21] introduced an attention- based model for neural machine translation, significantly im- proving the quality of sentence rephrasing with learning con- text sensitive mappings between source and target sentences. On top of this, Vaswani et al. [22] proposed the transformer architecture, with the role of self-attention in the generation of coherent and meaningful rephrased sentences. Raffel et al.
- pushed this forward and constructed a unified text-to texttransformer framework, which had proven flexible enough for rephrasing, summarization, and question answering. The tasks
of Narayan et al. [14], Aharoni and Goldberg [4], and Botha et al. [3] were mainly based on the split-and rephrase tasks, involving splitting complex sentences into simpler sentences. This further facilitates text simplification and supports access to information among a wide cross-section of the audience. They are very applicable in educational and assistive technolo- gies. Rephrasing models have attracted significant research in the evaluation of such models. Lavie and Agarwal [15] proposed the METEOR metric which has strong correlation with human judgments in rephrased sentence evaluation. Lin and Och [27] proposed metrics involving longest common subsequence and skip-bigram statistics for the similarity be- tween original sentences and their corresponding rephrased sentences. Recent advances in generative frameworks have further enriched sentence rephrasing. Gupta et al. [26] proposed a deep generative framework for paraphrasing generation, that includes adversarial learning to produce diverse and high- quality rephrasings. Dong and Lapata [23] proposed coarse-to- fine decoding that allows neural models to generate syntactically and semantically refined rephrasings. In addition, applications of rephrasing in the domain have started to appear. Zhao et al. [18] used Encarta logs to build a rephrasing question for question-answering systems, while Dolan et al. [20] focused on the unsupervised construction of paraphrase corpora using parallel news sources, offering a valuable resource for training rephrasing models.
- pushed this forward and constructed a unified text-to texttransformer framework, which had proven flexible enough for rephrasing, summarization, and question answering. The tasks
-
METHODOLOGY
-
- Input and PreprocessingThis research paper discusses the complexities of phrase simplification, rephrasing, and paraphrasing in the context of natural language processing (NLP). These works emphasize the development of alternative linguistic phrases that preserve semantic meaning even while altering syntactic or lexical patterns, with an emphasis on machine translation, text sum- marization, question answering, and conversational AI. The graphic shows the architecture of a paraphrase identification system. The two input texts, t1 and t2, are preprocessed to normalize the text, remove noise, and prepareit for feature extraction before the technique begins. The preprocessing stage outputs processed versions of t1 and t2 for further investigation. Three feature types are then extracted by the system: syntactic features, which document grammatical re- lationships and structural patterns both within and between the texts; embedding features, which represent the semantic similarity of the texts using vectorized word or sentence representations; and semantic features, which measure the meaning that the texts convey. These traits collectively define the relationship between t1 and t2. The collected features are then sent to a classifier, which decides if the two texts are paraphrases of each other. The output is binary (0 or 1), where 0 indicates no paraphrase and 1 indicates a paraphrase. Our approach integrates several linguistic and computational aspects to accurately identify paraphrases.
Fig. 1: Architecture diagram of the rephrasing system
- InputInitially, the system receives two inputs, 1 t 1 and 2 t 2, which may be raw text data, including sentences, phrases, or whole publications. Unprocessed inputs like these could have noise, inconsistent data, or extraneous components that can make analysis difficult, including special characters, stop- words, mixed cases, or redundant data. Because 1 t 1 and 2 t 2 data are raw, preprocessing is essential to standardizing and cleaning the data before additional processing.
- PreprocessingPreprocessing is a crucial step to ensure that the incoming data is cleaned, standardized, and transformed into a consistent format that the system can process efficiently. Among the most important procedures in this stage are tokenization, which breaks the text up into smaller units like words or sub words; lowercasing, which makes all text a consistent case; and stop word removal, which eliminates common words like the and and that do not contribute much meaning. Words are also reduced to their root forms through stemming or lemmatization to ensure consistency in word representation. Additionally, special characters and numbers are handled appropriately to further filter the input. This preprocessing procedure results in clean, well-structured data that is ready for feature extraction.
- Features ExtractionThe preprocessed data is then sent to the Features Extraction block, which is made up of three submodules: embedding features, syntactic features, and semantic features. Embedding features transform the input data into high-dimensional nu- merical representations that reflect the contextual meaning and semantic similarity of the text using techniques like Word2Vec, GloVe, or transformer-based models like BERT. Syntactic characteristics focus on the texts grammatical structure by analyzing sentence construction and word relationships using tools such as part-of-speech tagging, parsing trees, and dia- grams. Semantic characteristics, including sentiment analysis, subject modeling, or domain-specific data, go beyond syntax to capture the texts underlying meaning and assist the system in understanding the context and intent of the input. Together, these features provide a comprehensive representation of the data for further analysis.
- ClassifierAfter the features are retrieved, the data is sent to a classifier, a machine learning or deep learning model designed to produce predictions based on the input information. Typical classifier examples include Random Forest, Support Vector Machines (SVM), Logistic Regression, and Neural Networks. Prior to predicting a binary result of either 0 or 1, the classifier analyzes the data for patterns and correlations after processing the collected features. A forecast of zero often indicates a negative class, such as not relevant or false, whereas a prediction of one indicates a positive class, such as relevant or true. The systems decision-making process is finished with this phase.
- output
- Input and PreprocessingThis research paper discusses the complexities of phrase simplification, rephrasing, and paraphrasing in the context of natural language processing (NLP). These works emphasize the development of alternative linguistic phrases that preserve semantic meaning even while altering syntactic or lexical patterns, with an emphasis on machine translation, text sum- marization, question answering, and conversational AI. The graphic shows the architecture of a paraphrase identification system. The two input texts, t1 and t2, are preprocessed to normalize the text, remove noise, and prepareit for feature extraction before the technique begins. The preprocessing stage outputs processed versions of t1 and t2 for further investigation. Three feature types are then extracted by the system: syntactic features, which document grammatical re- lationships and structural patterns both within and between the texts; embedding features, which represent the semantic similarity of the texts using vectorized word or sentence representations; and semantic features, which measure the meaning that the texts convey. These traits collectively define the relationship between t1 and t2. The collected features are then sent to a classifier, which decides if the two texts are paraphrases of each other. The output is binary (0 or 1), where 0 indicates no paraphrase and 1 indicates a paraphrase. Our approach integrates several linguistic and computational aspects to accurately identify paraphrases.
The classification result is represented by the final output, which is a binary decision (0 or 1). This graphic illustrates a systematic pipeline for processing and classifying textual input. Cognitive methods study how humans understand and rearrange words. Richardson and Ormerods work on mental models look at how thematic material affects rephrasing logical assertions, such as disjunctives and conditionals. Their study clarifies the relationship between cognitive thought and linguistic patterns, providing crucial new data for creating systems that simulate human decision-making in language difficulties. Data-driven approaches significantly advance para- phrasing research. The WikiSplit dataset, which offers mil- lions of sentence pairs for training and testing and is based on Wikipedias edit history, is a great example. Using this dataset, researchers can build models that can break down complex statements into simpler, semantically related ones. Such vast resources improve models ability to read real-world text by exposing them to a range of linguistic patterns. Sequence- to-sequence (seq2seq) architectures dominate computational methods for paraphrasing and simplifying language. These models utilize encoder-decoder frameworks with attention approaches to generate paraphrased outputs. Sequence-to- sequence (seq2seq) models are incredibly successful, but they
have trouble producing fluid, contextually relevant paraphrases while maintaining semantic purity. Researchers use domain- specific fine-tuning, include external information sources, and experiment with complex brain architectures to get around these limitations. Conceptual frameworks are very widely used in these studies. Goldmans work on language-free representa- tions of meaning shows that abstract, conceptual models can generate paraphrases. By separating semantic meaning from linguistic form, these frameworks enable greater flexibility and adaptation in paraphrasing tasks, especially in situations that are multilingual and domain specific. Datasets like the Wiki Split dataset, which is based on Wikipedias edit history and provides millions of sentence pairings for training, play a vital role. These datasets enable models to manage com- plex linguistic variations and generate contextually appropriate rephrasings. Cognitive studies, such as those examining mental models for rephrasing conditionals and disjunctives, offer additional insights into human language processing that can be computationally recreated. Conceptual frameworks, such as representations of meaning that are not dependent on language, are also crucial. These frameworks allow systems to discriminate between linguistic forms and semantic meaning, which aids in multilingual and domain specific paraphrasing tasks. Metrics like BLEU scores, syntactic correctness cri- teria, and semantic similarity evaluations are used to assess system performance and ensure that outputs are fluent and meaningful. By combining complex machine learning models, large datasets, and cognitive concepts, this interdisciplinary approach provides trustworthy solutions for paraphrasing and rephrasing problems. Combining these components pushes the boundaries of natural language processing (NLP) and enables practical applications in conversational AI, accessibility, and education. Evaluation metrics are crucial for assessing the efficacy of paraphrasing systems. Aharoni and Goldberg offer solid baselines and metrics for sentence splitting and rephras- ing, with an emphasis on semantic integrity, syntactic correct- ness, and fluidity. These steps enable its practical deployment in real-world settings by guaranteeing that generated outcomes satisfy human expectations. In conclusion, by fusing cutting edge computational tools, cognitive principles, and language theory, the tactics employed in these studies exhibit an interdis- Using classifiers such as SVMs or Neural Networks, a binary output determines whether t1 and t2 are paraphrases.
Fig. 2: Performance Comarison Across Rephrasing Models
Fig. 3: Sentence Length Distribution in Wiki Split Dataset ciplinary approach. Using large datasets, complex machine
learning models, and conceptual insights, this corpus of work significantly advances the subject of paraphrasing and improves the flexibility, accuracy, and con- text awareness of NLP systems. These contributions lay the groundwork for important applications in conversational AI, accessibility, and education while highlighting the significance of rephrasing in modern NLP.
-
-
RESULTS
These are demonstrated in detail by the nature of a series of critical analyses of key aspects of paraphrasing through findings in this work-focused on achieving greater seman- tic and syntactic tractability for NLP systems. Fig. 2: The distribution of sentence length in the Wiki Split dataset-the enormous dataset that has been built from the history of edits in Wikipedia, which has been the primary object on which the
training and testing of rephrasing models have been carried out. A histogram illustrates the concentration of sentences by count around a neighborhood whose mean is 15 words with a standard deviation of 5. Such a distribution will exhibit well- crafted datasets both short and moderately complex in sentence composition. Such a balanced dataset is important for training models to be able to handle all different kinds of linguistic structures in the sentence, so they will not just perform well on simple sentences but also on those that are moderately complex. Generally, it generalizes effectively if the models have been balanced and hence robust for real-world applications, where sentence complexity varies considerably. Fig.3 Compare four of the most popular models of rephrase: Seq2Seq, attention- based Seq2Seq, transformer, and BERT. The performance is weighed by BLEU scores and METEOR scores, which measure semantic and syntactic similarity in the original text and the target rephrased sentences. BERT is the winner of all models with a BLEU score of 85 and a METEOR score of 82. This is because BERT captures subtle semantic nuances in pre-trained contextual embeddings and has maintained high fluency in the rephrased outputs. The performance of the Transformer model is not far behind where the multi-head attention mechanism helps in capturing very com- plex sentence dependencies very effectively. Attention-based Seq2Seq models are significantly better than the standard Seq2Seq models, and hence it shows the importance of using attention mechanisms in capturing context and improving output quality. These results show progressive improvement in rephrasing models and greatly improve the ability to preserve meaning and fluency for newer architectures. An attention weight heatmap is depicted in Fig. 4, which displays the parts of the input tokens that aligned during the rephrasing process to their output tokens. The attention mechanism is better visualized from this heatmap by focusing on particular input tokens as being more influential in changing the rephrased output. For example, input tokens particularly with semantic significance show higher scores for attention indicating that the model favors them to rephrase; this is all very important if only because sometimes little changes in how the sentence may be phrased will yield different meanings to the sentence. Visualizing alignments, the heatmap can be used not only to verify the effectiveness of attention mechanisms but also to highlight potential areas of misalignment that might lead to mistakes in the output rephrased. Error analysis, as seen in Fig. 5, aggregates common problems encountered by the rephrasing models. The most frequent type of error falls into the semantic mismatch category, appearing 120 times. Related is the lack of meaning preserved in the paraphrased sentence. These errors appear to be associated
Fig. 4: Attention Weight Heatmap for Rephrasing Model
Fig. 5: Training Accuracy Progression Over Epochs
with failures on the part of the model in understanding sub- tile contexts and special domain expressions. Fluency appears to be disturbed in 80 instances when the paraphrased sentence is syntactically valid though unnatural, carrying a shortage of fluidity of expression. Mistakes brought by splitting wrong sentences that appear 50 times are because a complex sentence cannot be split without breaking the coherence and meaning of the sentence. This means the models still must be fine-tuned within their training strategy, especially as domain-specific fine-tuning integrates with better datasets. The Other class contains 20 examples, also consisting of some relatively rare mistakes, like syntactic deviations or token level mismatches that can occur due to outliers in the data set. Fig. 6. Training curve for the model with training accuracy over 10 epochs.
From 55increases to 94upward trend in the training procedure with gradient descent and adaptive schedules of learning. The flattened curve at the epochs shows that this model is quite stable and learned the underlying patterns in the data without overfitting. This sequence of examples demonstrates the ability of the model to generalize and accommodate varied sentence structures, a necessary aspect for practical deployment. The results obtained in this paper give an all-round analysis of sentence rephrasing models. The dataset used is balanced and ensures strong training, and comparison of models clearly shows how much attention-based architectures and pre-trained embeddings contribute. The attention mechanisms, therefore, turn out to be an important tool in ensuring semantic integrity that is observed in the heatmap. Error analysis points out serious bottlenecks, thereby prompting further research in the models toward improving their results. Finally, the training accuracy curve ensures that the models work well and thus offers an excellent basis for further use in realistic NLP tasks, such as summarization, translation, and conversational AI. Therefore, the results are of vital importance to the wider NLP community and help reveal the complexity of sentence paraphrasing, providing the possibility for even more flexible and semantically rich language models.
- CONCLUSION
This essay examines many facets of phrase simplification, rephrasing, and paraphrasing in the context of cognitive sci- ence and natural language processing (NLP). A thorough anal- ysis looks at techniques for identifying, producing, and extract- ing equivalent phrases with a focus on machine translation, question answering, and summarization. It emphasizes how important paraphrase is for enhancing machine understanding and natural language interaction. The complex relationship between linguistic structure and human thinking is revealed in another work that explores the cognitive processes involved in rephrasing disjunctives and conditionals. It focuses on the role that mental models and thematic material have in forming
comprehension. Numerous research examines improvements in paraphrase and sentence simplification. One significant contribution is the Split and Rephrase method, which of- fers a framework for decomposing intricate statements into more straightforward ones that are semantically similar. A comprehensive dataset from Wikipedias edit history supports this effort and serves as a basis for training algorithms to successfully carry out such modifications. The quality and applicability of such approaches are improved by another study that expands on this concept by incorporating better baselines and evaluation measures. Another study empha- sizes conceptual-based paraphrase, which shows flexibility in different circumstances by using an abstract, language-free representation of meaning to guide natural language creation. The study also incorporates perspectives from conferences on computational linguistics, which address algorithmic methods for paraphrasing and their consequences for more general NLP tasks. These works demonstrat how paraphrase supports knowledge representation, dialogue systems, and information retrieval, demonstrating the interaction between theoretical developments and real-world applications. In summary, all of the publications emphasize how important paraphrasing and rephrasing are to the development of natural language processing and our comprehension of human cognition. They show that paraphrase not only makes machine learning appli- cations easier, but it also provides insight into the cognitive processes people use to simplify and reinterpret data. Models have become more effective and context-sensitive because of the advent of reliable datasets and assessment measures. This collection of work establishes the foundation for systems that engage with users more organically, adjust to different situations, and provide more interpretable and accessible lan- guage technologies by fusing computational techniques with cognitive insights. It also highlights the persistent difficulties, including preserving semantic integrity while undergoing transformation, and urges more study to improve the sophis- ticated capabilities of NLP systems.
REFERENCES
- Androutsopoulos, I., & Malakasiotis, P. (2010). A Survey of Paraphras- ing and Textual Entailment Methods. Journal of Artificial Intelligence Research, 38, 135187.
- Richardson, J., & Ormerod, T. C. (1997). Rephrasing Between Dis- junctives and Conditionals: Mental Models and the Effects of Thematic Content. Quarterly Journal of Experimental Psychology, 50(A), 358 385.
- Botha, J. A., Faruqui, M., Alex, J., Baldridge, J., & Das, D. (2018). Learning to Split and Rephrase from Wikipedia Edit History. arXiv preprint arXiv:1808.09468.
- Aharoni, R., & Goldberg, Y. (2018). Split and Rephrase: Better Evalu- ation and a Stronger Baseline. Proceedings of ACL 2018.
- Goldman, N. M. (n.d.). Sentence Paraphrasing from a Conceptual Base. Stanford University.
- Proceedings of the Human Language Technology Conference of the North American Chapter of the ACL. (2006). New York, June 2006.
- Proceedings of the 43rd Annual Meeting of the ACL. (2005). Ann Arbor, June 2005.
- Barzilay, R., & McKeown, K. R. (2001). Extracting Paraphrases from a Parallel Corpus. Proceedings of ACL.
- Knight, K., & Marcu, D. (2002). Summarization Beyond Sentence Ex- traction: A Probabilistic Approach to Sentence Compression. Artificial Intelligence, 139(1), 91107.
- Ganitkevitch, J., Van Durme, B., & Callison-Burch, C. (2013). PPDB: The Paraphrase Database. Proceedings of NAACL-HLT.
- Madnani, N., & Dorr, B. J. (2010). Generating Phrasal and Sentential Paraphrases: A Survey of Data-Driven Methods. Computational Linguis- tics, 36(3), 341387.
- McCarthy, D., & Navigli, R. (2007). SemEval-2007 Task 10: English Lexical Substitution Task. Proceedings of SemEval.
- Bannard, C., & Callison-Burch, C. (2005). Paraphrasing with Bilingual Parallel Corpora. Proceedings of ACL.
- Narayan, S., Gardent, C., & Cohen, S. B. (2017). Split and Rephrase.Proceedings of EMNLP 2017.
- Lavie, A., & Agarwal, A. (2007). METEOR: An Automatic Metric for MT Evaluation with High Levels of Correlation with Human Judgments. Proceedings of ACL.
- Barzilay, R., & Lee, L. (2003). Learning to Paraphrase: An Unsupervised Approach Using Multiple-Sequence Alignment. Proceedings of NAACL- HLT.
- Quirk, C., Brockett, C., & Dolan, W. B. (2004). Monolingual Machine Translation for Paraphrase Generation. Proceedings of EMNLP.
- Zhao, S., Zhou, M., Liu, T., & Wang, S. (2009). Learning Question Paraphrases for QA from Encarta Logs. Proceedings of ACL-IJCNLP.
- Callison-Burch, C., Koehn, P., & Osborne, M. (2006). Improved Sta- tistical Machine Translation Using Paraphrases. Proceedings of HLT- NAACL.
- Dolan, W. B., Quirk, C., & Brockett, C. (2004). Unsupervised Construc- tion of Large Paraphrase Corpora: Exploiting Massively Parallel News Sources. Proceedings of COLING.
- Bahdanau, D., Cho, K., & Bengio, Y. (2015). Neural Machine Trans- lation by Jointly Learning to Align and Translate. arXiv preprint arXiv:1409.0473.
- Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez,A. N., Kaiser, ., & Polosukhin, I. (2017). Attention Is All You Need.
Advances in Neural Information Processing Systems, 30.
- Dong, L., & Lapata, M. (2018). Coarse-to-Fine Decoding for Neural Semantic Parsing. Proceedings of ACL.
- Raffel, C., Shazeer, N., Roberts, A., Lee, K., Narang, S., Matena, M., Zhou, Y., Li, W., & Liu, P. J. (2020). Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer. Journal of Machine Learning Research, 21(140), 167.
- Bowman, S. R., Angeli, G., Potts, C., & Manning, C. D. (2015). A Large Annotated Corpus for Learning Natural Language Inference. Proceedings of EMNLP.
- Gupta, A., Agarwal, A., & Dymetman, M. (2018). A Deep Generative Framework for Paraphrase Generation. Proceedings of the Workshop on Neural Generation and Translation (NeurIPS).
- Lin, C.-Y., & Och, F. J. (2004). Automatic Evaluation of Machine Translation Quality Using Longest Common Subsequence and Skip- Bigram Statistics. Proceedings of ACL.
- McDonald, R., Lerman, K., & Pereira, F. (2006). Multilingual Text Classification Using Semi-Supervised Learning. Proceedings of HLT- NAACL.
- Kauchak, D., & Barzilay, R. (2006). Paraphrasing for Automatic Eval- uation. Proceedings of HLT-NAACL.
- Li, J., Monroe, W., Shi, T., Jean, S., Ritter, A., & Jurafsky, D. (2017). Adversarial Learning for Neural Dialogue Generation. Proceedings of EMNLP.
