@Article{cmc.2021.016517, AUTHOR = {Yuan Sun, Chaofan Chen, Andong Chen, Xiaobing Zhao}, TITLE = {Tibetan Question Generation Based on Sequence to Sequence Model}, JOURNAL = {Computers, Materials \& Continua}, VOLUME = {68}, YEAR = {2021}, NUMBER = {3}, PAGES = {3203--3213}, URL = {http://www.techscience.com/cmc/v68n3/42477}, ISSN = {1546-2226}, ABSTRACT = {As the dual task of question answering, question generation (QG) is a significant and challenging task that aims to generate valid and fluent questions from a given paragraph. The QG task is of great significance to question answering systems, conversational systems, and machine reading comprehension systems. Recent sequence to sequence neural models have achieved outstanding performance in English and Chinese QG tasks. However, the task of Tibetan QG is rarely mentioned. The key factor impeding its development is the lack of a public Tibetan QG dataset. Faced with this challenge, the present paper first collects 425 articles from the Tibetan Wikipedia website and constructs 7,234 question–answer pairs through crowdsourcing. Next, we propose a Tibetan QG model based on the sequence to sequence framework to generate Tibetan questions from given paragraphs. Secondly, in order to generate answer-aware questions, we introduce an attention mechanism that can capture the key semantic information related to the answer. Meanwhile, we adopt a copy mechanism to copy some words in the paragraph to avoid generating unknown or rare words in the question. Finally, experiments show that our model achieves higher performance than baseline models. We also further explore the attention and copy mechanisms, and prove their effectiveness through experiments.}, DOI = {10.32604/cmc.2021.016517} }