Abstract
This paper presents a web-based automatic question generation (AQG) system to generate reading comprehension questions and multiple-choice (MC) questions on grammar from a given English text. Such system saves teachers’ time on setting questions and facilitates students and their parents to prepare self-learning exercises. Our web-based system can automatically generate Wh-questions (i.e., what, who, when, where, why, and how) and MC grammar questions of selected sentences. Wh-questions can also be generated from user-specified answer phrases. The generation of Wh-questions exploits the pre-trained natural language understanding model, Text-To-Text Transfer Transformer (T5), and an adapted version of the SQuAD 2.0 machine reading comprehension dataset. The generation of MC questions involves identifying regular verbs in a text and using the verb’s lexemes as the answer choices. Our system takes an average time of about 1 s to generate a Wh-question and it generates a MC question almost instantly. User evaluation indicated that our system is easy-to-use and satisfactory in usefulness, usability, and quality, revealing the effectiveness of our system for teachers and parents.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Bahdanau, D., Cho, K., Bengio, Y.: Neural machine translation by jointly learning to align and translate. In: 3rd International Conference on Learning Representations, ICLR 2015, January 2015
Bury, J., Oka, T.: Undergraduate students’ perceptions of the importance of English in the tourism and hospitality industry. J. Teach. Travel Tour. 17(3), 173–188 (2017)
Cartwright, K.B.: Cognitive development and reading: the relation of reading-specific multiple classification skill to reading comprehension in elementary school children. J. Educ. Psychol. 94(1), 56 (2002)
De Smedt, T., Daelemans, W.: Pattern for Python. J. Mach. Learn. Res. 13, 2031–2035 (2012)
Devlin, J., Chang, M.W., Lee, K., Toutanova, K.: BERT: pre-training of deep bidirectional transformers for language understanding. In: NAACL-HLT (1), January 2019
Du, X., Shao, J., Cardie, C.: Learning to ask: neural question generation for reading comprehension. In: Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pp. 1342–1352, July 2017
Flor, M., Riordan, B.: A semantic role-based approach to open-domain automatic question generation. In: Proceedings of the Thirteenth Workshop on Innovative Use of NLP for Building Educational Applications, pp. 254–263, June 2018
Heilman, M., Smith, N.A.: Question generation via overgenerating transformations and ranking (No. CMU-LTI-09–013). Carnegie-Mellon Univ Pittsburgh pa language technologies insT (2009)
Hosking, T., Riedel, S.: Evaluating rewards for question generation models. In: Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), pp. 2278–2283, June 2019
Kurdi, G., Leo, J., Parsia, B., Sattler, U., Al-Emari, S.: A systematic review of automatic question generation for educational purposes. Int. J. Artif. Intell. Educ. 30(1), 121–204 (2020)
Qi, G.Y.: The importance of English in primary school education in China: perceptions of students. Multiling. Educ. 6(1), 1–18 (2016). https://doi.org/10.1186/s13616-016-0026-0
Liu, T., Wei, B., Chang, B., Sui, Z.: Large-scale simple question generation by template-based seq2seq learning. In: Huang, X., Jiang, J., Zhao, D., Feng, Y., Hong, Yu. (eds.) NLPCC 2017. LNCS (LNAI), vol. 10619, pp. 75–87. Springer, Cham (2018). https://doi.org/10.1007/978-3-319-73618-1_7
Mitkov, R., Le An, H., Karamanis, N.: A computer-aided environment for generating multiple-choice test items. Nat. Lang. Eng. 12(2), 177 (2006)
Nagy, W.E.: Teaching vocabulary to improve reading comprehension. National Council of Teachers of English, Urbana, IL.; ERIC Clearinghouse on Reading and Communication Skills, Urbana, IL.; International Reading Association, Newark, DE (1988)
Radford, A., Wu, J., Child, R., Luan, D., Amodei, D., Sutskever, I.: Language models are unsupervised multitask learners. OpenAI Blog 1(8), 9 (2019)
Raffel, C., et al.: Exploring the limits of transfer learning with a unified text-to-text transformer (2019). arXiv preprint arXiv:1910.10683
Rajpurkar, P., Jia, R., Liang, P.: Know what you don’t know: unanswerable questions for SQuAD. In: Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers), pp. 784–789, July 2018
Sutskever, I., Vinyals, O., Le, Q.V.: Sequence to sequence learning with neural networks. In: Advances in Neural Information Processing Systems, pp. 3104–3112 (2014)
Vaswani, A., et al.: Attention is all you need. In: Advances in Neural Information Processing Systems, pp. 5998–6008 (2017)
Yao, X., Bouma, G., Zhang, Y.: Semantics-based question generation and implementation. Dialogue Discourse 3(2), 11–42 (2012)
Yuan, X., et al.:Machine comprehension by text-to-text neural question generation. In: Proceedings of the 2nd Workshop on Representation Learning for NLP, pp. 15–25, August 2017
Zhao, Y., Ni, X., Ding, Y., Ke, Q.: Paragraph-level neural question generation with maxout pointer and gated self-attention networks. In: Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, pp. 3901–3910, October 2018
Zhou, Q., Yang, N., Wei, F., Tan, C., Bao, H., Zhou, M.: Neural question generation from text: a preliminary study. In: Huang, X., Jiang, J., Zhao, D., Feng, Y., Hong, Yu. (eds.) NLPCC 2017. LNCS (LNAI), vol. 10619, pp. 662–671. Springer, Cham (2018). https://doi.org/10.1007/978-3-319-73618-1_56
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2020 Springer Nature Singapore Pte Ltd.
About this paper
Cite this paper
Fung, YC., Kwok, J.CW., Lee, LK., Chui, K.T., U, L.H. (2020). Automatic Question Generation System for English Reading Comprehension. In: Lee, LK., U, L.H., Wang, F.L., Cheung, S.K.S., Au, O., Li, K.C. (eds) Technology in Education. Innovations for Online Teaching and Learning. ICTE 2020. Communications in Computer and Information Science, vol 1302. Springer, Singapore. https://doi.org/10.1007/978-981-33-4594-2_12
Download citation
DOI: https://doi.org/10.1007/978-981-33-4594-2_12
Published:
Publisher Name: Springer, Singapore
Print ISBN: 978-981-33-4593-5
Online ISBN: 978-981-33-4594-2
eBook Packages: Computer ScienceComputer Science (R0)