Abstract
Question answering (QA) is an important natural language processing (NLP) task and has received much attention in academic research and industry communities. Existing QA studies assume that questions are raised by humans and answers are generated by machines. Nevertheless, in many real applications, machines are also required to determine human needs or perceive human states. In such scenarios, machines may proactively raise questions and humans supply answers. Subsequently, machines should attempt to understand the true meaning of these answers. This new QA approach is called reverse-QA (rQA) throughout this paper. In this work, the human answer understanding problem is investigated and solved by classifying the answers into predefined answer-label categories (e.g., True, False, Uncertain). To explore the relationships between questions and answers, we use the interactive attention network (IAN) model and propose an improved structure called semi-interactive attention network (Semi-IAN). Two Chinese data sets for rQA are compiled. We evaluate several conventional text classification models for comparison, and experimental results indicate the promising performance of our proposed models.
This is a preview of subscription content, log in via an institution.
Buying options
Tax calculation will be finalised at checkout
Purchases are for personal use only
Learn about institutional subscriptionsNotes
- 1.
These two data sets have been uploaded to Github and the URL is provided after anonymous review.
References
Kumar, A., et al.: Ask me anything: dynamic memory networks for natural language processing. In: ICML, pp. 1378–1387 (2016)
Hixon, B., Clark, P., Hajishirzi, H.: Learning knowledge graphs for question answering through conversational dialog. In: NAACL-HLT 2015, pp. 851–861 (2015)
Tan, M., Santos, C., Xiang, B., Zhou, B.: LSTM-based deep learning models for non-factoid answer selection. arXiv preprint arXiv:1511.04108 (2015)
Xiong, C., Zhong, V., Socher, R.: Dynamic coattention networks for question answering. In: ICLR (2017)
Richardson, M., Burges, C.J.C., Renshaw, E.: MCTest: a challenge dataset for the open-domain machine comprehension of text. In: EMNLP, pp. 1532–1543 (2014)
Wang, H., Bansal, M., Gimpel, K., Mcallester, D.: Machine comprehension with syntax, frames, and semantics. In: ACL & IJNLP, pp. 700–706 (2015)
Chen, D., Bolton, J., Manning, C.D.: A thorough examination of the CNN/Daily mail reading comprehension task. In: ACL (2016)
Hill, F., Bordes, A., Chopra, S., Weston, J.: The Goldilocks principle: reading children’s books with explicit memory representations. In: ICLR (2016)
Kadlec, R., Schmid, M., Bajgar, O., Kleindienst, J.: Text understanding with the attention sum reader network. arXiv preprint arXiv:1603.01547 (2016)
Bao, J., Duan, N., Yan, Z., Zhou, M., Zhao, T.: Constraint-based question answering with knowledge graph. In: COLING, pp. 2503–2514 (2016)
Malinowski, M., Rohrbach, M., Fritz, M.: Ask your neurons: a neural-based approach to answering questions about images. In: ICCV, pp. 1–9 (2015)
Sasaki, M., Kita, K.: Rule-based text categorization using hierarchical categories. In: IEEE International Conference on SMC, pp. 2827–2830 (1998)
Kiritchenko, S., Zhu, X., Cherry, C., Mohammad, S.M.: NRC-Canada-2014: detecting aspects and sentiment in customer reviews. In: SemEval, pp. 437–442 (2014)
Deng, Z., Zhu, X., Cheng, D., Zong, M., Zhang, S.: Efficient kNN classification algorithm for big data. Neurocomputing 195(26), 143–148 (2016)
Lipps, O., Pekari, N., Roberts, C.: Undercoverage and nonresponse in a list-sampled telephone election survey. J. Eur. Surv. Res. Assoc. 9(2), 71–82 (2015)
Zhang, X., Zhao, J., Cun, Y.L.: Character-level convolutional networks for text classification. In: NIPS, pp. 649–657 (2015)
Tang, D., Qin, B., Liu, T.: Document modeling with gated recurrent neural network for sentiment classication. In: EMNLP 2015, pp. 1422–1432 (2015)
Ma, D., Li, S., Zhang, X., Wang, H.: Interactive attention networks for aspect-level sentiment classification. In: IJCAI, pp. 4068–4074 (2017)
Zhang, L., Wang, S, Liu, B.: Deep learning for sentiment analysis: a survey. WIREs: Data Min. Knowl. Disc., 25 (2018)
Mullen, T., Collier, N.: Sentiment analysis using support vector machines with diverse information sources. In: EMNLP, pp. 412–418 (2004)
Pennington, J., Socher, R., Manning, C.D.: Glove: global vectors for word representation. In: EMNLP, pp. 1532–1543 (2014)
Wu, O., Yang, T., Yang, M., Li, M.: \(\rho \)-hot lexical embedding-based two-level LSTM for sentiment analysis. arXiv preprint arXiv: 1803.07771 (2018)
Wang, B., Lu, W.: Learning latent opinions for aspect-level sentiment classification. In: AAAI (2018)
Acknowledgments
This work is partially supported by NSFC (61673377 and 61732011), and Tianjin AI Funding (17ZXRGGX00150).
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2019 Springer Nature Switzerland AG
About this paper
Cite this paper
Yin, Q., Luo, G., Zhu, X., Hu, Q., Wu, O. (2019). Semi-interactive Attention Network for Answer Understanding in Reverse-QA. In: Yang, Q., Zhou, ZH., Gong, Z., Zhang, ML., Huang, SJ. (eds) Advances in Knowledge Discovery and Data Mining. PAKDD 2019. Lecture Notes in Computer Science(), vol 11440. Springer, Cham. https://doi.org/10.1007/978-3-030-16145-3_1
Download citation
DOI: https://doi.org/10.1007/978-3-030-16145-3_1
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-16144-6
Online ISBN: 978-3-030-16145-3
eBook Packages: Computer ScienceComputer Science (R0)