L2R-QA: An Open-Domain Question Answering Framework
Open-domain question answering has always being a challenging task. It involves information retrieval, natural language processing, machine learning, and so on. In this work, we try to explore some comparable methods in improving the precision of open-domain question answering. In detail, we bring in the topic model in the phase of document retrieval, in the hope of exploiting more hidden semantic information of a document. Also, we incorporate the learning to rank model into the LSTM to train more available features for the ranking of candidate paragraphs. Specifically, we combine the results from both LSTM and learning to rank model, which lead to a more precise understanding of questions, as well as the paragraphs. We conduct an extensive set of experiments to evaluate the efficacy of our proposed framework, which proves to be superior.
KeywordsQuestion answering Learning to rank Topic model
The work is supported in part by the National Key Research and Development Program of China (2016YFC0800805) and the National Natural Science Foundation of China (61772014).
- 2.Berant, J., Chou, A., Frostig, R., Liang, P.: Semantic parsing on freebase from question-answer pairs. In: Proceedings of EMNLP (2013)Google Scholar
- 3.Bordes, A., Chopra, S., Weston, J.: Question answering with subgraph embeddings. Comput. Sci. (2014) Google Scholar
- 4.Chen, D., Fisch, A., Weston, J., Bordes, A.: Reading Wikipedia to answer open-domain questions, pp. 1870–1879 (2017)Google Scholar
- 5.Chu-Carroll, J., Fan, J.: Leveraging Wikipedia characteristics for search and candidate generation in question answering. In: AAAI Conference on Artificial Intelligence, pp. 872–877 (2011)Google Scholar
- 7.Dong, L., Wei, F., Zhou, M., Xu, K.: Question answering over freebase with multi-column convolutional neural networks. In: Meeting of the Association for Computational Linguistics and the International Joint Conference on Natural Language Processing, pp. 260–269 (2015)Google Scholar
- 9.Jurczyk, T., Deshmane, A., Choi, J.D.: Analysis of Wikipedia-based corpora for question answering (2018)Google Scholar
- 10.Manning, C.D., Surdeanu, M., Bauer, J., Finkel, J., Bethard, S.J., Mcclosky, D.: The Stanford CoreNLP natural language processing toolkit. In: Meeting of the Association for Computational Linguistics: System Demonstrations (2014)Google Scholar
- 11.Řuřek, R., Sojka, P.: Gensim–statistical semantics in python (2011)Google Scholar
- 12.Rajpurkar, P., Zhang, J., Lopyrev, K., Liang, P.: Squad: 100,000+ questions for machine comprehension of text, pp. 2383–2392 (2016)Google Scholar
- 14.Yan, X., Mou, L., Li, G., Chen, Y., Peng, H., Jin, Z.: Classifying relations via long short term memory networks along shortest dependency path. Comput. Sci. 42(1), 56–61 (2015)Google Scholar
- 15.Yao, X., Durme, B.V.: Information extraction over structured data: question answering with freebase. In: Meeting of the Association for Computational Linguistics, pp. 956–966 (2014)Google Scholar
- 16.Yih, W.T., Chang, M.W., He, X., Gao, J.: Semantic parsing via staged query graph generation: question answering with knowledge base. In: Meeting of the Association for Computational Linguistics and the International Joint Conference on Natural Language Processing, pp. 1321–1331 (2015)Google Scholar