skip to main content
research-article
Open access

Bi-directional Long Short-Term Memory Model with Semantic Positional Attention for the Question Answering System

Published: 30 June 2021 Publication History

Abstract

The intelligent question answering system aims to provide quick and concise feedback on the questions of users. Although the performance of phrase-level and numerous attention models have been improved, the sentence components and position information are not emphasized enough. This article combines Ci-Lin and word2vec to divide all of the words in the question-answer pairs into groups according to the semantics and select one kernel word in each group. The remaining words are common words and realize the semantic mapping mechanism between kernel words and common words. With this Chinese semantic mapping mechanism, the common words in all questions and answers are replaced by the semantic kernel words to realize the normalization of the semantic representation. Meanwhile, based on the bi-directional LSTM model, this article introduces a method of the combination of semantic role labeling and positional context, dividing the sentence into multiple semantic segments according to semantic logic. The weight is given to the neighboring words in the same semantic segment and propose semantic role labeling position attention based on the bi-directional LSTM model (BLSTM-SRLP). The good performance of the BLSTM-SRLP model has been demonstrated in comparative experiments on the food safety field dataset (FS-QA).

References

[1]
A. Abdi, N. Idris, and Z. Ahmad. 2018. QAPD: An ontology-based question answering system in the physics domain. Soft Computing 22, 1 (2018), 213.
[2]
A. Figueroa and G. Neumann. 2016. Context-aware semantic classification of search queries for browsing community question-answering archives. Knowledge-Based Systems 96 (2016), 1–13.
[3]
G. Zhou, Y. Zhou, T. He, and W. Wu. 2016. Learning semantic representation with neural networks for community question answering retrieval. Knowledge-Based Systems 93 (2016), 75–83.
[4]
T. Hao, W. Xie, Q. Wu, H. Weng, and Y. Qu. 2017. Leveraging question target word features through semantic relation expansion for answer type classification. Knowledge-Based Systems 133 (2017), 43–52.
[5]
C.-H. Wu, C.-H. Liu, and P.-H. Su. 2015. Sentence extraction with topic modeling for question-answer pair generation. Soft Computing 19, 1 (2015), 39–46.
[6]
D. Bahdanau, K. Cho, and Y. Bengio. 2014. Neural machine translation by jointly learning to align and translate. arXiv:1409.0473.
[7]
K. Cho, B. V. Merrienboer, C. Gulcehre, D. Bahdanau, F. Bougares, H. Schwenk, and Y. Bengio. 2014. Learning phrase representations using RNN encoder-decoder for statistical machine translation. arXiv:1406.1078.
[8]
I. Sutskever, O. Vinyals, and Q. V. Le. 2014. Sequence to sequence learning with neural networks. In Advances in Neural Information Processing Systems, Vol. 3104. Curran Associates, Red Hook, NY, 1–9.
[9]
J. Zhao, J. X. Huang, and B. He. 2011. CRTER: Using cross terms to enhance probabilistic information retrieval. In Proceedings of the 34th International ACM SIGIR Conference on Research and Development in Information Retrieval. ACM, New York, NY, 155–164.
[10]
M. Wang, N. A. Smith, and T. Mitamura. 2007. What is the Jeopardy model? A quasi-synchronous grammar for QA. In Proceedings of the 2007 Joint Conference on Empirical Methods in Natural Language Processing and Computational Natural Language Learning (EMNLP-CoNLL’07). 1–11.
[11]
M. Heilman and N. A. Smith. 2010. Tree edit models for recognizing textual entailments, paraphrases, and answers to questions. In Proceedings of the 2010 Annual Conference of the North American Chapter of the Association for Computational Linguistics. 1011–1019.
[12]
J. Jeon, W. B. Croft, and J. H. Lee. 2005. Finding similar questions in large question and answer archives. In Proceedings of the 14th ACM International Conference on Information and Knowledge Management. ACM, New York, NY, 84–90.
[13]
G. Zhou, L. Cai, J. Zhao, and K. Liu. 2011. Phrase-based translation model for question retrieval in community question answer archives. In Proceedings of the 49th Annual Meeting of the Association for Computational Linguistics: Human Language Technologies. 653–662.
[14]
C. D. Santos, M. Tan, B. Xiang, and B. Zhou. Attentive pooling networks. arXiv:1602.03609.
[15]
M. Tan, C. Santos, B. Xiang, and B. Zhou. 2015. LSTM-based deep learning models for non-factoid answer selection. arXiv:1511.04108.
[16]
B. Wang, K. Liu, and J. Zhao. 2016. Inner attention based recurrent neural networks for answer selection. In Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers). 1288–1297.
[17]
Z. Yang, D. Yang, C. Dyer, X. He, A. Smola, and E. Hovy. 2016. Hierarchical attention networks for document classification. In Proceedings of the 2016 Conference of the Association for Computational Linguistics: Human Language Technologies. 1480–1489.
[18]
H. Li, M. Min, Y. Ge, and A. Kadav. 2017. A context-aware attention network for interactive question answering. arXiv:1612.07411.
[19]
B. Liu, X. An, and J. X. Huang. 2015. Using term location information to enhance probabilistic information retrieval. In Proceedings of the 38th International ACM SIGIR Conference on Research and Development in Information Retrieval. 883–886.
[20]
Y. Zhang, Y. Wang, G. Zhou, J. Jin, B. Wang, and X. Wang. 2018. Multi-kernel extreme learning machine for EEG classification in brain-computer interfaces. Expert Systems with Applications 96 (2018), 302–310
[21]
D. Wang and E. Nyberg. 2015. A long short-term memory model for answer sentence selection in question answering. In Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing (Volume 2: Short Papers). 707–712.
[22]
M. D. Zeiler. 2012. ADADELTA: An adaptive learning rate method. arXiv:1212.5701.

Cited By

View all
  • (2024)RNN-LSTMJournal of King Saud University - Computer and Information Sciences10.1016/j.jksuci.2024.10206836:5Online publication date: 24-Jul-2024
  • (2023)On solving textual ambiguities and semantic vagueness in MRC based question answering using generative pre-trained transformersPeerJ Computer Science10.7717/peerj-cs.14229(e1422)Online publication date: 24-Jul-2023
  • (2023)KisanQRSComputers and Electronics in Agriculture10.1016/j.compag.2023.108180213:COnline publication date: 1-Oct-2023
  • Show More Cited By

Index Terms

  1. Bi-directional Long Short-Term Memory Model with Semantic Positional Attention for the Question Answering System

    Recommendations

    Comments

    Please enable JavaScript to view thecomments powered by Disqus.

    Information & Contributors

    Information

    Published In

    cover image ACM Transactions on Asian and Low-Resource Language Information Processing
    ACM Transactions on Asian and Low-Resource Language Information Processing  Volume 20, Issue 5
    September 2021
    320 pages
    ISSN:2375-4699
    EISSN:2375-4702
    DOI:10.1145/3467024
    Issue’s Table of Contents
    Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    Published: 30 June 2021
    Accepted: 01 November 2020
    Revised: 01 November 2020
    Received: 01 August 2020
    Published in TALLIP Volume 20, Issue 5

    Permissions

    Request permissions for this article.

    Check for updates

    Author Tags

    1. Question answering
    2. BLSTM model
    3. semantic positional-based attention
    4. Chinese semantic mapping mechanism

    Qualifiers

    • Research-article
    • Refereed

    Funding Sources

    • National Key Technology R&D Program of China
    • National Natural Science Foundation of China
    • Beijing Natural Science Foundation

    Contributors

    Other Metrics

    Bibliometrics & Citations

    Bibliometrics

    Article Metrics

    • Downloads (Last 12 months)141
    • Downloads (Last 6 weeks)20
    Reflects downloads up to 24 Oct 2024

    Other Metrics

    Citations

    Cited By

    View all
    • (2024)RNN-LSTMJournal of King Saud University - Computer and Information Sciences10.1016/j.jksuci.2024.10206836:5Online publication date: 24-Jul-2024
    • (2023)On solving textual ambiguities and semantic vagueness in MRC based question answering using generative pre-trained transformersPeerJ Computer Science10.7717/peerj-cs.14229(e1422)Online publication date: 24-Jul-2023
    • (2023)KisanQRSComputers and Electronics in Agriculture10.1016/j.compag.2023.108180213:COnline publication date: 1-Oct-2023
    • (2022)Low-Power Feature-Attention Chinese Keyword Spotting Framework with Distillation LearningACM Transactions on Asian and Low-Resource Language Information Processing10.1145/355800222:2(1-14)Online publication date: 27-Dec-2022
    • (2022)Machine reading comprehension combined with semantic dependency for Chinese zero pronoun resolutionArtificial Intelligence Review10.1007/s10462-022-10364-556:8(7597-7612)Online publication date: 15-Dec-2022

    View Options

    View options

    PDF

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    HTML Format

    View this article in HTML Format.

    HTML Format

    Get Access

    Login options

    Full Access

    Media

    Figures

    Other

    Tables

    Share

    Share

    Share this Publication link

    Share on social media