[go: up one dir, main page]

Skip to main content

Interactive Mongolian Question Answer Matching Model Based on Attention Mechanism in the Law Domain

  • Conference paper
  • First Online:
Chinese Computational Linguistics (CCL 2022)

Abstract

Mongolian question answer matching task is challenging, since Mongolian is a kind of low-resource language and its complex morphological structures lead to data sparsity. In this work, we propose an Interactive Mongolian Question Answer Matching Model (IMQAMM) based on attention mechanism for Mongolian question answering system. The key parts of the model are interactive information enhancement and max-mean pooling matching. Interactive information enhancement contains sequence enhancement and multi-cast attention. Sequence enhancement aims to provide a subsequent encoder with an enhanced sequence representation, and multi-cast attention is designed to generate scalar features through multiple attention mechanisms. Max-Mean pooling matching is to obtain the matching vectors for aggregation. Moreover, we introduce Mongolian morpheme representation to better learn the semantic feature. The model experimented on the Mongolian corpus, which contains question-answer pairs of various categories in the law domain. Experimental results demonstrate that our proposed Mongolian question answer matching model significantly outperforms baseline models.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Subscribe and save

Springer+ Basic
$34.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 39.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Similar content being viewed by others

References

  1. Bowman, S., Angeli, G., Potts, C., Manning, C.: A large annotated corpus for learning natural language inference. arXiv preprint arXiv:1508.05326 (2015)

  2. Chen, Q., Zhu, X., Ling, Z., Wei, S., Jiang, H., Inkpen, D.: Enhanced LSTM for natural language inference. arXiv preprint arXiv:1609.06038 (2016)

  3. Mirakyan, M., Hambardzumyan, K., Khachatrian, H.: Natural language inference over interaction space. arXiv preprint arXiv:1709.04348 (2018)

  4. Huang, P., He, X., Gao, J., Deng, L., Acero, A., Heck, L.: Learning deep structured semantic models for web search using clickthrough data. In: Proceedings of the 22nd ACM International Conference on Information & Knowledge Management, pp. 2333–2338 (2013)

    Google Scholar 

  5. Kim, S., Kang, I., Kwak, N.: Semantic sentence matching with densely-connected recurrent and co-attentive information. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 33, no. 01, pp. 6586–6593 (2019)

    Google Scholar 

  6. Kingma, D., Ba, J.: Adam: a method for stochastic optimization. arXiv preprint arXiv:1412.6980 (2014)

  7. Lu, J., Yang, J., Batra, D., Parikh, D.: Hierarchical question-image co-attention for visual question answering. arXiv preprint arXiv:1606.00061 (2016)

  8. Lu, M., Bao, F., Gao, G., Wang, W., Zhang, H.: An automatic spelling correction method for classical Mongolian. In: Douligeris, C., Karagiannis, D., Apostolou, D. (eds.) KSEM 2019. LNCS (LNAI), vol. 11776, pp. 201–214. Springer, Cham (2019). https://doi.org/10.1007/978-3-030-29563-9_19

    Chapter  Google Scholar 

  9. Mikolov, T., Sutskever, I., Chen, K., Corrado, G., Dean, J.: Distributed representations of words and phrases and their compositionality. In: Advances in Neural Information Processing Systems, pp. 3111–3119 (2013)

    Google Scholar 

  10. Mueller, J., Thyagarajan, A.: Siamese recurrent architectures for learning sentence similarity, vol. 30, no. 1 (2015)

    Google Scholar 

  11. Palangi, H., et al.: Semantic modelling with long-short-term memory for information retrieval. arXiv preprint arXiv:1412.6629 (2014)

  12. Parikh, A., Täckström, O., Das, D., Uszkoreit, J.: A decomposable attention model for natural language inference. arXiv preprint arXiv:1606.01933 (2016)

  13. Rendle, S.: Factorization machines. In: IEEE International Conference on Data Mining, pp. 995–1000 (2010)

    Google Scholar 

  14. Shen, Y., He, X., Gao, J., Deng, L., Mesnil, G.: A latent semantic model with convolutional-pooling structure for information retrieval. In: Proceedings of the 23rd ACM International Conference on Conference on Information and Knowledge Management, pp. 101–110 (2014)

    Google Scholar 

  15. Shen, T., Zhou, T., Long, G., Jiang, J., Pan, S., Zhang, C.: DiSAN: directional self-attention network for RNN/CNN-free language understanding. In: Thirty-Second AAAI Conference on Artificial Intelligence, vol. 32, no. 1 (2018)

    Google Scholar 

  16. Tan, M., Dos Santos, C., Xiang, B., Zhou, B.: Improved representation learning for question answer matching. In: Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pp. 464–473 (2016)

    Google Scholar 

  17. Tay, Y., Tuan, L., Hui, S.: Compare, compress and propagate: enhancing neural architectures with alignment factorization for natural language inference. arXiv preprint arXiv:1801.00102 (2017)

  18. Tay, Y., Tuan, L., Hui, S.: Multi-cast attention networks for retrieval-based question answering and response prediction. arXiv preprint arXiv:1806.00778 (2018)

  19. Wang, W., Bao, F., Gao, G.: Mongolian named entity recognition using suffixes segmentation. In: 2015 International Conference on Asian Language Processing (IALP), pp. 169–172 (2015)

    Google Scholar 

  20. Wang, S., Jiang, J.: A compare-aggregate model for matching text sequences. In: Proceedings of 5th International Conference on Learning Representations (2017)

    Google Scholar 

  21. Wang, Z., Hamza W., Florian, R.: Bilateral multi-perspective matching for natural language sentences. arXiv preprint arXiv:1702.03814 (2017)

  22. Wang, W., Bao, F., Gao, G.: Learning morpheme representation for Mongolian named entity recognition. Neural Process. Lett. 50(3), 2647–2664 (2019). https://doi.org/10.1007/s11063-019-10044-6

    Article  Google Scholar 

  23. Yang, K., Kao, H.: Generalize sentence representation with self-inference. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 34, pp. 9394–9401 (2020)

    Google Scholar 

  24. Yin, W., Schütze, H., Xiang, B., Zhou, B.: ABCNN: attention-based convolutional neural network for modeling sentence pairs. In: Transactions of the Association for Computational Linguistics, vol. 4, pp. 259–272 (2016)

    Google Scholar 

Download references

Acknowledgements

This work is supported by National Key R &D Program (Nos. 2018YFE0122900); National Natural Science Foundation of China (Nos. 62066033, 61773224); Inner Mongolia Applied Technology Research and Development Fund Project (Nos. 2019GG372, 2020GG0046, 2021GG0158, 2020PT0002); Inner Mongolia Achievement Transformation Project (Nos. 2019CG028); Inner Mongolia Natural Science Foundation (2020BS06001); Inner Mongolia Autonomous Region Higher Education Science and Technology Research Project (NJZY20008); Inner Mongolia Autonomous Region Overseas Students Innovation and Entrepreneurship Startup Program; Inner Mongolia Discipline Inspection and Supervision Big Data Laboratory Open Project. We are grateful for the useful suggestions from the anonymous reviewers.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Weihua Wang .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2022 The Author(s), under exclusive license to Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Peng, Y., Wang, W., Bao, F. (2022). Interactive Mongolian Question Answer Matching Model Based on Attention Mechanism in the Law Domain. In: Sun, M., et al. Chinese Computational Linguistics. CCL 2022. Lecture Notes in Computer Science(), vol 13603. Springer, Cham. https://doi.org/10.1007/978-3-031-18315-7_15

Download citation

  • DOI: https://doi.org/10.1007/978-3-031-18315-7_15

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-031-18314-0

  • Online ISBN: 978-3-031-18315-7

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics