Skip to main content
Top
Published in: BMC Medical Informatics and Decision Making 1/2023

Open Access 01-12-2023 | Research

Similarity matching of medical question based on Siamese network

Authors: Qing Li, Song He

Published in: BMC Medical Informatics and Decision Making | Issue 1/2023

Login to get access

Abstract

Background

With the rapid development of the medical industry and the gradual increase in people’s awareness of their health, the use of the Internet for medical question and answer, to obtain more accurate medical answers. It is necessary to first calculate the similarity of the questions asked by users, which further matches professional medical answers. Improving the efficiency of online medical question and answer sessions will not only reduce the burden on doctors, but also enhance the patient’s experience of online medical diagnosis.

Method

This paper focuses on building a bidirectional gated recurrent unit(BiGRU) deep learning model based on Siamese network for medical interrogative similarity matching, using Word2Vec word embedding tool for word vector processing of ethnic-medical corpus, and introducing an attention mechanism and convolutional neural network. Bidirectional gated recurrent unit extracts contextual semantic information and long-distance dependency features of interrogative sentences; Similar ethnic medicine interrogatives vary in length and structure, and the key information in the interrogative is crucial to similarity identification. By introducing an attention mechanism higher weight can be given to the keywords in the question, further improving the recognition of similar words in the question. Convolutional neural network takes into account the local information of interrogative sentences and can capture local position invariance, allowing feature extraction for words of different granularity through convolutional operations; By comparing the Euclidean distance, cosine distance and Manhattan distance to calculate the spatial distance of medical interrogatives, the Manhattan distance produced the best similarity result.

Result

Based on the ethnic medical question dataset constructed in this paper, the accuracy and F1-score reached 97.24% and 97.98%, which is a significant improvement compared to several other models.

Conclusion

By comparing with other models, the model proposed in this paper has better performance and achieve accurate matching of similar semantic question data of ethnic medicine.
Literature
1.
go back to reference Alqifari R. Question answering systems approaches and challenges. Proc Stud Res Workshop Assoc RANLP. 2019;2019:69–75. Alqifari R. Question answering systems approaches and challenges. Proc Stud Res Workshop Assoc RANLP. 2019;2019:69–75.
2.
3.
go back to reference Harispe Sébastien, et al. Semantic similarity from natural language and ontology analysis. Synth Lect Hum Lang Technol. 2015;8.1:1–254.CrossRef Harispe Sébastien, et al. Semantic similarity from natural language and ontology analysis. Synth Lect Hum Lang Technol. 2015;8.1:1–254.CrossRef
4.
go back to reference Lu W, Huang H, Zhu C. Feature words selection for knowledge-based word sense disambiguation with syntactic parsing. Przeglad Elektrotechniczny. 2012;88(1b):82–7. Lu W, Huang H, Zhu C. Feature words selection for knowledge-based word sense disambiguation with syntactic parsing. Przeglad Elektrotechniczny. 2012;88(1b):82–7.
5.
go back to reference Aliguliyev RM. A new sentence similarity measure and sentence based extractive technique for automatic text summarization. Expert Syst Appl. 2009;36(4):7764–72.CrossRef Aliguliyev RM. A new sentence similarity measure and sentence based extractive technique for automatic text summarization. Expert Syst Appl. 2009;36(4):7764–72.CrossRef
6.
go back to reference Thangaraj M, Sivakami M. Text classification techniques: a literature review. Interdiscip J Inf Knowl Manag. 2018;13:117. Thangaraj M, Sivakami M. Text classification techniques: a literature review. Interdiscip J Inf Knowl Manag. 2018;13:117.
7.
go back to reference Chiong R, Budhi GS, Dhakal S, et al. A textual-based featuring approach for depression detection using machine learning classifiers and social media texts. Comput Biol Med. 2021;135:104499.CrossRefPubMed Chiong R, Budhi GS, Dhakal S, et al. A textual-based featuring approach for depression detection using machine learning classifiers and social media texts. Comput Biol Med. 2021;135:104499.CrossRefPubMed
8.
go back to reference Amir S, Tanasescu A, Zighed DA. Sentence similarity based on semantic kernels for intelligent text retrieval. J Intell Inf Syst. 2017;48(3):675–89.CrossRef Amir S, Tanasescu A, Zighed DA. Sentence similarity based on semantic kernels for intelligent text retrieval. J Intell Inf Syst. 2017;48(3):675–89.CrossRef
9.
go back to reference Sarrouti M, El Alaoui SO. SemBioNLQA: A semantic biomedical question answering system for retrieving exact and ideal answers to natural language questions. Artif Intell Med. 2020;102:101767.CrossRefPubMed Sarrouti M, El Alaoui SO. SemBioNLQA: A semantic biomedical question answering system for retrieving exact and ideal answers to natural language questions. Artif Intell Med. 2020;102:101767.CrossRefPubMed
10.
go back to reference Yih SW, Chang MW, Meek C, et al. Question answering using enhanced lexical semantic models. In: Proceedings of the 51st Annual Meeting of the Association for Computational Linguistics. 2013. Yih SW, Chang MW, Meek C, et al. Question answering using enhanced lexical semantic models. In: Proceedings of the 51st Annual Meeting of the Association for Computational Linguistics. 2013.
11.
go back to reference Bär D, Biemann C, Gurevych I, et al. Ukp: Computing semantic textual similarity by combining multiple content similarity measures* SEM 2012: The First Joint Conference on Lexical and Computational Semantics–Volume 1: Proceedings of the main conference and the shared task, and Volume 2: Proceedings of the Sixth International Workshop on Semantic Evaluation (SemEval 2012). 2012. p. 435–40. Bär D, Biemann C, Gurevych I, et al. Ukp: Computing semantic textual similarity by combining multiple content similarity measures* SEM 2012: The First Joint Conference on Lexical and Computational Semantics–Volume 1: Proceedings of the main conference and the shared task, and Volume 2: Proceedings of the Sixth International Workshop on Semantic Evaluation (SemEval 2012). 2012. p. 435–40.
12.
go back to reference Jimenez S, Becerra C, Gelbukh A. Soft cardinality: A parameterized similarity function for text comparison* SEM 2012: The First Joint Conference on Lexical and Computational Semantics–Volume 1: Proceedings of the main conference and the shared task, and Volume 2: Proceedings of the Sixth International Workshop on Semantic Evaluation (SemEval 2012). 2012. p. 449–53. Jimenez S, Becerra C, Gelbukh A. Soft cardinality: A parameterized similarity function for text comparison* SEM 2012: The First Joint Conference on Lexical and Computational Semantics–Volume 1: Proceedings of the main conference and the shared task, and Volume 2: Proceedings of the Sixth International Workshop on Semantic Evaluation (SemEval 2012). 2012. p. 449–53.
13.
go back to reference Qaiser S, Ali R. Text mining: use of TF-IDF to examine the relevance of words to documents. Int J Comput Appl. 2018;181(1):25–9. Qaiser S, Ali R. Text mining: use of TF-IDF to examine the relevance of words to documents. Int J Comput Appl. 2018;181(1):25–9.
14.
go back to reference Kondrak G. N-gram similarity and distance[C]//International symposium on string processing and information retrieval. Berlin: Springer; 2005. p. 115–26.CrossRef Kondrak G. N-gram similarity and distance[C]//International symposium on string processing and information retrieval. Berlin: Springer; 2005. p. 115–26.CrossRef
15.
go back to reference Sadowski C, Levin G. Simhash: Hash-based similarity detection. 2007. Sadowski C, Levin G. Simhash: Hash-based similarity detection. 2007.
16.
go back to reference Niwattanakul S, Singthongchai J, Naenudorn E, et al. Using of Jaccard coefficient for keywords similarity. Proc Int Multiconf Eng Comput Sci. 2013;1(6):380–4. Niwattanakul S, Singthongchai J, Naenudorn E, et al. Using of Jaccard coefficient for keywords similarity. Proc Int Multiconf Eng Comput Sci. 2013;1(6):380–4.
17.
go back to reference He H, Gimpel K, Lin J. Multi-perspective sentence similarity modeling with convolutional neural networks. In: Proceedings of the 2015 conference on empirical methods in natural language processing. 2015. p. 1576–86.CrossRef He H, Gimpel K, Lin J. Multi-perspective sentence similarity modeling with convolutional neural networks. In: Proceedings of the 2015 conference on empirical methods in natural language processing. 2015. p. 1576–86.CrossRef
18.
go back to reference Shi-ying F, Wen-tin H, et al. Accelerating recurrent neural network training based on speech recognition model. J Chin Comput Syst. 2018;39(12):3–7. Shi-ying F, Wen-tin H, et al. Accelerating recurrent neural network training based on speech recognition model. J Chin Comput Syst. 2018;39(12):3–7.
19.
20.
go back to reference Huang PS, He X, Gao J, et al. Learning deep structured semantic models for web search using clickthrough data. In: Proceedings of the 22nd ACM international conference on Information & Knowledge Management. 2013. p. 2333–8.CrossRef Huang PS, He X, Gao J, et al. Learning deep structured semantic models for web search using clickthrough data. In: Proceedings of the 22nd ACM international conference on Information & Knowledge Management. 2013. p. 2333–8.CrossRef
21.
go back to reference Bromley J, Bentz J, Bottou L, Guyon I, Lecun Y, Moore C, Sackinger E, Shah R. Signature Verification using a "Siamese" Time Delay Neural Network[J]. International Journal of Pattern Recognition and Artificial Intelligence. 1993;7:25. Bromley J, Bentz J, Bottou L, Guyon I, Lecun Y, Moore C, Sackinger E, Shah R. Signature Verification using a "Siamese" Time Delay Neural Network[J]. International Journal of Pattern Recognition and Artificial Intelligence. 1993;7:25.
22.
go back to reference Shen Y, He X, Gao J, et al. A latent semantic model with convolutional-pooling structure for information retrieval. In: Proceedings of the 23rd ACM international conference on conference on information and knowledge management. 2014. p. 101–10.CrossRef Shen Y, He X, Gao J, et al. A latent semantic model with convolutional-pooling structure for information retrieval. In: Proceedings of the 23rd ACM international conference on conference on information and knowledge management. 2014. p. 101–10.CrossRef
23.
go back to reference Hu B T, Lu Z D, Li H, Chen Q C. Convolutional Neural Network Architectures for Matching Natural Language Sentences[C]. 28th Conference on Neural Information Processing Systems (NIPS). 2014:2042–50. Hu B T, Lu Z D, Li H, Chen Q C. Convolutional Neural Network Architectures for Matching Natural Language Sentences[C]. 28th Conference on Neural Information Processing Systems (NIPS). 2014:2042–50.
24.
go back to reference Palangi H, Deng L, Shen Y, et al. Semantic modelling with long-short-term memory for information retrieval. arXiv preprint arXiv:1412.6629, 2014. Palangi H, Deng L, Shen Y, et al. Semantic modelling with long-short-term memory for information retrieval. arXiv preprint arXiv:​1412.​6629, 2014.
25.
go back to reference Mueller J, Thyagarajan A, Aaai. Siamese Recurrent Architectures for Learning Sentence Similarity[C]. 30th Association-for-the-Advancement-of-Artificial-Intelligence (AAAI) Conference on Artificial Intelligence. 2016:2786–92. Mueller J, Thyagarajan A, Aaai. Siamese Recurrent Architectures for Learning Sentence Similarity[C]. 30th Association-for-the-Advancement-of-Artificial-Intelligence (AAAI) Conference on Artificial Intelligence. 2016:2786–92.
26.
go back to reference Neysiani B S, Babamir S M, IEEE. New Methodology for Contextual Features Usage in Duplicate Bug Reports Detection[C]. 5th International Conference on Web Research (ICWR). 2019:178–83. Neysiani B S, Babamir S M, IEEE. New Methodology for Contextual Features Usage in Duplicate Bug Reports Detection[C]. 5th International Conference on Web Research (ICWR). 2019:178–83.
27.
go back to reference Neculoiu P, Versteegh M, Rotaru M. Learning text similarity with siamese recurrent networks. In: Proceedings of the 1st Workshop on Representation Learning for NLP. 2016. p. 148–57.CrossRef Neculoiu P, Versteegh M, Rotaru M. Learning text similarity with siamese recurrent networks. In: Proceedings of the 1st Workshop on Representation Learning for NLP. 2016. p. 148–57.CrossRef
28.
go back to reference Chung J, Gulcehre C, Cho K H, et al. Empirical evaluation of gated recurrent neural networks on sequence modeling. arXiv preprint arXiv:1412.3555, 2014. Chung J, Gulcehre C, Cho K H, et al. Empirical evaluation of gated recurrent neural networks on sequence modeling. arXiv preprint arXiv:​1412.​3555, 2014.
29.
go back to reference Srivastava Nitish, et al. Dropout: a simple way to prevent neural networks from overfitting. J Mach Learn Res. 2014;15.1:1929–58. Srivastava Nitish, et al. Dropout: a simple way to prevent neural networks from overfitting. J Mach Learn Res. 2014;15.1:1929–58.
30.
go back to reference Semeniuta S, Barth E. Image Classification with Recurrent Attention Models[C]. IEEE Symposium Series on Computational Intelligence (IEEE SSCI). 2016:1–7. Semeniuta S, Barth E. Image Classification with Recurrent Attention Models[C]. IEEE Symposium Series on Computational Intelligence (IEEE SSCI). 2016:1–7.
31.
go back to reference Bertinetto L, Valmadre J, Henriques JF, et al. Fully-convolutional siamese networks for object tracking. In: European conference on computer vision. Cham: Springer; 2016. p. 850–65. Bertinetto L, Valmadre J, Henriques JF, et al. Fully-convolutional siamese networks for object tracking. In: European conference on computer vision. Cham: Springer; 2016. p. 850–65.
32.
go back to reference Che W, Li Z, Liu T. Ltp: A chinese language technology platform. In: Coling 2010: Demonstrations. 2010. p. 13–6. Che W, Li Z, Liu T. Ltp: A chinese language technology platform. In: Coling 2010: Demonstrations. 2010. p. 13–6.
34.
go back to reference Levy O, Goldberg Y. Neural Word Embedding as Implicit Matrix Factorization[C]. 28th Conference on Neural Information Processing Systems (NIPS). 2014. Levy O, Goldberg Y. Neural Word Embedding as Implicit Matrix Factorization[C]. 28th Conference on Neural Information Processing Systems (NIPS). 2014.
35.
go back to reference Sarzynska-Wawer J, Wawer A, Pawlak A, Szymanowska J, Stefaniak I, Jarkiewicz M, Okruszek L. Detecting formal thought disorder by deep contextualized word representations[J]. Psychiatry Research. 2021;304:114135. Sarzynska-Wawer J, Wawer A, Pawlak A, Szymanowska J, Stefaniak I, Jarkiewicz M, Okruszek L. Detecting formal thought disorder by deep contextualized word representations[J]. Psychiatry Research. 2021;304:114135.
36.
go back to reference Pennington J, Socher R, Manning CD. Glove: Global vectors for word representation. In: Proceedings of the 2014 conference on empirical methods in natural language processing (EMNLP). 2014. p. 1532–43.CrossRef Pennington J, Socher R, Manning CD. Glove: Global vectors for word representation. In: Proceedings of the 2014 conference on empirical methods in natural language processing (EMNLP). 2014. p. 1532–43.CrossRef
37.
go back to reference Devlin J, Chang M W, Lee K, et al. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805, 2018. Devlin J, Chang M W, Lee K, et al. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:​1810.​04805, 2018.
38.
go back to reference Mikolov T, Sutskever I, Chen K, Corrado G, Dean J. Distributed representations ofwords and phrases and their compositionality[C]. 27th Annual Conference on Neural Information Processing Systems, (NIPS). 2013. Mikolov T, Sutskever I, Chen K, Corrado G, Dean J. Distributed representations ofwords and phrases and their compositionality[C]. 27th Annual Conference on Neural Information Processing Systems, (NIPS). 2013.
39.
go back to reference Glorot X, Bordes A, Bengio Y. Deep sparse rectifier neural networks[C]. Proceedings of the fourteenth international conference on artificial intelligence and statistics. 2011:315–23. Glorot X, Bordes A, Bengio Y. Deep sparse rectifier neural networks[C]. Proceedings of the fourteenth international conference on artificial intelligence and statistics. 2011:315–23.
40.
go back to reference Yin W, Schütze H, Xiang B, et al. Abcnn: Attention-based convolutional neural network for modeling sentence pairs. Trans Assoc Comput Linguist. 2016;4:259–72.CrossRef Yin W, Schütze H, Xiang B, et al. Abcnn: Attention-based convolutional neural network for modeling sentence pairs. Trans Assoc Comput Linguist. 2016;4:259–72.CrossRef
42.
go back to reference Graves A, Schmidhuber J. Framewise phoneme classification with bidirectional LSTM and other neural network architectures. Neural Netw. 2005;18(5–6):602–10.CrossRefPubMed Graves A, Schmidhuber J. Framewise phoneme classification with bidirectional LSTM and other neural network architectures. Neural Netw. 2005;18(5–6):602–10.CrossRefPubMed
43.
go back to reference Liao H, Xu Z. Approaches to manage hesitant fuzzy linguistic information based on the cosine distance and similarity measures for HFLTSs and their application in qualitative decision making. Expert Syst Appl. 2015;42(12):5328–36.CrossRef Liao H, Xu Z. Approaches to manage hesitant fuzzy linguistic information based on the cosine distance and similarity measures for HFLTSs and their application in qualitative decision making. Expert Syst Appl. 2015;42(12):5328–36.CrossRef
44.
go back to reference Elmore KL, Richman MB. Euclidean distance as a similarity metric for principal component analysis. Mon Weather Rev. 2001;129(3):540–9.CrossRef Elmore KL, Richman MB. Euclidean distance as a similarity metric for principal component analysis. Mon Weather Rev. 2001;129(3):540–9.CrossRef
45.
go back to reference Wang J, Cao Z W. Chinese Text Sentiment Analysis Using LSTM Network Based on L2 and Nadam[C]. IEEE 17th International Conference on Communication Technology (ICCT). 2017:1891–95. Wang J, Cao Z W. Chinese Text Sentiment Analysis Using LSTM Network Based on L2 and Nadam[C]. IEEE 17th International Conference on Communication Technology (ICCT). 2017:1891–95.
46.
go back to reference Zhang Z. Improved adam optimizer for deep neural networks[C]. 2018 IEEE/ACM 26th international symposium on quality of service (IWQoS). 2018:1-2. Zhang Z. Improved adam optimizer for deep neural networks[C]. 2018 IEEE/ACM 26th international symposium on quality of service (IWQoS). 2018:1-2.
47.
go back to reference Babu DV, Karthikeyan C, Kumar A. Performance analysis of cost and accuracy for whale swarm and rmsprop optimizer[C]//IOP Conference Series: Materials Science and Engineering. IOP Publishing. 2020;993(1):012080. Babu DV, Karthikeyan C, Kumar A. Performance analysis of cost and accuracy for whale swarm and rmsprop optimizer[C]//IOP Conference Series: Materials Science and Engineering. IOP Publishing. 2020;993(1):012080.
Metadata
Title
Similarity matching of medical question based on Siamese network
Authors
Qing Li
Song He
Publication date
01-12-2023
Publisher
BioMed Central
Published in
BMC Medical Informatics and Decision Making / Issue 1/2023
Electronic ISSN: 1472-6947
DOI
https://doi.org/10.1186/s12911-023-02161-z

Other articles of this Issue 1/2023

BMC Medical Informatics and Decision Making 1/2023 Go to the issue