Skip to main content
Top
Published in: BMC Medical Informatics and Decision Making 1/2024

Open Access 01-12-2024 | Research

Hybrid architecture based intelligent diagnosis assistant for GP

Authors: Ruibin Wang, Kavisha Jayathunge, Rupert Page, Hailing Li, Jian Jun Zhang, Xiaosong Yang

Published in: BMC Medical Informatics and Decision Making | Issue 1/2024

Login to get access

Abstract

As the first point of contact for patients, General Practitioners (GPs) play a crucial role in the National Health Service (NHS). An accurate primary diagnosis from the GP can alleviate the burden on specialists and reduce the time needed to re-confirm the patient’s condition, allowing for more efficient further examinations. However, GPs have broad but less specialized knowledge, which limits the accuracy of their diagnosis. Therefore, it is imperative to introduce an intelligent system to assist GPs in making decisions. This paper introduces two data augmentation methods, the Complaint Symptoms Integration Method and Symptom Dot Separating Method, to integrate essential information into the Integration dataset. Additionally, it proposes a hybrid architecture that fuses the features of words from different representation spaces. Experiments demonstrate that, compared to commonly used pre-trained attention-based models, our hybrid architecture delivers the best classification performance for four common neurological diseases on the enhanced Integration dataset. For example, the classification accuracy of the BERT+CNN hybrid architecture is 0.897, which is a 5.1% improvement over both BERT and CNN with 0.846. Finally, this paper develops an AI diagnosis assistant web application that leverages the superior performance of this architecture to help GPs complete primary diagnosis efficiently and accurately.
Literature
1.
go back to reference O’Donnell CA. Variation in GP referral rates: what can we learn from the literature? Fam Pract. 2000;17(6):462–71.CrossRefPubMed O’Donnell CA. Variation in GP referral rates: what can we learn from the literature? Fam Pract. 2000;17(6):462–71.CrossRefPubMed
3.
go back to reference Shim H, Hwang SJ, Yang E. Joint active feature acquisition and classification with variable-size set encoding. NeurIPS; 2018. p. 31. Shim H, Hwang SJ, Yang E. Joint active feature acquisition and classification with variable-size set encoding. NeurIPS; 2018. p. 31.
8.
go back to reference Mousa A, Schuller B. Contextual bidirectional long short-term memory recurrent neural network language models: a generative approach to sentiment analysis. In: EACL. Valencia; 2017. p. 1023–1032. https://aclanthology.org/E17-1096. Mousa A, Schuller B. Contextual bidirectional long short-term memory recurrent neural network language models: a generative approach to sentiment analysis. In: EACL. Valencia; 2017. p. 1023–1032. https://​aclanthology.​org/​E17-1096.
9.
go back to reference Vaswani A, Shazeer N, Parmar N, Uszkoreit J, Jones L, Gomez AN, Kaiser Ł, Polosukhin I. Attention is all you need. NeurIPS; 2017. p. 30. Vaswani A, Shazeer N, Parmar N, Uszkoreit J, Jones L, Gomez AN, Kaiser Ł, Polosukhin I. Attention is all you need. NeurIPS; 2017. p. 30.
11.
go back to reference Sanh V, Debut L, Chaumond J, Wolf T. DistilBERT, a distilled version of BERT: smaller, faster, cheaper and lighter. arXiv preprint arXiv:1910.01108 . 2019. Sanh V, Debut L, Chaumond J, Wolf T. DistilBERT, a distilled version of BERT: smaller, faster, cheaper and lighter. arXiv preprint arXiv:1910.01108 . 2019.
12.
go back to reference Yang Z, Dai Z, Yang Y, Carbonell J, Salakhutdinov RR. Le QV. Xlnet: generalized autoregressive pretraining for language understanding. NeurIPS; 2019. p. 32. Yang Z, Dai Z, Yang Y, Carbonell J, Salakhutdinov RR. Le QV. Xlnet: generalized autoregressive pretraining for language understanding. NeurIPS; 2019. p. 32.
13.
go back to reference Liu Y, Ott M, Goyal N, Du J, Joshi M, Chen D, Levy O, Lewis M, Zettlemoyer L, Stoyanov V. Roberta: a robustly optimized bert pretraining approach. arXiv preprint arXiv:1907.11692 . 2019. Liu Y, Ott M, Goyal N, Du J, Joshi M, Chen D, Levy O, Lewis M, Zettlemoyer L, Stoyanov V. Roberta: a robustly optimized bert pretraining approach. arXiv preprint arXiv:1907.11692 . 2019.
15.
go back to reference Kononenko I. Inductive and Bayesian learning in medical diagnosis. Appl Artif Intell Int J. 1993;7(4):317–37.CrossRef Kononenko I. Inductive and Bayesian learning in medical diagnosis. Appl Artif Intell Int J. 1993;7(4):317–37.CrossRef
16.
go back to reference Kononenko I. Machine learning for medical diagnosis: history, state of the art and perspective. Artif Intell Med. 2001;23(1):89–109.CrossRefPubMed Kononenko I. Machine learning for medical diagnosis: history, state of the art and perspective. Artif Intell Med. 2001;23(1):89–109.CrossRefPubMed
17.
go back to reference Xu Z, Kusner M, Weinberger K, Chen M. Cost-sensitive tree of classifiers. In: International conference on machine learning. PMLR; 2013. p. 133–141. Xu Z, Kusner M, Weinberger K, Chen M. Cost-sensitive tree of classifiers. In: International conference on machine learning. PMLR; 2013. p. 133–141.
19.
go back to reference Nan F, Wang J, Saligrama V. Feature-budgeted random forest. In: International conference on machine learning. PMLR; 2015. p. 1983–1991. Nan F, Wang J, Saligrama V. Feature-budgeted random forest. In: International conference on machine learning. PMLR; 2015. p. 1983–1991.
20.
go back to reference Hayashi Y. A neural expert system with automated extraction of fuzzy if-then rules and its application to medical diagnosis. NeurIPS; 1990. p. 3. Hayashi Y. A neural expert system with automated extraction of fuzzy if-then rules and its application to medical diagnosis. NeurIPS; 1990. p. 3.
21.
go back to reference Genkin A, Lewis DD, Madigan D. Large-scale Bayesian logistic regression for text categorization. Technometrics. 2007;49(3):291–304.CrossRef Genkin A, Lewis DD, Madigan D. Large-scale Bayesian logistic regression for text categorization. Technometrics. 2007;49(3):291–304.CrossRef
24.
go back to reference Zhang W, Yoshida T, Tang X. TFIDF, LSI and multi-word in information retrieval and text categorization. In: IEEE SMC. IEEE; 2008. p. 108–113. Zhang W, Yoshida T, Tang X. TFIDF, LSI and multi-word in information retrieval and text categorization. In: IEEE SMC. IEEE; 2008. p. 108–113.
25.
go back to reference Mikolov T, Sutskever I, Chen K, Corrado GS, Dean J. Distributed representations of words and phrases and their compositionality. NeurIPS; 2013. p. 26. Mikolov T, Sutskever I, Chen K, Corrado GS, Dean J. Distributed representations of words and phrases and their compositionality. NeurIPS; 2013. p. 26.
26.
go back to reference Zhang X, Zhao J, LeCun Y. Character-level convolutional networks for text classification. NeurIPS; 2015. p. 28. Zhang X, Zhao J, LeCun Y. Character-level convolutional networks for text classification. NeurIPS; 2015. p. 28.
27.
go back to reference Kim Y, Lee H, Jung K. Attnconvnet at semeval-2018 task 1: Attention-based convolutional neural networks for multi-label emotion classification. arXiv preprint arXiv:1804.00831 . 2018. Kim Y, Lee H, Jung K. Attnconvnet at semeval-2018 task 1: Attention-based convolutional neural networks for multi-label emotion classification. arXiv preprint arXiv:1804.00831 . 2018.
29.
go back to reference Lee J, Yoon W, Kim S, Kim D, Kim S, So CH, Kang J. BioBERT: a pre-trained biomedical language representation model for biomedical text mining. Bioinformatics. 2020;36(4):1234–40.CrossRefPubMed Lee J, Yoon W, Kim S, Kim D, Kim S, So CH, Kang J. BioBERT: a pre-trained biomedical language representation model for biomedical text mining. Bioinformatics. 2020;36(4):1234–40.CrossRefPubMed
30.
go back to reference Zhang Y, Zhang Y, Qi P, Manning CD, Langlotz CP. Biomedical and clinical English model packages for the Stanza Python NLP library. JAMIA. 2021;28(9):1892–9.PubMedPubMedCentral Zhang Y, Zhang Y, Qi P, Manning CD, Langlotz CP. Biomedical and clinical English model packages for the Stanza Python NLP library. JAMIA. 2021;28(9):1892–9.PubMedPubMedCentral
32.
go back to reference Mikolov T, Grave E, Bojanowski P, Puhrsch C, Joulin A. Advances in pre-training distributed word representations. arXiv preprint arXiv:1712.09405 . 2017. Mikolov T, Grave E, Bojanowski P, Puhrsch C, Joulin A. Advances in pre-training distributed word representations. arXiv preprint arXiv:1712.09405 . 2017.
33.
go back to reference Kingma DP, Ba J. Adam: A Method for Stochastic Optimization. In: 3rd ICLR; 2015. arXiv:1412.6980 . Kingma DP, Ba J. Adam: A Method for Stochastic Optimization. In: 3rd ICLR; 2015. arXiv:1412.6980 .
Metadata
Title
Hybrid architecture based intelligent diagnosis assistant for GP
Authors
Ruibin Wang
Kavisha Jayathunge
Rupert Page
Hailing Li
Jian Jun Zhang
Xiaosong Yang
Publication date
01-12-2024
Publisher
BioMed Central
Published in
BMC Medical Informatics and Decision Making / Issue 1/2024
Electronic ISSN: 1472-6947
DOI
https://doi.org/10.1186/s12911-023-02398-8

Other articles of this Issue 1/2024

BMC Medical Informatics and Decision Making 1/2024 Go to the issue