Skip to main content
Top
Published in: BMC Medical Informatics and Decision Making 2/2018

Open Access 01-07-2018 | Research

Evaluating semantic relations in neural word embeddings with biomedical and general domain knowledge bases

Authors: Zhiwei Chen, Zhe He, Xiuwen Liu, Jiang Bian

Published in: BMC Medical Informatics and Decision Making | Special Issue 2/2018

Login to get access

Abstract

Background

In the past few years, neural word embeddings have been widely used in text mining. However, the vector representations of word embeddings mostly act as a black box in downstream applications using them, thereby limiting their interpretability. Even though word embeddings are able to capture semantic regularities in free text documents, it is not clear how different kinds of semantic relations are represented by word embeddings and how semantically-related terms can be retrieved from word embeddings.

Methods

To improve the transparency of word embeddings and the interpretability of the applications using them, in this study, we propose a novel approach for evaluating the semantic relations in word embeddings using external knowledge bases: Wikipedia, WordNet and Unified Medical Language System (UMLS). We trained multiple word embeddings using health-related articles in Wikipedia and then evaluated their performance in the analogy and semantic relation term retrieval tasks. We also assessed if the evaluation results depend on the domain of the textual corpora by comparing the embeddings of health-related Wikipedia articles with those of general Wikipedia articles.

Results

Regarding the retrieval of semantic relations, we were able to retrieve semanti. Meanwhile, the two popular word embedding approaches, Word2vec and GloVe, obtained comparable results on both the analogy retrieval task and the semantic relation retrieval task, while dependency-based word embeddings had much worse performance in both tasks. We also found that the word embeddings trained with health-related Wikipedia articles obtained better performance in the health-related relation retrieval tasks than those trained with general Wikipedia articles.

Conclusion

It is evident from this study that word embeddings can group terms with diverse semantic relations together. The domain of the training corpus does have impact on the semantic relations represented by word embeddings. We thus recommend using domain-specific corpus to train word embeddings for domain-specific text mining tasks.
Literature
2.
go back to reference Mikolov T, Sutskever I, Chen K, Corrado GS, Dean J. Distributed representations of words and phrases and their compositionality. In: Advances in Neural Information Processing Systems. Curran Associates, Inc.: 2013. p. 3111–9. Mikolov T, Sutskever I, Chen K, Corrado GS, Dean J. Distributed representations of words and phrases and their compositionality. In: Advances in Neural Information Processing Systems. Curran Associates, Inc.: 2013. p. 3111–9.
3.
go back to reference Pennington J, Socher R, Manning CD. Glove: Global vectors for word representation. In: EMNLP, vol 14. Association for Computational Linguistics: 2014. p. 1532–43. Pennington J, Socher R, Manning CD. Glove: Global vectors for word representation. In: EMNLP, vol 14. Association for Computational Linguistics: 2014. p. 1532–43.
4.
go back to reference Tang D, Wei F, Yang N, Zhou M, Liu T, Qin B. Learning sentiment-specific word embedding for twitter sentiment classification. In: ACL (1). Association for Computational Linguistics: 2014. p. 1555–65. Tang D, Wei F, Yang N, Zhou M, Liu T, Qin B. Learning sentiment-specific word embedding for twitter sentiment classification. In: ACL (1). Association for Computational Linguistics: 2014. p. 1555–65.
5.
go back to reference Li C, Wang H, Zhang Z, Sun A, Ma Z. Topic modeling for short texts with auxiliary word embeddings. In: Proceedings of the 39th International ACM SIGIR Conference on Research and Development in Information Retrieval. New York: ACM: 2016. p. 165–74. Li C, Wang H, Zhang Z, Sun A, Ma Z. Topic modeling for short texts with auxiliary word embeddings. In: Proceedings of the 39th International ACM SIGIR Conference on Research and Development in Information Retrieval. New York: ACM: 2016. p. 165–74.
6.
go back to reference Kim Y. Convolutional neural networks for sentence classification. Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP). Doha: 2014. p. 1746–51. Kim Y. Convolutional neural networks for sentence classification. Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP). Doha: 2014. p. 1746–51.
7.
go back to reference Tang D, Qin B, Liu T. Document modeling with gated recurrent neural network for sentiment classification. In: EMNLP. Association for Computational Linguistics: 2015. p. 1422–32. Tang D, Qin B, Liu T. Document modeling with gated recurrent neural network for sentiment classification. In: EMNLP. Association for Computational Linguistics: 2015. p. 1422–32.
10.
go back to reference Levy O, Goldberg Y. Dependency-based word embeddings. In: ACL (2). Stroudsburg: Citeseer: 2014. p. 302–8. Levy O, Goldberg Y. Dependency-based word embeddings. In: ACL (2). Stroudsburg: Citeseer: 2014. p. 302–8.
12.
go back to reference Bodenreider O. Biomedical ontologies in action: role in knowledge management, data integration and decision support In: Geissbuhler A, Kulikowski C, editors. IMIA Yearbook of Medical Informatics. IMIA, the Netherlands, Methods Inf Med. 2008;47(Suppl 1):67–79. Bodenreider O. Biomedical ontologies in action: role in knowledge management, data integration and decision support In: Geissbuhler A, Kulikowski C, editors. IMIA Yearbook of Medical Informatics. IMIA, the Netherlands, Methods Inf Med. 2008;47(Suppl 1):67–79.
13.
go back to reference Miller GA. Wordnet: a lexical database for english. Commun ACM. 1995; 38(11):39–41.CrossRef Miller GA. Wordnet: a lexical database for english. Commun ACM. 1995; 38(11):39–41.CrossRef
14.
go back to reference Lindberg DA, Humphreys BL, McCray AT, et al.The unified medical language system. the Netherlands: Yearbook, IMIA; 1993. p. 41–51. Lindberg DA, Humphreys BL, McCray AT, et al.The unified medical language system. the Netherlands: Yearbook, IMIA; 1993. p. 41–51.
15.
go back to reference Chen Z, He Z, Liu X, Bian J. An exploration of semantic relations in neural word embeddings using extrinsic knowledge. In: Bioinformatics and Biomedicine (BIBM), 2017 IEEE International Conference On. Piscataway: IEEE: 2017. p. 1246–51. Chen Z, He Z, Liu X, Bian J. An exploration of semantic relations in neural word embeddings using extrinsic knowledge. In: Bioinformatics and Biomedicine (BIBM), 2017 IEEE International Conference On. Piscataway: IEEE: 2017. p. 1246–51.
16.
go back to reference Lund K, Burgess C. Producing high-dimensional semantic spaces from lexical co-occurrence. Behav Res Methods. 1996; 28(2):203–8.CrossRef Lund K, Burgess C. Producing high-dimensional semantic spaces from lexical co-occurrence. Behav Res Methods. 1996; 28(2):203–8.CrossRef
17.
go back to reference Bengio Y, Ducharme R, Vincent P, Jauvin C. A neural probabilistic language model. J Mach Learn Res. 2003; 3(Feb):1137–55. Bengio Y, Ducharme R, Vincent P, Jauvin C. A neural probabilistic language model. J Mach Learn Res. 2003; 3(Feb):1137–55.
18.
go back to reference Mikolov T, Karafiát M, Burget L, Cernockỳ J, Khudanpur S. Recurrent neural network based language model. In: Interspeech, vol 2. International Speech Communication Association: 2010. p. 3. Mikolov T, Karafiát M, Burget L, Cernockỳ J, Khudanpur S. Recurrent neural network based language model. In: Interspeech, vol 2. International Speech Communication Association: 2010. p. 3.
19.
20.
go back to reference Levy O, Goldberg Y. Neural word embedding as implicit matrix factorization. In: Advances in Neural Information Processing Systems. Curran Associates, Inc.: 2014. p. 2177–85. Levy O, Goldberg Y. Neural word embedding as implicit matrix factorization. In: Advances in Neural Information Processing Systems. Curran Associates, Inc.: 2014. p. 2177–85.
21.
go back to reference Arora S, Li Y, Liang Y, Ma T, Risteski A. A latent variable model approach to pmi-based word embeddings. Trans Assoc Comput Linguist. 2016; 4:385–99. Arora S, Li Y, Liang Y, Ma T, Risteski A. A latent variable model approach to pmi-based word embeddings. Trans Assoc Comput Linguist. 2016; 4:385–99.
22.
go back to reference Finkelstein L, Gabrilovich E, Matias Y, Rivlin E, Solan Z, Wolfman G, Ruppin E. Placing search in context: The concept revisited. In: Proceedings of the 10th International Conference on World Wide Web. New York: ACM: 2001. p. 406–14. Finkelstein L, Gabrilovich E, Matias Y, Rivlin E, Solan Z, Wolfman G, Ruppin E. Placing search in context: The concept revisited. In: Proceedings of the 10th International Conference on World Wide Web. New York: ACM: 2001. p. 406–14.
23.
go back to reference Ono M, Miwa M, Sasaki Y. Word embedding-based antonym detection using thesauri and distributional information. In: HLT-NAACL. Association for Computational Linguistics: 2015. p. 984–9. Ono M, Miwa M, Sasaki Y. Word embedding-based antonym detection using thesauri and distributional information. In: HLT-NAACL. Association for Computational Linguistics: 2015. p. 984–9.
24.
go back to reference Schnabel T, Labutov I, Mimno DM, Joachims T. Evaluation methods for unsupervised word embeddings. In: EMNLP. Association for Computational Linguistics: 2015. p. 298–307. Schnabel T, Labutov I, Mimno DM, Joachims T. Evaluation methods for unsupervised word embeddings. In: EMNLP. Association for Computational Linguistics: 2015. p. 298–307.
25.
go back to reference Baroni M, Dinu G, Kruszewski G. Don’t count, predict! a systematic comparison of context-counting vs. context-predicting semantic vectors. In: ACL (1). Association for Computational Linguistics: 2014. p. 238–47. Baroni M, Dinu G, Kruszewski G. Don’t count, predict! a systematic comparison of context-counting vs. context-predicting semantic vectors. In: ACL (1). Association for Computational Linguistics: 2014. p. 238–47.
26.
go back to reference Levy O, Goldberg Y, Dagan I. Improving distributional similarity with lessons learned from word embeddings. Trans Assoc Comput Linguist. 2015; 3:211–25. Levy O, Goldberg Y, Dagan I. Improving distributional similarity with lessons learned from word embeddings. Trans Assoc Comput Linguist. 2015; 3:211–25.
28.
go back to reference Liu S, Bremer P-T, Thiagarajan JJ, Srikumar V, Wang B, Livnat Y, Pascucci V. Visual exploration of semantic relationships in neural word embeddings. IEEE Trans Vis Comput Graph. 2018; 24(1):553–62.CrossRefPubMed Liu S, Bremer P-T, Thiagarajan JJ, Srikumar V, Wang B, Livnat Y, Pascucci V. Visual exploration of semantic relationships in neural word embeddings. IEEE Trans Vis Comput Graph. 2018; 24(1):553–62.CrossRefPubMed
31.
go back to reference Maaten Lvd, Hinton G. Visualizing data using t-sne. J Mach Learn Res. 2008; 9(Nov):2579–605. Maaten Lvd, Hinton G. Visualizing data using t-sne. J Mach Learn Res. 2008; 9(Nov):2579–605.
34.
go back to reference Loper E, Bird S. Nltk: The natural language toolkit. In: Proceedings of the ACL-02 Workshop on Effective Tools and Methodologies for Teaching Natural Language Processing and Computational Linguistics - vol 1. ETMTNLP ’02. Stroudsburg: Association for Computational Linguistics: 2002. p. 63–70. https://doi.org/10.3115/1118108.1118117. Loper E, Bird S. Nltk: The natural language toolkit. In: Proceedings of the ACL-02 Workshop on Effective Tools and Methodologies for Teaching Natural Language Processing and Computational Linguistics - vol 1. ETMTNLP ’02. Stroudsburg: Association for Computational Linguistics: 2002. p. 63–70. https://​doi.​org/​10.​3115/​1118108.​1118117.
38.
go back to reference Duchi J, Hazan E, Singer Y. Adaptive subgradient methods for online learning and stochastic optimization. J Mach Learn Res. 2011; 12(Jul):2121–59. Duchi J, Hazan E, Singer Y. Adaptive subgradient methods for online learning and stochastic optimization. J Mach Learn Res. 2011; 12(Jul):2121–59.
41.
go back to reference He Z, Chen Z, Oh S, Hou J, Bian J. Enriching consumer health vocabulary through mining a social q&a site: A similarity-based approach. J Biomed Inform. 2017; 69:75–85.CrossRefPubMedPubMedCentral He Z, Chen Z, Oh S, Hou J, Bian J. Enriching consumer health vocabulary through mining a social q&a site: A similarity-based approach. J Biomed Inform. 2017; 69:75–85.CrossRefPubMedPubMedCentral
Metadata
Title
Evaluating semantic relations in neural word embeddings with biomedical and general domain knowledge bases
Authors
Zhiwei Chen
Zhe He
Xiuwen Liu
Jiang Bian
Publication date
01-07-2018
Publisher
BioMed Central
DOI
https://doi.org/10.1186/s12911-018-0630-x

Other articles of this Special Issue 2/2018

BMC Medical Informatics and Decision Making 2/2018 Go to the issue