Advertisement

Using Deep Learning Based Natural Language Processing Techniques for Clinical Decision-Making with EHRs

  • Runjie ZhuEmail author
  • Xinhui TuEmail author
  • Jimmy HuangEmail author
Chapter
Part of the Studies in Big Data book series (SBD, volume 68)

Abstract

Natural language processing (NLP) is an interdisciplinary domain of research that focuses on the interactions between human languages and computers. There has been a recent trend of solving the NLP problems using deep learning approach. The applications of deep learning in the healthcare sector are mostly considered to be related to canonical examples of applying image processing and computer vision techniques to medical scans for disease diagnoses. Electronic Health Record (EHR) is another source of data often being neglected, equally if not more important than medical scans, that can change the way we learn useful features and information from the medical records of patients. These text-based information stored within the EHR are data-rich by nature, but are often not well-understood due to its characteristics of high volume, variety, velocity and complexity. However, these specific characteristics fit right to the nature of deep learning. Therefore, we believe it is the right time to summarize the current status, to review and learn from the state-of-the-art medical-based NLP techniques. Different from the existing reviews, we examine and categorize the current deep learning-based NLP techniques in medical domain into three major purposes: representation learning, information extraction and clinical predictions. Meanwhile, we discuss whether the application of deep learning methods has tackled the problems differently and transformed these tasks revolutionarily. Based on the results, we find that the distance to revolutionize the existing healthcare sector using deep learning methods still remains long. However, the recent progress made by these proposed methods have already made a promising good start. Furthermore, we state some of the legal and ethical considerations, present the status quo of the healthcare industry applications, and provide several possible directions of future research.

Keywords

Deep learning Electronic health records Natural language processing Representation learning Information extraction Clinical predictions 

Notes

Acknowledgements

This work is supported by the Natural Sciences and Engineering Research Council (NSERC) of Canada, an NSERC CREATE award in ADERSIM,1 the York Research Chairs (YRC) program and an ORF-RE (Ontario Research Fund-Research Excellence) award in BRAIN Alliance.2

References

  1. 1.
    Hinton, G.E., Mcclelland, J.L., Rumelhart, D.E.: Distributed representation. https://web.stanford.edu/jlmcc/papers/PDP/Chapter3.pdf
  2. 2.
    Harris, Z.S.: Distributional structure. Word (1954)Google Scholar
  3. 3.
    Mikolov, T., Sutskever, I., Chen, K., Corrado, G.S., Dean, J.: Distributed representations of words and phrases and their compositionality. In: Advances in Neural Information Processing Systems, pp. 3111–3119 (2013)Google Scholar
  4. 4.
    Mikolov, T., Chen, K., Corrado, G., Dean, J.: Efficient estimation of word representations in vector space. In: arXiv preprint. arXiv:1301.3781 (2013)
  5. 5.
    Chalapathy, R., Borzeshi, E.Z., Piccardi, M.: Bidirectional LSTM-CRF for clinical concept extraction. arXiv. https://arxiv.org/abs/1611.08373v1. (2016)
  6. 6.
    Pennington, J., Socher, R., Manning, C.D.: GloVe: global vectors for word representation. In: European Conference on Machine Learning (ECML), pp. 1532–1543 (2014)Google Scholar
  7. 7.
    Kiela, D., Grace, E., Joulin, A., Mikolov, T.: Efficient large scale multi-modal classfication. arXiv. http://arxiv.org/pdf/1802/02892.pdf
  8. 8.
    Krizhevsky, A., Sutskever, I., Hinton, G.: ImageNet classification with deep convolutional neural networks. NIPS. http://papers.nips.cc/paper/4824-imagenet-classification-with-deep-convolutional-neural-networks.pdf
  9. 9.
    Mikolov, T.: Statistical language models based on neural networks. Ph.D. thesis, Brno University of Technology (2012)Google Scholar
  10. 10.
    Cho, K., van Merrienboer, B., Gulcehre, C., Bahdanau, D., Bougares, F, Schwenk, H., Bengio, Y.: Learning phrase representations using RNN encoder-decoder for statistical machine translation. In: Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP), pp. 1724–1734, Doha, Qatar. Association for Computational Linguistics, Oct 2014b.  https://doi.org/10.3115/v1/d14-1179 (2014)
  11. 11.
    Hochreiter, S., Schmidhuber, J.: Long short-term memory. Neural Comput. 9(8), 1735–1780.  https://doi.org/10.1162/neco.1997.9.8.1735 (1997)CrossRefGoogle Scholar
  12. 12.
    Devlin, J., Chang, M.W., Lee, K., Toutanova, K.: BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. arXiv. https://arxiv.org/pdf/1810.04805.pdf
  13. 13.
    Better Language Models and Their Implications. https://openai.com/blog/better-language-models/
  14. 14.
    Hinton, G.E.: Learning distributed representations of concepts. In: Proceedings of the Eighth Conference Cognitive Science Society, pp. 1–12. (1986)Google Scholar
  15. 15.
    Bengio, Y., Ducharme, R., Vincent, P., Jauvin, C.: A neural probabilistic language model. J. Mach. Learn. Res. 3, 137–1155 (2003)zbMATHGoogle Scholar
  16. 16.
    Bengio, Y.: Neural net language models. Scholarpedia 3(1), 3881 (2008)CrossRefGoogle Scholar
  17. 17.
    Bengio, Y., Courville, A., Vincent, P.: Representation learning: a review and new perspectives. IEEE Trans. Patt. Anal. Mach. Intell. 35(8), 1798–1828 (2013)CrossRefGoogle Scholar
  18. 18.
    Hinton, G.E., Osindero, S., Teh, Y.: A fast learning algorithm for deep belief nets. Neural Comput. 18, 1527–1554 (2006)MathSciNetCrossRefGoogle Scholar
  19. 19.
    Bengio, Y., Lamblin, P., Popovici, D., Larochelle, H.: Greedy layer-wise training of deep networks. In: Proceedings of the Neural Information and Processing Systems (2006)Google Scholar
  20. 20.
    Ranzato, M., Poultney, C., Chopra, S., LeCun, Y.: Efficient learning of sparse representations with an energy-based model. In: Proceedings of the Neural Information and Processing Systems (2006)Google Scholar
  21. 21.
    Lee, H., Ekanadham, C., Ng, A.: Sparse deep belief net model for visual area V2. In: Proceedings of the Neural Information and Processing Systems (2007)Google Scholar
  22. 22.
    Bengio, Y.: Learning Deep Architectures for AI. Foundations and Trends in Machine Learning 2(1), 1–127 (2009)MathSciNetCrossRefGoogle Scholar
  23. 23.
    Gong, J.J., Naumann, T., Szolovits, P., Guttag, J.V.: Predicting clinical outcomes across changing electronic health record systems. In: International Conference on Knowledge Discovery and Data Mining (KDD). ACM, pp. 1497–1505 (2017)Google Scholar
  24. 24.
    Choi, T., Xiao, C., Stewart, W.F., Sun, J.: MiME: multilevel medical embedding of electronic health records for predictive healthcare. arXiv. https://arxiv.org/pdf/1810.09593.pdf
  25. 25.
    Escudie, J.-B., Saade, A., Coucke, A., Lelarge, M.: Deep representation for patient visits from electronic health records. arXiv. https://arxiv.org/pdf/1803.09533.pdf
  26. 26.
    Choi, E., Schuetz, A., Steward, W.F., Sun, J.: Medical concept representation learning from electronic health records and its application on heart failure prediction. arXiv. https://arxiv.org/abs/1602.03686 (2017)
  27. 27.
    De Vine, L., Zuccon, G., Koopman, B., Sitbon, L., Bruza, P.: Medical semantic similarity with a neural language model. In: Proceedings of the 23rd ACM International conference on Information and Knowledge Management-CIKM ‘14, 3–7 Nov 2014, Shanghai, China, pp. 1819–1822. ACM, New York, NY, USAGoogle Scholar
  28. 28.
    Choi, E., Chiu, C.Y., Sontag, D.: Learning low-dimensional representations of medical concepts. https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5001761/pdf/2381736.pdf (2016)
  29. 29.
    Minarro-Gim ́enez, J.A., Mar ́ın-Alonso, O., Samwald, M.: Exploring the application of deep learning techniques on medical text corpora. Studies in health technology and informatics (2013)Google Scholar
  30. 30.
    Liu, J., Zhang, Z., Razavian, N.: Deep EHR: chronic disease prediction using medical notes. arXiv. https://arxiv.org/pdf/1808.04928.pdf (2018)
  31. 31.
    Choi, E., Bahadori, M.T., Schuetz, A., Stewart, W.F., Sun, J.: Doctor AI: predicting clinical events via recurrent neural networks. arXiv. https://arxiv.org/abs/1511.05942 (2016)
  32. 32.
    Choi, E., Bahadori, M.T., Searles, E., Coffey, C., Thompson, M., Bost, J., Tejedor-Sojo, J., Sun, J.: Multi-layer representation learning for medical concepts. In: Proceedings of the 22nd ACM SIGKDD International Conference Knowledge Discovery and Data Mining—KDD ’16’, 13–17 Aug 2016, San Francisco, CA, USA, pp. 1495–1504. ACM, New York, NY, USA (2016)Google Scholar
  33. 33.
    Li, C., Song, R., Liakata, M., Vlachos, A., Seneff, S., Zhang, X.: Using word embedding for bio-event extraction. In: Proceedings of the 2015 Workshop on Biomedical Natural Language Processing (BioNLP 2015), Beijing, China, 30 July 2015, pp. 121–126. Association for Computational Linguistics, Stroudsburg, PA (2015)Google Scholar
  34. 34.
    Tang, B., Cao, H., Wang, X., Chen, Q., Xu, H.: Evaluating word representation features in biomedical named entity recognition tasks. Biomed. Res. Int. 2014, 1–6 (2014).  https://doi.org/10.1155/2014/240403CrossRefGoogle Scholar
  35. 35.
    Miotto, R., Li, L., Kidd, B.A., Dudley, J.T.: Deep patient: an unsupervised representation to predict the future of patients from the electronic health records. Sci. Rep. 6, 26094 (2016).  https://doi.org/10.1038/srep26094CrossRefGoogle Scholar
  36. 36.
    Dligach, D., Miller, T.: Learning patient representations from text. ARXIV. https://arxiv.org/pdf/1805.02096.pdf
  37. 37.
    Zhang, Z., Kowsari, K., Harrison, J.H., Lobo, J.M., Barnes, L.E.: Patient2Vec: a personalized interpretable deep representation of the longitudinal electronic health record. arXiv. https://arxiv.org/pdf/1810.04793.pdf
  38. 38.
    Denaxas, S., Stenetorp, P., Riedel, S., Pikoula, M., Dobson, R., Hemingway, H.: Application of clinical concept embeddings for heart failure prediction in UK EHR data. arXiv. https://arxiv.org/pdf/1811.11005.pdf
  39. 39.
    Wei, X., Eickhoff, C.: Embedding electronic health records for clinical information retrieval. arXiv. https://arxiv.org/pdf/1811.05402.pdf
  40. 40.
    Zhu, Z., Yin, C., Qian, B., Cheng, Y., Wei, J., Wang, F., Measuring patient similarities via a deep architecture with medical concept embedding. arXiv. https://arxiv.org/pdf/1902.03376.pdf
  41. 41.
    Liu, L., Li, H., Hu, Z., Shi, H., Wang, Z., Tang, Z., Zhang, M.: Learning hierarchical representations of electronic health records for clinical outcome prediction. arXiv. https://arxiv.org/pdf/1903.08652.pdf
  42. 42.
    Liu, Y., Ge, T., Mathews, K., Ji, H., McGuinness, D.: Exploiting task-oriented resources to learn word embeddings for clinical abbreviation expansion. In: Proceedings of the 2015 Workshop on Biomedical Natural Language Processing (BioNLP 2015), Beijing, China, 30 July 2015. Association for Computational Linguistics, Stroudsburg, PA, pp. 92–97 (2015)Google Scholar
  43. 43.
    Wu, Y., Xu, J., Zhang, Y., Xu, H.: Clinical abbreviation disambiguation using neural word embeddings. In: Proceedings of the 2015 Workshop on Biomedical Natural Language Processing (BioNLP 2015), Beijing, China, 30 July 2015. Association for Computational Linguistics, Stroudsburg, PA, pp. 171–176 (2015)Google Scholar
  44. 44.
    Li, C., Ji, L., et al.: Acronym disambiguation using word embedding. In: Proceedings of the 29th AAAI Conference on Artificial Intelligence (2014)Google Scholar
  45. 45.
    Gligic, L., Kormilitzin, A., Goldberg, P., Nevado-Holgado, A.: Named entity recognition in electronic health records using transfer learning bootstrapped neural networks. arXiv. https://arxiv.org/pdf/1901.01592.pdf
  46. 46.
    Sachan, D.S., Xie, P., Sachan, M., Xing, E.P.: Effective use of bidirectional language modeling for transfer learning in biomedical named entity recognition. arXiv https://arxiv.org/pdf/1711.07908.pdf (2018)
  47. 47.
    Gorinski, P.J., Wu, H., Grover, C., Tobin, R., Talbot, C., Whalley, H., Sudlow, C., Whiteley, W., Alex, B.: Named entity recognition for electronic health records: a comparison of rule-based and machine learning approaches. arXiv. https://arxiv.org/pdf/1903.03985.pdf
  48. 48.
    Yin, X., Huang, X.J., Li, Z., Zhou, X.: A survival modeling approach to biomedical search result diversification using wikipedia. IEEE Trans. Knowl. Data Eng. (TKDE) 25(6), 1201–1212CrossRefGoogle Scholar
  49. 49.
    Huang, X., Zhong, M., Si, X.: York University at TREC 2005: genomics track. In: Proceedings of the Fourteenth Text REtrieval Conference (TREC), Gaithersburg, Maryland, USA, 15–18 Nov (2005)Google Scholar
  50. 50.
    Huang, X., Hu, Q.: A bayesian learning approach to promoting diversity in ranking for biomedical information retrieval. In: Proceedings of the 32nd Annual International ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR), pp. 307–314. Boston, MA, USA, 19–23 July (2009)Google Scholar
  51. 51.
    An, X., Huang, X., geNov: a new metric for measuring novelty and relevancy in biomedical information retrieval (Special Issue on Biomedical Information Retrieval). Nov 2017, 68(11), 2620–2635 (2017)CrossRefGoogle Scholar
  52. 52.
    Li, F., Zhang, M., Fu, G., Ji, D.: A neural joint model for entity and relation extraction from biomedical text. BMC Bioinformatics 18, 1 (2017).  https://doi.org/10.1186/s12859-017-1609-9CrossRefGoogle Scholar
  53. 53.
    Mehryary, F., Bjo ̈rne, J., Pyysalo, S., Salakoski, T., Ginter, F.: Deep learning with minimal training data: TurkuNLP entry in the BioNLP shared task 2016. In Proceedings of the 4th BioNLP Shared Task Workshop, 13 Aug 2016, Berlin, Germany, pp. 73–81. Association for Computational Linguistics, Stroudsburg, PA (2016)Google Scholar
  54. 54.
    Quan, C., Hua, L., Sun, X., Bai, W.: Multichannel convolutional neural network for biological relation extraction. Biomed. Res. Int. 2016, 1–10 (2016).  https://doi.org/10.1155/2016/1850404CrossRefGoogle Scholar
  55. 55.
    Pyysalo, S., Ginter, F., Moen, F., Salakoski, T.: Distributional semantics resources for biomedical text processing. In: Proceedings of the Languages in Biology and Medicine (LBM ’13), pp. 39–44, Tokyo, Japan, Dec 2013 (2013)Google Scholar
  56. 56.
    Cheng, Y., Wang, F., Zhang, P., Hu, J.: Risk prediction with electric health record: a deep learning approach. SDM 2016. https://astro.temple.edu/tua87106/sdm16.pdf (2016)
  57. 57.
    Zhang, Z., Roy, A., Li, X., Espino, S., Clara, S., Khan, S., Luo, Y.: Using clinical narratives and structured data to identify distant recurrences in breast cancer. arXiv. https://arxiv.org/pdf/1806.04818.pdf
  58. 58.
    Galk´o, F., Eickhof, C.: Biomedical question answering via weighted neural network passage retrieval. arXiv. https://arxiv.org/pdf/1801.02832.pdf
  59. 59.
    Li, H., Zhang, J., Wang, J., Lin, H., Yang, Z.: DUTIR in BioNLP-ST 2016: utilizing convolutional network and distributed representation to extract complicate relations. In: Proceedings of the 4th BioNLP Shared Task Workshop, 13 Aug 2016, Berlin, Germany, pp. 93–100. Association for Computational Linguistics, Stroudsburg, PA (2016)Google Scholar
  60. 60.
    Rahul, P.V.S.S., Sahu, S.K., Anand, A.: Biomedical event trigger identification using bidirectional recurrent neural network based models. arXiv. https://arxiv.org/abs/1705.09516v1 (2017)
  61. 61.
    Jagannatha, A.N., Yu, H.: Bidirectional RNN for medical event detection in electronic health records. In: Proceedings of the Conference Association for Computational Linguistics. North American Chapter. Meeting. See https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5119627/
  62. 62.
    Cho, K., van Merrienboer, B., Bahdanau, D., Bengio, Y.: On the properties of neural machine translation: encoder-decoder approaches. arXiv e-prints. 2014 Sep. 1409:arXiv:1409.1259 (2014)
  63. 63.
    Choi, E., Biswal, S., Malin, B., Duke, J., Stewart, W.F., Sun, J.: Generating multi-label discrete electronic health records using generative adversarial networks. arXiv. https://arxiv.org/abs/1703.06490v1 (2017)
  64. 64.
    Lee, S.: Natural language generation for electronic health records. arXiv. https://arxiv.org/pdf/1806.01353.pdf
  65. 65.
    Vinyals, O., Toshev, A., Bengio, S., Erhan, D.: Show and tell: a neural image caption generator. In: Computer Vision and Pattern Recognition (CVPR), 2015 IEEE Conference on 2015 Jun 7, pp. 3156-3164. IEEE (2015)Google Scholar
  66. 66.
    Liu, X., Xu, K., Xie, P., Xing, E.: Unsupervised pseudo-labeling for extractive summarization on electronic health records. arXiv. https://arxiv.org/pdf/1811.08040.pdf
  67. 67.
    Datta, S., Bernstam, S.V., Roberts, K.: A frame semantic overview of NLP-based information extraction for cancer-related EHR notes. arXiv. https://arxiv.org/pdf/1904.01655.pdf
  68. 68.
    Zeng, Z., Deng, Y., Li, X., Naumann, T., Luo, Y.: Natural language processing for EHR-based computational phenotyping. arXiv. https://arxiv.org/pdf/1806.04820.pdf
  69. 69.
    Rajkomar, A., Oren, E., Chen, K., Dai, A.M., Hajaj, N., Liu, P.J., Liu, X., Sun, M., Sundberg, P., Yee, H., et al.: Scalable and accurate deep learning for electronic health records. arXiv preprint. arXiv:1801.07860 (2018)
  70. 70.
    Zhang, X.S., Tang, F., Dodge, H., Zhou, J., Wang, F.: MetaPred: meta-learning for clinical risk prediction with limited patient electronic health records. arXiv. https://arxiv.org/pdf/1905.03218.pdf
  71. 71.
    Hosseini, A., Chen, T., Wu, W., Sun, Y., Sarrafzadeh, M.: HeteroMed: heterogeneous information network for medicaldiagnosis. arXiv., https://arxiv.org/pdf/1804.08052.pdf
  72. 72.
    Avati, A., Duan, T., Jung, K., Shah, N.H., Ng, A.: Countdown regression: sharp and calibrated survival predictions. arXiv. https://arxiv.org/pdf/1806.08324.pdf
  73. 73.
    Chung, I., Kim, S., Lee, J., Hwang, S.J., Yang, E.: Mixed effect composite RNN-GP: a personalized and reliable prediction model for healthcare. arXiv. https://arxiv.org/pdf/1806.01551.pdf
  74. 74.
    Heo, J., Lee, H.B., Kim, S., Lee, J., Kim, K.J., Yang, K., Hwang, S.J.: Uncertainty-aware attention for reliable interpretation and prediction. arXiv. https://arxiv.org/pdf/1805.09653.pdf
  75. 75.
    Wang, L., Zhang, W., He, X., Zha, H.: Supervised reinforcement learning with recurrent neural network for dynamic treatment recommendation. arXiv. https://arxiv.org/pdf/1807.01473.pdf
  76. 76.
    Pham, T., Tran, T., Phung, D., Venkatesh, S.: DeepCare: a deep dynamic memory model for predictive medicine. arXiv. https://arxiv.org/abs/1602.00357v2 (2016)
  77. 77.
    Ma, F., Gao, J., Suo, Q., You, Q., Zhou, J., Zhang, A.: 2018 risk prediction on electronic health records with prior medical knowledge. In: KDD ’18: The 24th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, 19–23 Aug 2018, London, United Kingdom. ACM, New York, NY, USA, p. 10.  https://doi.org/10.1145/3219819.3220020
  78. 78.
    Suresh, H., Hunt, N., Johnson, A., Celi, L.A., Szolovits, P., Ghassemi, M.: Clinical intervention prediction and understanding with deep neural networks. In: Machine Learning for Healthcare Conference, pp. 322–337 (2017)Google Scholar
  79. 79.
    Lasko, T.A., Denny, J.C., Levy, M.A.: Computational phenotype discovery using unsupervised feature learning over noisy, sparse, and irregular clinical data. PLoS ONE 8, e66341 (2013).  https://doi.org/10.1371/journal.pone.0066341CrossRefGoogle Scholar
  80. 80.
    Liang, Z., Liu, J., Ou, A., Zhang, H., Li, Z., Huang, X.: Deep generative learning for automated EHR diagnosis of traditional Chinese medicine. Comput. Methods Progr. Biomed. 174, 17–23 (2019)CrossRefGoogle Scholar
  81. 81.
    Liang, Z., Zhang, G., Huang, X., Hu, Q.: Deep learning for healthcare decision making with EMRs. In: Proceedings of 2014 IEEE International Conference on Bioinformatics and Biomedicine (BIBM), pp. 556–559Google Scholar
  82. 82.
    Mei, j., Zhao, S., Jin, F., Xia, E., Liu, H., Li, X.: Deep diabetologist: learning to prescribe hypoglycemia medications with hierarchical recurrent neural networks. arXiv. https://arxiv.org/pdf/1810.07692.pdf
  83. 83.
    Sousa, R.T., Pereira, L.A., Soares, A.S.: Predicting diabetes disease evolution using financial records and recurrent neural networks. arXiv. https://arxiv.org/pdf/1811.09350.pdf
  84. 84.
    Kale, D.C, Che, Z., Bahadori, M.T., Li, W., Liu, Y., Wetzel, R.: Causal phenotype discovery via deep networks. AMIA Annual Symposium Proceedings https://www.ncbi.nlm.nih.gov/pmc/articles/PMC4765623/ (2015)
  85. 85.
    Ghassemi, M., Naumann, T., Schulam, P., Beam, A.L., Ranganath, R.: Opportunities in machine learning for healthcare. arXiv. https://arxiv.org/pdf/1806.00388.pdf (2018)
  86. 86.
    Lyu, X., Huser, M., Hyland, S.L., Zerveas, G., Ratsch, G.: Improving clinical predictions through unsupervised time series representation learning. arXiv https://arxiv.org/pef/1812.00490.pdf (2018)
  87. 87.
    Nickel, M., Kiela, D.: Poincar\’e embeddings for learning hierarchical representations. arXiv preprint arXiv:1705.08039 (2017)
  88. 88.
    Greenland, S., Robins, J.M., Pearl, J.: Confounding and collapsibility in causal inference. Stat. Sci., pp. 29–46 (1999)Google Scholar
  89. 89.
    Miotto, R., Wang, F., Wang, S., Jiang, Z., Dudley, J.T.: Deep learning for healthcare: review, opportunities and challenges. Brief. Bioinform. 375, 4 (2017).  https://doi.org/10.1093/bib/bbx044CrossRefGoogle Scholar
  90. 90.
    Wei, C.-H., Harris, B.R., Kao, H.-Y., Lu, Z.: tmVar: a text mining approach for extracting sequence variants in biomedical literature. Bioinformatics 29, 1433–1439 (2013).  https://doi.org/10.1093/bioinformatics/btt156CrossRefGoogle Scholar
  91. 91.
    Liu, S., Tang, B., Chen, Q., Wang, X.: Effects of semantic features on machine learning-based drug name recognition systems: word embeddings vs. manually constructed dictionaries. Information 6, 848–865 (2015).  https://doi.org/10.3390/info6040848CrossRefGoogle Scholar
  92. 92.
    Mohan, S., Fiorini, N., Kim, S., Lu, Z.: Deep learning for biomedical information retrieval: learning textual relevance from click logs. In: Proceedings of the BioNLP 2017 Workshop, Vancouver, Canada, 4 Aug 2017, pp. 222–231. Association for Computational Linguistics Stroudsburg, PA (2017)Google Scholar
  93. 93.
    Ohno-Machado, L.: Realizing the full potential of electronic health records: the role of natural language processing. J. Am. Med. Inform. Assoc. 18, 539 (2011).  https://doi.org/10.1136/amiajnl-2011-000501CrossRefGoogle Scholar
  94. 94.
    Bruijn, Bd, Cherry, C., Kiritchenko, S., Martin, J., Zhu, X.: Machine-learned solutions for three stages of clinical information extraction: the state of the art at i2b2 2010. J. Am. Med. Inform. Assoc. 18, 557–562 (2011).  https://doi.org/10.1136/amiajnl-2011-000150CrossRefGoogle Scholar
  95. 95.
    Yoon, H.-J., Ramanathan, A., Tourassi, G.: Multi-task deep neural networks for automated extraction of primary site and laterality information from cancer pathology reports. In: Advances in big data, INNS 2016, 23–25 Oct 2016, Thessaloniki, Greece; Angelov, P., Manolopoulos, Y., Iliadis, L., Roy, A., Vellasco, M. (eds.)Advances in Intelligent Systems and Computing, vol. 529. Springer, Cham (2016)Google Scholar
  96. 96.
    Beaulieu-Jones, B.K., Greene, C.S.: Semi- supervised learning of the electronic health record for phenotype stratification. J. Biomed. Inform. 64, 168–178 (2016).  https://doi.org/10.1016/j.jbi.2016.10.007CrossRefGoogle Scholar
  97. 97.
    Bowman, S.: Impact of electronic health record systems on information integrity: quality and safety implications. Perspect. Health Inf. Manag. 10, 1c (2013)Google Scholar
  98. 98.
    Beaulieu-Jones, B.K., Wu, Z.S., Williams, C., Byrd, J.B., Greene, C.S.: Privacy-preserving generative deep neural networks support clinical data sharing. bioRxiv  https://doi.org/10.1101/159756 (2017)
  99. 99.
    Letham, B., Rudin, C., McCormick, T.H., Madigan, D., et al.: Interpretable classifiers using rules and bayesian analysis: building a better stroke prediction model. Ann. Appl. Stat. 9(3), 1350–1371 (2015)MathSciNetCrossRefGoogle Scholar
  100. 100.
    Robins, J.M.: Robust estimation in sequentially ignorable missing data and causal inference models. Proc. Am. Stat. Assoc. 1999, 6–10 (2000)Google Scholar
  101. 101.
    Robins, J.M., Rotnitzky, A., Scharfstein, D.O.: Sensitivity analysis for selection bias and unmeasured confounding in missing data and causal inference models. In: Statistical models in epidemiology, the environment, and clinical trials. Springer, pp 1–94 (2000)Google Scholar
  102. 102.
    Papernot, N., McDaniel, P., Sinha, A., Wellman, M.: Towards the science of security and privacy in machine learning. arXiv. https://arxiv.org/abs/1611.03814v1 (2016)
  103. 103.
    Xu, Z., Chou, J., Zhang, X.S., Luo, Y., Isakova, T., et al.: Identification of predictive sub-phenotypes of acute kidney injury using structured and unstructured electronic health record data with memory networks. arXiv. https://arxiv.org/pdf/1904.04990.pdf
  104. 104.
    Chou, E., Nguyen, T., Beal, J., Haque, A., Fei-Fei, L.: A fully private pipeline for deep learning on electronic health records. arXiv. https://arxiv.org/pdf/1811.09951.pdf
  105. 105.
    Banerjee, I., Gensheimer, M.F., Wood, D.J., Henry, S., Chang, D., Rubin, D.L.: Probabilistic prognostic estimates of survival in metastatic cancer patients (PPES-Met) utilizing free-text clinical narratives. arXiv. https://arxiv.org/pdf/1801.03058.pdf
  106. 106.
    Kayali, I.: Expert system for diagnosis of chest diseases using neural networks. arXiv. https://arxiv.org/pdf/1802.06866.pdf
  107. 107.
    de la Torre, J., Valls, A., Puig, D.: A deep learning interpretable classifier for diabetic retinopathy disease grading. arXiv. https://arxiv.org/pdf/1712.08107.pdf
  108. 108.
    Holzinger, A., Malle, B., Kieseberg, P., Roth, P.M., M¨uller, H., Reihs, R., Zatloukal, K.: Towards the augmented pathologist: challenges of explainable-ai in digital pathology. arXiv. https://arxiv.org/pdf/1712.06657.pdf

Copyright information

© Springer Nature Switzerland AG 2020

Authors and Affiliations

  1. 1.Information Retrieval and Knowledge Management Research Lab, Department of Electrical Engineering and Computer ScienceYork UniversityTorontoCanada
  2. 2.School of Computer ScienceCentral China Normal UniversityWuhanChina
  3. 3.Information Retrieval and Knowledge Management Research Lab, School of Information TechnologyYork UniversityTorontoCanada

Personalised recommendations