was successfully added to your cart.

Sending BERT to Med School – Injecting Medical Knowledge into BERT

General NLP research has greatly advanced over the past several years thanks to large pre-trained neural language models such as BERT and GPT. But in the healthcare domain, this pre-training on unlabeled clinical texts can take us only so far, due to the special properties of medical texts and the lack of labeled data. One possible solution to this problem is to utilize medical ontologies in order to enrich pre-trained neural networks with medical knowledge.

This approach is not limited to the healthcare domain ⁠— recent studies have shown that knowledge graphs, such as WordNet and ConceptNet, can be used to infuse common sense and world knowledge into neural language models. In this lecture we will explore similar approaches using medical knowledge structures such as UMLS, ICD and MeSH, and examine how structured knowledge can be used in a neural network setup to solve various healthcare information retrieval challenges.

This lecture is intended for people interested in the challenges of healthcare language processing and for people in other specialized domains who want to get their structured knowledge sources out of the attic.

Application of Spark NLP for Development of Multi-Modal Prediction Model from EHR

Different data sources such as structured data, clinical notes, laboratory measurements capture information about the human body at different time-scales. There is...