Show simple item record

dc.contributor.authorExternal author(s) only
dc.date.accessioned2020-07-15T15:58:36Z
dc.date.available2020-07-15T15:58:36Z
dc.date.issued2020-04
dc.identifier.citationAndrey Kormilitzina, Nemanja Vacia, Qiang Liua, Alejo Nevado-Holgado. Med7: a transferable clinical natural language processing model for electronic health records (arxiv.org - preprint)en
dc.identifier.issn2331-8422
dc.identifier.urihttps://oxfordhealth-nhs.archive.knowledgearc.net/handle/123456789/517
dc.description.abstractThe field of clinical natural language processing has been advanced significantly since the introduction of deep learning models. The self-supervised representation learning and the transfer learning paradigm became the methods of choice in many natural language processing application, in particular in the settings with the dearth of high quality manually annotated data. Electronic health record systems are ubiquitous and the majority of patients’ data are now being collected electronically and in particular in the form of free text. Identification of medical concepts and information extraction is a challenging task, yet important ingredient for parsing unstructured data into structured and tabulated format for downstream analytical tasks. In this work we introduced a named-entity recognition model for clinical natural language processing. The model is trained to recognise seven categories: drug names, route, frequency, dosage, strength, form, duration. The model was first self-supervisedly pre-trained by predicting the next word, using a collection of 2 million free-text patients’ records from MIMIC-III corpora and then fine-tuned on the named-entity recognition task. The model achieved a lenient (strict) micro-averaged F1 score of 0.957 (0.893) across all seven categories. Additionally, we evaluated the transferability of the developed model using the data from the Intensive Care Unit in the US to secondary care mental health records (CRIS) in the UK. A direct application of the trained NER model to CRIS data resulted in reduced performance of F1=0.762, however after fine-tuning on a small sample from CRIS, the model achieved a reasonable performance of F1=0.944. This demonstrated that despite a close similarity between the data sets and the NER tasks, it is essential to fine-tune on the target domain data in order to achieve more accurate results.en
dc.description.urihttps://arxiv.org/pdf/2003.01271.pdfen
dc.language.isoenen
dc.subjectNatural Language Processingen
dc.subjectCRISen
dc.titleMed7: a transferable clinical natural language processing model for electronic health recordsen
dc.typePreprinten


Files in this item

FilesSizeFormatView

There are no files associated with this item.

This item appears in the following Collection(s)

Show simple item record