首页 | 本学科首页   官方微博 | 高级检索  
     


MT-clinical BERT: scaling clinical information extraction with multitask learning
Authors:Andriy Mulyar  Ozlem Uzuner  Bridget McInnes
Affiliation:1.Computer Science Department, Virginia Commonwealth University, Richmond, Virginia, USA;2.Information Sciences and Technology, George Mason University, Fairfax, Virginia, USA
Abstract:
ObjectiveClinical notes contain an abundance of important, but not-readily accessible, information about patients. Systems that automatically extract this information rely on large amounts of training data of which there exists limited resources to create. Furthermore, they are developed disjointly, meaning that no information can be shared among task-specific systems. This bottleneck unnecessarily complicates practical application, reduces the performance capabilities of each individual solution, and associates the engineering debt of managing multiple information extraction systems.Materials and MethodsWe address these challenges by developing Multitask-Clinical BERT: a single deep learning model that simultaneously performs 8 clinical tasks spanning entity extraction, personal health information identification, language entailment, and similarity by sharing representations among tasks.ResultsWe compare the performance of our multitasking information extraction system to state-of-the-art BERT sequential fine-tuning baselines. We observe a slight but consistent performance degradation in MT-Clinical BERT relative to sequential fine-tuning.DiscussionThese results intuitively suggest that learning a general clinical text representation capable of supporting multiple tasks has the downside of losing the ability to exploit dataset or clinical note-specific properties when compared to a single, task-specific model.ConclusionsWe find our single system performs competitively with all state-the-art task-specific systems while also benefiting from massive computational benefits at inference.
Keywords:multitask learning   natural language processing   clinical natural language processing   named entity recognition   textual entailment   semantic text similarity
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号