Contrastive Representations Pre-Training for Enhanced Discharge Summary BERT

Cited 1 time in webofscience Cited 0 time in scopus
  • Hit : 60
  • Download : 0
Recently BERT has shown tremendous improvement in performance for various NLP tasks. BERT has been applied to many domains including biomedical field. Especially clinical domain, the semantic relationship between sentences is very important to understand patient's medical record and health history in physical examination. However, in current Clinical BERT model, the pre-training method is difficult to capture sentence level semantics. To address this problem, we propose a contrastive representations pre-training (CRPT), which can enhance contextual meanings between sentences by replacing cross-entropy loss to contrastive loss in next sentence prediction (NSP) task. Also we tried to improve the performance by changing random masking technique to whole word masking (WWM) for masked language model (MLM). Especially, we focus on enhancing language representations of BERT model by pre-training with discharge summaries to optimize clinical studies. We demonstrate that our CRPT strategy yields performance improvements on clinical NLP task in BLUE (Biomedical Language Understanding Evaluation) Benchmark dataset.
Publisher
IEEE
Issue Date
2021-08
Language
English
Citation

2021 IEEE 9th International Conference on Healthcare Informatics (ICHI), pp.507 - 508

ISSN
2575-2634
DOI
10.1109/ichi52183.2021.00093
URI
http://hdl.handle.net/10203/312266
Appears in Collection
CS-Conference Papers(학술회의논문)
Files in This Item
There are no files associated with this item.
This item is cited by other documents in WoS
⊙ Detail Information in WoSⓡ Click to see webofscience_button
⊙ Cited 1 items in WoS Click to see citing articles in records_button

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0