BioBERT: a pre-trained biomedical language representation model for biomedical text mining BioBERT: a pre-trained biomedical language representation model for biomedical text mining
Paper summary The paper discusses the idea of using BERT pre-trained network in bio-medical domain NLP tasks. It lays a path for future applications of Bert in different business verticals. Sharing some points from the review article I wrote on BioBERT on medium (https://medium.com/@raghudeep/biobert-insights-b4c66fde8fa7). The major contribution is a pre-trained bio-medical language representation model for various bio-medical text mining tasks. Tasks such as NER from Bio-medical data, relation extraction, question & answer in the biomedical field. For pretraining the model, authors have used a combination of general & medical corpora, which has shown interesting results. They have compared the results of all their combinations.
arxiv.org
scholar.google.com
BioBERT: a pre-trained biomedical language representation model for biomedical text mining
Lee, Jinhyuk and Yoon, Wonjin and Kim, Sungdong and Kim, Donghyeon and Kim, Sunkyu and So, Chan Ho and Kang, Jaewoo
arXiv e-Print archive - 2019 via Local Bibsonomy
Keywords: dblp


[link]
Summary by wanderer 3 months ago
Loading...
Your comment:


ShortScience.org allows researchers to publish paper summaries that are voted on and ranked!
About

Sponsored by: and