Publication:
KU_ai at MEDIQA 2019: domain-specific pre-training and transfer learning for medical NLI

dc.contributor.departmentDepartment of Computer Engineering
dc.contributor.departmentGraduate School of Sciences and Engineering
dc.contributor.kuauthorCengiz, Cemil
dc.contributor.kuauthorSert, Ulaş
dc.contributor.kuauthorYüret, Deniz
dc.contributor.schoolcollegeinstituteCollege of Engineering
dc.contributor.schoolcollegeinstituteGRADUATE SCHOOL OF SCIENCES AND ENGINEERING
dc.date.accessioned2024-11-09T23:26:19Z
dc.date.issued2019
dc.description.abstractin this paper, we describe our system and results submitted for the Natural Language inference (NLI) track of the MEDIQa 2019 Shared Task (Ben abacha et al., 2019). as KU ai team, we used BERT (Devlin et al., 2018) as our baseline model and pre-processed the MedNLI dataset to mitigate the negative impact of de-identification artifacts. Moreover, we investigated different pre-training and transfer learning approaches to improve the performance. We show that pre-training the language model on rich biomedical corpora has a significant effect in teaching the model domain-specific language. in addition, training the model on large NLI datasets such as MultiNLI and SNLI helps in learning task-specific reasoning. Finally, we ensembled our highest-performing models, and achieved 84.7% accuracy on the unseen test dataset and ranked 10th out of 17 teams in the official results.
dc.description.indexedbyWOS
dc.description.openaccessNO
dc.description.publisherscopeInternational
dc.description.sponsoredbyTubitakEuN/A
dc.description.sponsorshipHuawei Turkey R&D Center through the Huawei Graduate Research Support Scholarship Cemil Cengiz is supported by Huawei Turkey R&D Center through the Huawei Graduate Research Support Scholarship.
dc.identifier.isbn978-1-950737-28-4
dc.identifier.quartileN/A
dc.identifier.urihttps://hdl.handle.net/20.500.14288/11535
dc.identifier.wos521946800045
dc.language.isoeng
dc.publisherassoc Computational Linguistics-acl
dc.relation.ispartofSigbiomed Workshop on Biomedical Natural Language Processing (Bionlp 2019)
dc.subjectComputer science
dc.subjectArtificial intelligence
dc.subjectMedical informatics
dc.titleKU_ai at MEDIQA 2019: domain-specific pre-training and transfer learning for medical NLI
dc.typeConference Proceeding
dspace.entity.typePublication
local.contributor.kuauthorCengiz, Cemil
local.contributor.kuauthorSert, Ulaş
local.contributor.kuauthorYüret, Deniz
local.publication.orgunit1GRADUATE SCHOOL OF SCIENCES AND ENGINEERING
local.publication.orgunit1College of Engineering
local.publication.orgunit2Department of Computer Engineering
local.publication.orgunit2Graduate School of Sciences and Engineering
relation.isOrgUnitOfPublication89352e43-bf09-4ef4-82f6-6f9d0174ebae
relation.isOrgUnitOfPublication3fc31c89-e803-4eb1-af6b-6258bc42c3d8
relation.isOrgUnitOfPublication.latestForDiscovery89352e43-bf09-4ef4-82f6-6f9d0174ebae
relation.isParentOrgUnitOfPublication8e756b23-2d4a-4ce8-b1b3-62c794a8c164
relation.isParentOrgUnitOfPublication434c9663-2b11-4e66-9399-c863e2ebae43
relation.isParentOrgUnitOfPublication.latestForDiscovery8e756b23-2d4a-4ce8-b1b3-62c794a8c164

Files