{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2024,10,18]],"date-time":"2024-10-18T04:28:12Z","timestamp":1729225692376,"version":"3.27.0"},"reference-count":0,"publisher":"IOS Press","isbn-type":[{"value":"9781643685489","type":"electronic"}],"license":[{"start":{"date-parts":[[2024,10,16]],"date-time":"2024-10-16T00:00:00Z","timestamp":1729036800000},"content-version":"unspecified","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc\/4.0\/"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2024,10,16]]},"abstract":"Most existing biomedical language models are trained on plain text with general learning goals such as random word infilling, failing to capture the knowledge in the biomedical corpus sufficiently. Since biomedical articles usually contain many tables summarising the main entities and their relations, in the paper, we propose a Tabular knowledge enhanced bioMedical pretrained language model, called TabMedBERT. Specifically, we align entities between table cells, and article text spans with pre-defined rules. Then we add two table-related self-supervised tasks to integrate tabular knowledge into the language model: Entity Infilling (EI) and Table Cloze Test (TCT). While EI masks tokens within aligned entities in the article, TCT converts aligned entities in the table layout into a cloze text by erasing one entity and prompts the model to extract the appropriate span to fill in the blank. Experimental results demonstrate that TabMedBERT surpasses all competing language models without adding additional parameters, establishing a new state-of-the-art performance of 85.59% (+1.29%) on the BLURB biomedical NLP benchmark and 7 additional information extraction datasets. Moreover, the model architecture for TCT provides a straightforward solution to revise information extraction with paired entities.<\/jats:p>","DOI":"10.3233\/faia240674","type":"book-chapter","created":{"date-parts":[[2024,10,17]],"date-time":"2024-10-17T13:04:35Z","timestamp":1729170275000},"source":"Crossref","is-referenced-by-count":0,"title":["TabMedBERT: A Tabular Knowledge Enhanced Biomedical Pretrained Language Model"],"prefix":"10.3233","author":[{"given":"Xu","family":"Yan","sequence":"first","affiliation":[{"name":"Institute of Artificial Intelligence, Soochow University"}]},{"given":"Lei","family":"Geng","sequence":"additional","affiliation":[{"name":"Institute of Artificial Intelligence, Soochow University"}]},{"given":"Ziqiang","family":"Cao","sequence":"additional","affiliation":[{"name":"Institute of Artificial Intelligence, Soochow University"}]},{"given":"Juntao","family":"Li","sequence":"additional","affiliation":[{"name":"Institute of Artificial Intelligence, Soochow University"}]},{"given":"Wenjie","family":"Li","sequence":"additional","affiliation":[{"name":"The Hong Kong Polytechnic University"}]},{"given":"Sujian","family":"Li","sequence":"additional","affiliation":[{"name":"Peking University"}]},{"given":"Xinjie","family":"Zhou","sequence":"additional","affiliation":[{"name":"Pharmcube Inc."}]},{"given":"Yang","family":"Yang","sequence":"additional","affiliation":[{"name":"Pharmcube Inc."}]},{"given":"Jun","family":"Zhang","sequence":"additional","affiliation":[{"name":"Changping Laboratory"}]}],"member":"7437","container-title":["Frontiers in Artificial Intelligence and Applications","ECAI 2024"],"original-title":[],"link":[{"URL":"https:\/\/ebooks.iospress.nl\/pdf\/doi\/10.3233\/FAIA240674","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,10,17]],"date-time":"2024-10-17T13:04:36Z","timestamp":1729170276000},"score":1,"resource":{"primary":{"URL":"https:\/\/ebooks.iospress.nl\/doi\/10.3233\/FAIA240674"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,10,16]]},"ISBN":["9781643685489"],"references-count":0,"URL":"https:\/\/doi.org\/10.3233\/faia240674","relation":{},"ISSN":["0922-6389","1879-8314"],"issn-type":[{"value":"0922-6389","type":"print"},{"value":"1879-8314","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,10,16]]}}}