{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,19]],"date-time":"2025-03-19T13:28:16Z","timestamp":1742390896584,"version":"3.37.3"},"publisher-location":"New York, NY, USA","reference-count":43,"publisher":"ACM","funder":[{"DOI":"10.13039\/https:\/\/doi.org\/10.13039\/100000002","name":"NIH (National Institutes of Health)","doi-asserted-by":"publisher","award":["K25DK135913"],"id":[{"id":"10.13039\/https:\/\/doi.org\/10.13039\/100000002","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2024,7,10]]},"DOI":"10.1145\/3626772.3657904","type":"proceedings-article","created":{"date-parts":[[2024,7,11]],"date-time":"2024-07-11T16:40:05Z","timestamp":1720716005000},"page":"2589-2593","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":2,"title":["PromptLink: Leveraging Large Language Models for Cross-Source Biomedical Concept Linking"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0009-0001-3241-9418","authenticated-orcid":false,"given":"Yuzhang","family":"Xie","sequence":"first","affiliation":[{"name":"Emory University, Atlanta, GA, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-9052-6951","authenticated-orcid":false,"given":"Jiaying","family":"Lu","sequence":"additional","affiliation":[{"name":"Emory University, Atlanta, GA, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-9168-3916","authenticated-orcid":false,"given":"Joyce","family":"Ho","sequence":"additional","affiliation":[{"name":"Emory University, Atlanta, GA, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-1710-9881","authenticated-orcid":false,"given":"Fadi","family":"Nahab","sequence":"additional","affiliation":[{"name":"Emory University, Atlanta, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-9478-5571","authenticated-orcid":false,"given":"Xiao","family":"Hu","sequence":"additional","affiliation":[{"name":"Emory University, Atlanta, GA, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-9145-4531","authenticated-orcid":false,"given":"Carl","family":"Yang","sequence":"additional","affiliation":[{"name":"Emory University, Atlanta, GA, USA"}]}],"member":"320","published-online":{"date-parts":[[2024,7,11]]},"reference":[{"key":"e_1_3_2_1_1_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.cell.2019.02.039"},{"key":"e_1_3_2_1_2_1","first-page":"72","article-title":"Publicly Available Clinical BERT Embeddings","volume":"2019","author":"Alsentzer Emily","year":"2019","unstructured":"Emily Alsentzer, John R Murphy, Willie Boag, Wei-Hung Weng, Di Jin, Tristan Naumann, WA Redmond, and Matthew BA McDermott. 2019. Publicly Available Clinical BERT Embeddings. NAACL HLT 2019 (2019), 72.","journal-title":"NAACL HLT"},{"key":"e_1_3_2_1_3_1","doi-asserted-by":"publisher","DOI":"10.1136\/jamia.2009.002733"},{"key":"e_1_3_2_1_4_1","volume-title":"Translating embeddings for modeling multi-relational data. Advances in neural information processing systems","author":"Bordes Antoine","year":"2013","unstructured":"Antoine Bordes, Nicolas Usunier, Alberto Garcia-Duran, Jason Weston, and Oksana Yakhnenko. 2013. Translating embeddings for modeling multi-relational data. Advances in neural information processing systems , Vol. 26 (2013)."},{"key":"e_1_3_2_1_5_1","volume-title":"Yuanzhi Li, Scott Lundberg, et al.","author":"Bubeck S\u00e9bastien","year":"2023","unstructured":"S\u00e9bastien Bubeck, Varun Chandrasekaran, Ronen Eldan, Johannes Gehrke, Eric Horvitz, Ece Kamar, Peter Lee, Yin Tat Lee, Yuanzhi Li, Scott Lundberg, et al. 2023. Sparks of artificial general intelligence: Early experiments with gpt-4. arXiv preprint arXiv:2303.12712 (2023)."},{"key":"e_1_3_2_1_6_1","unstructured":"Kevin Donnelly et al. 2006. SNOMED-CT: The advanced terminology and coding system for eHealth. Studies in health technology and informatics Vol. 121 (2006) 279."},{"key":"e_1_3_2_1_7_1","doi-asserted-by":"publisher","DOI":"10.3115\/v1\/P15-2049"},{"key":"e_1_3_2_1_8_1","volume-title":"Accessed","author":"Centers for Disease Control.","year":"2007","unstructured":"Centers for Disease Control. 2007. International Classification of Diseases-9-CM. Available at http:\/\/www.cdc.gov\/nchs\/icd.htm. Accessed Feb, 2024."},{"key":"e_1_3_2_1_9_1","volume-title":"Proceedings of the AMIA Symposium. American Medical Informatics Association, 189","author":"Friedman Carol","year":"2001","unstructured":"Carol Friedman, Hongfang Liu, Lyudmila Shagina, Stephen Johnson, and George Hripcsak. 2001. Evaluating the UMLS as a source of lexical knowledge for medical language processing.. In Proceedings of the AMIA Symposium. American Medical Informatics Association, 189."},{"key":"e_1_3_2_1_10_1","doi-asserted-by":"publisher","DOI":"10.1145\/2939672.2939754"},{"key":"e_1_3_2_1_11_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-642-25073-6_18"},{"key":"e_1_3_2_1_12_1","volume-title":"Leo Anthony Celi, and Roger G Mark","author":"Johnson Alistair EW","year":"2016","unstructured":"Alistair EW Johnson, Tom J Pollard, Lu Shen, Li-wei H Lehman, Mengling Feng, Mohammad Ghassemi, Benjamin Moody, Peter Szolovits, Leo Anthony Celi, and Roger G Mark. 2016. MIMIC-III, a freely accessible critical care database. Scientific data, Vol. 3, 1 (2016), 1--9."},{"key":"e_1_3_2_1_13_1","doi-asserted-by":"publisher","DOI":"10.1136\/amiajnl-2012-001173"},{"key":"e_1_3_2_1_14_1","volume-title":"Bilinear attention networks. Advances in neural information processing systems","author":"Kim Jin-Hwa","year":"2018","unstructured":"Jin-Hwa Kim, Jaehyun Jun, and Byoung-Tak Zhang. 2018. Bilinear attention networks. Advances in neural information processing systems , Vol. 31 (2018)."},{"key":"e_1_3_2_1_15_1","doi-asserted-by":"publisher","DOI":"10.2196\/22219"},{"key":"e_1_3_2_1_16_1","doi-asserted-by":"publisher","DOI":"10.1093\/bioinformatics\/btz682"},{"key":"e_1_3_2_1_17_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2021.naacl-main.334"},{"key":"e_1_3_2_1_18_1","doi-asserted-by":"publisher","DOI":"10.1145\/3485447.3511945"},{"key":"e_1_3_2_1_19_1","doi-asserted-by":"publisher","DOI":"10.1145\/3539618.3591997"},{"key":"e_1_3_2_1_20_1","volume-title":"BioGPT: generative pre-trained transformer for biomedical text generation and mining. Briefings in bioinformatics","author":"Luo Renqian","year":"2022","unstructured":"Renqian Luo, Liai Sun, Yingce Xia, Tao Qin, Sheng Zhang, Hoifung Poon, and Tie-Yan Liu. 2022. BioGPT: generative pre-trained transformer for biomedical text generation and mining. Briefings in bioinformatics, Vol. 23, 6 (2022), bbac409."},{"key":"e_1_3_2_1_21_1","doi-asserted-by":"publisher","DOI":"10.1145\/3219819.3220020"},{"key":"e_1_3_2_1_22_1","volume-title":"Jerry Tworek, Qiming Yuan, Nikolas Tezak, Jong Wook Kim, Chris Hallacy, et al.","author":"Neelakantan Arvind","year":"2022","unstructured":"Arvind Neelakantan, Tao Xu, Raul Puri, Alec Radford, Jesse Michael Han, Jerry Tworek, Qiming Yuan, Nikolas Tezak, Jong Wook Kim, Chris Hallacy, et al. 2022. Text and code embeddings by contrastive pre-training. arXiv preprint arXiv:2201.10005 (2022)."},{"key":"e_1_3_2_1_23_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/D19-1005"},{"key":"e_1_3_2_1_24_1","doi-asserted-by":"publisher","DOI":"10.1109\/34.682181"},{"key":"e_1_3_2_1_25_1","doi-asserted-by":"publisher","DOI":"10.1561\/1500000019"},{"key":"e_1_3_2_1_26_1","volume-title":"Bioinformatics","volume":"39","author":"Rohanian Omid","year":"2023","unstructured":"Omid Rohanian, Mohammadmahdi Nouriborji, Samaneh Kouchaki, and David A Clifton. 2023. On the effectiveness of compact biomedical transformers. Bioinformatics, Vol. 39, 3 (2023), btad103."},{"key":"e_1_3_2_1_27_1","doi-asserted-by":"publisher","DOI":"10.1136\/jamia.2009.001560"},{"key":"e_1_3_2_1_28_1","first-page":"217","article-title":"The UMLS Metathesaurus: representing different views of biomedical concepts","volume":"81","author":"Schuyler Peri L","year":"1993","unstructured":"Peri L Schuyler, William T Hole, Mark S Tuttle, and David D Sherertz. 1993. The UMLS Metathesaurus: representing different views of biomedical concepts. Bulletin of the Medical Library Association, Vol. 81, 2 (1993), 217.","journal-title":"Bulletin of the Medical Library Association"},{"key":"e_1_3_2_1_29_1","doi-asserted-by":"publisher","DOI":"10.3233\/SW-222986"},{"key":"e_1_3_2_1_30_1","doi-asserted-by":"publisher","DOI":"10.1007\/s11280-023-01144-4"},{"key":"e_1_3_2_1_31_1","volume-title":"Nathan Scales, Ajay Tanwani, Heather Cole-Lewis, Stephen Pfohl, et al.","author":"Singhal Karan","year":"2023","unstructured":"Karan Singhal, Shekoofeh Azizi, Tao Tu, S Sara Mahdavi, Jason Wei, Hyung Won Chung, Nathan Scales, Ajay Tanwani, Heather Cole-Lewis, Stephen Pfohl, et al. 2023. Large language models encode clinical knowledge. Nature, Vol. 620, 7972 (2023), 172--180."},{"key":"e_1_3_2_1_32_1","volume-title":"Zehra Abedi, Haotan Zhang, Zilong Bai, Anthony Cuturrufo, Winston Guo, et al.","author":"Su Chang","year":"2023","unstructured":"Chang Su, Yu Hou, Manqi Zhou, Suraj Rajendran, Jacqueline RMA Maasch, Zehra Abedi, Haotan Zhang, Zilong Bai, Anthony Cuturrufo, Winston Guo, et al. 2023. Biomedical discovery through the integrative biomedical knowledge hub (iBKH). Iscience, Vol. 26, 4 (2023)."},{"key":"e_1_3_2_1_33_1","doi-asserted-by":"publisher","DOI":"10.1155\/2018\/4302425"},{"key":"e_1_3_2_1_34_1","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1145\/3611651","article-title":"b. Pre-trained language models in biomedical domain: A systematic survey","volume":"56","author":"Wang Benyou","year":"2023","unstructured":"Benyou Wang, Qianqian Xie, Jiahuan Pei, Zhihong Chen, Prayag Tiwari, Zhao Li, and Jie Fu. 2023 b. Pre-trained language models in biomedical domain: A systematic survey. Comput. Surveys, Vol. 56, 3 (2023), 1--52.","journal-title":"Comput. Surveys"},{"key":"e_1_3_2_1_35_1","volume-title":"2023 a. Exploring the in-context learning ability of large language model for biomedical concept linking. arXiv preprint arXiv:2307.01137","author":"Wang Qinyong","year":"2023","unstructured":"Qinyong Wang, Zhenxiang Gao, and Rong Xu. 2023 a. Exploring the in-context learning ability of large language model for biomedical concept linking. arXiv preprint arXiv:2307.01137 (2023)."},{"key":"e_1_3_2_1_36_1","volume-title":"Self-Consistency Improves Chain of Thought Reasoning in Language Models. In The Eleventh International Conference on Learning Representations.","author":"Wang Xuezhi","year":"2022","unstructured":"Xuezhi Wang, Jason Wei, Dale Schuurmans, Quoc V Le, Ed H Chi, Sharan Narang, Aakanksha Chowdhery, and Denny Zhou. 2022. Self-Consistency Improves Chain of Thought Reasoning in Language Models. In The Eleventh International Conference on Learning Representations."},{"key":"e_1_3_2_1_37_1","unstructured":"William E Winkler. 1990. String comparator metrics and enhanced decision rules in the Fellegi-Sunter model of record linkage. (1990)."},{"key":"e_1_3_2_1_38_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.acl-main.748"},{"key":"e_1_3_2_1_39_1","unstructured":"Ran Xu Yue Yu Chao Zhang Mohammed K Ali Joyce C Ho and Carl Yang. 2022. Counterfactual and factual reasoning over hypergraphs for interpretable clinical predictions on ehr. In Machine Learning for Health. PMLR 259--278."},{"key":"e_1_3_2_1_40_1","volume-title":"How language model hallucinations can snowball. arXiv preprint arXiv:2305.13534","author":"Zhang Muru","year":"2023","unstructured":"Muru Zhang, Ofir Press, William Merrill, Alisa Liu, and Noah A Smith. 2023. How language model hallucinations can snowball. arXiv preprint arXiv:2305.13534 (2023)."},{"key":"e_1_3_2_1_41_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2022.findings-emnlp.61"},{"key":"e_1_3_2_1_42_1","volume-title":"Graph neural networks: A review of methods and applications. AI open","author":"Zhou Jie","year":"2020","unstructured":"Jie Zhou, Ganqu Cui, Shengding Hu, Zhengyan Zhang, Cheng Yang, Zhiyuan Liu, Lifeng Wang, Changcheng Li, and Maosong Sun. 2020. Graph neural networks: A review of methods and applications. AI open , Vol. 1 (2020), 57--81."},{"key":"e_1_3_2_1_43_1","volume-title":"Ziwen Han, Keiran Paster, Silviu Pitis, Harris Chan, and Jimmy Ba.","author":"Zhou Yongchao","year":"2023","unstructured":"Yongchao Zhou, Andrei Ioan Muresanu, Ziwen Han, Keiran Paster, Silviu Pitis, Harris Chan, and Jimmy Ba. 2023. Large Language Models Are Human-Level Prompt Engineers.(2023). ProQuest Number: INFORMATION TO ALL USERS, Vol. 30490868 (2023)."}],"event":{"name":"SIGIR 2024: The 47th International ACM SIGIR Conference on Research and Development in Information Retrieval","sponsor":["SIGIR ACM Special Interest Group on Information Retrieval"],"location":"Washington DC USA","acronym":"SIGIR 2024"},"container-title":["Proceedings of the 47th International ACM SIGIR Conference on Research and Development in Information Retrieval"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3626772.3657904","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,10,18]],"date-time":"2024-10-18T05:25:14Z","timestamp":1729229114000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3626772.3657904"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,7,10]]},"references-count":43,"alternative-id":["10.1145\/3626772.3657904","10.1145\/3626772"],"URL":"https:\/\/doi.org\/10.1145\/3626772.3657904","relation":{},"subject":[],"published":{"date-parts":[[2024,7,10]]},"assertion":[{"value":"2024-07-11","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}