{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,2,21]],"date-time":"2025-02-21T14:25:27Z","timestamp":1740147927832,"version":"3.37.3"},"reference-count":32,"publisher":"Institute of Electrical and Electronics Engineers (IEEE)","issue":"3","license":[{"start":{"date-parts":[[2024,4,1]],"date-time":"2024-04-01T00:00:00Z","timestamp":1711929600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2024,4,1]],"date-time":"2024-04-01T00:00:00Z","timestamp":1711929600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2024,4,1]],"date-time":"2024-04-01T00:00:00Z","timestamp":1711929600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62371411"],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"name":"Research Grants Council of the Hong Kong SAR","award":["GRF 11217823"]},{"name":"InnoHK initiative, the Government of the HKSAR"},{"name":"Laboratory for AI-Powered Financial Technologies"},{"name":"Meituan Robotics Research Institute"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEEE J. Sel. Top. Signal Process."],"published-print":{"date-parts":[[2024,4]]},"DOI":"10.1109\/jstsp.2024.3376962","type":"journal-article","created":{"date-parts":[[2024,4,3]],"date-time":"2024-04-03T17:43:26Z","timestamp":1712166206000},"page":"529-539","source":"Crossref","is-referenced-by-count":1,"title":["A Language Model-Based Fine-Grained Address Resolution Framework in UAV Delivery System"],"prefix":"10.1109","volume":"18","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-8753-9137","authenticated-orcid":false,"given":"Sichun","family":"Luo","sequence":"first","affiliation":[{"name":"Department of Computer Science, City University of Hong Kong, Kowloon, Hong Kong"}]},{"ORCID":"https:\/\/orcid.org\/0009-0009-3955-7272","authenticated-orcid":false,"given":"Yuxuan","family":"Yao","sequence":"additional","affiliation":[{"name":"Department of Computer Science, City University of Hong Kong, Kowloon, Hong Kong"}]},{"ORCID":"https:\/\/orcid.org\/0009-0003-6084-1522","authenticated-orcid":false,"given":"Haohan","family":"Zhao","sequence":"additional","affiliation":[{"name":"Department of Computer Science, City University of Hong Kong, Kowloon, Hong Kong"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-2756-4984","authenticated-orcid":false,"given":"Linqi","family":"Song","sequence":"additional","affiliation":[{"name":"Department of Computer Science, City University of Hong Kong, Kowloon, Hong Kong"}]}],"member":"263","reference":[{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1038\/nature14542"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.3390\/rs9050459"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1002\/9781118396780"},{"year":"2021","key":"ref4","article-title":"Food delivery giant meituan unveils drones for delivery service, offering new user experience"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.3390\/ijgi7030091"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1002\/rob.22231"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1109\/GLOBECOM48099.2022.10001285"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1109\/TrustCom53373.2021.00142"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1007\/s10708-008-9181-5"},{"key":"ref10","first-page":"4171","article-title":"BERT: Pre-training of deep bidirectional transformers for language understanding","volume-title":"Proc. NAACL-HLT (1)","author":"Devlin","year":"2019"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2019.2909530"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-72322-4_196-1"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.1016\/j.asoc.2022.109225"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.1109\/IROS.2016.7759739"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.1051\/matecconf\/201815104011"},{"key":"ref16","first-page":"2145","article-title":"A survey on recent advances in named entity recognition from deep learning models","volume-title":"Proc. COLING","author":"Yadav","year":"2018"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1109\/TKDE.2020.2981314"},{"issue":"2","key":"ref18","first-page":"3","article-title":"Conditional random fields: Probabilistic models for segmenting and labeling sequence data","volume-title":"Proc. Icml","volume":"1","author":"Lafferty","year":"2001"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.1023\/A:1007558221122"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.1162\/neco.1997.9.8.1735"},{"article-title":"Bidirectional LSTM-CRF models for sequence tagging","year":"2015","author":"Huang","key":"ref21"},{"article-title":"Improving language understanding by generative pre-training","year":"2018","author":"Radford","key":"ref22"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.48550\/ARXIV.1706.03762"},{"issue":"5","key":"ref24","first-page":"547","article-title":"Comparative study of floral distribution in a portion of the alps and jura","volume":"37","author":"Jaccard","year":"1901","journal-title":"Company Vaudoise Bull. Natural Sci."},{"key":"ref25","article-title":"ELECTRA: Pre-training text encoders as discriminators rather than generators","volume-title":"Proc. ICLR","author":"Clark","year":"2020"},{"article-title":"Roberta: A robustly optimized bert pretraining approach","year":"2019","author":"Liu","key":"ref26"},{"key":"ref27","first-page":"5754","article-title":"XLNet: Generalized autoregressive pretraining for language understanding","volume-title":"Proc. NeurIPS","author":"Yang","year":"2019"},{"key":"ref28","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/P19-1139"},{"article-title":"LERT: A linguistically-motivated pre-trained language model","year":"2022","author":"Cui","key":"ref29"},{"article-title":"PERT: Pre-training BERT with permuted language model","year":"2022","author":"Cui","key":"ref30"},{"key":"ref31","first-page":"29","article-title":"L3Cube-MahaNER: A. marathi named entity recognition dataset and bert models","volume-title":"Proc. WILDRE-6 Workshop 13th Lang. Resour. Eval. Conf.","author":"Litake","year":"2022"},{"article-title":"A survey of large language models","year":"2023","author":"Zhao","key":"ref32"}],"container-title":["IEEE Journal of Selected Topics in Signal Processing"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/4200690\/10665905\/10490097.pdf?arnumber=10490097","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,2,6]],"date-time":"2025-02-06T18:42:39Z","timestamp":1738867359000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/10490097\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,4]]},"references-count":32,"journal-issue":{"issue":"3"},"URL":"https:\/\/doi.org\/10.1109\/jstsp.2024.3376962","relation":{},"ISSN":["1932-4553","1941-0484"],"issn-type":[{"type":"print","value":"1932-4553"},{"type":"electronic","value":"1941-0484"}],"subject":[],"published":{"date-parts":[[2024,4]]}}}