{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,26]],"date-time":"2025-03-26T18:24:11Z","timestamp":1743013451670},"publisher-location":"Stroudsburg, PA, USA","reference-count":0,"publisher":"Association for Computational Linguistics","content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2023]]},"DOI":"10.18653\/v1\/2023.acl-long.869","type":"proceedings-article","created":{"date-parts":[[2023,8,5]],"date-time":"2023-08-05T00:57:42Z","timestamp":1691197062000},"source":"Crossref","is-referenced-by-count":9,"title":["Pre-trained Language Models Can be Fully Zero-Shot Learners"],"prefix":"10.18653","author":[{"given":"Xuandong","family":"Zhao","sequence":"first","affiliation":[]},{"given":"Siqi","family":"Ouyang","sequence":"additional","affiliation":[]},{"given":"Zhiguo","family":"Yu","sequence":"additional","affiliation":[]},{"given":"Ming","family":"Wu","sequence":"additional","affiliation":[]},{"given":"Lei","family":"Li","sequence":"additional","affiliation":[]}],"member":"1643","event":{"name":"Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)","location":"Toronto, Canada","start":{"date-parts":[[2023,7]]},"end":{"date-parts":[[2023,7]]}},"container-title":["Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)"],"original-title":[],"deposited":{"date-parts":[[2023,8,5]],"date-time":"2023-08-05T01:21:26Z","timestamp":1691198486000},"score":1,"resource":{"primary":{"URL":"https:\/\/aclanthology.org\/2023.acl-long.869"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023]]},"references-count":0,"URL":"https:\/\/doi.org\/10.18653\/v1\/2023.acl-long.869","relation":{},"subject":[],"published":{"date-parts":[[2023]]}}}