{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2023,11,2]],"date-time":"2023-11-02T00:43:59Z","timestamp":1698885839743},"reference-count":0,"publisher":"Association for the Advancement of Artificial Intelligence (AAAI)","issue":"11","content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["AAAI"],"abstract":"In dialogue state tracking (DST), the exploitation of dialogue history is a crucial research direction, and the existing DST models can be divided into two categories: full-history models and partial-history models. Since the \u201cselect first, use later\u201d mechanism explicitly filters the distracting information being passed to the downstream state prediction, the partial-history models have recently achieved a performance advantage over the full-history models. However, besides the redundant information, some critical dialogue context information was inevitably filtered out by the partial-history models simultaneously. To reconcile the contextual consideration with avoiding the introduction of redundant information, we propose DICE-DST, a model-agnostic module widely applicable to the partial-history DST models, which aims to strengthen the ability of context exploitation for the encoder of each DST model. Specifically, we first construct a teacher encoder and devise two contextual reasoning tasks to train it to acquire extensive dialogue contextual knowledge. Then we transfer the contextual knowledge from the teacher encoder to the student encoder via a novel turn-level attention-alignment distillation. Experimental results show that our approach extensively improves the performance of partial-history DST models and thereby achieves new state-of-the-art performance on multiple mainstream datasets while keeping high efficiency.<\/jats:p>","DOI":"10.1609\/aaai.v37i11.26510","type":"journal-article","created":{"date-parts":[[2023,6,27]],"date-time":"2023-06-27T18:08:29Z","timestamp":1687889309000},"page":"12845-12853","source":"Crossref","is-referenced-by-count":1,"title":["Learning to Imagine: Distillation-Based Interactive Context Exploitation for Dialogue State Tracking"],"prefix":"10.1609","volume":"37","author":[{"given":"Jinyu","family":"Guo","sequence":"first","affiliation":[]},{"given":"Kai","family":"Shuang","sequence":"additional","affiliation":[]},{"given":"Kaihang","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Yixuan","family":"Liu","sequence":"additional","affiliation":[]},{"given":"Jijie","family":"Li","sequence":"additional","affiliation":[]},{"given":"Zihan","family":"Wang","sequence":"additional","affiliation":[]}],"member":"9382","published-online":{"date-parts":[[2023,6,26]]},"container-title":["Proceedings of the AAAI Conference on Artificial Intelligence"],"original-title":[],"link":[{"URL":"https:\/\/ojs.aaai.org\/index.php\/AAAI\/article\/download\/26510\/26282","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/ojs.aaai.org\/index.php\/AAAI\/article\/download\/26510\/26282","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,6,27]],"date-time":"2023-06-27T18:08:29Z","timestamp":1687889309000},"score":1,"resource":{"primary":{"URL":"https:\/\/ojs.aaai.org\/index.php\/AAAI\/article\/view\/26510"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,6,26]]},"references-count":0,"journal-issue":{"issue":"11","published-online":{"date-parts":[[2023,6,27]]}},"URL":"https:\/\/doi.org\/10.1609\/aaai.v37i11.26510","relation":{},"ISSN":["2374-3468","2159-5399"],"issn-type":[{"value":"2374-3468","type":"electronic"},{"value":"2159-5399","type":"print"}],"subject":[],"published":{"date-parts":[[2023,6,26]]}}}