{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,2,21]],"date-time":"2025-02-21T01:39:00Z","timestamp":1740101940891,"version":"3.37.3"},"publisher-location":"New York, NY, USA","reference-count":13,"publisher":"ACM","funder":[{"name":"Fonds de recherche du Qu\u00e9bec \u00f0 Nature et technologies","award":["304302"]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2022,12,16]]},"DOI":"10.1145\/3582768.3582776","type":"proceedings-article","created":{"date-parts":[[2023,6,27]],"date-time":"2023-06-27T23:48:32Z","timestamp":1687909712000},"page":"16-20","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":0,"title":["Preventing RNN from Using Sequence Length as a Feature"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-6921-431X","authenticated-orcid":false,"given":"Jean-Thomas","family":"Baillargeon","sequence":"first","affiliation":[{"name":"Computer Science and Software Engineering, University; Laval, Canada"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-2111-7545","authenticated-orcid":false,"given":"Helene","family":"Cossette","sequence":"additional","affiliation":[{"name":"School of Actuarial Sciences, University; Laval, Canada"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-0255-5117","authenticated-orcid":false,"given":"Luc","family":"Lamontagne","sequence":"additional","affiliation":[{"name":"Computer Science and Software Engineering, University; Laval, Canada"}]}],"member":"320","published-online":{"date-parts":[[2023,6,27]]},"reference":[{"key":"e_1_3_2_1_1_1","volume-title":"Longformer: The long-document transformer. arXiv preprint arXiv:2004.05150(2020).","author":"Beltagy Iz","year":"2020","unstructured":"Iz Beltagy , Matthew\u00a0 E Peters , and Arman Cohan . 2020 . Longformer: The long-document transformer. arXiv preprint arXiv:2004.05150(2020). Iz Beltagy, Matthew\u00a0E Peters, and Arman Cohan. 2020. Longformer: The long-document transformer. arXiv preprint arXiv:2004.05150(2020)."},{"key":"e_1_3_2_1_2_1","unstructured":"Mengting Fang Zhenglong Zhou Sharon Chen and Jay McClelland. 2018. Can a recurrent neural network learn to count things?. In CogSci. 360\u2013365. Mengting Fang Zhenglong Zhou Sharon Chen and Jay McClelland. 2018. Can a recurrent neural network learn to count things?. In CogSci. 360\u2013365."},{"key":"e_1_3_2_1_3_1","doi-asserted-by":"publisher","DOI":"10.1109\/IJCNN.2000.861302"},{"key":"e_1_3_2_1_4_1","volume-title":"Long short-term memory. Neural computation 9, 8","author":"Hochreiter Sepp","year":"1997","unstructured":"Sepp Hochreiter and J\u00fcrgen Schmidhuber . 1997. Long short-term memory. Neural computation 9, 8 ( 1997 ), 1735\u20131780. Sepp Hochreiter and J\u00fcrgen Schmidhuber. 1997. Long short-term memory. Neural computation 9, 8 (1997), 1735\u20131780."},{"key":"e_1_3_2_1_5_1","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2020.2992013"},{"key":"e_1_3_2_1_6_1","unstructured":"Stephen Merity Nitish\u00a0Shirish Keskar and Richard Socher. 2017. Regularizing and optimizing LSTM language models. arXiv preprint arXiv:1708.02182(2017). Stephen Merity Nitish\u00a0Shirish Keskar and Richard Socher. 2017. Regularizing and optimizing LSTM language models. arXiv preprint arXiv:1708.02182(2017)."},{"key":"e_1_3_2_1_7_1","volume-title":"International conference on machine learning. PMLR, 1310\u20131318","author":"Pascanu Razvan","year":"2013","unstructured":"Razvan Pascanu , Tomas Mikolov , and Yoshua Bengio . 2013 . On the difficulty of training recurrent neural networks . In International conference on machine learning. PMLR, 1310\u20131318 . Razvan Pascanu, Tomas Mikolov, and Yoshua Bengio. 2013. On the difficulty of training recurrent neural networks. In International conference on machine learning. PMLR, 1310\u20131318."},{"key":"e_1_3_2_1_8_1","doi-asserted-by":"publisher","DOI":"10.1080\/095400999116340"},{"key":"e_1_3_2_1_9_1","volume-title":"Visualizing data using t-SNE.Journal of machine learning research 9, 11","author":"Maaten Laurens Van\u00a0der","year":"2008","unstructured":"Laurens Van\u00a0der Maaten and Geoffrey Hinton . 2008. Visualizing data using t-SNE.Journal of machine learning research 9, 11 ( 2008 ). Laurens Van\u00a0der Maaten and Geoffrey Hinton. 2008. Visualizing data using t-SNE.Journal of machine learning research 9, 11 (2008)."},{"key":"e_1_3_2_1_10_1","doi-asserted-by":"crossref","unstructured":"Du\u0161an Vari\u0161 and Ond\u0159ej Bojar. 2021. Sequence length is a domain: Length-based overfitting in transformer models. arXiv preprint arXiv:2109.07276(2021). Du\u0161an Vari\u0161 and Ond\u0159ej Bojar. 2021. Sequence length is a domain: Length-based overfitting in transformer models. arXiv preprint arXiv:2109.07276(2021).","DOI":"10.18653\/v1\/2021.emnlp-main.650"},{"key":"e_1_3_2_1_11_1","volume-title":"Attention is all you need. Advances in neural information processing systems 30","author":"Vaswani Ashish","year":"2017","unstructured":"Ashish Vaswani , Noam Shazeer , Niki Parmar , Jakob Uszkoreit , Llion Jones , Aidan\u00a0 N Gomez , \u0141ukasz Kaiser , and Illia Polosukhin . 2017. Attention is all you need. Advances in neural information processing systems 30 ( 2017 ). Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan\u00a0N Gomez, \u0141ukasz Kaiser, and Illia Polosukhin. 2017. Attention is all you need. Advances in neural information processing systems 30 (2017)."},{"key":"e_1_3_2_1_12_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/N16-1174"},{"key":"e_1_3_2_1_13_1","volume-title":"Character-level convolutional networks for text classification. Advances in neural information processing systems 28","author":"Zhang Xiang","year":"2015","unstructured":"Xiang Zhang , Junbo Zhao , and Yann LeCun . 2015. Character-level convolutional networks for text classification. Advances in neural information processing systems 28 ( 2015 ). Xiang Zhang, Junbo Zhao, and Yann LeCun. 2015. Character-level convolutional networks for text classification. Advances in neural information processing systems 28 (2015)."}],"event":{"name":"NLPIR 2022: 2022 6th International Conference on Natural Language Processing and Information Retrieval","acronym":"NLPIR 2022","location":"Bangkok Thailand"},"container-title":["Proceedings of the 2022 6th International Conference on Natural Language Processing and Information Retrieval"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3582768.3582776","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,12,16]],"date-time":"2023-12-16T11:11:19Z","timestamp":1702725079000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3582768.3582776"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,12,16]]},"references-count":13,"alternative-id":["10.1145\/3582768.3582776","10.1145\/3582768"],"URL":"https:\/\/doi.org\/10.1145\/3582768.3582776","relation":{},"subject":[],"published":{"date-parts":[[2022,12,16]]},"assertion":[{"value":"2023-06-27","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}