{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2024,9,8]],"date-time":"2024-09-08T15:29:37Z","timestamp":1725809377224},"publisher-location":"New York, NY, USA","reference-count":21,"publisher":"ACM","license":[{"start":{"date-parts":[[2019,6,2]],"date-time":"2019-06-02T00:00:00Z","timestamp":1559433600000},"content-version":"vor","delay-in-days":0,"URL":"http:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2019,6,2]]},"DOI":"10.1145\/3316781.3317783","type":"proceedings-article","created":{"date-parts":[[2019,5,23]],"date-time":"2019-05-23T18:07:13Z","timestamp":1558634833000},"page":"1-6","update-policy":"http:\/\/dx.doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":18,"title":["BiScaled-DNN"],"prefix":"10.1145","author":[{"given":"Shubham","family":"Jain","sequence":"first","affiliation":[{"name":"School of Electrical and Computer Engineering, Purdue University"}]},{"given":"Swagath","family":"Venkataramani","sequence":"additional","affiliation":[{"name":"IBM T.J. Watson Research Center, Yorktown Heights, NY"}]},{"given":"Vijayalakshmi","family":"Srinivasan","sequence":"additional","affiliation":[{"name":"IBM T.J. Watson Research Center, Yorktown Heights, NY"}]},{"given":"Jungwook","family":"Choi","sequence":"additional","affiliation":[{"name":"IBM T.J. Watson Research Center, Yorktown Heights, NY"}]},{"given":"Kailash","family":"Gopalakrishnan","sequence":"additional","affiliation":[{"name":"IBM T.J. Watson Research Center, Yorktown Heights, NY"}]},{"given":"Leland","family":"Chang","sequence":"additional","affiliation":[{"name":"IBM T.J. Watson Research Center, Yorktown Heights, NY"}]}],"member":"320","published-online":{"date-parts":[[2019,6,2]]},"reference":[{"volume-title":"Neural Networks with Few Multiplications. CoRR, abs","year":"2015","author":"Lin Z","key":"e_1_3_2_1_1_1","unstructured":"Z Lin Neural Networks with Few Multiplications. CoRR, abs , 2015 . Z Lin et al. Neural Networks with Few Multiplications. CoRR, abs, 2015."},{"volume-title":"Ternary Weight Networks. CoRR, abs\/1605.04711","year":"2016","author":"Li F.","key":"e_1_3_2_1_2_1","unstructured":"F. Li and B. Liu . Ternary Weight Networks. CoRR, abs\/1605.04711 , 2016 . F. Li and B. Liu. Ternary Weight Networks. CoRR, abs\/1605.04711, 2016."},{"volume-title":"DoReFa-Net: Training Low Bitwidth Convolutional Neural Networks with Low Bitwidth Gradients. CoRR, abs\/1606.06160","year":"2016","author":"Zhou S","key":"e_1_3_2_1_3_1","unstructured":"S Zhou DoReFa-Net: Training Low Bitwidth Convolutional Neural Networks with Low Bitwidth Gradients. CoRR, abs\/1606.06160 , 2016 . S Zhou et al. DoReFa-Net: Training Low Bitwidth Convolutional Neural Networks with Low Bitwidth Gradients. CoRR, abs\/1606.06160, 2016."},{"key":"e_1_3_2_1_4_1","doi-asserted-by":"publisher","DOI":"10.1145\/2627369.2627613"},{"volume-title":"PACT: parameterized clipping activation for quantized neural networks. CoRR, abs\/1805.06085","year":"2018","author":"J.","key":"e_1_3_2_1_5_1","unstructured":"J. Choi et. al. PACT: parameterized clipping activation for quantized neural networks. CoRR, abs\/1805.06085 , 2018 . J. Choi et. al. PACT: parameterized clipping activation for quantized neural networks. CoRR, abs\/1805.06085, 2018."},{"volume-title":"WRPN: wide reduced-precision networks. CoRR, abs\/1709.01134","year":"2017","author":"Mishra K.","key":"e_1_3_2_1_6_1","unstructured":"Asit K. Mishra , Eriko Nurvitadhi, Jeffrey J. Cook , and Debbie Marr. WRPN: wide reduced-precision networks. CoRR, abs\/1709.01134 , 2017 . Asit K. Mishra, Eriko Nurvitadhi, Jeffrey J. Cook, and Debbie Marr. WRPN: wide reduced-precision networks. CoRR, abs\/1709.01134, 2017."},{"key":"e_1_3_2_1_7_1","doi-asserted-by":"publisher","DOI":"10.1145\/3061639.3062259"},{"key":"e_1_3_2_1_8_1","doi-asserted-by":"publisher","DOI":"10.1109\/ISCA.2018.00063"},{"key":"e_1_3_2_1_9_1","doi-asserted-by":"publisher","DOI":"10.1145\/3195970.3196012"},{"key":"e_1_3_2_1_10_1","doi-asserted-by":"publisher","DOI":"10.1007\/s11263-015-0816-y"},{"volume-title":"BinaryNet: Training Deep Neural Networks with Weights and Activations Constrained to +1 or -1. CoRR, abs\/1602.02830","year":"2016","author":"Courbariaux M.","key":"e_1_3_2_1_11_1","unstructured":"M. Courbariaux BinaryNet: Training Deep Neural Networks with Weights and Activations Constrained to +1 or -1. CoRR, abs\/1602.02830 , 2016 . M. Courbariaux et al. BinaryNet: Training Deep Neural Networks with Weights and Activations Constrained to +1 or -1. CoRR, abs\/1602.02830, 2016."},{"volume-title":"XNOR-Net: ImageNet Classification Using Binary Convolutional Neural Networks. CoRR, abs\/1603.05279","year":"2016","author":"Rastegari M","key":"e_1_3_2_1_12_1","unstructured":"M Rastegari XNOR-Net: ImageNet Classification Using Binary Convolutional Neural Networks. CoRR, abs\/1603.05279 , 2016 . M Rastegari et al. XNOR-Net: ImageNet Classification Using Binary Convolutional Neural Networks. CoRR, abs\/1603.05279, 2016."},{"key":"e_1_3_2_1_13_1","doi-asserted-by":"publisher","DOI":"10.1145\/2925426.2926294"},{"key":"e_1_3_2_1_14_1","doi-asserted-by":"publisher","DOI":"10.5555\/3130379.3130725"},{"volume-title":"Deep compression: Compressing deep neural network with pruning, trained quantization and huffman coding. CoRR, abs","year":"2015","author":"Han S.","key":"e_1_3_2_1_15_1","unstructured":"S. Han Deep compression: Compressing deep neural network with pruning, trained quantization and huffman coding. CoRR, abs , 2015 . S. Han et al. Deep compression: Compressing deep neural network with pruning, trained quantization and huffman coding. CoRR, abs, 2015."},{"key":"e_1_3_2_1_16_1","doi-asserted-by":"publisher","DOI":"10.1145\/3079856.3080244"},{"key":"e_1_3_2_1_17_1","doi-asserted-by":"publisher","DOI":"10.1109\/ISCA.2016.32"},{"volume-title":"Proc. VLSI Symposium","year":"2018","author":"B.","key":"e_1_3_2_1_18_1","unstructured":"B. Fleischer et. al. A scalable multi-teraops deep learning processor core for ai training and inference . In Proc. VLSI Symposium , June 2018 . B. Fleischer et. al. A scalable multi-teraops deep learning processor core for ai training and inference. In Proc. VLSI Symposium, June 2018."},{"volume-title":"Sparce: Sparsity aware general purpose core extensions to accelerate deep neural networks. CoRR, abs\/1711.06315","year":"2017","author":"Sen S","key":"e_1_3_2_1_19_1","unstructured":"S Sen Sparce: Sparsity aware general purpose core extensions to accelerate deep neural networks. CoRR, abs\/1711.06315 , 2017 . S Sen et al. Sparce: Sparsity aware general purpose core extensions to accelerate deep neural networks. CoRR, abs\/1711.06315, 2017."},{"volume-title":"Caffe: Convolutional Architecture for Fast Feature Embedding. arXiv preprint arXiv:1408.5093","year":"2014","author":"Jia Y","key":"e_1_3_2_1_20_1","unstructured":"Y Jia Caffe: Convolutional Architecture for Fast Feature Embedding. arXiv preprint arXiv:1408.5093 , 2014 . Y Jia et al. Caffe: Convolutional Architecture for Fast Feature Embedding. arXiv preprint arXiv:1408.5093, 2014."},{"volume-title":"Proc. PACT","year":"2017","author":"Poster S.","key":"e_1_3_2_1_21_1","unstructured":"S. Venkataramani et. al. Poster : Design space exploration for performance optimization of deep neural networks on shared memory accelerators . In Proc. PACT , 2017 . S. Venkataramani et. al. Poster: Design space exploration for performance optimization of deep neural networks on shared memory accelerators. In Proc. PACT, 2017."}],"event":{"name":"DAC '19: The 56th Annual Design Automation Conference 2019","sponsor":["SIGDA ACM Special Interest Group on Design Automation","IEEE-CEDA","SIGBED ACM Special Interest Group on Embedded Systems"],"location":"Las Vegas NV USA","acronym":"DAC '19"},"container-title":["Proceedings of the 56th Annual Design Automation Conference 2019"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3316781.3317783","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,1,6]],"date-time":"2023-01-06T03:52:02Z","timestamp":1672977122000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3316781.3317783"}},"subtitle":["Quantizing Long-tailed Datastructures with Two Scale Factors for Deep Neural Networks"],"short-title":[],"issued":{"date-parts":[[2019,6,2]]},"references-count":21,"alternative-id":["10.1145\/3316781.3317783","10.1145\/3316781"],"URL":"https:\/\/doi.org\/10.1145\/3316781.3317783","relation":{},"subject":[],"published":{"date-parts":[[2019,6,2]]},"assertion":[{"value":"2019-06-02","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}