{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2024,10,30]],"date-time":"2024-10-30T05:23:47Z","timestamp":1730265827103,"version":"3.28.0"},"reference-count":33,"publisher":"IEEE","license":[{"start":{"date-parts":[[2019,7,1]],"date-time":"2019-07-01T00:00:00Z","timestamp":1561939200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2019,7,1]],"date-time":"2019-07-01T00:00:00Z","timestamp":1561939200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2019,7,1]],"date-time":"2019-07-01T00:00:00Z","timestamp":1561939200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2019,7]]},"DOI":"10.1109\/ijcnn.2019.8852111","type":"proceedings-article","created":{"date-parts":[[2019,10,1]],"date-time":"2019-10-01T03:44:32Z","timestamp":1569901472000},"page":"1-8","source":"Crossref","is-referenced-by-count":2,"title":["Cropout: A General Mechanism for Reducing Overfitting on Convolutional Neural Networks"],"prefix":"10.1109","author":[{"given":"Wenbo","family":"Hou","sequence":"first","affiliation":[]},{"given":"Wenhai","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Ruo-Ze","family":"Liu","sequence":"additional","affiliation":[]},{"given":"Tong","family":"Lu","sequence":"additional","affiliation":[]}],"member":"263","reference":[{"journal-title":"arXiv preprint arXiv 1707 01083","article-title":"Shufflenet: An extremely efficient convolutional neural network for mobile devices","year":"2017","author":"zhang","key":"ref33"},{"key":"ref32","doi-asserted-by":"publisher","DOI":"10.5244\/C.30.87"},{"journal-title":"arXiv preprint arXiv 1705 02887","article-title":"Generative cooperative net for image generation and data augmentation","year":"2017","author":"xu","key":"ref31"},{"journal-title":"arXiv preprint arXiv 1611 05431","article-title":"Aggregated residual transformations for deep neural networks","year":"2016","author":"xie","key":"ref30"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-46493-0_39"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.496"},{"key":"ref12","first-page":"448","article-title":"Batch normalization: Accelerating deep network training by reducing internal covariate shift","author":"ioffe","year":"2015","journal-title":"International Conference on Machine Learning"},{"journal-title":"Learning multiple layers of features from tiny images","year":"2009","author":"krizhevsky","key":"ref13"},{"key":"ref14","first-page":"1097","article-title":"Imagenet classification with deep convolutional neural networks","author":"krizhevsky","year":"2012","journal-title":"Advances in neural information processing systems"},{"journal-title":"arXiv preprint arXiv 1605 07648","article-title":"Fractalnet: Ultra-deep neural networks without residuals","year":"2016","author":"larsson","key":"ref15"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.1162\/neco.1989.1.4.541"},{"key":"ref17","first-page":"562","article-title":"Deeply-supervised nets","author":"lee","year":"2015","journal-title":"Artificial Intelligence and Statistics"},{"journal-title":"arXiv preprint arXiv 1312 4400","article-title":"Network in network","year":"2013","author":"lin","key":"ref18"},{"key":"ref19","first-page":"2234","article-title":"Improved techniques for training gans","author":"salimans","year":"2016","journal-title":"Advances in neural information processing systems"},{"key":"ref28","doi-asserted-by":"publisher","DOI":"10.24963\/ijcai.2018\/391"},{"journal-title":"Computer Science","article-title":"High-performance neural networks for visual object classification","year":"2011","author":"cirean","key":"ref4"},{"key":"ref27","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2015.7298594"},{"journal-title":"arXiv preprint arXiv 1610 02357","article-title":"Xception: Deep learning with depthwise separable convolutions","year":"2016","author":"chollet","key":"ref3"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.90"},{"key":"ref29","article-title":"Deep image: Scaling up image recognition","volume":"7","author":"wu","year":"2015","journal-title":"arXiv preprint arXiv 1501 02876"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2009.5206848"},{"journal-title":"arXiv preprint arX-iv 1704 04861","article-title":"Mobilenets: Efficient convolutional neural networks for mobile vision applications","year":"2017","author":"howard","key":"ref8"},{"key":"ref7","first-page":"630","article-title":"Identity mappings in deep residual networks","author":"he","year":"2016","journal-title":"European Conference on Computer Vision"},{"journal-title":"arXiv preprint arXiv 1707 01629","article-title":"Dual path networks","year":"2017","author":"chen","key":"ref2"},{"journal-title":"arXiv preprint arXiv 1608 06993","article-title":"Densely connected convolutional networks","year":"2016","author":"huang","key":"ref9"},{"journal-title":"arXiv preprint arXiv 1607 06450","article-title":"Layer normalization","year":"2016","author":"ba","key":"ref1"},{"key":"ref20","first-page":"901","article-title":"Weight normalization: A simple reparameterization to accelerate training of deep neural networks","author":"salimans","year":"2016","journal-title":"Advances in neural information processing systems"},{"journal-title":"arXiv preprint arXiv 1612 07828","article-title":"Learning from simulated and unsupervised images through adversarial training","year":"2016","author":"shrivastava","key":"ref22"},{"key":"ref21","first-page":"2534","article-title":"Patch reordering: A novelway to achieve rotation and translation invariance in convolutional neural networks","author":"shen","year":"2017","journal-title":"AAAI"},{"journal-title":"arXiv preprint arX-iv 1412 6806","article-title":"Striving for simplicity: The all convolutional net","year":"2014","author":"springenberg","key":"ref24"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.1109\/ICDAR.2003.1227801"},{"key":"ref26","first-page":"2377","article-title":"Training very deep networks","author":"srivastava","year":"2015","journal-title":"Advances in neural information processing systems"},{"key":"ref25","first-page":"1929","article-title":"Dropout: a simple way to prevent neural networks from overfitting","volume":"15","author":"srivastava","year":"2014","journal-title":"Journal of Machine Learning Research"}],"event":{"name":"2019 International Joint Conference on Neural Networks (IJCNN)","start":{"date-parts":[[2019,7,14]]},"location":"Budapest, Hungary","end":{"date-parts":[[2019,7,19]]}},"container-title":["2019 International Joint Conference on Neural Networks (IJCNN)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/8840768\/8851681\/08852111.pdf?arnumber=8852111","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,7,17]],"date-time":"2022-07-17T21:50:59Z","timestamp":1658094659000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/8852111\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2019,7]]},"references-count":33,"URL":"https:\/\/doi.org\/10.1109\/ijcnn.2019.8852111","relation":{},"subject":[],"published":{"date-parts":[[2019,7]]}}}