{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2024,9,8]],"date-time":"2024-09-08T04:50:18Z","timestamp":1725771018513},"reference-count":27,"publisher":"IEEE","license":[{"start":{"date-parts":[[2022,6,1]],"date-time":"2022-06-01T00:00:00Z","timestamp":1654041600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2022,6,1]],"date-time":"2022-06-01T00:00:00Z","timestamp":1654041600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2022,6]]},"DOI":"10.1109\/cvprw56347.2022.00382","type":"proceedings-article","created":{"date-parts":[[2022,8,23]],"date-time":"2022-08-23T15:52:53Z","timestamp":1661269973000},"page":"3386-3394","source":"Crossref","is-referenced-by-count":5,"title":["Does Federated Dropout actually work?"],"prefix":"10.1109","author":[{"given":"Gary","family":"Cheng","sequence":"first","affiliation":[{"name":"Stanford University"}]},{"given":"Zachary","family":"Charles","sequence":"additional","affiliation":[{"name":"Google Research"}]},{"given":"Zachary","family":"Garrett","sequence":"additional","affiliation":[{"name":"Google Research"}]},{"given":"Keith","family":"Rush","sequence":"additional","affiliation":[{"name":"Google Research"}]}],"member":"263","reference":[{"key":"ref10","article-title":"Fedboost: A communication-efficient algorithm for federated learning","author":"hamer","year":"2020","journal-title":"ICML"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-44781-0_9"},{"article-title":"Federated learning for mobile keyboard prediction","year":"2018","author":"hard","key":"ref12"},{"article-title":"Fjord: Fair and accurate federated learning under heterogeneous targets with ordered dropout","year":"2021","author":"horvath","key":"ref13"},{"article-title":"Advances and open problems in federated learning","year":"2019","author":"kairouz","key":"ref14"},{"article-title":"Scaling laws for neural language models","year":"2020","author":"kaplan","key":"ref15"},{"article-title":"Federated learning: Strategies for improving communication efficiency","year":"2016","author":"konecn\u00fd","key":"ref16"},{"article-title":"Think locally, act globally: Federated learning with local and global representations","year":"2020","author":"liang","key":"ref17"},{"key":"ref18","article-title":"Communication-efficient learning of deep networks from decentralized data","author":"mcmahan","year":"2017","journal-title":"Proceedings of the 20th International Conference on Artificial Intelligence and Statistics"},{"key":"ref19","article-title":"Adaptive federated optimization","author":"reddi","year":"2021","journal-title":"Proceedings of International Conference on Learning Representations"},{"article-title":"Leaf: A benchmark for federated settings","year":"2018","author":"caldas","key":"ref4"},{"key":"ref27","article-title":"Is local SGD better than minibatch SGD?","author":"woodworth","year":"2020","journal-title":"Proceedings of the 37th International Conference on Machine Learning"},{"article-title":"Expanding the reach of federated learning by reducing client resource requirements","year":"2018","author":"caldas","key":"ref3"},{"article-title":"Exploiting shared representations for personalized federated learning","year":"2021","author":"collins","key":"ref6"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1145\/3369583.3392686"},{"article-title":"Enabling on-device training of speech recognition models with federated dropout","year":"2021","author":"guliani","key":"ref8"},{"article-title":"Dropout as a bayesian approximation: Representing model uncertainty in deep learning","year":"2016","author":"gal","key":"ref7"},{"article-title":"Language models are few-shot learners","year":"2020","author":"brown","key":"ref2"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1016\/j.jnca.2018.05.003"},{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1109\/INFOCOMWKSHPS51825.2021.9484526"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.1002\/widm.1249"},{"article-title":"Efficient and private federated learning with partially trainable networks","year":"2021","author":"sidahmed","key":"ref22"},{"article-title":"Fed-ensemble: Improving generalization through model ensembling in federated learning","year":"2021","author":"shi","key":"ref21"},{"key":"ref24","first-page":"1929","article-title":"Dropout: a simple way to prevent neural networks from overfitting","author":"srivastava","year":"2014","journal-title":"J Mach Learn Res"},{"article-title":"Federated reconstruction: Partially local federated learning","year":"2021","author":"singhal","key":"ref23"},{"article-title":"Splitfed: When federated learning meets split learning","year":"2020","author":"thapa","key":"ref26"},{"article-title":"Measuring robustness to natural distribution shifts in image classification","year":"2020","author":"taori","key":"ref25"}],"event":{"name":"2022 IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)","start":{"date-parts":[[2022,6,19]]},"location":"New Orleans, LA, USA","end":{"date-parts":[[2022,6,20]]}},"container-title":["2022 IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/9856930\/9856648\/09857371.pdf?arnumber=9857371","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,9,12]],"date-time":"2022-09-12T16:06:52Z","timestamp":1662998812000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/9857371\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,6]]},"references-count":27,"URL":"https:\/\/doi.org\/10.1109\/cvprw56347.2022.00382","relation":{},"subject":[],"published":{"date-parts":[[2022,6]]}}}