{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2024,10,30]],"date-time":"2024-10-30T20:39:11Z","timestamp":1730320751457,"version":"3.28.0"},"publisher-location":"New York, NY, USA","reference-count":33,"publisher":"ACM","license":[{"start":{"date-parts":[[2020,1,27]],"date-time":"2020-01-27T00:00:00Z","timestamp":1580083200000},"content-version":"vor","delay-in-days":365,"URL":"http:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"funder":[{"DOI":"10.13039\/100000001","name":"NSF","doi-asserted-by":"publisher","award":["CCF-1763299"],"id":[{"id":"10.13039\/100000001","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2019,1,27]]},"DOI":"10.1145\/3306618.3314287","type":"proceedings-article","created":{"date-parts":[[2019,7,10]],"date-time":"2019-07-10T08:10:59Z","timestamp":1562746259000},"page":"247-254","update-policy":"http:\/\/dx.doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":94,"title":["Multiaccuracy"],"prefix":"10.1145","author":[{"given":"Michael P.","family":"Kim","sequence":"first","affiliation":[{"name":"Stanford University, Stanford, CA, USA"}]},{"given":"Amirata","family":"Ghorbani","sequence":"additional","affiliation":[{"name":"Stanford University, Stanford, CA, USA"}]},{"given":"James","family":"Zou","sequence":"additional","affiliation":[{"name":"Stanford University, Stanford, CA, USA"}]}],"member":"320","published-online":{"date-parts":[[2019,1,27]]},"reference":[{"key":"e_1_3_2_1_1_1","volume-title":"Machine bias: There's software used across the country to predict future criminals. and it's biased against blacks. ProPublica","author":"Angwin J.","year":"2016","unstructured":"Angwin , J. , Larson , J. , Mattu , S. , and Kirchner , L . Machine bias: There's software used across the country to predict future criminals. and it's biased against blacks. ProPublica ( 2016 ). Angwin, J., Larson, J., Mattu, S., and Kirchner, L. Machine bias: There's software used across the country to predict future criminals. and it's biased against blacks. ProPublica (2016)."},{"key":"e_1_3_2_1_2_1","first-page":"77","volume-title":"Conference on Fairness, Accountability and Transparency","author":"Buolamwini J.","year":"2018","unstructured":"Buolamwini , J. , and Gebru , T . Gender shades: Intersectional accuracy disparities in commercial gender classification . In Conference on Fairness, Accountability and Transparency ( 2018 ), pp. 77 -- 91 . Buolamwini, J., and Gebru, T. Gender shades: Intersectional accuracy disparities in commercial gender classification. In Conference on Fairness, Accountability and Transparency (2018), pp. 77--91."},{"key":"e_1_3_2_1_3_1","volume-title":"Fair prediction with disparate impact: A study of bias in recidivism prediction instruments. Big Data","author":"Chouldechova A.","year":"2017","unstructured":"Chouldechova , A. Fair prediction with disparate impact: A study of bias in recidivism prediction instruments. Big Data ( 2017 ). Chouldechova, A. Fair prediction with disparate impact: A study of bias in recidivism prediction instruments. Big Data (2017)."},{"key":"e_1_3_2_1_4_1","volume-title":"Fairer and more accurate, but for whom? FATML","author":"Chouldechova A.","year":"2016","unstructured":"Chouldechova , A. , and G'Sell , M. Fairer and more accurate, but for whom? FATML ( 2016 ). Chouldechova, A., and G'Sell, M. Fairer and more accurate, but for whom? FATML (2016)."},{"key":"e_1_3_2_1_5_1","doi-asserted-by":"publisher","DOI":"10.1145\/3097983.3098095"},{"key":"e_1_3_2_1_6_1","doi-asserted-by":"publisher","DOI":"10.1126\/science.aaa9375"},{"key":"e_1_3_2_1_7_1","doi-asserted-by":"publisher","DOI":"10.1145\/2090236.2090255"},{"key":"e_1_3_2_1_8_1","volume-title":"Decoupled classifiers for fair and efficient machine learning. arXiv preprint arXiv:1707.06613","author":"Dwork C.","year":"2017","unstructured":"Dwork , C. , Immorlica , N. , Kalai , A. T. , and Leiserson , M . Decoupled classifiers for fair and efficient machine learning. arXiv preprint arXiv:1707.06613 ( 2017 ). Dwork, C., Immorlica, N., Kalai, A. T., and Leiserson, M. Decoupled classifiers for fair and efficient machine learning. arXiv preprint arXiv:1707.06613 (2017)."},{"key":"e_1_3_2_1_9_1","volume-title":"Proceedings of the First Symposium on Innovations in Computer Science'10","author":"Feldman V.","year":"2010","unstructured":"Feldman , V. Distribution-specific agnostic boosting . In Proceedings of the First Symposium on Innovations in Computer Science'10 ( 2010 ). Feldman, V. Distribution-specific agnostic boosting. In Proceedings of the First Symposium on Innovations in Computer Science'10 (2010)."},{"key":"e_1_3_2_1_10_1","volume-title":"Greedy function approximation: a gradient boosting machine. Annals of statistics","author":"Friedman J. H.","year":"2001","unstructured":"Friedman , J. H. Greedy function approximation: a gradient boosting machine. Annals of statistics ( 2001 ), 1189--1232. Friedman, J. H. Greedy function approximation: a gradient boosting machine. Annals of statistics (2001), 1189--1232."},{"key":"e_1_3_2_1_11_1","first-page":"3315","volume-title":"Advances in Neural Information Processing Systems","author":"Hardt M.","year":"2016","unstructured":"Hardt , M. , Price , E. , and Srebro , N . Equality of opportunity in supervised learning . In Advances in Neural Information Processing Systems ( 2016 ), pp. 3315 -- 3323 . Hardt, M., Price, E., and Srebro, N. Equality of opportunity in supervised learning. In Advances in Neural Information Processing Systems (2016), pp. 3315--3323."},{"key":"e_1_3_2_1_12_1","volume-title":"Fairness without demographics in repeated loss minimization. ICML","author":"Hashimoto T. B.","year":"2018","unstructured":"Hashimoto , T. B. , Srivastava , M. , Namkoong , H. , and Liang , P . Fairness without demographics in repeated loss minimization. ICML ( 2018 ). Hashimoto, T. B., Srivastava, M., Namkoong, H., and Liang, P. Fairness without demographics in repeated loss minimization. ICML (2018)."},{"key":"e_1_3_2_1_13_1","volume-title":"Calibration for the (computationally-identifiable) masses. ICML","author":"H\u00e9","year":"2018","unstructured":"H\u00e9 bert-Johnson, \u00da ., Kim , M. P. , Reingold , O. , and Rothblum , G. N . Calibration for the (computationally-identifiable) masses. ICML ( 2018 ). H\u00e9 bert-Johnson, \u00da ., Kim, M. P., Reingold, O., and Rothblum, G. N. Calibration for the (computationally-identifiable) masses. ICML (2018)."},{"key":"e_1_3_2_1_14_1","doi-asserted-by":"publisher","DOI":"10.1145\/1374376.1374466"},{"key":"e_1_3_2_1_15_1","doi-asserted-by":"publisher","DOI":"10.1145\/293347.293351"},{"key":"e_1_3_2_1_16_1","volume-title":"Preventing fairness gerrymandering: Auditing and learning for subgroup fairness. arXiv preprint arXiv:1711.05144v3","author":"Kearns M.","year":"2017","unstructured":"Kearns , M. , Neel , S. , Roth , A. , and Wu , Z. S . Preventing fairness gerrymandering: Auditing and learning for subgroup fairness. arXiv preprint arXiv:1711.05144v3 ( 2017 ). Kearns, M., Neel, S., Roth, A., and Wu, Z. S. Preventing fairness gerrymandering: Auditing and learning for subgroup fairness. arXiv preprint arXiv:1711.05144v3 (2017)."},{"key":"e_1_3_2_1_17_1","doi-asserted-by":"publisher","DOI":"10.1007\/BF00993468"},{"key":"e_1_3_2_1_18_1","volume-title":"Multiaccuracy: Black-box post-processing for fairness in classification (full version). arXiv preprint","author":"Kim M. P.","year":"1805","unstructured":"Kim , M. P. , Ghorbani , A. , and Zou , J . Multiaccuracy: Black-box post-processing for fairness in classification (full version). arXiv preprint 1805 .12317 (2018). Kim, M. P., Ghorbani, A., and Zou, J. Multiaccuracy: Black-box post-processing for fairness in classification (full version). arXiv preprint 1805.12317 (2018)."},{"key":"e_1_3_2_1_19_1","volume-title":"Fairness through computationally-bounded awareness. arXiv Preprint","author":"Kim M. P.","year":"1803","unstructured":"Kim , M. P. , Reingold , O. , and Rothblum , G. N . Fairness through computationally-bounded awareness. arXiv Preprint 1803 .03239 (2018). Kim, M. P., Reingold, O., and Rothblum, G. N. Fairness through computationally-bounded awareness. arXiv Preprint 1803.03239 (2018)."},{"key":"e_1_3_2_1_20_1","volume-title":"Inherent trade-offs in the fair determination of risk scores. ITCS","author":"Kleinberg J.","year":"2017","unstructured":"Kleinberg , J. , Mullainathan , S. , and Raghavan , M . Inherent trade-offs in the fair determination of risk scores. ITCS ( 2017 ). Kleinberg, J., Mullainathan, S., and Raghavan, M. Inherent trade-offs in the fair determination of risk scores. ITCS (2017)."},{"key":"e_1_3_2_1_21_1","first-page":"202","volume-title":"KDD","volume":"96","author":"Kohavi R.","year":"1996","unstructured":"Kohavi , R. Scaling up the accuracy of naive-bayes classifiers: a decision-tree hybrid . In KDD ( 1996 ), vol. 96 , Citeseer , pp. 202 -- 207 . Kohavi, R. Scaling up the accuracy of naive-bayes classifiers: a decision-tree hybrid. In KDD (1996), vol. 96, Citeseer, pp. 202--207."},{"key":"e_1_3_2_1_22_1","volume-title":"Detecting and correcting for label shift with black box predictors. ICML","author":"Lipton Z. C.","year":"2018","unstructured":"Lipton , Z. C. , Wang , Y. , and Smola , A. J . Detecting and correcting for label shift with black box predictors. ICML ( 2018 ). Lipton, Z. C., Wang, Y., and Smola, A. J. Detecting and correcting for label shift with black box predictors. ICML (2018)."},{"key":"e_1_3_2_1_23_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2015.425"},{"key":"e_1_3_2_1_24_1","first-page":"512","volume-title":"Advances in neural information processing systems","author":"Mason L.","year":"2000","unstructured":"Mason , L. , Baxter , J. , Bartlett , P. L. , and Frean , M. R . Boosting algorithms as gradient descent . In Advances in neural information processing systems ( 2000 ), pp. 512 -- 518 . Mason, L., Baxter, J., Bartlett, P. L., and Frean, M. R. Boosting algorithms as gradient descent. In Advances in neural information processing systems (2000), pp. 512--518."},{"key":"e_1_3_2_1_25_1","volume-title":"On fairness and calibration. NIPS","author":"Pleiss G.","year":"2017","unstructured":"Pleiss , G. , Raghavan , M. , Wu , F. , Kleinberg , J. , and Weinberger , K. Q . On fairness and calibration. NIPS ( 2017 ). Pleiss, G., Raghavan, M., Wu, F., Kleinberg, J., and Weinberger, K. Q. On fairness and calibration. NIPS (2017)."},{"key":"e_1_3_2_1_26_1","volume-title":"Probably approximately metric-fair learning. ICML","author":"Rothblum G. N.","year":"2018","unstructured":"Rothblum , G. N. , and Yona , G . Probably approximately metric-fair learning. ICML ( 2018 ). Rothblum, G. N., and Yona, G. Probably approximately metric-fair learning. ICML (2018)."},{"key":"e_1_3_2_1_27_1","volume-title":"How much does your data exploration overfit? controlling bias via information usage. AISTATS","author":"Russo D.","year":"2016","unstructured":"Russo , D. , and Zou , J . How much does your data exploration overfit? controlling bias via information usage. AISTATS ( 2016 ). Russo, D., and Zou, J. How much does your data exploration overfit? controlling bias via information usage. AISTATS (2016)."},{"key":"e_1_3_2_1_28_1","doi-asserted-by":"crossref","DOI":"10.7551\/mitpress\/8291.001.0001","volume-title":"Boosting: Foundations and algorithms","author":"Schapire R. E.","year":"2012","unstructured":"Schapire , R. E. , and Freund , Y . Boosting: Foundations and algorithms . MIT press , 2012 . Schapire, R. E., and Freund, Y. Boosting: Foundations and algorithms. MIT press, 2012."},{"key":"e_1_3_2_1_29_1","volume-title":"Facenet: A unified embedding for face recognition and clustering. CoRR abs\/1503.03832","author":"Schroff F.","year":"2015","unstructured":"Schroff , F. , Kalenichenko , D. , and Philbin , J . Facenet: A unified embedding for face recognition and clustering. CoRR abs\/1503.03832 ( 2015 ). Schroff, F., Kalenichenko, D., and Philbin, J. Facenet: A unified embedding for face recognition and clustering. CoRR abs\/1503.03832 (2015)."},{"key":"e_1_3_2_1_30_1","volume-title":"Uk biobank: an open access resource for identifying the causes of a wide range of complex diseases of middle and old age. PLoS medicine 12, 3","author":"Sudlow C.","year":"2015","unstructured":"Sudlow , C. , Gallacher , J. , Allen , N. , Beral , V. , Burton , P. , Danesh , J. , Downey , P. , Elliott , P. , Green , J. , Landray , M. , Uk biobank: an open access resource for identifying the causes of a wide range of complex diseases of middle and old age. PLoS medicine 12, 3 ( 2015 ), e1001779. Sudlow, C., Gallacher, J., Allen, N., Beral, V., Burton, P., Danesh, J., Downey, P., Elliott, P., Green, J., Landray, M., et al. Uk biobank: an open access resource for identifying the causes of a wide range of complex diseases of middle and old age. PLoS medicine 12, 3 (2015), e1001779."},{"key":"e_1_3_2_1_31_1","first-page":"12","volume-title":"AAAI","volume":"4","author":"Szegedy C.","year":"2017","unstructured":"Szegedy , C. , Ioffe , S. , Vanhoucke , V. , and Alemi , A. A . Inception-v4, inception-resnet and the impact of residual connections on learning . In AAAI ( 2017 ), vol. 4 , p. 12 . Szegedy, C., Ioffe, S., Vanhoucke, V., and Alemi, A. A. Inception-v4, inception-resnet and the impact of residual connections on learning. In AAAI (2017), vol. 4, p. 12."},{"key":"e_1_3_2_1_32_1","doi-asserted-by":"publisher","DOI":"10.1109\/CCC.2009.41"},{"key":"e_1_3_2_1_33_1","first-page":"325","volume-title":"Proceedings of the 30th International Conference on Machine Learning (ICML-13)","author":"Zemel R.","year":"2013","unstructured":"Zemel , R. , Wu , Y. , Swersky , K. , Pitassi , T. , and Dwork , C . Learning fair representations . In Proceedings of the 30th International Conference on Machine Learning (ICML-13) ( 2013 ), pp. 325 -- 333 . Zemel, R., Wu, Y., Swersky, K., Pitassi, T., and Dwork, C. Learning fair representations. In Proceedings of the 30th International Conference on Machine Learning (ICML-13) (2013), pp. 325--333."}],"event":{"name":"AIES '19: AAAI\/ACM Conference on AI, Ethics, and Society","sponsor":["SIGAI ACM Special Interest Group on Artificial Intelligence","AAAI American Association for Artificial Intelligence"],"location":"Honolulu HI USA","acronym":"AIES '19"},"container-title":["Proceedings of the 2019 AAAI\/ACM Conference on AI, Ethics, and Society"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3306618.3314287","content-type":"application\/pdf","content-version":"vor","intended-application":"syndication"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3306618.3314287","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,1,10]],"date-time":"2023-01-10T08:17:58Z","timestamp":1673338678000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3306618.3314287"}},"subtitle":["Black-Box Post-Processing for Fairness in Classification"],"short-title":[],"issued":{"date-parts":[[2019,1,27]]},"references-count":33,"alternative-id":["10.1145\/3306618.3314287","10.1145\/3306618"],"URL":"https:\/\/doi.org\/10.1145\/3306618.3314287","relation":{},"subject":[],"published":{"date-parts":[[2019,1,27]]},"assertion":[{"value":"2019-01-27","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}