{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2024,11,21]],"date-time":"2024-11-21T05:40:35Z","timestamp":1732167635290,"version":"3.28.0"},"publisher-location":"New York, NY, USA","reference-count":146,"publisher":"ACM","content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2024,6,3]]},"DOI":"10.1145\/3630106.3658964","type":"proceedings-article","created":{"date-parts":[[2024,6,5]],"date-time":"2024-06-05T13:14:21Z","timestamp":1717593261000},"page":"1174-1186","update-policy":"http:\/\/dx.doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":1,"title":["Should Users Trust Advanced AI Assistants? Justified Trust As a Function of Competence and Alignment"],"prefix":"10.1145","author":[{"ORCID":"http:\/\/orcid.org\/0000-0001-7710-8974","authenticated-orcid":false,"given":"Arianna","family":"Manzini","sequence":"first","affiliation":[{"name":"Google DeepMind, United Kingdom"}]},{"ORCID":"http:\/\/orcid.org\/0000-0003-3251-4981","authenticated-orcid":false,"given":"Geoff","family":"Keeling","sequence":"additional","affiliation":[{"name":"Google Research, United Kingdom"}]},{"ORCID":"http:\/\/orcid.org\/0000-0002-8518-3840","authenticated-orcid":false,"given":"Nahema","family":"Marchal","sequence":"additional","affiliation":[{"name":"Google DeepMind, United Kingdom"}]},{"ORCID":"http:\/\/orcid.org\/0000-0002-4412-1686","authenticated-orcid":false,"given":"Kevin R.","family":"McKee","sequence":"additional","affiliation":[{"name":"Google DeepMind, United Kingdom"}]},{"ORCID":"http:\/\/orcid.org\/0000-0001-6117-4395","authenticated-orcid":false,"given":"Verena","family":"Rieser","sequence":"additional","affiliation":[{"name":"Google DeepMind, United Kingdom"}]},{"ORCID":"http:\/\/orcid.org\/0000-0002-7552-4576","authenticated-orcid":false,"given":"Iason","family":"Gabriel","sequence":"additional","affiliation":[{"name":"Google DeepMind, United Kingdom"}]}],"member":"320","published-online":{"date-parts":[[2024,6,5]]},"reference":[{"key":"e_1_3_2_1_1_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1"},{"key":"e_1_3_2_1_2_1","unstructured":"Lize Alberts and Max\u00a0Van Kleek. 2023. Computers as Bad Social Actors: Dark Patterns and Anti-Patterns in Interfaces that Act Socially. arxiv:2302.04720\u00a0[cs.HC]"},{"key":"e_1_3_2_1_3_1","unstructured":"Dario Amodei Chris Olah Jacob Steinhardt Paul Christiano John Schulman and Dan Man\u00e9. 2016. Concrete Problems in AI Safety. arxiv:1606.06565\u00a0[cs.AI]"},{"key":"e_1_3_2_1_4_1","unstructured":"Markus Anderljung Joslyn Barnhart Anton Korinek Jade Leung Cullen O\u2019Keefe Jess Whittlestone Shahar Avin Miles Brundage Justin Bullock Duncan Cass-Beggs Ben Chang Tantum Collins Tim Fist Gillian Hadfield Alan Hayes Lewis Ho Sara Hooker Eric Horvitz Noam Kolt Jonas Schuett Yonadav Shavit Divya Siddarth Robert Trager and Kevin Wolf. 2023. Frontier AI Regulation: Managing Emerging Risks to Public Safety. arxiv:2307.03718\u00a0[cs.CY]"},{"key":"e_1_3_2_1_5_1","unstructured":"Anthropic. 2023. Core Views on AI Safety: When Why What and How. https:\/\/www.anthropic.com\/index\/core-views-on-ai-safety."},{"key":"e_1_3_2_1_6_1","unstructured":"Anthropic. 2023. Frontier Threats Red Teaming for AI Safety. https:\/\/www.anthropic.com\/index\/frontier-threats-red-teaming-for-ai-safety."},{"key":"e_1_3_2_1_7_1","unstructured":"Carolyn Ashurst. 2023. How to Regulate Foundation Models Can We Do Better than the EU AI Act?https:\/\/www.turing.ac.uk\/research\/interest-groups\/fairness-transparency-privacy\/how-to-regulate-foundation-models."},{"key":"e_1_3_2_1_8_1","unstructured":"Yuntao Bai Andy Jones Kamal Ndousse Amanda Askell Anna Chen Nova DasSarma Dawn Drain Stanislav Fort Deep Ganguli Tom Henighan Nicholas Joseph Saurav Kadavath Jackson Kernion Tom Conerly Sheer El-Showk Nelson Elhage Zac Hatfield-Dodds Danny Hernandez Tristan Hume Scott Johnston Shauna Kravec Liane Lovitt Neel Nanda Catherine Olsson Dario Amodei Tom Brown Jack Clark Sam McCandlish Chris Olah Ben Mann and Jared Kaplan. 2022. Training a Helpful and Harmless Assistant with Reinforcement Learning from Human Feedback. arxiv:2204.05862\u00a0[cs.CL]"},{"key":"e_1_3_2_1_9_1","volume-title":"Constitutional AI: Harmlessness from AI Feedback. arxiv:2212.08073\u00a0[cs.CL]","author":"Bai Yuntao","year":"2022","unstructured":"Yuntao Bai, Saurav Kadavath, Sandipan Kundu, Amanda Askell, Jackson Kernion, Andy Jones, Anna Chen, Anna Goldie, Azalia Mirhoseini, Cameron McKinnon, Carol Chen, Catherine Olsson, Christopher Olah, Danny Hernandez, Dawn Drain, Deep Ganguli, Dustin Li, Eli Tran-Johnson, Ethan Perez, Jamie Kerr, Jared Mueller, Jeffrey Ladish, Joshua Landau, Kamal Ndousse, Kamile Lukosuite, Liane Lovitt, Michael Sellitto, Nelson Elhage, Nicholas Schiefer, Noemi Mercado, Nova DasSarma, Robert Lasenby, Robin Larson, Sam Ringer, Scott Johnston, Shauna Kravec, Sheer\u00a0El Showk, Stanislav Fort, Tamera Lanham, Timothy Telleen-Lawton, Tom Conerly, Tom Henighan, Tristan Hume, Samuel\u00a0R. Bowman, Zac Hatfield-Dodds, Ben Mann, Dario Amodei, Nicholas Joseph, Sam McCandlish, Tom Brown, and Jared Kaplan. 2022. Constitutional AI: Harmlessness from AI Feedback. arxiv:2212.08073\u00a0[cs.CL]"},{"key":"e_1_3_2_1_10_1","doi-asserted-by":"crossref","first-page":"995","DOI":"10.1038\/s41591-020-0838-z","article-title":"The human side of health data","volume":"26","author":"Banner F","year":"2020","unstructured":"Natalie\u00a0F Banner. 2020. The human side of health data. Nature Medicine 26, 7 (2020), 995\u2013995.","journal-title":"Nature Medicine"},{"key":"e_1_3_2_1_11_1","doi-asserted-by":"publisher","DOI":"10.1145\/3551624.3555290"},{"key":"e_1_3_2_1_12_1","volume-title":"2021. On the Opportunities and Risks of Foundation Models. CoRR abs\/2108.07258","author":"Bommasani Rishi","year":"2021","unstructured":"Rishi Bommasani, Drew\u00a0A. Hudson, Ehsan Adeli, Russ\u00a0B. Altman, Simran Arora, Sydney von Arx, Michael\u00a0S. Bernstein, Jeannette Bohg, Antoine Bosselut, Emma Brunskill, Erik Brynjolfsson, Shyamal Buch, Dallas Card, Rodrigo Castellon, Niladri\u00a0S. Chatterji, Annie\u00a0S. Chen, Kathleen Creel, Jared\u00a0Quincy Davis, Dorottya Demszky, Chris Donahue, Moussa Doumbouya, Esin Durmus, Stefano Ermon, John Etchemendy, Kawin Ethayarajh, Li Fei-Fei, Chelsea Finn, Trevor Gale, Lauren Gillespie, Karan Goel, Noah\u00a0D. Goodman, Shelby Grossman, Neel Guha, Tatsunori Hashimoto, Peter Henderson, John Hewitt, Daniel\u00a0E. Ho, Jenny Hong, Kyle Hsu, Jing Huang, Thomas Icard, Saahil Jain, Dan Jurafsky, Pratyusha Kalluri, Siddharth Karamcheti, Geoff Keeling, Fereshte Khani, Omar Khattab, Pang\u00a0Wei Koh, Mark\u00a0S. Krass, Ranjay Krishna, Rohith Kuditipudi, and et al.2021. On the Opportunities and Risks of Foundation Models. CoRR abs\/2108.07258 (2021). arXiv:2108.07258https:\/\/arxiv.org\/abs\/2108.07258"},{"volume-title":"Accountability and European governance","author":"Bovens Mark","key":"e_1_3_2_1_13_1","unstructured":"Mark Bovens. 2014. Two Concepts of Accountability: Accountability as a Virtue and as a Mechanism. In Accountability and European governance. Routledge, 18\u201339."},{"key":"e_1_3_2_1_14_1","unstructured":"Miles Brundage Shahar Avin Jasmine Wang Haydn Belfield Gretchen Krueger Gillian Hadfield Heidy Khlaaf Jingying Yang Helen Toner Ruth Fong Tegan Maharaj Pang\u00a0Wei Koh Sara Hooker Jade Leung Andrew Trask Emma Bluemke Jonathan Lebensold Cullen O\u2019Keefe Mark Koren Th\u00e9o Ryffel JB Rubinovitz Tamay Besiroglu Federica Carugati Jack Clark Peter Eckersley Sarah de Haas Maritza Johnson Ben Laurie Alex Ingerman Igor Krawczuk Amanda Askell Rosario Cammarota Andrew Lohn David Krueger Charlotte Stix Peter Henderson Logan Graham Carina Prunkl Bianca Martin Elizabeth Seger Noa Zilberman Se\u00e1n\u00a0\u00d3 h\u00c9igeartaigh Frens Kroeger Girish Sastry Rebecca Kagan Adrian Weller Brian Tse Elizabeth Barnes Allan Dafoe Paul Scharre Ariel Herbert-Voss Martijn Rasser Shagun Sodhani Carrick Flynn Thomas\u00a0Krendl Gilbert Lisa Dyer Saif Khan Yoshua Bengio and Markus Anderljung. 2020. Toward Trustworthy AI Development: Mechanisms for Supporting Verifiable Claims. arxiv:2004.07213\u00a0[cs.CY]"},{"key":"e_1_3_2_1_15_1","unstructured":"Alignment\u00a0Research Centre. 2023. Update on ARC\u2019s Recent Eval Efforts. https:\/\/evals.alignment.org\/blog\/2023-03-18-update-on-recent-evals\/."},{"key":"e_1_3_2_1_16_1","doi-asserted-by":"publisher","DOI":"10.1145\/3593013.3594033"},{"key":"e_1_3_2_1_17_1","doi-asserted-by":"publisher","DOI":"10.2196\/47184"},{"key":"e_1_3_2_1_18_1","first-page":"1","article-title":"Palm: Scaling language modeling with pathways","volume":"24","author":"Chowdhery Aakanksha","year":"2023","unstructured":"Aakanksha Chowdhery, Sharan Narang, Jacob Devlin, Maarten Bosma, Gaurav Mishra, Adam Roberts, Paul Barham, Hyung\u00a0Won Chung, Charles Sutton, Sebastian Gehrmann, 2023. Palm: Scaling language modeling with pathways. Journal of Machine Learning Research 24, 240 (2023), 1\u2013113.","journal-title":"Journal of Machine Learning Research"},{"key":"e_1_3_2_1_19_1","doi-asserted-by":"publisher","DOI":"10.1145\/3593013.3594073"},{"key":"e_1_3_2_1_20_1","volume-title":"Can we trust robots?Ethics and information technology 14","author":"Coeckelbergh Mark","year":"2012","unstructured":"Mark Coeckelbergh. 2012. Can we trust robots?Ethics and information technology 14 (2012), 53\u201360."},{"volume-title":"The intentional stance","author":"Dennett C","key":"e_1_3_2_1_21_1","unstructured":"Daniel\u00a0C Dennett. 1989. The intentional stance. MIT press."},{"key":"e_1_3_2_1_22_1","volume-title":"Innovation and Technology","author":"Department for Science","year":"2023","unstructured":"Department for Science, Innovation and Technology. 2023. A Pro-Innovation Approach to AI Regulation. https:\/\/www.gov.uk\/government\/publications\/ai-regulation-a-pro-innovation-approach\/white-paper."},{"key":"e_1_3_2_1_23_1","volume-title":"Decision and Control","volume":"184","author":"Devitt S","year":"2018","unstructured":"S Devitt. 2018. Trustworthiness of autonomous systems. Foundations of trusted autonomy (Studies in Systems, Decision and Control, Volume 117) (2018), 161\u2013184."},{"key":"e_1_3_2_1_24_1","first-page":"4","article-title":"From the \u2018digital divide\u2019 to \u2018digital inequality\u2019: Studying Internet use as penetration increases. Princeton: Center for Arts and Cultural Policy Studies, Woodrow Wilson School","volume":"4","author":"DiMaggio Paul","year":"2001","unstructured":"Paul DiMaggio, Eszter Hargittai, 2001. From the \u2018digital divide\u2019 to \u2018digital inequality\u2019: Studying Internet use as penetration increases. Princeton: Center for Arts and Cultural Policy Studies, Woodrow Wilson School, Princeton University 4, 1 (2001), 4\u20132.","journal-title":"Princeton University"},{"key":"e_1_3_2_1_25_1","unstructured":"Connor Dunlop. 2023. An EU AI Act That Works for People and Society. https:\/\/www.adalovelaceinstitute.org\/policy-briefing\/eu-ai-act-trilogues\/."},{"key":"e_1_3_2_1_26_1","doi-asserted-by":"publisher","DOI":"10.1007\/s11023-018-9481-6"},{"key":"e_1_3_2_1_27_1","unstructured":"Tyna Eloundou Sam Manning Pamela Mishkin and Daniel Rock. 2023. GPTs are GPTs: An Early Look at the Labor Market Impact Potential of Large Language Models. arxiv:2303.10130\u00a0[econ.GN]"},{"key":"e_1_3_2_1_28_1","unstructured":"European Commission. 2021. The EU AI Act. https:\/\/www.europarl.europa.eu\/news\/en\/headlines\/society\/20230601STO93804\/eu-ai-act-first-regulation-on-artificial-intelligence."},{"key":"e_1_3_2_1_29_1","volume-title":"Proceedings of the AAAI Conference on Artificial Intelligence, Vol.\u00a035","author":"Everitt Tom","year":"2021","unstructured":"Tom Everitt, Ryan Carey, Eric\u00a0D Langlois, Pedro\u00a0A Ortega, and Shane Legg. 2021. Agent incentives: A causal perspective. In Proceedings of the AAAI Conference on Artificial Intelligence, Vol.\u00a035. 11487\u201311495."},{"key":"e_1_3_2_1_30_1","doi-asserted-by":"publisher","DOI":"10.1007\/s13347-019-00378-3"},{"key":"e_1_3_2_1_31_1","first-page":"1","article-title":"A Unified Framework of Five Principles for AI in Society","volume":"1","author":"Floridi Luciano","year":"2019","unstructured":"Luciano Floridi and Josh Cowls. 2019. A Unified Framework of Five Principles for AI in Society. Harvard Data Science Review 1, 1 (jul 1 2019). https:\/\/hdsr.mitpress.mit.edu\/pub\/l0jsh9d1.","journal-title":"Harvard Data Science Review"},{"key":"e_1_3_2_1_32_1","doi-asserted-by":"publisher","DOI":"10.1007\/s43681-022-00241-w"},{"key":"e_1_3_2_1_33_1","volume-title":"Artificial intelligence, values, and alignment. Minds and machines 30, 3","author":"Gabriel Iason","year":"2020","unstructured":"Iason Gabriel. 2020. Artificial intelligence, values, and alignment. Minds and machines 30, 3 (2020), 411\u2013437."},{"volume-title":"Oxford Handbook of Digital Ethics, Carissa V\u00e9liz (Ed.)","author":"Gabriel Iason","key":"e_1_3_2_1_34_1","unstructured":"Iason Gabriel and Vafa Ghazavi. 2023. The Challenge of Value Alignment: From Fairer Algorithms to AI Safety. In Oxford Handbook of Digital Ethics, Carissa V\u00e9liz (Ed.). Oxford University Press, Chapter\u00a018, 336\u2013355."},{"key":"e_1_3_2_1_35_1","volume-title":"William Isaac, and James Manyika.","author":"Gabriel Iason","year":"2024","unstructured":"Iason Gabriel, Arianna Manzini, Geoff Keeling, Lisa\u00a0Anne Hendricks, Verena Rieser, Hasan Iqbal, Nenad Toma\u0161ev, Ira Ktena, Zachary Kenton, Mikel Rodriguez, Seliem El-Sayed, Sasha Brown, Canfer Akbulut, Andrew Trask, Edward Hughes, A.\u00a0Stevie Bergman, Renee Shelby, Nahema Marchal, Conor Griffin, Juan Mateos-Garcia, Laura Weidinger, Winnie Street, Benjamin Lange, Alex Ingerman, Alison Lentz, Reed Enger, Andrew Barakat, Victoria Krakovna, John\u00a0Oliver Siy, Zeb Kurth-Nelson, Amanda McCroskery, Vijay Bolina, Harry Law, Murray Shanahan, Lize Alberts, Borja Balle, Sarah de Haas, Yetunde Ibitoye, Allan Dafoe, Beth Goldberg, S\u00e9bastien Krier, Alexander Reese, Sims Witherspoon, Will Hawkins, Maribeth Rauh, Don Wallace, Matija Franklin, Josh\u00a0A. Goldstein, Joel Lehman, Michael Klenk, Shannon Vallor, Courtney Biles, Meredith\u00a0Ringel Morris, Helen King, Blaise\u00a0Ag\u00fcera y Arcas, William Isaac, and James Manyika. 2024. The Ethics of Advanced AI Assistants. arxiv:2404.16244\u00a0[cs.CY]"},{"key":"e_1_3_2_1_36_1","unstructured":"Maximilian Gahntz. 2023. The EU\u2019s AI Act and Foundation Models: The Final Stretch. https:\/\/foundation.mozilla.org\/en\/blog\/the-eus-ai-act-and-foundation-models-the-final-stretch\/."},{"key":"e_1_3_2_1_37_1","doi-asserted-by":"publisher","DOI":"10.30658\/hmc.1.5"},{"key":"e_1_3_2_1_38_1","doi-asserted-by":"publisher","DOI":"10.1145\/3531146.3533229"},{"key":"e_1_3_2_1_39_1","doi-asserted-by":"publisher","DOI":"10.5465\/annals.2018.0057"},{"volume-title":"AI Principles Progress Update","year":"2023","key":"e_1_3_2_1_40_1","unstructured":"Google. 2023. AI Principles Progress Update 2023. https:\/\/ai.google\/static\/documents\/ai-principles-2023-progress-update.pdf."},{"key":"e_1_3_2_1_41_1","unstructured":"Google. 2023. Assistant with Bard: A step toward a more personal assistant. https:\/\/blog.google\/products\/assistant\/google-assistant-bard-generative-ai\/."},{"key":"e_1_3_2_1_42_1","unstructured":"Google DeepMind. 2022. How Our Principles Helped Define AlphaFold\u2019s Release. https:\/\/deepmind.google\/discover\/blog\/how-our-principles-helped-define-alphafolds-release\/."},{"key":"e_1_3_2_1_43_1","unstructured":"Google DeepMind. 2023. Transforming The Future of Music Creation. https:\/\/deepmind.google\/discover\/blog\/transforming-the-future-of-music-creation\/."},{"key":"e_1_3_2_1_44_1","volume-title":"Data for sale: trust, confidence and sharing health data with commercial companies. Journal of Medical Ethics","author":"Graham Mackenzie","year":"2021","unstructured":"Mackenzie Graham. 2021. Data for sale: trust, confidence and sharing health data with commercial companies. Journal of Medical Ethics (2021)."},{"key":"e_1_3_2_1_45_1","volume-title":"Trust and the Goldacre Review: Why trusted research environments are not about trust. Journal of Medical Ethics","author":"Graham Mackenzie","year":"2022","unstructured":"Mackenzie Graham, Richard Milne, Paige Fitzsimmons, and Mark Sheehan. 2022. Trust and the Goldacre Review: Why trusted research environments are not about trust. Journal of Medical Ethics (2022)."},{"key":"e_1_3_2_1_46_1","doi-asserted-by":"publisher","DOI":"10.1111\/nous.12000"},{"key":"e_1_3_2_1_47_1","unstructured":"Dan Hendrycks Nicholas Carlini John Schulman and Jacob Steinhardt. 2022. Unsolved Problems in ML Safety. arxiv:2109.13916\u00a0[cs.LG]"},{"key":"e_1_3_2_1_48_1","unstructured":"High-Level Expert Group on Artificial Intelligence. 2019. Ethics Guidelines for Trustworthy AI. https:\/\/ec.europa.eu\/futurium\/en\/ai-alliance-consultation.1.html."},{"key":"e_1_3_2_1_49_1","volume-title":"Volkswagen: The Scandal Explained. https:\/\/www.bbc.co.uk\/news\/business-34324772.","author":"Hotten Russell","year":"2015","unstructured":"Russell Hotten. 2015. Volkswagen: The Scandal Explained. https:\/\/www.bbc.co.uk\/news\/business-34324772."},{"key":"e_1_3_2_1_50_1","volume-title":"Translation: Measures for the Management of Generative Artificial Intelligence Services (Draft for Comment) \u2013","author":"Huang Seaton","year":"2023","unstructured":"Seaton Huang, Helen Toner, Zac Haluza, Rogier Creemers, and Graham Webster. 2023. Translation: Measures for the Management of Generative Artificial Intelligence Services (Draft for Comment) \u2013 April 2023. https:\/\/digichina.stanford.edu\/work\/translation-measures-for-the-management-of-generative-artificial-intelligence-services-draft-for-comment-april-2023\/."},{"key":"e_1_3_2_1_51_1","unstructured":"Inflection. 2023. I\u2019m Pi Your personal AI. https:\/\/inflection.ai\/."},{"key":"e_1_3_2_1_52_1","unstructured":"Geoffrey Irving Paul Christiano and Dario Amodei. 2018. AI safety via debate. arxiv:1805.00899\u00a0[stat.ML]"},{"key":"e_1_3_2_1_53_1","doi-asserted-by":"publisher","DOI":"10.1145\/3442188.3445923"},{"key":"e_1_3_2_1_54_1","doi-asserted-by":"publisher","DOI":"10.1086\/233694"},{"key":"e_1_3_2_1_55_1","unstructured":"Eirini Kalliamvakou. 2022. Research: Quantifying GitHub Copilot\u2019s Impact on Developer Productivity and Happiness. https:\/\/github.blog\/2022-09-07-research-quantifying-github-copilots-impact-on-developer-productivity-and-happiness\/."},{"key":"e_1_3_2_1_56_1","volume-title":"Trust in artificial intelligence: Meta-analytic findings. Human factors 65, 2","author":"Kaplan D","year":"2023","unstructured":"Alexandra\u00a0D Kaplan, Theresa\u00a0T Kessler, J\u00a0Christopher Brill, and PA Hancock. 2023. Trust in artificial intelligence: Meta-analytic findings. Human factors 65, 2 (2023), 337\u2013359."},{"key":"e_1_3_2_1_57_1","doi-asserted-by":"crossref","unstructured":"Zachary Kenton Ramana Kumar Sebastian Farquhar Jonathan Richens Matt MacDermott and Tom Everitt. 2022. Discovering Agents. arxiv:2208.08345\u00a0[cs.AI]","DOI":"10.1016\/j.artint.2023.103963"},{"key":"e_1_3_2_1_58_1","doi-asserted-by":"crossref","first-page":"43","DOI":"10.1007\/s11019-016-9721-6","article-title":"Trust me, I\u2019ma researcher!: The role of trust in biomedical research","volume":"20","author":"Kerasidou Angeliki","year":"2017","unstructured":"Angeliki Kerasidou. 2017. Trust me, I\u2019ma researcher!: The role of trust in biomedical research. Medicine, Health Care and Philosophy 20, 1 (2017), 43\u201350.","journal-title":"Medicine, Health Care and Philosophy"},{"key":"e_1_3_2_1_59_1","volume-title":"Before and beyond trust: Reliance in medical AI. Journal of medical ethics","author":"Kerasidou Charalampia\u00a0Xaroula","year":"2022","unstructured":"Charalampia\u00a0Xaroula Kerasidou, Angeliki Kerasidou, Monika Buscher, and Stephen Wilkinson. 2022. Before and beyond trust: Reliance in medical AI. Journal of medical ethics (2022)."},{"key":"e_1_3_2_1_60_1","doi-asserted-by":"publisher","DOI":"10.1145\/3593013.3593978"},{"key":"e_1_3_2_1_61_1","volume-title":"Specification Gaming: The Flip Side of AI Ingenuity. https:\/\/deepmind.google\/discover\/blog\/specification-gaming-the-flip-side-of-ai-ingenuity\/.","author":"Krakovna Victoria","year":"2020","unstructured":"Victoria Krakovna, Jonathan Uesato, Vladimir Mikulik, Matthew Rahtz, Tom Everitt, Ramana Kumar, Zac Kenton, Jan Leike, and Shane Legg. 2020. Specification Gaming: The Flip Side of AI Ingenuity. https:\/\/deepmind.google\/discover\/blog\/specification-gaming-the-flip-side-of-ai-ingenuity\/."},{"key":"e_1_3_2_1_62_1","volume-title":"Too human and not human enough: A grounded theory analysis of mental health harms from emotional dependence on the social chatbot Replika","author":"Laestadius Linnea","year":"2022","unstructured":"Linnea Laestadius, Andrea Bishop, Michael Gonzalez, Diana Illen\u010d\u00edk, and Celeste Campos-Castillo. 2022. Too human and not human enough: A grounded theory analysis of mental health harms from emotional dependence on the social chatbot Replika. New Media & Society (2022), 14614448221142007."},{"key":"e_1_3_2_1_63_1","doi-asserted-by":"publisher","DOI":"10.17705\/1jais.00411"},{"key":"e_1_3_2_1_64_1","volume-title":"Trustworthy artificial intelligence and the European Union AI act: On the conflation of trustworthiness and acceptability of risk. Regulation & Governance","author":"Laux Johann","year":"2023","unstructured":"Johann Laux, Sandra Wachter, and Brent Mittelstadt. 2023. Trustworthy artificial intelligence and the European Union AI act: On the conflation of trustworthiness and acceptability of risk. Regulation & Governance (2023)."},{"key":"e_1_3_2_1_65_1","doi-asserted-by":"publisher","DOI":"10.1126\/science.adi8982"},{"key":"e_1_3_2_1_66_1","doi-asserted-by":"publisher","DOI":"10.1145\/3531146.3533182"},{"key":"e_1_3_2_1_67_1","doi-asserted-by":"crossref","unstructured":"Alexandra\u00a0Sasha Luccioni Yacine Jernite and Emma Strubell. 2023. Power Hungry Processing: Watts Driving the Cost of AI Deployment?arxiv:2311.16863\u00a0[cs.LG]","DOI":"10.1145\/3630106.3658542"},{"key":"e_1_3_2_1_68_1","volume-title":"Thinking about trust: People, process, and place. Patterns 1, 3","author":"Marsh Stephen","year":"2020","unstructured":"Stephen Marsh, Tosan Atele-Williams, Anirban Basu, Natasha Dwyer, Peter\u00a0R Lewis, Hector Miller-Bakewell, and Jeremy Pitt. 2020. Thinking about trust: People, process, and place. Patterns 1, 3 (2020)."},{"key":"e_1_3_2_1_69_1","doi-asserted-by":"publisher","DOI":"10.2307\/258792"},{"key":"e_1_3_2_1_70_1","doi-asserted-by":"publisher","DOI":"10.2307\/256727"},{"key":"e_1_3_2_1_71_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.isci.2023.107256"},{"key":"e_1_3_2_1_72_1","unstructured":"Meta. 2023. Introducing New AI Experiences Across Our Family of Apps and Devices. https:\/\/about.fb.com\/news\/2023\/09\/introducing-ai-powered-assistants-characters-and-creative-tools\/."},{"key":"e_1_3_2_1_73_1","doi-asserted-by":"publisher","DOI":"10.1145\/3593013.3594092"},{"volume-title":"Copilot: Your everyday companion. https:\/\/copilot.microsoft.com\/.","year":"2023","key":"e_1_3_2_1_74_1","unstructured":"Microsoft. 2023. Copilot: Your everyday companion. https:\/\/copilot.microsoft.com\/."},{"key":"e_1_3_2_1_75_1","doi-asserted-by":"publisher","DOI":"10.1080\/10447318.2020.1807710"},{"key":"e_1_3_2_1_76_1","doi-asserted-by":"publisher","DOI":"10.1145\/3287560.3287596"},{"key":"e_1_3_2_1_77_1","volume-title":"Principles alone cannot guarantee ethical AI. Nature machine intelligence 1, 11","author":"Mittelstadt Brent","year":"2019","unstructured":"Brent Mittelstadt. 2019. Principles alone cannot guarantee ethical AI. Nature machine intelligence 1, 11 (2019), 501\u2013507."},{"key":"e_1_3_2_1_78_1","volume-title":"Auditing large language models: A three-layered approach. AI and Ethics","author":"M\u00f6kander Jakob","year":"2023","unstructured":"Jakob M\u00f6kander, Jonas Schuett, Hannah\u00a0Rose Kirk, and Luciano Floridi. 2023. Auditing large language models: A three-layered approach. AI and Ethics (2023), 1\u201331."},{"key":"e_1_3_2_1_79_1","unstructured":"Sarah Myers\u00a0West. 2023. General Purpose AI Poses Serious Risks Should Not Be Excluded From the EU\u2019s AI Act | Policy Brief. https:\/\/ainowinstitute.org\/publication\/gpai-is-high-risk-should-not-be-excluded-from-eu-ai-act."},{"key":"e_1_3_2_1_80_1","unstructured":"Arvind Narayanan. 2019. How to Recognize AI Snake Oil. https:\/\/www.cs.princeton.edu\/\u00a0arvindn\/talks\/MIT-STS-AI-snakeoil.pdf."},{"key":"e_1_3_2_1_81_1","unstructured":"Jessica Newman. 2023. A Taxonomy of Trustworthiness for Artificial Intelligence: Connecting Properties of Trustworthiness with Risk Management and the AI Lifecycle. https:\/\/cltc.berkeley.edu\/publication\/a-taxonomy-of-trustworthiness-for-artificial-intelligence\/."},{"key":"e_1_3_2_1_82_1","doi-asserted-by":"publisher","DOI":"10.1111\/phpr.12823"},{"key":"e_1_3_2_1_83_1","doi-asserted-by":"crossref","first-page":"429","DOI":"10.1007\/s12130-010-9124-6","article-title":"Can we make sense of the notion of trustworthy technology?Knowledge","volume":"23","author":"Nickel J","year":"2010","unstructured":"Philip\u00a0J Nickel, Maarten Franssen, and Peter Kroes. 2010. Can we make sense of the notion of trustworthy technology?Knowledge, Technology & Policy 23 (2010), 429\u2013444.","journal-title":"Technology & Policy"},{"key":"e_1_3_2_1_84_1","volume-title":"Accountability in a computerized society. Science and engineering ethics 2","author":"Nissenbaum Helen","year":"1996","unstructured":"Helen Nissenbaum. 1996. Accountability in a computerized society. Science and engineering ethics 2 (1996), 25\u201342."},{"key":"e_1_3_2_1_85_1","doi-asserted-by":"publisher","unstructured":"NIST. 2023. Artificial Intelligence Risk Management Framework. https:\/\/doi.org\/10.6028\/nist.ai.100-1.","DOI":"10.6028\/nist.ai.100-1"},{"key":"e_1_3_2_1_86_1","volume-title":"Proceedings of the Human Factors and Ergonomics Society Annual Meeting, Vol.\u00a063","author":"Noah Benjamin","year":"2019","unstructured":"Benjamin Noah and Arathi Sethumadhavan. 2019. Generational differences in trust in digital assistants. In Proceedings of the Human Factors and Ergonomics Society Annual Meeting, Vol.\u00a063. SAGE Publications Sage CA: Los Angeles, CA, 206\u2013210."},{"key":"e_1_3_2_1_87_1","unstructured":"OECD. 2021. Tools for trustworthy AI: A framework to compare implementation tools for trustworthy AI systems. https:\/\/www.oecd.org\/science\/tools-for-trustworthy-ai-008232ec-en.htm."},{"key":"e_1_3_2_1_88_1","unstructured":"Office of Science and Technology Policy. 2022. Blueprint for an AI Bill of Rights. https:\/\/www.whitehouse.gov\/ostp\/ai-bill-of-rights\/."},{"volume-title":"The Logic of Collective Action","author":"Olson Mancur","key":"e_1_3_2_1_89_1","unstructured":"Mancur Olson. 1965. The Logic of Collective Action. Harvard University Press."},{"key":"e_1_3_2_1_90_1","unstructured":"OpenAI. 2019. GPT-4 System Card. https:\/\/cdn.openai.com\/papers\/gpt-4-system-card.pdf."},{"key":"e_1_3_2_1_91_1","unstructured":"OpenAI. 2023. Assistants API. https:\/\/platform.openai.com\/docs\/assistants\/overview."},{"volume-title":"Autonomy and Trust in Bioethics","author":"O\u2019Neill Onora","key":"e_1_3_2_1_92_1","unstructured":"Onora O\u2019Neill. 2002. Autonomy and Trust in Bioethics. Cambridge University Press."},{"key":"e_1_3_2_1_93_1","doi-asserted-by":"publisher","DOI":"10.1080\/09672559.2018.1454637"},{"key":"e_1_3_2_1_94_1","unstructured":"Onora O\u2019Neill and James Bardrick. 2017. Trust Trustworthiness And Transparency. https:\/\/www.thebritishacademy.ac.uk\/documents\/2563\/Future-of-the-corporation-Trust-trustworthiness-transparency.pdf."},{"key":"e_1_3_2_1_95_1","volume-title":"ART: Automatic multi-step reasoning and tool-use for large language models. arxiv:2303.09014\u00a0[cs.CL]","author":"Paranjape Bhargavi","year":"2023","unstructured":"Bhargavi Paranjape, Scott Lundberg, Sameer Singh, Hannaneh Hajishirzi, Luke Zettlemoyer, and Marco\u00a0Tulio Ribeiro. 2023. ART: Automatic multi-step reasoning and tool-use for large language models. arxiv:2303.09014\u00a0[cs.CL]"},{"key":"e_1_3_2_1_96_1","unstructured":"Peter\u00a0S. Park Simon Goldstein Aidan O\u2019Gara Michael Chen and Dan Hendrycks. 2023. AI Deception: A Survey of Examples Risks and Potential Solutions. arxiv:2308.14752\u00a0[cs.CY]"},{"key":"e_1_3_2_1_97_1","unstructured":"Sida Peng Eirini Kalliamvakou Peter Cihon and Mert Demirer. 2023. The Impact of AI on Developer Productivity: Evidence from GitHub Copilot. arxiv:2302.06590\u00a0[cs.SE]"},{"key":"e_1_3_2_1_98_1","doi-asserted-by":"publisher","DOI":"10.1002\/mar.21457"},{"key":"e_1_3_2_1_99_1","doi-asserted-by":"crossref","first-page":"445","DOI":"10.1007\/s12130-010-9125-5","article-title":"It\u2019s not about technology","volume":"23","author":"Pitt C","year":"2010","unstructured":"Joseph\u00a0C Pitt. 2010. It\u2019s not about technology. Knowledge, Technology & Policy 23 (2010), 445\u2013454.","journal-title":"Knowledge, Technology & Policy"},{"key":"e_1_3_2_1_100_1","doi-asserted-by":"crossref","first-page":"325","DOI":"10.1007\/s11019-013-9539-4","article-title":"The value and pitfalls of speculation about science and technology in bioethics: the case of cognitive enhancement","volume":"17","author":"Racine Eric","year":"2014","unstructured":"Eric Racine, Tristana Martin\u00a0Rubio, Jennifer Chandler, Cynthia Forlini, and Jayne Lucke. 2014. The value and pitfalls of speculation about science and technology in bioethics: the case of cognitive enhancement. Medicine, Health Care and Philosophy 17 (2014), 325\u2013337.","journal-title":"Medicine, Health Care and Philosophy"},{"key":"e_1_3_2_1_101_1","doi-asserted-by":"publisher","DOI":"10.1145\/3531146.3533158"},{"key":"e_1_3_2_1_102_1","doi-asserted-by":"publisher","DOI":"10.1145\/3351095.3372873"},{"key":"e_1_3_2_1_103_1","doi-asserted-by":"publisher","DOI":"10.1145\/3514094.3534181"},{"key":"e_1_3_2_1_104_1","doi-asserted-by":"publisher","DOI":"10.1177\/000276426601000303"},{"key":"e_1_3_2_1_105_1","doi-asserted-by":"publisher","DOI":"10.1007\/s43681-022-00200-5"},{"key":"e_1_3_2_1_106_1","volume-title":"Mapping the Stony Road toward Trustworthy AI. Machines we trust: Perspectives on dependable AI","author":"Rieder Gernot","year":"2021","unstructured":"Gernot Rieder, Judith Simon, and Pak-Hang Wong. 2021. Mapping the Stony Road toward Trustworthy AI. Machines we trust: Perspectives on dependable AI (2021), 27."},{"key":"e_1_3_2_1_107_1","volume-title":"Agency Laundering and Algorithmic Decision Systems. In Information in Contemporary Society: 14th International Conference, iConference 2019, Washington, DC, USA, March 31\u2013April 3, 2019, Proceedings 14","author":"Rubel Alan","year":"2019","unstructured":"Alan Rubel, Adam Pham, and Clinton Castro. 2019. Agency Laundering and Algorithmic Decision Systems. In Information in Contemporary Society: 14th International Conference, iConference 2019, Washington, DC, USA, March 31\u2013April 3, 2019, Proceedings 14. Springer, 590\u2013598."},{"key":"e_1_3_2_1_108_1","doi-asserted-by":"publisher","DOI":"10.1007\/s11948-020-00228-y"},{"key":"e_1_3_2_1_109_1","volume-title":"Toward Transparent AI: A Survey on Interpreting the Inner Structures of Deep Neural Networks. arxiv:2207.13243\u00a0[cs.LG]","author":"R\u00e4uker Tilman","year":"2023","unstructured":"Tilman R\u00e4uker, Anson Ho, Stephen Casper, and Dylan Hadfield-Menell. 2023. Toward Transparent AI: A Survey on Interpreting the Inner Structures of Deep Neural Networks. arxiv:2207.13243\u00a0[cs.LG]"},{"key":"e_1_3_2_1_110_1","unstructured":"Lucile Saulnier Siddharth Karamcheti Hugo Lauren\u00e7on Leo Tronchon Thomas Wang Victor Sanh and Amanpreet Singh. 2022. Putting Ethical Principles at the Core of the Research Lifecycle. https:\/\/huggingface.co\/blog\/ethical-charter-multimodal."},{"key":"e_1_3_2_1_111_1","unstructured":"Rylan Schaeffer Brando Miranda and Sanmi Koyejo. 2023. Are Emergent Abilities of Large Language Models a Mirage?arxiv:2304.15004\u00a0[cs.AI]"},{"key":"e_1_3_2_1_112_1","doi-asserted-by":"publisher","DOI":"10.1145\/3593013.3593994"},{"key":"e_1_3_2_1_113_1","volume-title":"Toolformer: Language Models Can Teach Themselves to Use Tools. arxiv:2302.04761\u00a0[cs.CL]","author":"Schick Timo","year":"2023","unstructured":"Timo Schick, Jane Dwivedi-Yu, Roberto Dess\u00ec, Roberta Raileanu, Maria Lomeli, Luke Zettlemoyer, Nicola Cancedda, and Thomas Scialom. 2023. Toolformer: Language Models Can Teach Themselves to Use Tools. arxiv:2302.04761\u00a0[cs.CL]"},{"key":"e_1_3_2_1_114_1","doi-asserted-by":"publisher","DOI":"10.1007\/s11023-021-09581-8"},{"key":"e_1_3_2_1_115_1","doi-asserted-by":"publisher","DOI":"10.1145\/3287560.3287598"},{"key":"e_1_3_2_1_116_1","doi-asserted-by":"publisher","DOI":"10.1098\/rsta.2017.0362"},{"key":"e_1_3_2_1_117_1","volume-title":"Goal Misgeneralization: Why Correct Specifications Aren\u2019t Enough For Correct Goals. arxiv:2210.01790\u00a0[cs.LG]","author":"Shah Rohin","year":"2022","unstructured":"Rohin Shah, Vikrant Varma, Ramana Kumar, Mary Phuong, Victoria Krakovna, Jonathan Uesato, and Zac Kenton. 2022. Goal Misgeneralization: Why Correct Specifications Aren\u2019t Enough For Correct Goals. arxiv:2210.01790\u00a0[cs.LG]"},{"key":"e_1_3_2_1_118_1","volume-title":"Role play with large language models. Nature 623, 7987","author":"Shanahan Murray","year":"2023","unstructured":"Murray Shanahan, Kyle McDonell, and Laria Reynolds. 2023. Role play with large language models. Nature 623, 7987 (2023), 493\u2013498."},{"key":"e_1_3_2_1_119_1","unstructured":"Yonadav Shavit Sandhini Agarwal Miles Brundage Steven Adler Cullen O\u2019Keefe Rosie Campbell Teddy Lee Pamela Mishkin Tyna Eloundou Alan Hickey Katarina Slama Lama Ahmad Paul McMillan Alex Beutel Alexandre Passos and David\u00a0G. Robinson. 2023. Practices for Governing Agentic AI Systems. https:\/\/openai.com\/research\/practices-for-governing-agentic-ai-systems."},{"key":"e_1_3_2_1_120_1","doi-asserted-by":"publisher","DOI":"10.1136\/medethics-2019-106048"},{"key":"e_1_3_2_1_121_1","volume-title":"Adaptable Robots, Ethics, and Trust: A Qualitative and Philosophical Exploration of the Individual Experience of Trustworthy AI","author":"Sheir Stephanie","year":"2024","unstructured":"Stephanie Sheir, Arianna Manzini, Helen Smith, and Jonathan\u00a0CS Ives. 2024. Adaptable Robots, Ethics, and Trust: A Qualitative and Philosophical Exploration of the Individual Experience of Trustworthy AI. AI and Society (2024)."},{"key":"e_1_3_2_1_122_1","unstructured":"Toby Shevlane Sebastian Farquhar Ben Garfinkel Mary Phuong Jess Whittlestone Jade Leung Daniel Kokotajlo Nahema Marchal Markus Anderljung Noam Kolt Lewis Ho Divya Siddarth Shahar Avin Will Hawkins Been Kim Iason Gabriel Vijay Bolina Jack Clark Yoshua Bengio Paul Christiano and Allan Dafoe. 2023. Model evaluation for extreme risks. arxiv:2305.15324\u00a0[cs.AI]"},{"key":"e_1_3_2_1_123_1","first-page":"2","article-title":"Trustworthy artificial intelligence","volume":"1","author":"Simion Mona","year":"2023","unstructured":"Mona Simion and Christoph Kelp. 2023. Trustworthy artificial intelligence. Asian Journal of Philosophy 1, 2 (March 3 2023). https:\/\/link.springer.com\/article\/10.1007\/s44204-023-00063-5.","journal-title":"Asian Journal of Philosophy"},{"key":"e_1_3_2_1_124_1","unstructured":"Sangeeta Singh-Kurtz. 2023. The Man of Your Dreams. https:\/\/www.thecut.com\/article\/ai-artificial-intelligence-chatbot-replika-boyfriend.html."},{"key":"e_1_3_2_1_125_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.ijhcs.2022.102903"},{"key":"e_1_3_2_1_126_1","doi-asserted-by":"publisher","DOI":"10.1145\/3597512.3600207"},{"key":"e_1_3_2_1_127_1","doi-asserted-by":"publisher","DOI":"10.1145\/3593013.3593981"},{"key":"e_1_3_2_1_128_1","unstructured":"Merlin Stein and Connor Dunlop. 2023. Safe Before Sale. https:\/\/www.adalovelaceinstitute.org\/report\/safe-before-sale\/."},{"key":"e_1_3_2_1_129_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.respol.2013.05.008"},{"key":"e_1_3_2_1_130_1","unstructured":"The Adaptive Agents Group. 2021. The Shibboleth Rule for Artificial Agents. https:\/\/hai.stanford.edu\/news\/shibboleth-rule-artificial-agents."},{"key":"e_1_3_2_1_131_1","unstructured":"The White House. 2023. Ensuring Safe Secure and Trustworthy AI. https:\/\/www.whitehouse.gov\/briefing-room\/statements-releases\/2023\/07\/21\/fact-sheet-biden-harris-administration-secures-voluntary-commitments-from-leading-artificial-intelligence-companies-to-manage-the-risks-posed-by-ai\/."},{"key":"e_1_3_2_1_132_1","unstructured":"The White House. 2023. Executive Order on the Safe Secure and Trustworthy Development and Use of Artificial Intelligence. https:\/\/www.whitehouse.gov\/briefing-room\/presidential-actions\/2023\/10\/30\/executive-order-on-the-safe-secure-and-trustworthy-development-and-use-of-artificial-intelligence\/."},{"key":"e_1_3_2_1_133_1","doi-asserted-by":"publisher","DOI":"10.1007\/s12525-020-00441-4"},{"key":"e_1_3_2_1_134_1","doi-asserted-by":"publisher","DOI":"10.1145\/3531146.3533196"},{"key":"e_1_3_2_1_135_1","doi-asserted-by":"publisher","DOI":"10.1145\/3351095.3372834"},{"key":"e_1_3_2_1_136_1","unstructured":"Andrew Trask Emma Bluemke Ben Garfinkel Claudia\u00a0Ghezzou Cuervas-Mons and Allan Dafoe. 2020. Beyond Privacy Trade-offs with Structured Transparency. arxiv:2012.08347\u00a0[cs.CR]"},{"volume-title":"Court Issues Order Requiring Cigarette Companies to Post Corrective Statements","author":"US Department of Justice. 2022.","key":"e_1_3_2_1_137_1","unstructured":"US Department of Justice. 2022. Court Issues Order Requiring Cigarette Companies to Post Corrective Statements; Resolves Historic RICO Tobacco Litigation. https:\/\/www.justice.gov\/opa\/pr\/court-issues-order-requiring-cigarette-companies-post-corrective-statements-resolves-historic."},{"key":"e_1_3_2_1_138_1","unstructured":"Boxin Wang Weixin Chen Hengzhi Pei Chulin Xie Mintong Kang Chenhui Zhang Chejian Xu Zidi Xiong Ritik Dutta Rylan Schaeffer Sang\u00a0T. Truong Simran Arora Mantas Mazeika Dan Hendrycks Zinan Lin Yu Cheng Sanmi Koyejo Dawn Song and Bo Li. 2024. DecodingTrust: A Comprehensive Assessment of Trustworthiness in GPT Models. arxiv:2306.11698\u00a0[cs.CL]"},{"key":"e_1_3_2_1_139_1","unstructured":"Francis\u00a0Rhys Ward Francesco Belardinelli Francesca Toni and Tom Everitt. 2023. Honesty Is the Best Policy: Defining and Mitigating AI Deception. arxiv:2312.01350\u00a0[cs.AI]"},{"key":"e_1_3_2_1_140_1","unstructured":"Jason Wei Yi Tay Rishi Bommasani Colin Raffel Barret Zoph Sebastian Borgeaud Dani Yogatama Maarten Bosma Denny Zhou Donald Metzler Ed\u00a0H. Chi Tatsunori Hashimoto Oriol Vinyals Percy Liang Jeff Dean and William Fedus. 2022. Emergent Abilities of Large Language Models. arxiv:2206.07682\u00a0[cs.CL]"},{"key":"e_1_3_2_1_141_1","unstructured":"Laura Weidinger Maribeth Rauh Nahema Marchal Arianna Manzini Lisa\u00a0Anne Hendricks Juan Mateos-Garcia Stevie Bergman Jackie Kay Conor Griffin Ben Bariach Iason Gabriel Verena Rieser and William Isaac. 2023. Sociotechnical Safety Evaluation of Generative AI Systems. arxiv:2310.11986\u00a0[cs.AI]"},{"key":"e_1_3_2_1_142_1","doi-asserted-by":"publisher","DOI":"10.1145\/3531146.3533088"},{"key":"e_1_3_2_1_143_1","unstructured":"Jess Whittlestone and Jack Clark. 2021. Why and How Governments Should Monitor AI Development. arxiv:2108.12427\u00a0[cs.CY]"},{"key":"e_1_3_2_1_144_1","doi-asserted-by":"publisher","DOI":"10.1145\/3351095.3372833"},{"volume-title":"Information asymmetry in online advertising","author":"Wiktor W","key":"e_1_3_2_1_145_1","unstructured":"Jan\u00a0W Wiktor and Katarzyna Sanak-Kosmowska. 2021. Information asymmetry in online advertising. Routledge."},{"key":"e_1_3_2_1_146_1","volume-title":"International Conference on Wirtschaftsinformatik (WI).-Potsdam, Germany.","author":"Zierau Naim","year":"2020","unstructured":"Naim Zierau, Christian Engel, Matthias S\u00f6llner, and Jan\u00a0Marco Leimeister. 2020. Trust in smart personal assistants: A systematic literature review and development of a research agenda. In International Conference on Wirtschaftsinformatik (WI).-Potsdam, Germany."}],"event":{"name":"FAccT '24: The 2024 ACM Conference on Fairness, Accountability, and Transparency","acronym":"FAccT '24","location":"Rio de Janeiro Brazil"},"container-title":["The 2024 ACM Conference on Fairness, Accountability, and Transparency"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3630106.3658964","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,11,21]],"date-time":"2024-11-21T04:56:13Z","timestamp":1732164973000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3630106.3658964"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,6,3]]},"references-count":146,"alternative-id":["10.1145\/3630106.3658964","10.1145\/3630106"],"URL":"https:\/\/doi.org\/10.1145\/3630106.3658964","relation":{},"subject":[],"published":{"date-parts":[[2024,6,3]]},"assertion":[{"value":"2024-06-05","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}