Abstract
Reports of errors committed in public contexts by facial recognition systems based on machine learning techniques have multiplied. Still, these systems have been increasingly used by the Brazilian public administration. Consequently, the following key problem is established: how can errors committed by facial recognition systems be prevented or mitigated when these systems are used for the elaboration and implementation of public policies? Guided by the understanding that algorithmic transparency is key to preventing and mitigating these errors, we empirically analysed whether, or not, the Brazilian General Data Protection Law (Lei Geral de Proteção de Dados Pessoais – LGPD, in the Portuguese acronym) has been used to promote this kind of transparency in situations in which facial recognition systems are employed. We circumscribed our study to the public transportation sector of 30 large Brazilian municipalities. To gather information, we sent a questionnaire to the municipal public agencies responsible for the public transportation system with questions about how the LGPD works in this public policy area. We used the Access to Information Law to do that. Upon legal analyses, we built an algorithmic transparency scale and found that, in the sector studied, the level of transparency is “Very Low” in most municipalities. This research finding indicates that the risk of lack of control over errors made by facial recognition systems is high. It suggests that the Brazilian public administration does not know how to use the systems in question ethically, and that this lack of knowledge may apply to other Artificial Intelligence systems.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
Notes
- 1.
There are different definitions for the term “Artificial Intelligence”, which were mapped by Russell and Norvig (2016) and Sweeney (2003). As we will see throughout this section, few works have investigated the use of AI and its applications by the Brazilian public sector. In this article, we cannot analyse each of them. We only mention that, in all of them, the AI systems analysed seem to adhere to the OECD (Organization for Economic Co-operation and Development) definition. For this reason, we also adopted this definition.
- 2.
- 3.
Despite its name, the purpose of the LGPD is to protect the personal data subject, and not the personal data itself.
- 4.
A controller of personal data is any “natural or legal person […] who is responsible for decisions concerning the processing of personal data” (Article 5, Item VI).
- 5.
- 6.
In Campinas, João Pessoa, and Rio Branco, the use of FR systems in public transportation was suspended at times during the Covid-19 pandemic, as the use of masks negatively interfered with the functioning of the technology.
References
Ada Lovelace Institute: Algorithmic accountability for the public sector – learning from the first wave of policy implementation. Technical report. Ada Lovelace Institute, AI Now and Open Government Partnership. London, New York, and Washington, DC (2021)
Ananny, M., Crawford, K.: Seeing without knowing: limitations of the transparency ideal and its application to algorithmic accountability. New Media Soc. 20(3), 973–989 (2018)
Brandão, R., Oliveira, J.L.: Reconhecimento facial e viés algorítmico em grandes municípios brasileiros. In: Anais do II Workshop sobre as Implicações da Computação na Sociedade, pp. 122–127 (2021)
Brandão, R.: Artificial intelligence, human oversight, and public policies: facial recognition systems in Brazilian cities. In: Proceedings of the 35th Canadian Conference on Artificial Intelligence (2022)
Brandão, R., et al.: Reconhecimento facial, viés algorítmico e intervenção humana no transporte municipal (2022). Article under peer-review
Central Digital and Data Office’s homepage. https://www.gov.uk/government/collections/algorithmic-transparency-standard. Accessed 25 May 2022
Coelho, J., Burg, T.: Uso de inteligência artificial pelo poder público. Technical report. Transparência Brasil, Rio de Janeiro (2020)
Felzmann, H., et al.: Transparency you can trust: transparency requirements for artificial intelligence between legal norms and contextual concerns. Big Data Soc. 6(1), 1–14 (2019)
Floridi, L., Cowls, J., King, T.C., Taddeo, M.: How to Design AI for Social Good: Seven Essential Factors. Sci. Eng. Ethics 26(3), 1771–1796 (2020). https://doi.org/10.1007/s11948-020-00213-5
Francisco, P., et al.: Regulação do reconhecimento facial no setor público: avaliação de experiências internacionais. Instituto Igarapé and DataPrivacy BR Research. Technical report, Rio de Janeiro and São Paulo (2020)
Leslie, D.: Understanding artificial intelligence ethics and safety: a guide for the responsible design and implementation of AI systems in the public sector. Technical Report. The Alan Turing Institute, London (2019)
LGPD’s official link. http://www.planalto.gov.br/ccivil_03/_ato2015-2018/2018/lei/l13709.htm. Accessed 25 May 2022
Igarapé Institute’s 2019 homepage on facial recognition in Brazil. https://igarape.org.br/infografico-reconhecimento-facial-no-brasil/. Accessed 25 May 2022
Nunes, P.: Novas ferramentas, velhas práticas: reconhecimento facial e policiamento no Brasil. Retratos da Violência: cinco meses de monitoramento, análises e descobertas – Junho a Outubro 2019. Technical report. CeSEC – Centro de Estudos de Segurança e Cidadania, Rio de Janeiro (2019)
NIC.BR.: ICT Electronic Government – Survey on the Use of Information and Communication Technologies in Brazilian Public Sector. Technical Report, São Paulo (2020)
Nunes, P., et al.: Um Rio de olhos seletivos – uso de reconhecimento facial pela polícia fluminense. Digital book. CeSEC – Centro de Estudos de Segurança e Cidadania, Rio de Janeiro (2022)
OECD’s homepage on legal instruments. https://legalinstruments.oecd.org/en/instruments/oecd-legal-0449
Reis, C., et al.: Relatório sobre o uso de tecnologias de reconhecimento facial e câmeras de vigilância pela administração pública no Brasil. Technical report. LAPIN – Laboratório de Políticas Públicas e Internet, Brasília (2021)
Reisman, D., et al.: Algorithmic Impact Assessments: A Practical Framework for Public Agency Accountability. Technical Report. AI Now, New York (2018)
Russell, S., Norvig, P.: Artificial Intelligence: A Modern Approach. Pearson Education Limited, Malaysia (2016)
Shearer, E., et al.: Government AI Readiness Index 2020. Technical Report. IDRC & Oxford Insights, Canada & London (2020)
Silva, T.: Visão computacional e racismo algorítmico: branquitude e opacidade no aprendizado de máquina. Revista da ABPN 12(31), 428–448 (2020)
Sweeney, L.: That’s AI?: a history and critique of the field. Technical Report. Carnegie Mellon University, School of Computer Science, CMU-CS-03-106, Pittsburgh (2003)
Acknowledgements
The authors are grateful to Professor Glauco Arbix for his invaluable comments on previous versions of this work.
Funding
The authors of this work would like to thank the C4AI-USP and the support from the São Paulo Research Foundation (FAPESP grant #2019/07665-4) and from the IBM Corporation.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2022 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Brandão, R. et al. (2022). Artificial Intelligence, Algorithmic Transparency and Public Policies: The Case of Facial Recognition Technologies in the Public Transportation System of Large Brazilian Municipalities. In: Xavier-Junior, J.C., Rios, R.A. (eds) Intelligent Systems. BRACIS 2022. Lecture Notes in Computer Science(), vol 13653. Springer, Cham. https://doi.org/10.1007/978-3-031-21686-2_39
Download citation
DOI: https://doi.org/10.1007/978-3-031-21686-2_39
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-21685-5
Online ISBN: 978-3-031-21686-2
eBook Packages: Computer ScienceComputer Science (R0)