Abstract
Recent AI ethics has focused on applying abstract principles downward to practice. This paper moves in the other direction. Ethical insights are generated from the lived experiences of AI-designers working on tangible human problems, and then cycled upward to influence theoretical debates surrounding these questions: (1) Should AI as trustworthy be sought through explainability, or accurate performance? (2) Should AI be considered trustworthy at all, or is reliability a preferable aim? (3) Should AI ethics be oriented toward establishing protections for users, or toward catalyzing innovation? Specific answers are less significant than the larger demonstration that AI ethics is currently unbalanced toward theoretical principles, and will benefit from increased exposure to grounded practices and dilemmas.
Similar content being viewed by others
References
Argenziano G, Fabbrocini G, Carli P, De Giorgi V, Sammarco E, Delfino M (1998) Epiluminescence microscopy for the diagnosis of doubtful melanocytic skin lesions: comparison of the ABCD rule of dermatoscopy and a new 7 point checklist based on pattern analysis. Arch Dermatol 134(12):1563–1570. https://doi.org/10.1001/archderm.134.12.1563
Blomberg SN, Folke F, Ersbøll AK, Christensen HC, Torp-Pedersen C, Sayre MR, Counts CR, Lippert FK (2019) Machine learning as a supportive tool to recognize cardiac arrest in emergency calls. Resuscitation 138:322–329. https://doi.org/10.1016/j.resuscitation.2019.01.015
Blomberg SN, Christensen HC, Lippert F, Ersbøll AK, Torp-Petersen C, Sayre MR, Kudenchuk PJ, Folke F (2021) Effect of machine learning on dispatcher recognition of out-of-hospital cardiac arrest during calls to emergency medical services: a randomized clinical trial. JAMA Netw Open 4(1):e2032320. https://doi.org/10.1001/jamanetworkopen.2020.32320
Brinker TJ, Hekler A, Enk AH, Klode J, Hauschild A, Berking C, Schilling B, Haferkamp S, Schadendorf D, Holland-Letz T, Utikal JS, von Kalle C, Ludwig-Peitsch W, Sirokay J, Heinzerling L, Albrecht M, Baratella K, Bischof L, Chorti E, Schrüfer P (2019) Deep learning outperformed 136 of 157 dermatologists in a head-to-head dermoscopic melanoma image classification task. Eur J Cancer 113:47–54. https://doi.org/10.1016/j.ejca.2019.04.001
Brown T, Dandelion M, Aurko R, Martín A, Justin G (2017) Adversarial patch. 31st conference on neural information processing systems (NIPS 2017), Long Beach, CA, USA. https://arxiv.org/pdf/1712.09665.pdf
Brusseau, J (2020) What a philosopher learned at an AI ethics evaluation. AI Ethics J 1(1). Doi: https://doi.org/10.47289/AIEJ20201214
Bryson, J (2018) No one should trust AI. AI and global governance, united nations university centre for policy research. 2018-11-13. https://cpr.unu.edu/publications/articles/ai-global-governance-no-one-should-trust-ai.html
Chen, H, Lundberg S, Lee S-I (2021) Explaining models by propagating shapley values of local components. In: Shaban-Nejad A, Michalowski M, Buckeridge DL (Eds.) Explainable AI in healthcare and medicine. Studies in computational intelligence, Vol 914. Springer, Cham. Doi: https://doi.org/10.1007/978-3-030-53352-6_24
Derrida J (1998) Of grammatology. Johns Hopkins University Press, Baltimore
Frankfurt H (2005) On bullshit. Princeton University Press, Princeton
Havtorn JD, Latko J, Edin J, Borgholt L, Maaløe L, Belgrano L, Jacobsen NF, Sdun R, Agi (2020) MultiQT: multimodal learning for real-time question tracking in speech. Ar**v:2005.00812. [Cs, Eess]. http://arxiv.org/abs/2005.00812
HLEG (2019) High-level expert group on artificial intelligence (2019) Ethics guidelines for trustworthy AI [Text]. European Commission. https://ec.europa.eu/digital-single-market/en/news/ethics-guidelines-trustworthy-ai
Kant I (1987) Critique of judgment, trans. Werner Pluhar. Hackett, Indianapolis
Lambert F (2019) Tesla nerfs autopilot in Europe due to new regulations, Electrek. May 17. https://electrek.co/2019/05/17/tesla-nerfs-autopilot-europe-regulations/. https://www.cnet.com/roadshow/news/tesla-model-s-model-x-autopilot-europe-regulations/. https://electrek.co/2019/05/17/tesla-nerfs-autopilot-europe-regulations/
Lucieri A, Bajwa MN, Braun SA, Malik MI, Dengel A, Ahmed S (2020) On interpretability of deep learning based skin lesion classifiers using concept activation vectors. 2020 International joint conference on neural networks, IJCNN, pp. 1–10. Doi: https://doi.org/10.1109/IJCNN48605.2020.9206946
Maaløe L, Fraccaro M, Liévin V, Winther O (2019) BIVA: a very deep hierarchy of latent variables for generative modeling. Adv Neural Inf Process Syst 32:6551–6562
Morley J, Floridi L, Kinsey L et al (2020) From what to how: an initial review of publicly available ai ethics tools, methods and research to translate principles into practices. Sci Eng Ethics 26:2141–2168. https://doi.org/10.1007/s11948-019-00165-5
Murphy DJ, Burrows D, Santilli S, Kemp AW, Tenner S, Kreling B, Teno J (1994) The influence of the probability of survival on patients’ preferences regarding cardiopulmonary resuscitation. N Engl J Med 330(8):545–549. https://doi.org/10.1056/NEJM199402243300807
Roberts H, Cowls J, Hine E, Morley J, Taddeo M, Wang V, Floridi L (2021) Governing artificial intelligence in China and the European union: comparing aims and promoting ethical outcomes. Available at SSRN: https://ssrn.com/abstract=3811034
Ryan M (2020) In AI we trust: ethics, artificial intelligence, and reliability. Sci Eng Ethics 26:2749–2767. https://doi.org/10.1007/s11948-020-00228-y
Spinellis D (2008) A tale of four kernels. Proceedings of the 30th international conference on Software engineering. Association for computing machinery, New York, NY, USA, 381–390. Doi:https://doi.org/10.1145/1368088.1368140
Zicari RV et al (2021a) On assessing trustworthy AI in healthcare. Machine learning as a supportive tool to recognize cardiac arrest in emergency calls. Front Hum Dyn. https://doi.org/10.3389/fhumd.2021.673104
Zicari RV et al (2021b) Co-design of a trustworthy AI system in healthcare: deep learning based skin lesion classifier. Front Hum Dyn. https://doi.org/10.3389/fhumd.2021.688152
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflict of interest
Author declares that there is no conflict of interest or external funding.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
About this article
Cite this article
Brusseau, J. From the ground truth up: doing AI ethics from practice to principles. AI & Soc 38, 1651–1657 (2023). https://doi.org/10.1007/s00146-021-01336-4
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s00146-021-01336-4