Abstract
As automated decision systems (ADS) get more deeply embedded into business processes worldwide, there is a growing need for practical ways to establish meaningful transparency. Here we argue that universally perfect transparency is impossible to achieve. We introduce the concept of contextual transparency as an approach that integrates social science, engineering and information design to help improve ADS transparency for specific professions, business processes and stakeholder groups. We demonstrate the applicability of the contextual transparency approach by using it for a well-established ADS transparency tool: nutritional labels that display specific information about an ADS. Empirically, it focuses on the profession of recruiting. Presenting data from an ongoing study about ADS use in recruiting alongside a typology of ADS nutritional labels, we suggest a nutritional label prototype for ADS-driven rankers such as LinkedIn Recruiter before closing with directions for future work.
This is a preview of subscription content, access via your institution
Access options
Access Nature and 54 other Nature Portfolio journals
Get Nature+, our best-value online-access subscription
$29.99 per month
cancel any time
Subscribe to this journal
Receive 12 digital issues and online access to articles
$99.00 per year
only $8.25 per issue
Rent or buy this article
Get just this article for as long as you need it
$39.95
Prices may be subject to local taxes which are calculated during checkout




References
AI capabilities deployed in standard business processes 2020. Statista http://www.statista.com/statistics/1112998/ai-capabilities-standard-business-processes/ (2022).
Bailey, D. E. Emerging technologies at work: policy ideas to address negative consequences for work, workers, and society. ILR Rev. 75, 527–551 (2022).
Ajunwa, I., Crawford, K. & Schultz, J. Limitless worker surveillance. Preprint at SSRN https://ssrn.com/abstract=2746211 (2016).
Kizilcec, R. F. & Lee, H. Algorithmic fairness in education. Preprint at https://doi.org/10.48550/arXiv.2007.05443 (2021).
Baker, R. S. & Hawn, A. Algorithmic bias in education. Int. J. Artif. Intell. Educ. 32, 1052–1092 (2022).
Gipson Rankin, S. Technological tethereds: potential impact of untrustworthy artificial intelligence in criminal justice risk assessment instruments. Preprint at SSRN https://doi.org/10.2139/ssrn.3662761 (2020).
Angwin, J., Larson, J., Mattu, S. & Kirchner, L. Machine Bias. ProPublica (23 May 2016).
Executive Order on Improving the Nation’s Cybersecurity. The White House https://www.whitehouse.gov/briefing-room/presidential-actions/2021/05/12/executive-order-on-improving-the-nations-cybersecurity/ (12 May 2021).
Blueprint for an AI Bill of Rights. The White House https://www.whitehouse.gov/ostp/ai-bill-of-rights/ (2022).
AI RMF Playbook https://pages.nist.gov/AIRMF/ (NIST, 2022).
Peirce, C. S. Philosophical Writings of Peirce (Dover Publications, 1955).
James, W. Pragmatism, and Other Essays (Washington Square Press, 1963).
de Lange, F. P., Heilbron, M. & Kok, P. How do expectations shape perception? Trends Cognit. Sci. 22, 764–779 (2018).
Dretske, F. Knowledge and the Flow of Information (Basil Blackwell, 1981).
Glynn, D. In Corpus Methods for Semantics 7–38 (John Benjamins Publishing Company, 2014).
Otten, M., Seth, A. K. & Pinto, Y. A social Bayesian brain: How social knowledge can shape visual perception. Brain Cognit. 112, 69–77 (2017).
Snyder, J. S., Schwiedrzik, C. M., Vitela, A. D. & Melloni, L. How previous experience shapes perception in different sensory modalities. Front. Hum. Neurosci. 9, 594 (2015).
Baldauf, M., Dustdar, S. & Rosenberg, F. A survey on context-aware systems. Int. J. Ad Hoc Ubiquitous Comput. 2, 263–277 (2007).
Greenberg, S. Context as a dynamic construct. Hum. Comput. Interact. 16, 257–268 (2001).
Suchman, L. Human-Machine Reconfigurations: Plans and Situated Actions (Cambridge Univ. Press, 1987).
Dourish, P. What we talk about when we talk about context. Pers. Ubiquit. Comput. 8, 19–30 (2004).
Bazire, M. & Brézillon, P. In Modeling and Using Context (eds. Dey, A., Kokinov, B., Leake, D. & Turner, R.) Vol. 3554, 29–40 (Springer, 2005).
Bellotti, V. & Edwards, K. Intelligibility and accountability: human considerations in context-aware systems. Hum. Comput. Interact. 16, 193–212 (2001).
Sloane, M. & Moss, E. AI’s social sciences deficit. Nat. Mach. Intell. 1, 330–331 (2019).
Miller, T. Explanation in artificial intelligence: insights from the social sciences. Preprint at https://doi.org/10.48550/arXiv.1706.07269 (2018).
Hirsch, T., Merced, K., Narayanan, S., Imel, Z. E. & Atkins, D. C. Designing contestability: interaction design, machine learning, and mental health. in Proceedings of the 2017 Conference on Designing Interactive Systems 95–99 (Association for Computing Machinery, 2017); https://doi.org/10.1145/3064663.3064703
Lyons, H., Velloso, E. & Miller, T. Conceptualising contestability: perspectives on contesting algorithmic decisions. Proc. ACM Hum. Comput. Interact. 5, 106:1–106:25 (2021).
Stoyanovich, J., Van Bavel, J. J. & West, T. V. The imperative of interpretable machines. Nat. Mach. Intell. 2, 197–199 (2020).
Rudin, C. Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead. Nat. Mach. Intell. 1, 206–215 (2019).
Marcinkevičs, R. & Vogt, J. E. Interpretability and explainability: a machine learning zoo mini-tour. Preprint at https://doi.org/10.48550/arXiv.2012.01805 (2020).
Ribeiro, M. T., Singh, S. & Guestrin, C. ‘Why should i trust you?’: Explaining the predictions of any classifier. In Proc. 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining 1135–1144 (Association for Computing Machinery, 2016); https://doi.org/10.1145/2939672.2939778
Datta, A., Sen, S. & Zick, Y. Algorithmic transparency via quantitative input influence: theory and experiments with learning systems. In 2016 IEEE Symposium on Security and Privacy 598–617 (2016); https://doi.org/10.1109/SP.2016.42
Lundberg, S. M. & Lee, S.-I. A unified approach to interpreting model predictions. in Advances in Neural Information Processing Systems Vol. 30 (Curran Associates, 2017).
Weld, D. S. & Bansal, G. The challenge of crafting intelligible intelligence. Commun. ACM 62, 70–79 (2019).
Nissenbaum, H. Privacy in Context: Technology, Policy, and the Integrity of Social Life (Stanford Univ. Press, 2009); https://doi.org/10.1515/9780804772891
Zimmer, M. Addressing conceptual gaps in big data research ethics: an application of contextual integrity. Social Media Soc. 4, 2056305118768300 (2018).
Jacovi, A., Marasović, A., Miller, T. & Goldberg, Y. Formalizing trust in artificial intelligence: prerequisites, causes and goals of human trust in AI. in Proceedings of the 2021 ACM Conference on Fairness, Accountability, and Transparency 624–635 (Association for Computing Machinery, 2021); https://doi.org/10.1145/3442188.3445923
Celino, I. Who is this explanation for? Human intelligence and knowledge graphs for explainable AI. Preprint at https://doi.org/10.48550/arXiv.2005.13275 (2020)
Ajunwa, I. An auditing imperative for automated hiring. Preprint at SSRN https://doi.org/10.2139/ssrn.3437631 (2019).
Black, J. S. & van Esch, P. AI-enabled recruiting: what is it and how should a manager use it? Bus. Horiz. 63, 215–226 (2020).
Cumbo, L. A. NYC Local Law 144 of 2021 (The New York City Council, 2021).
Artificial Intelligence Video Interview Act (Illinois General Assembly, 2020).
Statista. Number of LinkedIn users worldwide from 2019 to 2021, by subscription type. Statista http://www.statista.com/statistics/1335947/linkedin-global-users-by-subscription/ (2022).
Burns, S. What is Boolean search? The New York Public Library https://www.nypl.org/blog/2011/02/22/what-boolean-search (2011).
Stoyanovich, J. Hiring and AI: let job candidates know why they were rejected. WSJ https://www.wsj.com/articles/hiring-job-candidates-ai-11632244313 (2021).
Wartella, E. A., Lichtenstein, A. H. & Boon, C. S. History of Nutrition Labeling. Front-of-Package Nutrition Rating Systems and Symbols: Phase I Report (National Academies Press, 2010).
Guidance for Industry: Food Labeling Guide. US Food and Drug Administration https://www.fda.gov/regulatory-information/search-fda-guidance-documents/guidance-industry-food-labeling-guide (2013).
Felzmann, H., Fosch-Villaronga, E., Lutz, C. & Tamò-Larrieux, A. Towards transparency by design for artificial intelligence. Sci. Eng. Ethics 26, 3333–3361 (2020).
Kay, J., Kuflik, T. & Rovatsos, M. Transparency by design (Dagstuhl seminar 21231). Dagstuhl Rep. 11, 1–22 (2021).
Zieglmeier, V. & Pretschner, A. Trustworthy transparency by design. Preprint at https://doi.org/10.48550/arXiv.2103.10769 (2021).
Cavoukian, A. The 7 Foundational Principles (Information and Privacy Commissioner of Ontario, 2009).
Pattakou, A., Mavroeidi, A.-G., Diamantopoulou, V., Kalloniatis, C. & Gritzalis, S. Towards the design of usable privacy by design methodologies. In 2018 IEEE 5th International Workshop on Evolving Security & Privacy Requirements Engineering (ESPRE) (2018); https://doi.org/10.1109/ESPRE.2018.00007
Romanou, A. The necessity of the implementation of privacy by design in sectors where data protection concerns arise. Comput. Law Secur. Rev. 34, 99–110 (2018).
Emami-Naeini, P., Agarwal, Y., Faith Cranor, L. & Hibshi, H. Ask the experts: what should be on an IoT privacy and security label? In 2020 IEEE Symposium on Security and Privacy 447–464 (2020); https://doi.org/10.1109/SP40000.2020.00043
Johansen, J. et al. A multidisciplinary definition of privacy labels. Inf. Comput. Secur. 30, 452–469 (2022).
Kelley, P. G., Cesca, L., Bresee, J. & Cranor, L. F. Standardizing privacy notices: an online study of the nutrition label approach. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems 1573–1582 (Association for Computing Machinery, 2010); https://doi.org/10.1145/1753326.1753561
Shen, Y. & Vervier, P.-A. in Privacy Technologies and Policy (eds. Naldi, M. et al.) 136–147 (Springer, 2019); https://doi.org/10.1007/978-3-030-21752-5_9
Kelley, P. G., Bresee, J., Cranor, L. F. & Reeder, R. W. A ‘nutrition label’ for privacy. in Proceedings of the 5th Symposium on Usable Privacy and Security (Association for Computing Machinery, 2009); https://doi.org/10.1145/1572532.1572538
Kollnig, K., Shuba, A., Van Kleek, M., Binns, R. & Shadbolt, N. Goodbye tracking? Impact of iOS app tracking transparency and privacy labels. In 2022 ACM Conference on Fairness, Accountability, and Transparency 508–520 (Association for Computing Machinery, 2022); https://doi.org/10.1145/3531146.3533116
Scoccia, G. L., Autili, M., Stilo, G. & Inverardi, P. An empirical study of privacy labels on the Apple iOS mobile app store. In Proceedings of the 9th IEEE/ACM International Conference on Mobile Software Engineering and Systems 114–124 (Association for Computing Machinery, 2022); https://doi.org/10.1145/3524613.3527813
Mascharka, D., Tran, P., Soklaski, R. & Majumdar, A. Transparency by design: closing the gap between performance and interpretability in visual reasoning. In 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition 4942–4950 (2018). https://doi.org/10.1109/CVPR.2018.00519
Bargh, M., van de Mosselaar, M., Rutten, P. & Choenni, S. On using privacy labels for visualizing the privacy practice of SMEs: challenges and research directions. In DG.O 2022: The 23rd Annual International Conference on Digital Government Research 166–175 (Association for Computing Machinery, 2022).
Yang, K. et al. A nutritional label for rankings. In Proceedings of the 2018 International Conference on Management of Data 1773–1776 (Association for Computing Machinery, 2018); https://doi.org/10.1145/3183713.3193568
Mitchell, M. et al. Model cards for model reporting. In Proceedings of the Conference on Fairness, Accountability, and Transparency 220–229 (Association for Computing Machinery, 2019); https://doi.org/10.1145/3287560.3287596
Wohlin, C. Guidelines for snowballing in systematic literature studies and a replication in software engineering. In Proceedings of the 18th International Conference on Evaluation and Assessment in Software Engineering (Association for Computing Machinery, 2014); https://doi.org/10.1145/2601248.2601268
Wnuk, K. & Garrepalli, T. Knowledge management in software testing: a systematic snowball literature review. E-Informatica Softw. Eng. J. 12, 51–78 (2018).
Stoyanovich, J. & Howe, B. Nutritional labels for data and models. In A Quarterly bulletin of the Computer Society of the IEEE Technical Committee on Data Engineering (2019).
Gebru, T. et al. Datasheets for datasets. Commun. ACM 64, 86–92 (2021).
Harkous, H. et al. Polisis: automated analysis and presentation of privacy policies using deep learning. Preprint at https://arxiv.org/abs/1802.02561 (2018).
Fang, L. & LeFevre, K. Privacy wizards for social networking sites. In Proceedings of the 19th International Conference on World Wide Web 351–360 (Association for Computing Machinery, 2010); https://doi.org/10.1145/1772690.1772727
Mateescu, A. & Elish, M. C. AI in context. Data & Society https://datasociety.net/library/ai-in-context/ (2019).
Volokhin, S., Collins, M., Rokhlenko, O. & Agichtein, E. Generating and validating contextually relevant justifications for conversational recommendation. In ACM SIGIR Conference on Human Information Interaction and Retrieval 284–289 (ACM, 2022); https://doi.org/10.1145/3498366.3505789
Balog, K. & Radlinski, F. Measuring recommendation explanation quality: the conflicting goals of explanations. In Proceedings of the 43rd International ACM SIGIR Conference on Research and Development in Information Retrieval 329–338 (Association for Computing Machinery, 2020); https://doi.org/10.1145/3397271.3401032
Sinha, R. & Swearingen, K. The role of transparency in recommender systems. In CHI ’02 Extended Abstracts on Human Factors in Computing Systems 830–831 (Association for Computing Machinery, 2002); https://doi.org/10.1145/506443.506619
2020/C 424/01 (European Union, 2020).
Bathaee, Y. The artificial intelligence black box and the failure of intent and causation. Harvard J. Law Technol. 31, 889–938 (2018).
Wagner, B. Liable, but not in control? Ensuring meaningful human agency in automated decision-making systems. Policy Internet 11, 104–122 (2019).
Susskind, R. E. & Susskind, D. The Future of the Professions: How Technology Will Transform the Work of Human Experts (Oxford Univ. Press, 2015).
Klegon, D. The sociology of professions: an emerging perspective. Sociol. Work Occup. 5, 259–283 (1978).
Abbott, A. The System of Professions: An Essay on The Division of Expert Labor (Univ. Chicago Press, 1988).
Wenger, E. Communities of Practice: Learning, Meaning, and Identity (Cambridge Univ. Press, 1999).
Kartikeya, A. in Intelligent Computing (ed. Arai, K.) 353–358 (Springer, 2022); https://doi.org/10.1007/978-3-031-10464-0_23
Schmidt, P., Biessmann, F. & Teubner, T. Transparency and trust in artificial intelligence systems. J. Decis. Syst. 29, 260–278 (2020).
Elia, J. Transparency rights, technology, and trust. Ethics Inf. Technol. 11, 145–153 (2009).
Ashwin, Agnew, W., Pajaro, U., Jethwani, H. & Subramonian, A. Rebuilding trust: queer in AI approach to artificial intelligence risk management. Preprint at https://doi.org/10.48550/arXiv.2110.09271 (2022).
LaRosa, E. & Danks, D. Impacts on trust of healthcare AI. in Proceedings of the 2018 AAAI/ACM Conference on AI, Ethics, and Society 210–215 (Association for Computing Machinery, 2018); https://doi.org/10.1145/3278721.3278771
Chmielinski, K. S. et al. The Dataset Nutrition Label (2nd Gen): Leveraging Context to Mitigate Harms in Artificial Intelligence. Preprint at https://arxiv.org/abs/2201.03954 (2022).
Stoyanovich, J., Abiteboul, S., Howe, B., Jagadish, H. V. & Schelter, S. Responsible data management. Commun. ACM 65, 64–74 (2022).
Arnold, M. et al. FactSheets: increasing trust in AI services through supplier’s declarations of conformity. Preprint at https://doi.org/10.48550/arXiv.1808.07261 (2019).
Sun, C., Asudeh, A., Jagadish, H. V., Howe, B. & Stoyanovich, J. MithraLabel: flexible dataset nutritional labels for responsible data science. In Proceedings of the 28th ACM International Conference on Information and Knowledge Management 2893–2896 (Association for Computing Machinery, 2019); https://doi.org/10.1145/3357384.3357853
Wachter, S., Mittelstadt, B. & Russell, C. Counterfactual explanations without opening the black box: automated decisions and the GDPR. Preprint at https://doi.org/10.48550/arXiv.1711.00399 (2018)
Byrne, R. M. J. Counterfactuals in explainable artificial intelligence (XAI): evidence from human reasoning. In Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence 6276–6282 (2019).
Acknowledgements
This research was supported in part by National Science Foundation awards 1916505, 1922658 and 1928627.
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Competing interests
The authors declare no competing interests.
Peer review
Peer review information
Nature Machine Intelligence thanks Aurelia Tamo-Larrieux and Silvia Milano for their contribution to the peer review of this work.
Additional information
Publisher’s note Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Sloane, M., Solano-Kamaiko, I.R., Yuan, J. et al. Introducing contextual transparency for automated decision systems. Nat Mach Intell 5, 187–195 (2023). https://doi.org/10.1038/s42256-023-00623-7
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1038/s42256-023-00623-7