Simpson, K. orcid.org/0000-0002-6883-4146, O'Hara, E. and Fringi, E. (2026) Creating an effective methodology for end-user engagement in AI auditing. In: Hart, E., Horvath, T., Tan, Z. and Thomson, S., (eds.) Advances in Computational Intelligence Systems. UKCI 2025. UK Workshop on Computational Intelligence, 03-05 Sep 2025, Edinburgh, United Kingdom. Advances in Intelligent Systems and Computing, 1468. Springer Nature, pp. 256-262. ISBN: 9783032079374. ISSN: 2194-5357. EISSN: 2194-5365.
Abstract
A methodology explains the object of an AI-audit. This object has three loci: identifying significant events (harms or risks), governance (model is behaving as expected), and assurance (trust). The methodology in this paper is being developed as part of the PHAWM project (The Participatory Harm Auditing Workbenches and Methodologies project can be found at https://phawm.org), which seeks to design a workbench that supports inclusive, participant-led auditing of AI application across a range of domains. Project participants range from health service users, parents of school-aged children, to museum professionals and librarians. The project addresses a key gap in existing approaches: the absence of human-centred infrastructures that empower end-users to identify events (An event refers to an occurrence triggered by an AI application that may affect entities and has associated metrics. Each event can be assessed for likelihood, magnitude, and positive or negative valence. We avoid the term harm in our methodology due to its subjectivity, although we acknowledge its common use, including in our own project title, within AI auditing discourse), understand system behavior and participate meaningfully in audit processes.
Metadata
| Item Type: | Proceedings Paper |
|---|---|
| Authors/Creators: |
|
| Editors: |
|
| Copyright, Publisher and Additional Information: | © 2026 The Author(s). Except as otherwise noted, this author-accepted version of a conference paper published in Advances in Computational Intelligence Systems: Contributions Presented at The 24th UK Workshop on Computational Intelligence (UKCI 2025), September 3-5, 2025, Edinburgh, UK is made available via the University of Sheffield Research Publications and Copyright Policy under the terms of the Creative Commons Attribution 4.0 International License (CC-BY 4.0), which permits unrestricted use, distribution and reproduction in any medium, provided the original work is properly cited. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/ |
| Keywords: | Participatory Design; Explainable AI (XAI); Methodology; Trust and Ethics in AI systems; Audit and Accountability |
| Dates: |
|
| Institution: | The University of Sheffield |
| Academic Units: | The University of Sheffield > Faculty of Arts and Humanities (Sheffield) > School of History, Philosophy and Digital Humanities |
| Funding Information: | Funder Grant number RESPONSIBLE AI UK / RAI UK UNSPECIFIED RESPONSIBLE AI UK EP/Y009800/1 |
| Date Deposited: | 22 Jan 2026 09:45 |
| Last Modified: | 04 Feb 2026 14:38 |
| Status: | Published |
| Publisher: | Springer Nature |
| Series Name: | Advances in Intelligent Systems and Computing |
| Refereed: | Yes |
| Identification Number: | 10.1007/978-3-032-07938-1_22 |
| Open Archives Initiative ID (OAI ID): | oai:eprints.whiterose.ac.uk:236801 |
Download
Filename: UKCI_2025_Methodology_Paper_accepted.pdf
Licence: CC-BY 4.0

CORE (COnnecting REpositories)
CORE (COnnecting REpositories)