Abstract
Trusted Research Environments (TREs) provide secure access to personal and sensitive data, such as Electronic Healthcare Records (EHRs), for approved research. The increasing use of Artificial Intelligence (AI) and Machine Learning (ML) in health research introduces new challenges for managing privacy risks of individuals’ data. We present a comprehensive framework that embeds mitigation strategies throughout the entire AI project lifecycle, structured across six project phases: design, governance, development, evaluation, disclosure control, and release.
This framework aims to empower all stakeholders - researchers, project teams, output checkers, and TRE staff - with clear, phase-specific recommendations on which measures and checks are necessary before model release to help identify potential disclosure risks to data. It promotes early identification of risks with corresponding mitigations and ensures responsibilities are clearly assigned to relevant actors at each stage, from initial planning through to deployment and monitoring. Mitigation strategies include good AI/ML practices, both in terms of code and documentation, privacy-enhancing techniques during training and evaluation, restricting model access via secure query systems, licencing agreements, and adversarial attack testing using tools like SACRO-ML. This process highlights the need to train everyone involved appropriately with relevant role-specific material.
A novel tiering system for disclosure control is proposed, categorising AI projects based on the likelihood of attack and associated sensitive data leakage risks. By integrating a lifecycle-focused risk management process with a scalable disclosure control tiering system, this approach enables innovative AI research while maintaining rigorous data protection standards and public trust.
This framework aims to empower all stakeholders - researchers, project teams, output checkers, and TRE staff - with clear, phase-specific recommendations on which measures and checks are necessary before model release to help identify potential disclosure risks to data. It promotes early identification of risks with corresponding mitigations and ensures responsibilities are clearly assigned to relevant actors at each stage, from initial planning through to deployment and monitoring. Mitigation strategies include good AI/ML practices, both in terms of code and documentation, privacy-enhancing techniques during training and evaluation, restricting model access via secure query systems, licencing agreements, and adversarial attack testing using tools like SACRO-ML. This process highlights the need to train everyone involved appropriately with relevant role-specific material.
A novel tiering system for disclosure control is proposed, categorising AI projects based on the likelihood of attack and associated sensitive data leakage risks. By integrating a lifecycle-focused risk management process with a scalable disclosure control tiering system, this approach enables innovative AI research while maintaining rigorous data protection standards and public trust.
| Original language | English |
|---|---|
| Publication status | Published - 23 Sept 2025 |
| Event | Data and Digital Health: Scotland Cross-Sector Hub Event 3 - COSLA, Verity House, Edinburgh, United Kingdom Duration: 23 Sept 2025 → 23 Sept 2025 https://www.tickettailor.com/events/nhsresearchscotland/1786651 |
Workshop
| Workshop | Data and Digital Health |
|---|---|
| Country/Territory | United Kingdom |
| City | Edinburgh |
| Period | 23/09/25 → 23/09/25 |
| Internet address |
Keywords
- Trusted Research Environment
- Machine Learning
- Artificial Intelligence
- Release
- Disclosure Control
ASJC Scopus subject areas
- Health Informatics
Fingerprint
Dive into the research topics of 'RELEASE-AI: Protecting Sensitive Data Across The AI Lifecycle Disclosure Risks and Mitigations in Trusted Research Environments'. Together they form a unique fingerprint.Projects
- 1 Active
-
Practical guide to Artificial Intelligence risks and mitigations for Trusted Research Environments and the RELEASE-AI framework
Crespi Boixader, A., Li, S., Liley, J., Ward, L., Cole, C. & Smith, J., 15 Oct 2025.Research output: Contribution to conference › Poster
Open AccessFile -
SACRO: Semi-Automated Checking of Research Outputs
Cole, C., Smith, J., Albashir, M., Bacon, S., Butler Cole, B., Caldwell, J., Crespi Boixader, A., Green, E., Jefferson, E., Jones, Y., Krueger, S., Liley, J., McNeill, A., O'Sullivan, K., Oldfield, K., Preen, R., Robinson, L., Rogers, S., Stokes, P. & Tilbrok, A. & 1 others, , 6 Nov 2023Research output: Book/Report › Other report
-
GRAIMatter: Guidelines and Resources for AI Model Access from TrusTEd Research environments
Jefferson, E., Cole, C., Crespi Boixader, A., Rogers, S., Roche, M., Ritchie, F., Smith, J., Tava, F., Daly, A., Beggs, J. & Chuter, A., 25 Aug 2022, Conference Proceedings for International Population Data Linkage Conference 2022. 3 ed. Vol. 7.Research output: Chapter in Book/Report/Conference proceeding › Conference contribution
Open Access
Equipment
Cite this
- APA
- Author
- BIBTEX
- Harvard
- Standard
- RIS
- Vancouver