E. Jefferson, Christian Cole, Alba Crespi i Boixader, Simon Rogers, Maeve Malone, F. Ritchie, Jim Q. Smith, Francesco Tava, A. Daly, J. Beggs, Antony Chuter
{"title":"GRAIMatter:TrusTEd研究环境中的人工智能模型访问指南和资源(GRAIMatter)。","authors":"E. Jefferson, Christian Cole, Alba Crespi i Boixader, Simon Rogers, Maeve Malone, F. Ritchie, Jim Q. Smith, Francesco Tava, A. Daly, J. Beggs, Antony Chuter","doi":"10.23889/ijpds.v7i3.2005","DOIUrl":null,"url":null,"abstract":"ObjectivesTo assess a range of tools and methods to support Trusted Research Environments (TREs) to assess output from AI methods for potentially identifiable information, investigate the legal and ethical implications and controls, and produce a set of guidelines and recommendations to support all TREs with export controls of AI algorithms. \nApproachTREs provide secure facilities to analyse confidential personal data, with staff checking outputs for disclosure risk before publication. Artificial intelligence (AI) has high potential to improve the linking and analysis of population data, and TREs are well suited to supporting AI modelling. However, TRE governance focuses on classical statistical data analysis. The size and complexity of AI models presents significant challenges for the disclosure-checking process. Models may be susceptible to external hacking: complicated methods to reverse engineer the learning process to find out about the data used for training, with more potential to lead to re-identification than conventional statistical methods. \nResultsGRAIMatter is: \n \nQuantitatively assessing the risk of disclosure from different AI models exploring different models, hyper-parameter settings and training algorithms over common data types \nEvaluating a range of tools to determine effectiveness for disclosure control \nAssessing the legal and ethical implications of TREs supporting AI development and identifying aspects of existing legal and regulatory frameworks requiring reform. \nRunning 4 PPIE workshops to understand their priorities and beliefs around safeguarding and securing data \nDeveloping a set of recommendations including \n \nsuggested open-source toolsets for TREs to use to measure and reduce disclosure risk \ndescriptions of the technical and legal controls and policies TREs should implement across the 5 Safes to support AI algorithm disclosure control \ntraining implications for both TRE staff and how they validate researchers \n \n \n \nConclusionGRAIMatter is developing a set of usable recommendations for TREs to use to guard against the additional risks when disclosing trained AI models from TREs.","PeriodicalId":36483,"journal":{"name":"International Journal of Population Data Science","volume":" ","pages":""},"PeriodicalIF":1.6000,"publicationDate":"2022-08-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"GRAIMatter: Guidelines and Resources for AI Model Access from TrusTEd Research environments (GRAIMatter).\",\"authors\":\"E. Jefferson, Christian Cole, Alba Crespi i Boixader, Simon Rogers, Maeve Malone, F. Ritchie, Jim Q. Smith, Francesco Tava, A. Daly, J. Beggs, Antony Chuter\",\"doi\":\"10.23889/ijpds.v7i3.2005\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"ObjectivesTo assess a range of tools and methods to support Trusted Research Environments (TREs) to assess output from AI methods for potentially identifiable information, investigate the legal and ethical implications and controls, and produce a set of guidelines and recommendations to support all TREs with export controls of AI algorithms. \\nApproachTREs provide secure facilities to analyse confidential personal data, with staff checking outputs for disclosure risk before publication. Artificial intelligence (AI) has high potential to improve the linking and analysis of population data, and TREs are well suited to supporting AI modelling. However, TRE governance focuses on classical statistical data analysis. The size and complexity of AI models presents significant challenges for the disclosure-checking process. Models may be susceptible to external hacking: complicated methods to reverse engineer the learning process to find out about the data used for training, with more potential to lead to re-identification than conventional statistical methods. \\nResultsGRAIMatter is: \\n \\nQuantitatively assessing the risk of disclosure from different AI models exploring different models, hyper-parameter settings and training algorithms over common data types \\nEvaluating a range of tools to determine effectiveness for disclosure control \\nAssessing the legal and ethical implications of TREs supporting AI development and identifying aspects of existing legal and regulatory frameworks requiring reform. \\nRunning 4 PPIE workshops to understand their priorities and beliefs around safeguarding and securing data \\nDeveloping a set of recommendations including \\n \\nsuggested open-source toolsets for TREs to use to measure and reduce disclosure risk \\ndescriptions of the technical and legal controls and policies TREs should implement across the 5 Safes to support AI algorithm disclosure control \\ntraining implications for both TRE staff and how they validate researchers \\n \\n \\n \\nConclusionGRAIMatter is developing a set of usable recommendations for TREs to use to guard against the additional risks when disclosing trained AI models from TREs.\",\"PeriodicalId\":36483,\"journal\":{\"name\":\"International Journal of Population Data Science\",\"volume\":\" \",\"pages\":\"\"},\"PeriodicalIF\":1.6000,\"publicationDate\":\"2022-08-25\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"International Journal of Population Data Science\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.23889/ijpds.v7i3.2005\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q3\",\"JCRName\":\"HEALTH CARE SCIENCES & SERVICES\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"International Journal of Population Data Science","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.23889/ijpds.v7i3.2005","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"HEALTH CARE SCIENCES & SERVICES","Score":null,"Total":0}
GRAIMatter: Guidelines and Resources for AI Model Access from TrusTEd Research environments (GRAIMatter).
ObjectivesTo assess a range of tools and methods to support Trusted Research Environments (TREs) to assess output from AI methods for potentially identifiable information, investigate the legal and ethical implications and controls, and produce a set of guidelines and recommendations to support all TREs with export controls of AI algorithms.
ApproachTREs provide secure facilities to analyse confidential personal data, with staff checking outputs for disclosure risk before publication. Artificial intelligence (AI) has high potential to improve the linking and analysis of population data, and TREs are well suited to supporting AI modelling. However, TRE governance focuses on classical statistical data analysis. The size and complexity of AI models presents significant challenges for the disclosure-checking process. Models may be susceptible to external hacking: complicated methods to reverse engineer the learning process to find out about the data used for training, with more potential to lead to re-identification than conventional statistical methods.
ResultsGRAIMatter is:
Quantitatively assessing the risk of disclosure from different AI models exploring different models, hyper-parameter settings and training algorithms over common data types
Evaluating a range of tools to determine effectiveness for disclosure control
Assessing the legal and ethical implications of TREs supporting AI development and identifying aspects of existing legal and regulatory frameworks requiring reform.
Running 4 PPIE workshops to understand their priorities and beliefs around safeguarding and securing data
Developing a set of recommendations including
suggested open-source toolsets for TREs to use to measure and reduce disclosure risk
descriptions of the technical and legal controls and policies TREs should implement across the 5 Safes to support AI algorithm disclosure control
training implications for both TRE staff and how they validate researchers
ConclusionGRAIMatter is developing a set of usable recommendations for TREs to use to guard against the additional risks when disclosing trained AI models from TREs.