Part two
Here we get into the detail of each section of AIME, analysing and critiquing them against questions 5 to 9 of the consultation. Importantly, the response to question 8 below sets out gaps we have identified.
5. Are there any questions that you think are difficult to answer? If yes, what are they? Why are they difficult to answer?
Section 1: AI Systems Record.
The group which considered this section of AIME had the following responses to question 5:
- What is an AI systems record? Most SME organisations will not understand what an AI systems record is and the AIME (as drafted) is not likely to help them understand what it is. This will make AIME Q1.1 difficult to answer. Before providing the explanation as set out, it would be useful to contain commentary along the following lines: “It is very important that you have a list (also called a record) of all of the AI tools that your organisation uses. The reason you need to have such a record is to help you understand what AI systems require your analysis in AIME. It may also help you to categorise your AI system use into different tiers, such as high risk, medium risk and low risk. Having a categorisation like this will help you to know which AI systems to focus on first (i.e. high risk) and which to focus on last (i.e. low risk). If you’re not sure how to categorise AI systems, then [please consult our guidance] OR [we recommend that you categorise them all as high risk]."
- Resource issues. Smaller organisations or start-ups might lack the resources or organisational maturity to create and manage an AI inventory. Alternatively they may have a very rudimentary record which means their answer is not likely to be yes or no.
- Quality matters as much as quantity: Finally, as a general observation which recurs throughout part 2 of our comments, the AIME does not address issues of quality or utility. One organisation may have an excellent, comprehensive and well-structured AI systems record and the other may have a poorly drafted, incoherent AI systems record. Both would be able to pass this question of AIME, but it could not be said that both are following good AI governance processes. AIME would be improved significantly if it were possible for organisations to review and analyse whether their AI systems record is of sufficient quality, meeting minimum mandatory requirements to enable them to answer yes to AIME Q1.1. This could be achieved in various ways including (a) use if AI to scan and comment on AI policies (b) providing a summary of useful topics that a policy must cover with supporting guidance (c) providing template AI policies which users or developers can compare.
Section 2: AI Policy.
The group which considered this section of AIME believed that most of section 2 was clear and the questions were not difficult. However, they also had the following observations:
- What is an AI policy? Just like with the response above regarding AI systems record, further guidance is required to help SMEs understand exactly what an AI Policy is. Recognition also needs to be given to the fact that an AI Policy may be comprised of several documents not just one singular document called “AI Policy”.
- Who is the policy aimed at? The group’s overall impression on the section on AI Policy is that it is primarily inward-looking in nature i.e. towards employees or consultants of an organisation using AI. For example AIME Q2.2 asks whether an AI Policy is available and accessible to all employees. The group suggests an AI Policy should not just be accessible to employees and may be relevant to other parties who are external e.g. customer and suppliers. It may be the case that the AIME tool’s reference to AI policy is too narrow.
- Most difficult question in section 2: The group felt that AIME Q2.3 (“Does your AI policy help users evaluate whether the use of an AI is appropriate for a given function or task?”) was the hardest to answer because it could be widely interpreted. Guidance is needed on how to determine whether use of AI for a task is appropriate. Some case examples would be helpful.
Section 3: Fairness.
The group focussing on fairness spent a lot of time debating how this is likely to be applied in practice. It is clearly one of the more complex areas of AIME because it is easily misunderstood.
- The problem of direct impact. AIME Q3.1 is difficult because directness is a flexible concept which permits a high rate of error based entirely on the judgement of the user. The examples given are also quite widely drafted and the group were of the view that most AI use types would be caught in this description. AIME Q3.1 would actually be easier if the word direct was removed because most users will equate directness and impact as equal terms. In other words, for the average user an indirect impact is no impact at all. Have said that, we question whether it is right to delineate between direct and indirect impact. We think it may be better if the user is entitled to choose between different types of impact. For example, if part of the impact of the AI is to remove jobs from the workforce, this could (arguably) be an indirect impact within the scope of the AIME questions as posed but perhaps should still be considered as part of good baseline governance. An alternative would be to rephrase AIMEQ3.1 as, ‘Have you identified who will be impacted by the AI you are using, and how they will be impacted?’ To cover the full scope of potential unfairness, the list of examples should be broadened to encourage users to think about distribution of benefits arising from the AI use, and less direct impacts of the AI. The example of employees being affected by reduction in jobs, and the demographics of those affected, would prompt users to think about the wider impacts of AI use.
- Being clear about what fairness means. In the context of AI, our group considered that fairness relates to how AI impacts a particular user or group of users. It is mostly about how the organisation decides to implement the AI and what that means for those affected by the decision. In this sense it is very different to bias, which is a broader, more objective, concept relating to the input or training data and the effect that has on the AI model. To help clarify our point, we consider that a totally unbiased AI system can be used in an unfair manner by an organisation if it deliberately or unintentionally uses it to harm or discriminate against certain types of people; this use would be unfair and responsibility for it would rest only with the organisation. Conversely, an organisation may seek to deploy AI systems in a fair manner and have in place lots of checks and balances to ensure that it does so, but use an AI system trained on biased data. In this latter example a user or group of users is harmed, but not on account of unfair use. Instead, it is on account of failure to mitigate bias, responsibility for which may rest with many organisations in the AI ecosystem including AI developers and providers. The guidance to AIME Q3.2 says "We differentiate unfairness from bias, where bias is statistical phenomenon that is characteristic of a process such as decision-making, and unfairness is an outcome of a biased process being implemented in the real world." This is unnecessarily wordy and, whilst the intention may be similar to what we have described above, it falls short.
- Fairness continued: The group analysing fairness concluded that what matters most is that users know, before they start to use an AI system, what that system is designed to do, what sort of outcome it can give, and where they can access support if needed. If this sort of information is provided up front, the likelihood is that the AI system’s use will be fair. To achieve this, organisations should be encouraged to be transparent about the use of AI by providing notices or policies to users. The group acknowledged that there is some crossover here with AIME Q4.1 but would observe that this topic (the topic of impact on individuals) sits more neatly in AIME section 3.
- Monitoring fairness is tricky: The group considered that AIME Q3.4 will be difficult for organisations to implement. To successfully monitor fairness requires a clear understanding of what it is they are being asked to monitor. Many of the participants in the group reviewing this section of AIME thought that SMEs would just conflate monitoring fairness with complaint monitoring generally which may be the only method by which unfairness in AI systems’ use is captured. It is likely to be reactive unless tested voluntarily and pro-actively. It would be appropriate for the DSIT guidance to recommend that this task is designated to an appointed representative in each organisation.
Sections 4 & 5: Impact Assessment & Risk Assessment.
The group focussing on impact assessments and risk assessments considered at length the effect of these requirements on organisations of all sizes. They concluded that larger organisations with well-established governance processes will find these requirements easier than smaller organisations. In particular this group said:
- AIME Q4.1 is difficult to answer because it is not clear what impact could mean, especially as it follows AIME section 3 which appears to address more easily the topic of impact through the mechanism of fairness. AIME Q4.1.1 to Q4.1.3 are hard to answer for that reason and the group felt this section could be set out better, mainly by absorbing it into the section on fairness.
- AIME Q4.2 and 4.3 are likely to be misunderstood by most organisations. In our experience many participants confuse impact assessment with risk assessment and the group analysing this thinks that will probably be the case with most organisations.
- AIME Q5 is easier to understand and follow. The concept of a risk assessment is common to most businesses, although this group observed that not all organisations will appreciate why AI requires a risk assessment in the first place. In the commercial consciousness it may be the case that AI does not feature as highly as health and safety or other practices for which risk assessments are common. The AIME would be improved if it explained why a risk assessment is needed in the first place.
- Separate users and developers: The group was of the view that AIME Q5.3.1 would be challenging for most organisations that are not AI developers due to the lack of access to the underlying mechanism and controls. Most organisations will be able to answer AIME Q5.3.2 because that has a more direct link to outcomes (most organisations should be able to understand if the AI is performing as they expected) but general errors and failures may be too difficult for most organisations to appreciate.
Section 6: Data Management.
The group reviewing section 6 identified the following questions as difficult to answer:
- Separate users and developers: AIME Q6.1 this question is most likely targeted at AI developers and deployers. However, if that is the case, it is not obvious from the context and in fact most organisations will be involved in assisting in model development by continued use of the system. This group thought that sort of model training was not the target of AIME Q6.1 but, if that is the case, the question should be clearer.
- Unrealistic for SMEs: AIME Q6.6 (“Do you document details about the data preparation activities undertaken to develop your AI systems?”) is problematic for smaller organisations. We appreciate that the legal standard requires what is posed, but the question does not recognise the reality of life for small businesses. They generally struggle without legal departments to document contracts with all third parties who process personal data. The majority of staff would not be able to answer this question, as they are often unsure of who deals with contracts. In addition, AIME Q6.6 would be better suited to AIME section 8 (data protection).
Section 7: Bias Mitigation.
Bias easier to grapple with than fairness: the group reviewing section 7 considered the issue of bias mitigation in detail. This group considered that the topic of bias mitigation will be easier for most organisations to understand than fairness (a connected topic, at AIME section 3) but conversely is harder for most organisations to manage. This is because most organisations completing the AIME tool will not have access to or be able to influence training data sets for large AI models at a scale that could impact bias. The AIME tool does not appear to take this into account but it should do so. This is one of many good examples where separation between users and developers would be advantageous.
In particular the group picked out AIME Q7.4 “Do you have processes to ensure compliance with relevant bias mitigation measures stipulated by international or domestic regulation” which they believed would be too difficult for most AI user organisations to answer. This is because, based on the experience of this group, many providers of AI systems will not provide information relating to any bias in their systems. It is extremely difficult to eradicate bias and so it can be argued that these questions, whilst not inherently unnecessary or superfluous, are potentially inefficacious.
Too wide: This group was concerned that the scope of the bias topic is very large and the nature of the questions would yield imprecise answers on account of unfamiliarity with how to tackle the answer. For example, in answering AIME Q7.1 (“Do you take action to mitigate against foreseeable harmful or unfair bias related to the training data of your AI systems?”), organisation Y might take a single action of mitigation in relation to the limited AI systems it uses enabling it to tick yes to Q7.1 (a). In contrast, organisation Z might have a suite of mitigation measures it implements against unfair bias for a majority of its AI systems but will only be able to tick Q7.2 (b). Organisation Z has the more comprehensive AI management system but their answer will generate a lower score than organisation Y (on the presumption that’s how answers will feed into the rating). This means if an organisation genuinely wants to use the tool to understand the state of their AI management systems and practices, their score could be inaccurate unless they think deeply about their answers. If they don’t, AIME could generate a score which gives them unfounded confidence in their AI management systems and practices.
Section 8: Data Protection.
Clean bill of health: The group reviewing section 8 broadly felt that the questions were not difficult to answer, with the exception of AIME Q8.5. The term “interference from third parties” seems far too broad and could create confusion with users.
Section 9: Issue Reporting.
What does good look like? The group reviewing section 9 was of the view that generally AIME Q9 is not difficult to answer, but the questions could be widely interpreted. The challenge when reviewing AIME Q9 is understanding what is intended. The tool would benefit from guidance and examples, as is it not clear what ‘good’ and ‘better’ would look like and it is not clear what a reportable ‘issue’ may be.
For example:
- Who is an external third party for reporting purposes? AIME Q9.1 requires reporting mechanisms for all employees, users and external third parties. It is not clear in this context who an ‘external third party’ would be. Even in large organisation there is not likely to be a method for external third parties to conduct reporting of system failures.
- The poor DPO. Regarding AIME Q9.3 (“Have you identified who in your organisation will be responsible for addressing concerns when they are escalated?”) it would be helpful to provide guidance on who the responsible person should be and in particular what skills they should hold and how they would be expected to fulfil the role. In a small organisation there may not be sufficient resources or risk to justify a dedicated role, whereas a larger developer which provides tools with a more significant risk profile may require a dedicated role. One participant said: “I could be wrong but – as the person who is already DPO and other things besides – the business is just going to make me the point of escalation for AI issues. Is that what you think DSIT is getting at here? If not me then who?”
- Can one be unintentionally transparent? Whilst ‘transparency’ in AIME Q9.4 is defined, it is not clear what ‘meaningfully’ adds to that definition, and perhaps would benefit from widening the definition to include what is intended by ‘meaningfully’, with examples. As with answers to other questions, particularly those regarding fairness or AI policy, we think transparency is an important thread that should run through much of the AIME tool when encouraging good compliance.
- Time scales: When considering the time scale in responding to concerns (AIME Q9.5), the group was of the view that an appropriate timely response will very much depend on the severity of the issue being reported. The current requirement for all issues to be responded to within 72 hours does not address the fact that some issues will require a response than others. We appreciate the rationale for using 72 hours and, subject to what we say above, it provides a useful benchmark.
- What goes in the response? AIME Q9.5 is unclear about what a ‘response’ would entail. Would an acknowledgement be sufficient or is it intended that the issue will be resolved within 72 hours? Some concerns may just require an automated response, acknowledging receipt, but others may require a fuller response, and a more timely approach. Over reporting may lead to reputational damage, and so the required response should take into account the level of control the business has over the tool, such as whether a deployer or developer.
- Take action? Once a concern is reported, AIME Q9.6 refers to documenting the report, but perhaps this would be better to include a process to follow in addition to documenting the report. Merely recording the issue may be good practice, but considering what action may be needed following the report would nudge the user to ensure any mitigation etc was put in place.
Section 10: Third Party Communications.
The group which analysed AIME Q10 were of the view that the questions were extremely broad.
- Binary problem pops up again: The fact that organisations can answer “yes for some” means it is very difficult to identify what good practice would look like in relation to any of the questions.
- Non-technical documentation: Regarding AIME Q10.3 in particular it would be useful to have clearer examples of what should be included in non-technical documentation. Would this include environmental impact, workforce impact, societal impact etc?
- Problems for non-developer SMEs: Answering any of the questions in AIME section 10 may be challenging for organisations that are not a developer or don’t have somebody within a regulatory/legal role. Some examples would help clarify what the question aims to achieve/elicit. The way in which a user interprets the questions in section 10 could yield vastly different responses depending on their role. For example, a developer would very likely be able to explain in detail (and understand) the technical aspects of their AI system, but a deployer/end-user is much less likely be able to provide this type of information and/or understand it. AIME Q 10.1 and 10.2, seem as if they are very much aimed at an AI developer, but an AI end-user should perhaps be seeking this type of technical documentation from the developer of the AI tool they are considering using.
6. Are there any questions that you think are superfluous/unnecessary? If yes, what are they? Why are they helpful/appealing?
Section 1: AI Systems Record.
The group which considered this section of AIME had the following responses to question 6:
- Most of the questions seem appropriate but AIME Q1.3 could be superfluous if the user already maintains an AI system record per AIME Q1.1. These questions could perhaps be merged.
Section 2: AI Policy.
The group which considered this section of AIME had the following responses to question 6:
- Most of the questions seem appropriate but AIME Q2.2 could be redundant. Accessibility is the main concern that could be addressed as part of broader governance practices rather than requiring a dedicated question. It could perhaps be rephrased as “Is your policy accessible to those who need to access it in a clear, easy-to-use format?”
Section 3: Fairness.
The group which considered this section of AIME continued to hold the view that fairness and bias mitigation are conflated by AIME. None of the questions were obviously redundant but, continuing the theme, they observed that:
- Fairness is still tricky: The questions in AIME Q3 are thematically closer to questions of bias rather than fairness.
- What is fairness in AI? AIME Q3.2 requires a better explanation of fairness (this may form part of the guidance) which was tackled in the response to question 5 above. In our view fairness is more about how the AI’s use impacts the individual exposed to the AI’s operations. For example, bias (as AIME points out) is more closely aligned to a statistical phenomenon in the data sets on which the AI system operates. What matters most is that those using the AI system are aware they are doing so and what the most likely outcomes of that use are, so that they can make informed decisions about what happens next. The most obvious method of transparency is something like a published AI policy, but other more innovative options should be encouraged too. The means should not determine the ends; what matters most is that users know, before they start to use an AI system, what that system is designed to do, what sort of outcome it can give, and where they can access support if needed.
Section 4 & 5: Impact Assessment & Risk Assessment
Merge AIME Q4 and Q5: The group which considered these sections of AIME did not think any of the questions were obviously redundant but, continuing with previous themes, they observed that it is very easy to confuse section 4 (impact assessment) and section 5 (risk assessment). While the questions in AIME section 4 are important, this group thinks it makes sense to think about the impacts as associated with risks (e.g. see the repeated use of the word "potential" in this section). They suggest a framework tying these together in a more signposted way or, if they are intended to be different, AIME should make it very clear how they are different.
Section 6: Data Management.
The group which considered this section of AIME made the following observations:
- Separate user and developers: AIME section 6 is a useful example of a topic which is better suited to developers of AI than users of AI. Not all questions are relevant to all respondents and in that sense this section will be superfluous to most small-scale AI users. AIME section 6 is almost entirely for developers of ML (Machine Learning) solutions, for example. In all cases, a discussion of the reason or motivation for each question and how organisations might answer them, or assign them to qualified resources to answer, would help to resolve this question of relevance of questions, as well as assisting users.
- Merge AIME sections 6 and 8: In a similar vein, this group questioned the need for both AIME section 6 (Data management) and AIME section 8 (Data protection). The questions in section 6 are primarily targeted at companies training their own models except for 6.6 which is about personal data and could be more closely related to AIME section 8. Some of the questions in AIME section 6 are too detailed and technical for a tool aimed at SME organisational processes, such as 6.3 (about data quality) - which might subsume 6.4 (about representativeness). This group wished to encourage DSIT to revisit these questions.
Section 7: Bias Mitigation.
Clean bill of health: The group which considered this section of AIME did not consider that any of the questions were obviously superfluous, but reiterated the views from the response to question 5 above, meaning that the topic itself is likely to feel redundant to many SME organisations which are unable to influence bias in data sets where they are deployers of AI.
Section 8: Data Protection.
A little bit light: The group which considered this section of AIME did not consider that any of the questions were obviously superfluous. However, they did pose the question whether the right data protection questions are being asked here. In simplifying the area of data protection to a handful of core questions, the recommendation of the group is to focus on the concepts of lawfulness, fairness and transparency, rather than or in addition to some of the questions currently presented in AIME section 8. The introductory paragraph talks about an overarching data protection by design and default approach and, whilst the group agreed that it is important to understand that organisations have appropriate technical and organisational measures in place, they thought that the questions could have then focussed on some of the more fundamental elements of data protection law.
Section 9: Issue Reporting.
Interchangeable terms? The group reviewing section 9 of AIME was of the view that most of the questions in this section serve a useful purpose. However, section 9 uses both the phrase “reporting mechanism” and “reporting procedure” and there was some confusion about whether they were intended to mean different things. We would encourage DSIT to use one term or to clarify the difference between them. The current wording provides that if the answer to question 9.1 about having a “reporting mechanism” is ‘no’ the respondent is sent to question 9.4 asking if the “reporting procedures” are transparent (see our previous comments on this). Depending on whether the two phrases are used interchangeably (there are no definitions), it may suggest that if the answer to 9.1 is ‘no’ skipping to 9.4 is superfluous.
Section 10: Third party communications
Binary questions: The group reviewing section 10 of AIME was of the view that most of the questions in this section serve a useful purpose. However, some of the multiple-choice answers offer a “yes, for some” option which seems superfluous. What does this mean if the user has only provided “some” technical or non-technical documentation to “some” of their stakeholders? This feels like an unnecessary question as there can be no conclusion or inference with regards to whether the responder is doing the right thing with regards to their AI use.
Separate users and developers: In section 10 in particular AIME would benefit greatly if there were 2 question sets: one set aimed at developers and the other aimed at end-user organisations of AI. Alternatively, this could be resolved by a separation between these groups as we recommend at the outset of this consultation document. In both cases it would be helpful to have clarity around the purpose of the technical and non-technical documentation and the reasons for making this documentation available to stakeholders.
7. Are there any questions that you particularly liked or would find helpful for improving your internal processes? If yes, what are they? Why are they helpful/appealing?
Sections 1 and 2: AI Systems Record & AI Policy
The group which considered these sections of AIME was of the view that there was nothing further to add in response to this question, save for comments already made above.
Section 3: Fairness.
More of the same: The group which considered section 3 of AIME was of the view that it is helpful to prompt organisations to consider fairness in relation to AI as it recognises that fairness is different from bias and can arise in the absence of bias. Transparency and consultation over the use of AI are key to improving fairness and it would be helpful to see this referenced in AIME Q3.
Section 4: Impact Assessment
Broad coverage of issues here is good: The group which considered section 4 of AIME commented that AIME adequately asks respondents to consider a comprehensive breadth of the internal processes that they would expect to see in a startup or SME achieving good hygiene regarding the development and/or deployment of AI systems, but equally it was useful for larger organisations too. This section of AIME was commended for its consideration of a wide range of impact areas - particularly society, the environment, and end-users/customers. As organisations of any size have the potential to create significant impact and risk exposure through the development and/or deployment of AI systems, it is critical that everyone choosing to develop and/or deploy AI systems understands and actively considers the benefits, costs, trade-offs, and responsibilities associated with those choices. However, there was still a lot of challenge from the group as to what the difference was between AIME sections 4 and 5.
Section 5: Risk Assessment
Separate users and developers: The group which considered section 5 of AIME commented that AIME Q5.4 and 5.5 were considered particularly valuable because they ask respondents to consider their processes for responding to or repairing AI system failures and under what conditions they would stop the development or use of AI systems in their organisation. Such failures are inevitable and it is vital that those developing and/or deploying AI think critically about the limitations, fallibility, and vulnerabilities of such systems prior to failure occurring so they are able to detect failure early and respond effectively, intentionally, and consistently to minimise potential harms. Ultimately this is not relevant to users. If expectations about the depth and detail of responses are set, the questions cover sufficient breadth of the subject area that the responses could themselves be used to iteratively improve AIME.
Section 6: Data Management
Clean bill of health: The group which considered section 6 of AIME did not have any further comments to provide in response to question 7.
Section 7: Bias Mitigation
The group which considered section 7 of AIME commented that they found question 7.4 helpful because it may help focus internal stakeholders on their legal obligations and demonstrate the compliance incentives for effectively managing AI usage. However, as pointed out above, they went on to observe that organisations may struggle to effectively answer bias mitigation questions unless they are AI developers.
Section 8: Data Protection.
Lack of specificity: The group which considered section 8 of AIME did not have anything further to add regarding question 7 but they felt that the data protection questions lacked specificity because they didn’t address the areas (lawfulness, fairness and transparency) which are most relevant to data protection by design and default.
Section 9: Issue Reporting.
Useful checklist: The group which considered section 9 of AIME said that the overall structure of section 9 is positive because it provides a clear framework for addressing issue reporting within any organisation. It includes key questions such as whether there is a designated person responsible for handling AI-related issues and if there are established reporting mechanisms or procedures. This can serve as a useful checklist for SMEs, helping them to ensure that they have the necessary processes in place to manage and respond to AI-related issues effectively. The group also likes that the section emphasises the need for differentiation in reporting requirements based on the audience, such as whistleblowers, end users, and deployers/developers, but clarity on these roles would be useful. This is particularly helpful as it highlights the importance of tailoring responses based on the severity and nature of the issue. Additionally, it draws parallels to the role of DPO, suggesting a framework where the responsibilities scale with the volume of AI usage, sensitivity of decisions, and company size.
Whistleblowing anonymity should be encouraged: The group was of the view that AIME Q9.2 is particularly positive because it requires organisations to provide individuals with the option of reporting concerns on an anonymous or confidential basis (or both). This is important because, similar to a whistleblowing process, enabling concerns to be raised anonymously will increase the likelihood of serious issues being reported.
Section 10: Third Party Communication
Separate users and developers: The group which considered section 10 of AIME liked the inclusion of non-technical documentation. They felt this is useful because it requires businesses to acknowledge that AI tools may have wider impacts than their technical operation. But as noted in their response to question 5, group members recommend that DSIT provide more guidance in relation to what the non-technical documentation should include (environmental impact, workforce impact, societal impact etc.) and reconsider whether this was more for developers rather than users.
8. Are there any necessary conditions, statements or processes that you feel are missing that organisations should be implementing? What are they?
General comment about missing topics.
There were many occasions where the groups working on this consultation response identified that AIME seemed to omit important topics. Most participants felt that these missing topics should be included in the final AIME tool because they bear relevance to the issues which AIME is attempting to tackle. However, a minority of participants disagreed and were of the view that AIME is already quite long and challenging, particularly for SMEs, and so adding more content was undesirable. There was universal agreement in the group that if other recommendations in this consultation response were adopted (e.g. consolidating certain sections such as AIME 4 and 5 or AIME 6 and 8) then other topics could be added into the space created by question consolidation.
Most important missing topics
Missing Topic 1: Sustainability and AI
All participants were surprised that AIME did not have a section that tackled AI sustainability as a standalone topic. We acknowledge that it forms part of AIME Q4.1.3. However, we expected greater detail on this issue. Sustainability is a broad concept which should include how use of AI impacts the natural environment, community environment and social environment. It is our view that many organisations are concerned about how their use of AI will impact on these factors. Most organisations have spent time and resource understanding their sustainability impact and have manifest their concern in ESG commitments. There is some limited appreciation amongst the business community that the more they rely on AI, the greater harm they may do to their sustainability targets.
AIME could be improved by posing a question such as the following:
- “Have you considered how your use of AI may impact commitments regarding sustainability?” [This is could in turn link to guidance about sustainability]
- “Would your use of AI create societal issues, such as significant changes working patterns or worker displacement”?
Missing Topic 2: Transparency, Explainability and Interpretability of AI
On many occasions participants discussed issues of organisational transparency regarding AI use. There should be processes in place to help organisations address:
- transparency (meaning they communicate their use of AI in a way that is accurate, brief and clear and easily discernible);
- explainability (meaning they are able to communicate objectively how the AI system reaches its conclusions); and
- interpretability (meaning they are able to interpret the outcomes in each subjective use case).
AIME would be improved by requiring AI users to answer questions about these points, and to have in place processes or requirements for ensuring that AI systems use types and transparent and organisations provide explainable and interpretable results, especially for critical decision-making applications. This is important because they will help build trust with users and accountability with regulators, especially for high-stakes decisions (e.g., hiring, medical diagnosis, or credit scoring). A suggested additional question could be “Do you ensure that your AI systems provide explainable and interpretable outputs, particularly in high-impact use cases?”
Missing Topic 3: Intellectual Property Rights
All participants were of the view that AIME should, in some form or another, address the issue of intellectual property rights compliance. This is a very popular and complex topic in the domain of AI and not one we believe should be tackled in detail via AIME, but it would be sensible for organisations to stop and consider whether their potential use of AI could result in use of proprietary materials in an unlawful fashion. We appreciate that there is a separate consultation regarding copyright and generative AI co-led by the Intellectual Property Office but we do not think that should detract from a general question regarding this topic in AIME.
Additional missing topics
Missing Topic 4: Ethical decision-making processes
Some participants were of the view that a statement or process for ensuring ethical decision-making during AI system design, deployment, and use would be beneficial. AI ethics is a pervasive topic and one that continues to arise in professional and non-professional settings. Ethics is a cornerstone of responsible AI but is only indirectly addressed in the current AIME tool. Depending on the use cases, organisations should have a documented framework for resolving ethical dilemmas (e.g., prioritising transparency versus performance or balancing fairness against utility). This could be resolved by including a question such as: “Do you have an ethical decision-making framework for guiding AI system development and deployment?”
Missing Topic 5: Stakeholder involvement
Some participants felt that AIME would be complimented by including questions about involving diverse stakeholders (e.g., employees, end-users, affected communities) in the design, testing, and monitoring of AI systems. This is important because involving stakeholders can help ensure AI systems reflect a range of perspectives and minimise unintended harms. It may also help to increase trust and acceptance. This could be addressed by including a question such as: “Do you involve stakeholders in the design, testing, and evaluation of your AI systems?”
Missing Topic 6: Human oversight
Some participants were of the view that AIME should include a clear statement on the role of human oversight in AI decision-making, especially for critical or high-risk applications. AIME does not explicitly address whether a human is actively involved in overriding AI decisions or ensuring accountability for high-impact outcomes. This is important because human oversight ensures accountability and provides a failsafe for preventing harm or mitigating errors caused by AI. This could be addressed by including a question such as: “Do you define and implement processes for human oversight of AI systems in critical decision-making workflows?”
Sections 1 and 2: AI system record & AI policy
Clean bill of health: The group which considered sections 1 and 2 of AIME were not of the view that there were missing conditions, statements or processes regarding AI system records, other than observations made in previous answers.
Section 3: Fairness
Senior leadership to be involved in fairness: The group which considered section 3 of AIME were not of the view that there were missing conditions, statements or processes regarding AI policies, other than observations made in previous answers. However, they did add that the concept of fairness is one that should be considered by senior leadership of an organisation. The group recommend adding a question or guidance of the kind of questions the board or senior leadership team should think about, or asking the developer, supplier, or whoever in their organisation is responsible for implementing, managing, training and monitoring the AI - critically - before they start using it.
Section 4: Impact Assessment
Due diligence: The group which considered section 4 of AIME were of the view that it was missing questions or commentary about minimum standards of due diligence when selling or procuring AI systems, for example whether the supplier provided detailed specifications, an explanation of the full supply chain and a description of the security and ethical standards they adhere to.
SMEs: This group also said that their concerns were less about areas that are missing from this section and more about a lack of context and explanation for the items that are covered here. Small or medium sized businesses may find it hard to understand the questions and how to achieve compliance in the relevant areas because definitions and guidance are missing. In all likelihood, small businesses will have to answer “no” to most questions as their internal processes will not be as sophisticated to consider and manage all of the topics referred to in the questionnaire. It would be helpful to signpost businesses to advice that helps them better understand the requirements, which ones to prioritise and what steps to take to become compliant.
This group also suggested that users should either be prompted to complete the questionnaire separately for each AI use case, or explain the implications if business provide a single answer for all their AI systems and use cases.
Section 5: Risk Assessment
Clean bill of health: The group which considered section 5 of AIME were not of the view that there were missing conditions, statements or processes regarding AI system records, other than observations made in previous answers.
Section 6: Data Management
The group which considered section 6 of AIME were not of the view that there were significant missing conditions, statements or processes regarding AI system records, other than observations made in previous answers. However, they did observe that some organisations may be unclear about:
- what is meant by AI training in AIME Q6.1 and further guidance on this should be provided.
- what constitutes a written contract in AIME Q6.6. Organisations may only think of bespoke signed contracts, rather than the multitude of contract terms they frequently accept when using technology or online services.
Section 7: Bias Mitigation
Quality and utility: The group that reviewed section 7 of AIME considered that AIME Q7.1 omits several key steps in bias mitigation and does not address the quality of the mitigation. They observed that organisations must first take steps to detect bias mitigation which to some extent may be covered by section 4 (impact assessment) or 5 (risk assessment) but this was not obvious or overtly acknowledged in AIME. AIME section 7 should contain questions about measuring the effectiveness of the mitigating steps taken to help ensure they are effective.
Section 8: Data Protection
Lack of specificity: The group which reviewed section 8 were of the view that this section omits a number of data protection issues which they considered would be more helpful and more relevant to users than the current questions. For example:
- there is no mention in section 8 of AIME regarding lawfulness, fairness, transparency (e.g. privacy notices), data minimisation, accuracy, or storage limitation.
- there is no reference to having a lawful basis of processing for personal data processed using AI systems.
Some of this guidance could easily be replicated from existing ICO guidance on AI use. Including it in the AIME tool would enable users to understand some of these key points without having to open a new browser page for the ICO guidance which, whilst good, can be unnecessarily lengthy at times.
Section 9: Issue Reporting
The most useful point of escalation is… The group that reviewed section 9 of AIME considered that it was missing questions which are necessary to establish whether user organisations have the ability to escalate concerns to developers. Without an ability for concerns to be escalated to the developer, and obligations on developers to address those concerns, the deploying and using organisations will have little (if any) ability to actually fix the concerns which have been identified. They also observed that AIME Q9.3, which relates to who in an organisation is responsible for addressing concerns when they are escalated, could be limiting. Perhaps rewording this question to: “does your organisation have an internal procedure and identified contacts for addressing concerns when escalated” would provoke organisations to implement a proper escalation process rather than simply singling out one individual. AIME Q9.3 also says nothing about the necessary skills, qualities and competence of the appointed person(s) and how to pick such a person should the organisation not have made any such appointment.
Section 10: Third Party Communication
Proportionality: The group that reviewed section 10 of AIME considered that a key omission from this section is an explanation of the purpose of producing and disseminating technical and non-technical documentation. The need for, and content of, such documentation has to be proportionate and risk-based and not place additional transparency burdens on organisations that have no purpose or benefit to the end-user. It may be that this is dealt with in supplementary guidance which DSIT is producing. Clear guidance of expectations of the content of such documentation depending on the organisation’s role in developing, deploying or using AI would be helpful. For example, do all organisations need to declare what models are being used and/or what model has been developed by a third party and/or a first party (proprietary)?
In AIME Q10.3 and Q10.4, the definition of non-technical documentation appears to be aimed at a B2B scenario and is missing any reference to a generative AI notice for end-users, especially where the risk threshold is higher. For example, where AI is making automated decisions about end-users.
Transparency: Informing end-users of AI use where there are potential risks is an important part of third-party communication. Guidance as to how these risks should be communicated would be welcomed and it may be that this is contained in accompanying guidance being prepared by DSIT; this topic is closely connected to the issue of transparency which all groups identified at various points and which we have addressed earlier in this response. This transparency information could rightly sit within an existing privacy notice (as is already required for AI automated decisions) or in some cases, where the level of risk requires more a detailed explanation, within a generative AI notice in itself.
9. Is the tool overly burdensome or unrealistic for the target audience (i.e. organisations with limited resources to extensively engage with AI governance frameworks, for example, start-ups and SMEs)?
General Observations.
The various sub-groups considered this question regarding the specific sections of AIME allocated to them. Generally the feeling of all groups was similar: the AIME tool is well-designed for start-up and SME users and clearly it has been created with them in mind. The AIME tool is likely to be suitable for organisations which are smaller and starting to get to grips with AI in their organisations. Some participants thought that it could be further simplified or better tailored to address the resource constraints of start-ups and SMEs.
One participant said that: Why AIME Is Not Burdensome
“The tool is designed as a self-assessment, which avoids the need for external consultants or complex audits which is great. This approach is relatively accessible for smaller organisations. It is broken into clear sections (e.g., AI Policy, Risk Assessment, Bias Mitigation), making it manageable for organisations to address one area at a time rather than tackling everything at once. The inclusion of glossary terms and inline explanations is good and ensures that non-experts can engage with the content without extensive prior knowledge of AI governance. The tool is voluntary, which gives organisations the flexibility to adopt the recommendations incrementally rather than all at once. It targets baseline practices like maintaining AI system records and ensuring data quality, which are necessary for even minimal AI governance. While the tool itself is voluntary, implementing the recommended actions (e.g., conducting risk assessments or bias audits) can be time-consuming and resource-intensive for organisations with small teams or limited budgets but that is not the fault of AIME per se.”
But some participants questioned whether AIME was trying to be all things to all people and so losing its purpose along the way: Luke warm
“Fundamentally I think many of the concepts and questions have real merit and DSIT has included the right things, but I feel some questions worded in their current form may be beyond some businesses but I get the feeling that the AI experts in our group find the questions (or the wording of the questions) so basic as to be not particularly helpful either. I think DSIT have tried to find a middle ground which has ended up being perhaps too burdensome for beginners and too basic for developers.”
However, there was considerable debate about whether it would be suitable for larger organisations. The majority of participants considered that it would be of some use, but acknowledge that in some respects it creates more questions than answers for those organisations that already have comprehensive governance infrastructure. Some participants suggested that it would be avoided by larger organisations due to its simplicity.
Section 1: AI system record.
Quality and utility: The group that reviewed section 1 of AIME did not feel that this section was overly burdensome but did query whether it was realistic to expect organisations to have an AI systems record in place. There is the potential for great disparity between AI system records for all organisations, ranging from none at all through to complex system records. The group suggested that DSIT should work on the assumption that most SME organisations would not have such a record unless they were directly involved AI development or deployment. As we have said elsewhere, quality and utility matter are very important for good governance frameworks.
Section 2: AI policy
Clean bill of health: The group that reviewed section 2 of AIME did not feel that this section was overly burdensome or unrealistic for the target audience and had nothing further to add beyond previous comments about section 2 (a few of which overlap).
Section 3: Fairness
The group that reviewed section 3 of AIME thought that some questions may be overly burdensome for start-ups and SMEs deploying widely-available AI tools, although the accompanying DSIT guidance may provide clarity over this.
Sections 4, 5 and 6: Impact Assessment, Risk Assessment and Data Management.
The group that reviewed sections 4, 5 and 6 of AIME thought that some questions may be overly burdensome and commented as follows:
- Separate users and developers: the AIME tool may be overly burdensome simply because it asks questions in each of these sections that should probably just be posed to developers. It is unduly burdensome on SMEs to be expected to have in place a AI impact assessment or to understand what AI training involves. This group repeated earlier calls for AIME to segment users and developers.
- Missing guidance: Answering this question is also difficult because the group observed that they are looking at the tool without the associated guidance. It may be the case that, with the guidance, some of the concerns above for non-developer users may be allayed.
Section 7: Bias Mitigation
The group that reviewed section 7 of AIME had the following observations:
- Bias a burden for some SMEs: The AIME tool is relatively straightforward and not inherently burdensome or unrealistic, it does have the potential to be more burdensome to start-ups and SMEs. This is in part because resource limitation will play a factor in the scope of due diligence that organisations are able to carry out, as referred to in AIME Q7.1 and Q7.3. Additional difficulties may be encountered by small organisations who may not have established, or have the ability to maintain, the relevant processes and measures.
- SME resource issues: A lack of technology and/or legal literacy may also increase the burden on small organisations, particularly where they are simply AIaaS users. This is noticeable in relation to AIME Q7.3 and Q7.4, where recognising what constitutes appropriate due diligence, or understanding which international or domestic regulations should be consulted on the topic of bias mitigation, is likely to be prove difficult.
- AIME Q7.2 and 7.3 are unrealistic for the stated target audience of AIME, namely, start-ups and SMEs. Consequently, the inclusion of these questions is unlikely to provide any helpful insight into whether an organisation adopts responsible AI management practices.
Section 8: Data Protection
The group that reviewed section 8 of AIME had the following observations:
- the AIME tool in this section is not burdensome or unrealistic per se.
- a lack of technology and/or legal literacy may also increase the burden on small organisations, particularly where they are simply AIaaS users. This is noticeable in relation to AIME Q8.3 and Q8.4 that presuppose some form of prior data protection knowledge and they noted that many of the questions in AIME section 8 (8.2 – 8.4) do not specifically relate to AI, which may confuse some users.
Section 9: Issue Reporting
The group that reviewed section 9 of AIME had the following observations:
- The questions are generally not overly burdensome for organisations to answer but the lack of guidance may make it difficult to all organisations (regardless of size) to understand what is required in terms of having reporting mechanisms for external third parties to report concerns and system failures (per AIME Q9.1).
- The obligation to provide timely responses within 72 hours (per AIME Q9.5) may also be unduly burdensome depending on the nature of the issues reported and depending on what is required by way of response.
Section 10: Third Party Communications
The group that reviewed section 10 of AIME had the following observations:
- The ability to answer ‘yes for some’ to all of the questions in section 10 greatly reduces the burden because it doesn’t appear to necessitate a particular baseline of compliance in relation to the provision of documentation. However, this lack of specificity reduces the value of the responses because it does not demonstrate compliance.
- As noted in previous answers, some of the questions may be difficult for non-developers (both SMEs and large organisations) to answer because they won’t have access to (or sufficient understanding of) the necessary information.