Ethical, legal and privacy risks: Accounts for data privacy, fairness, bias and regulatory risks by assessing the use case’s utilization of AI model type, corporate intellectual property, and sensitive personal data.
Data, algorithmic and development risks: Accounts for technical complexity by assessing risks based on input data, tech stack, model utilized, model output and development methodologies.
Business risks: Accounts for potential harms to the organization by assessing impact to revenue, customer experience, regulatory compliance and corporate reputation.
The outcome of a use case’s risk tier classification will determine two things:
- Specific RAI controls and model monitors to be implemented with the AI system in production. Monitors can include tracking of performance accuracy, biases, end-user misuse, adversarial anomalies and more. RAI controls can include governance activities such as tracking use case compliance against data privacy, data retention, security, IT, and/or applicable legal policies.
- The ongoing frequency of AI system reviews and evaluation from various groups within the AI governance structure’s three lines of defense throughout the AI system’s lifespan in production.
Pillar 3 - Solution development lifecycle embedded Responsible AI
Every phase in the AI system development lifecycle has potential risks that should have proper oversight. Developing with responsible AI considerations across all projects will enable companies to produce inherently risk-adverse AI systems. There are various categories of risk that exist across the AI lifecycle:
- Use case initiation phase: design risks
- Data acquisition and preparation: data risks
- Model training, experimentation and validation: algorithmic risks
- Deployment and monitoring: performance risks
The EY Responsible AI framework guides development teams to check and analyze for the specific risks that could occur during each phase. Risk-mitigation activities should be incorporated as part of the standard solution development lifecycle to address the nuanced risks that arise for a specific business problem.
For example, if the data set required for training AI models is not readily available due to lack of digitization, disorganized data domains, or scattered ownership of data across various business units, etc., then analysis of how imbalanced a data set is before moving forward with model training is a critical part of mitigating data and algorithmic risks. It is also important to help ensure transparency into what kinds of data set was utilized for model training, as it informs the potential scenarios of bias and inaccuracy that can occur from the AI system’s outputs. Depending on the business problem the AI system is intended to solve, biased outputs could be statistical, resulting in performance risks such as inaccurate insights given to the business. Social biases can also occur with AI systems that utilize human demographic data, such as those commonly employed in HR functions.
Pillar 4 - Proactive monitoring and controls
As an AI program scales, risk mitigation monitors will need to be identified and streamlined. Specific monitors should be configured based on use case risks identified. Adhering to safe AI practices requires monitoring across several areas to account and adjust for potential risks, many of which have been heightened by the advent of GenAI. A selection of key risk categories are noted below:
Hallucination: Generation of outputs or conclusions by an AI system that are not grounded in its training data or input provided, leading to potentially incorrect, nonsensical or harmful responses.
- Deter the model from producing unfounded or imaginary content.
- Set parameters to screen content outside of prescribed use case.
- Create metrics to identify anomalies across various dimensions that might signal hallucination
Data Leakage: The unintentional exposure or sharing of sensitive or confidential data, either through the AI model’s training data, predictive outputs or metadata, which may lead to privacy violations or security threats.
- Safeguard user confidentiality by vigilantly monitoring and controlling data output.
- Prevent inadvertent revelation of sensitive information, fortifying user privacy and security.
Prompt Injection: The intentional manipulation of the instruction or query given to an AI model with the aim to trick it into producing harmful, misleading or inappropriate responses, bypassing built-in safeguards.
- Guard against attempts to manipulate the model into bypassing its own safety protocols.
- Provide a basis for refining the robustness of the model’s safeguards, helping ensure it remains impervious to exploitation.
- Helping ensure model functions around the prescribed use case parameters
Toxicity: Harmful or offensive content generated by an AI system, whether in response to a specific input or on its own, that could cause harm, distress or discomfort to individuals or groups.
- Proactively identify and mitigate “toxicity,” defined as the generation of harmful, offensive or inappropriate content.
- Systematic reinforcement of content moderation protocols.
- Preemptively neutralize content that could undermine user wellbeing or violate platform guidelines.
Automating the controls for the above categories of AI model risks can be done both proactively and retroactively. Proactive controls can include measures such as scanning and gating against toxic language, data leakage, prompt injection or setting thresholds for context variance from use case purpose. These risks can be monitored at both the input and the output level. Retroactive monitors can reveal past data and model performance to identify areas of improvement, reoccurring issues, and emerging risks. Automating responsible AI controls enhances the ability to provide continuous, efficient monitoring and can be done at the level of the AI platform architecture and in post launch data reviews.
The success of a responsible AI program lies in finding the intersection between technical controls and functional processes. To develop responsible, scalable and successful AI systems, data science and technology teams must follow guidelines and regulations set up by AI governance and business functional bodies. Meanwhile, governance bodies must collaborate with technical teams to continuously refine these rules, based on emerging risks and changing societal context. Through this collaborative model, AI can be utilized in an ethical and beneficial manner, contributing positively to business strategies, customer trust and brand reputation.