Management of bias and its potential harms of an AI system is critical to ensuring compliance with federal anti-discrimination legislation. Australia’s anti-discrimination law states:
…it is unlawful to discriminate on the basis of a number of protected attributes including age, disability, race, sex, intersex status, gender identity and sexual orientation in certain areas of public life, including education and employment.
Certain forms of bias, such as affirmative measures for disadvantaged or vulnerable groups, play a constructive role in aligning AI systems to human values, intentions, and ethical principles. At the same time, it’s important to identify and address biases that may lead to unintended or harmful consequences. A balanced approach to bias management ensures that beneficial biases are preserved while minimising the impact of problematic ones.
When integrating off-the-shelf AI products, it’s essential to ensure they deliver fair and equitable outcomes in the targe operating environment. Conducting thorough bias evaluations becomes especially important when documentation or supporting evidence is limited.
Criterion 15: Identify how bias could affect people, processes, data, and technologies involved in the AI system lifecycle.
Systemic biases: are rooted in societal and organisational culture, procedures, or practices that disadvantage or benefit specific cohorts. These biases manifest in datasets and in the processes throughout the AI lifecycle.
Human bias: can affect design decisions, data collection, labelling, test selection, or any process that require judgment throughout the AI lifecycle. They could be conscious (implicit) or unconscious (explicit).
Statistical and computational bias occurs when data used to train an AI system is not representative of the population. This is explored in more depth in the data section.
This includes:
Criterion 16: Assess the impact of bias on your use case.
This typically involves:
Criterion 17: Manage identified bias across the AI system lifecycle.
For off-the-shelf products, AI deployers should ensure that the AI system provides fair outcomes. Evaluating for bias will be critical where insufficient documentation from the off-the-shelf AI model supplier is provided.
This involves:
Version control is a process that tracks and manages changes to information such as data, models, and system code. This allows business and technical stakeholders to identify the state of an AI system when decisions are made, restore previous versions, and restore deleted or overwritten files.
AI system versioning can extend beyond traditional coding practices, which manages a package of identifiable code or configuration information. Version control for information such as training data, models, and hyperparameters will need to be considered.
Information across the AI lifecycle, that was used to generate a decision or outcome, must be captured. This applies to all AI products, including low code or no code third-party tools.
Criterion 18: Apply version management practices to the end-to-end development lifecycle.
Australian Government API guidelines mandate the use of semantic versioning. They should be enhanced to cater for AI related information and processes.
Version standards should clearly document the difference between production and non-production data, models and code.
This involves applying version management practices to:
Criterion 19: Use metadata in version control to distinguish between production and non-production data, models, and code.
This includes:
Criterion 20: Use a version control toolset to improve useability for users.
Version toolsets improve the usability for service delivery and business users, addressing activities such as appeals, Ministerial correspondence, executive briefs, court cases, audit, assurance, privacy, and legislative reviews
This includes:
Criterion 21: Record version control information in audit logs.
This includes:
AI watermarking can be used to embed visual or hidden markers into generated content, so that its creation details can be identified. It provides transparency, authenticity, and trust to content consumers.
Visual watermarks or disclosures provide a simple way for someone to know they are viewing content created by, or interacting with, an AI system. This may include generated media content or GenAI systems.
The Coalition for Content Provenance and Authenticity (C2PA) is developing an open technical standard for publishers, creators, and consumers to establish the origin and edits of digital content. Advice on the use of C2PA is out of scope for the standard.
Criterion 22: Apply visual watermarks and metadata to generated media content to provide transparency and provenance, including authorship.
This will only apply where AI generated content may directly impact a user. For instance, using AI to generate a team logo would not need to be watermarked.
Criterion 23: Apply watermarks and metadata that are WCAG compatible where relevant
Criterion 24: Apply visual and accessible content to indicate when a user is interacting with an AI system.
For example, this may include adding text to a GenAI interface so that users are aware they are interacting with an AI system rather than a human.
Criterion 25: For hidden watermarks, use watermarking tools based on the use case and content risk.
This includes:
Criterion 26: Assess watermarking risks and limitations.
This includes:
Criterion 27: Define the problem to be solved, its context, intended use, and impacted stakeholders.
This includes:
Criterion 28: Assess AI and non-AI alternatives.
This includes:
Criterion 29: Assess environmental impact and sustainability.
Developing and using AI systems may have corresponding trade-offs with electricity usage, water consumption, and carbon emissions.
Criterion 30: Perform cost analysis across all aspects of the AI system.
This includes:
Criterion 31: Analyse how the use of AI will impact the solution and its delivery.
This includes:
Criterion 32: Identify human values requirements.
Human values represent what people deem important in life such as autonomy, simplicity, tradition, achievement, and social recognition.
This includes:
Criterion 33: Establish a mechanism to inform users of AI interactions and output, as part of transparency.
Depending on use case this may include:
Criterion 34: Design AI systems to be inclusive, ethical, and meets accessibility standards using appropriate mechanisms.
This includes:
Criterion 35: Define feedback mechanisms.
This includes:
Criterion 36: Define human oversight and control mechanisms.
This includes:
Criterion 37: Involve users in the design process.
The intention is to promote better outcomes for managing inclusion and accessibility by setting expectations at the beginning of the AI system lifecycle.
This includes:
Criterion 38: Analyse and assess harms.
This includes:
Criterion 39: Mitigate harms by embedding mechanisms for prevention, detection, and intervention.
This includes:
Criterion 40: Design the system to allow calibration at deployment.
This includes:
Criterion 41: Identify, assess, and select metrics appropriate to the AI system.
Relying on a single metric could lead to false confidence, while tracking irrelevant metrics could lead to false incidents. To mitigate these risks, analyse the capabilities and limitations of each metric, select multiple complementary metrics, and implement methods to test assumptions and to find missing information.
Considerations for metrics includes:
After metrics have been identified, understand and assess the trade-offs between the metrics.
This includes:
Criterion 42: Reevaluate the selection of appropriate success metrics as the AI system moves through the AI lifecycle.
Criterion 43: Continuously verify correctness of the metrics.
Before relying on the metrics, verify the following:
Criterion 44: Create and collect data for the AI system and identify the purpose for its use.
It is important to identify:
Criterion 45: Plan for data archival and destruction.
Consider the following:
Criterion 46: Analyse data for use by mapping the data supply chain and ensuring traceability.
Mapping the data supply chain to the AI system involves capturing how data will be stored, shared, and processed, particularly at the training and testing stages, which involve regular injections of data. When mapping the data account for:
Ensuring traceability entails maintaining awareness of the flow of data across the AI system.
This includes:
Criterion 47: Implement practices to maintain and reuse data.
This involves determining ongoing mechanisms for ensuring data is protected, accessible, and available for use in line with the original consent parameters.
Any changes in data scope, including expansion in scope and usage patterns, would need to be monitored and addressed.