• Next page

    8. Transparency and explainability

  • Version 2.0

  • Next page

    9. Contestability

  • Version 2.0

  • Next page

    10. Human-centred values

  • Next page

    Implementation 

  • Version 2

  • Next page

    Strategy and oversight 

  • Next page

    Preparedness and operations

  • Next page

    AI use case impact assessment

  • Next page

    Appendices

  • Policy introduction

    The increasing adoption of artificial intelligence (AI) is reshaping the economy, society and government. Government agencies are embracing the AI opportunity to boost efficiency, enable better data analysis and evidence-based decisions, and improve service delivery for people and business.

    The Australian Government is committed to capturing the opportunities of AI in a way that facilitates innovation while mitigating harms. Government has an elevated responsibility to adopt AI safely and should be held to a high ethical standard of use.

    The December 2025 update to the policy strengthens government's approach to safe and responsible AI through new measures on AI governance. It introduces requirements for agencies to develop a strategic approach to adopting AI, establish an approach to operationalise the responsible use of AI, ensure designated accountability for AI use cases, and undertake risk-based use case-level actions.

    In recognition of the speed and scale of change in this area, the policy will evolve over time as the technology changes, leading practices emerge, the broader regulatory environment develops, and government's AI maturity improves.

    Policy overview

    This policy provides a framework to enable the accelerated and sustainable adoption of AI by agencies. It allows them to embrace AI while continuing to position government as an exemplar in responsible AI use, in line with the Australian community's expectations.

    The policy aims to create a coordinated approach to government's use of AI and has been designed to complement and strengthen – not duplicate – existing frameworks in use by the APS.

    The policy outlines principles and requirements under the following sections: strategy and oversight, preparedness and operations, and AI use case impact assessment.

    Each section of the policy is underpinned by principles and outlines the mandatory requirements agencies must follow. Recommended actions and considerations are also included where relevant.

  • have your say logo

    Consultation on the Digital Seller Underperformance Policy (DSUP)

    The DTA is seeking feedback from sellers and other industry stakeholders on the draft Policy to ensure the Policy is fair, clear, transparent and proportionate.

    - 9 February Have Your Say Now

  • Standard for accountability

    Supporting the policy for responsible use of AI in government

    The information in this standard supports your agency's implementation of the policy for responsible use of AI in government. It covers Accountable Officials, Accountable Use Case Owners and Internal Use Case Registers.

    Accountable officials

    Your responsibilities

    Agencies must designate accountability for implementing the policy to accountable official(s) (AOs), who must:

    • be accountable for implementation of the policy within their agencies
    • notify the Digital Transformation Agency (DTA) where the agency has identified a new high-risk use case by emailing ai@dta.gov.au
    • be a contact point for whole-of-government AI coordination and respond to DTA requests for information
    • engage in whole-of-government AI forums and processes
    • keep up to date with changing requirements as they evolve over time.

    An agency may decide to apply additional responsibilities to their chosen AOs.

    How to apply

    Choose a suitable accountable official

    Agencies may choose AOs who suit the agency context and structure.

    The responsibilities may be vested in an individual or in the chair of a body. The responsibilities may be split across officials or existing roles to suit agency preferences. For example, Chief Information Officer, Chief Technology Officer or Chief Data Officer.

    Implementation of the policy is not solely focused on technology, so AOs may also be selected from business or policy areas. AOs should have the authority and influence to effectively drive the policy's implementation in their agency.

    Agencies may choose to share AO responsibilities across multiple leadership positions.

    Agencies must notify the DTA of AO selection including the contact details of all their AOs at initial selection and when the accountable roles change. Notify DTA by emailing ai@dta.gov.au.

    Implementing the policy

    AOs are accountable for their agency's implementation of the policy. Agencies should implement the entire policy as soon as practical, considering the agency's context, size and function.

    The mandatory actions set out in the policy must be implemented within the specified timelines.

    The policy provides a coordinated approach for the use of AI across the Australian Government. It builds public trust by supporting the Australian Public Service (APS) to engage with AI in a responsible way. AOs should assist in delivering its aims by:

    • uplifting internal capability to support governance of AI adoption in their agency
    • embedding a culture that fairly balances AI risk management and innovation
    • enhancing the response and adaptation to AI policy changes in their agency
    • facilitating agency involvement in cross-government coordination and collaboration.

    AOs should also consider the following activities:

    • developing a policy implementation plan
    • monitoring and measuring the implementation of each policy requirement
    • strongly encouraging the implementation of the AI technical standard for Australian Government
    • strongly encouraging additional training for staff in consideration of their role and responsibilities, such as those responsible for the procurement, development, training and deployment of AI systems
    • establishing a mechanism for staff to seek advice about responsible AI use
    • encouraging the implementation of further actions suggested in the policy
    • reporting internally to relevant governance mechanisms in their agency
    • reviewing policy implementation regularly and provide feedback to the DTA.

    In line with the Standard for Transparency Statements, agencies must provide the DTA with a link to their agency transparency statement each time it is updated, by emailing ai@dta.gov.au.

    Reporting high-risk use cases

    In the event their agency has decided to deploy a new use case with an inherent high risk rating, AOs must notify the DTA, by emailing ai@dta.gov.au.

    AOs must also notify the DTA when an existing AI use case has been re-assessed as having an inherent high risk, or when use case is no longer high risk.

    The notification should include:

    • the type of AI
    • intended application
    • how the agency arrived at a 'high-risk' assessment
    • any sensitivities.

    This is not intended to prevent agencies from adopting the use case. Instead, it will help government develop risk mitigation approaches and maintain a whole-of-government view of high-risk use cases.

    Acting as the agency's contact point

    At times, the DTA will need to collect information and coordinate activities across government to mature the whole-of-government approach and policy.

    AOs are the primary point of contact within their agency. They must respond to DTA requests for information and facilitate connection to the appropriate internal areas for information collection and agency participation in these activities.

    Engaging with forums and processes

    AOs must participate in, or nominate a delegate for, whole-of-government forums and processes which support collaboration and coordination on current and emerging AI issues. These forums will be communicated to AOs as they emerge.

    Keeping up-to-date with changes

    The policy will evolve as technology, leading practices and the broader regulatory environment mature. While the DTA will communicate changes, AOs should keep themselves and stakeholders in their agency up to date on:

    • changes to the policy
    • impacts of policy requirements.
    Questions about policy implementation

    AOs can contact the DTA with questions about policy implementation by emailing ai@dta.gov.au.

    Accountable Use Case Owners

    Your responsibilities

    Agencies must ensure each AI use case that is in scope of the policy (see Appendix C) has designated accountability registered with the agency's AO(s) as an accountable use case owner. Accountable use case owners must:

    • ensure their use case is registered with the agency's AO
    • be accountable for applying the actions under the AI use case impact assessment requirements. These include:
      • conducting an AI use case impact assessment
      • regularly monitoring and evaluating the use case
      • re-validating the assessment when required
    • applying the high-risk use case actions if their use case has an inherent high-risk rating.

    Accountable use case owners should ensure records demonstrate transparency and accountability in the design, development, deployment and monitoring of AI systems related to their use case. This may include establishing documentation and traceability of decisions and changes, ensuring information accessibility and availability to assist with audits, and ensuring explainability of technical and non-technical information.

    How to apply

    Choose a suitable accountable use case owner

    In setting the role of accountable use case owner, accountability can:

    • Be split between roles, for example, between business and technology areas, depending on where accountability is best placed. It is recommended that agencies involve their technology area and consider sharing accountability.
    • Adapt over time based on the lifecycle of the AI use case.
    • Be designated in line with existing agency and whole of government accountability structures.

    Accountable use case owners should either have or be able to access appropriate skills and expertise to identify risks and emerging issues related to their AI use case. Accountable use case owners must also be familiar with Australia's AI Ethics Principles, the Australian Government Impact Assessment Tool and the policy.

    Accountable use case owners of high-risk use cases must have the ability to identify and manage risks and emerging issues.

    Accountable use case owner actions can be delegated to other suitable staff.

    Internal use case register

    Minimum fields for the register

    Agencies must create a register of AI use cases that are in scope of the policy to enable registration of an accountable use case owner with accountable official(s). At a minimum, the register must include the following fields:

    • use case name
    • agency identifier (reference number)
    • description (including what the AI does, its business objective and, if applicable, the underpinning product's name)
    • AI technology type (Generative AI, Machine Learning, Natural Language Processing and/or Computer Vision)
    • lifecycle stage (Discover, Operate or Retire)
    • use of Technical standard for government's use of artificial intelligence (not applied, partially applied or fully applied)
    • domain (see Standard for AI transparency statements)
    • usage pattern (see Standard for AI transparency statements)
    • accountable use case owner (name and email address)
    • what criteria the use case met to be in scope of the policy (see Appendix C of the policy)
    • inherent risk rating (determined through an AI use case impact assessment)
    • residual risk rating (determined through an AI use case impact assessment)
    • the date when the AI impact assessment was last updated (if applicable).

    For AI use cases with an inherent high-risk rating, the register must also include:

    • the last date of review
    • date for next review.

    Agencies should ensure that use case register are up to date via periodic review.

    Sharing the register with the DTA

    Agencies must share the register with the DTA every 6 months, commencing from when they create the register to meet the policy requirement. They can share the register by emailing ai@dta.gov.au or through a method pre-agreed with the DTA. The DTA may update the required method of submission during policy implementation.

  • Standard for AI transparency statements

    Supporting the policy for responsible use of AI in government

    Use the following information to support your agency's implementation of the policy for responsible use of AI in government.

    Your responsibilities

    Under the policy, agencies must make a publicly available statement that outlines their approach to AI adoption, as directed by the Digital Transformation Agency (DTA).

    Agencies must follow this standard, which sets the direction for AI transparency statements including expectations and formatting. It establishes a consistent format and expectation for AI transparency statements in the Australian Government. Clear and consistent transparency statements build public trust and make it easier to understand and compare how government agencies adopt AI.

    At a minimum, agencies must provide the following information regarding their use of AI in their transparency statement:

    • the intentions behind why the agency uses AI or is considering its adoption
    • classification of AI use according to usage patterns and domains, as listed at Attachment A
    • classification of use where the public may directly interact with, or be significantly impacted by, AI or its outputs without human review
    • measures to monitor the effectiveness of deployed AI systems and protect the public against negative impacts
    • overview of compliance with the requirements under the Policy for responsible use of AI in government
    • compliance with applicable legislation and regulation
    • when the statement was most recently updated.

    Statements must use clear, plain language1 that avoids technical jargon and is consistent with the Australian Government Style Manual. They must also provide or direct to a contact email for further public enquiries.

    Agencies must publish transparency statements on their public facing website. It's recommended that a link to the statement is placed in a global menu, aligned to the approach often taken for privacy policies.

    Transparency statements must be reviewed and updated at these junctures:

    • at least once a year
    • when making a significant change to the agency's approach to AI
    • when any new factor materially impacts the existing statement's accuracy.

    How to apply

    Implementing the AI transparency statements

    The policy provides a coordinated approach for the use of AI across the Australian Government. It builds public trust by supporting the Australian Public Service (APS) to engage with AI in a responsible way.

    Transparency is critical to public trust and is an important aim of the policy and broader APS Reform agenda2. The public should have confidence that agencies monitor the effectiveness of deployed AI systems and have measures to protect against negative impacts.

    AI transparency statements help agencies meet these aims by providing a foundational level of transparency on their use of AI. They publicly disclose:

    • how AI is used and managed by the agency
    • a commitment to safe and responsible use
    • compliance with the policy.

    Agency transparency statements are intended to provide a high-level overview of agency AI use and management in line with the policy intent.

    Agencies are not required to list individual use cases or provide use case level detail. However, agencies may choose to provide detail beyond the requirements to publicly explain their approach to AI.

    Agencies must send the DTA a link to the statement when it is published or updated by emailing ai@dta.gov.au.

    Questions about implementation

    Accountable officials can contact the DTA with questions about implementing the transparency statements by emailing ai@dta.gov.au.

    Footnotes

    1. Australian Government Style Manual, Plain language and word choice
    2. APS Reform, Priority one: An APS that embodies integrity in everything it does. "The APS acts with integrity and fairness and is accountable and transparent in everything it does. This will build public trust and strengthen standards of integrity in our federal government." https://www.apsreform.gov.au/about-aps-reform/our-focus-areas
  • Version 2.0

  • Guidance for staff training on AI

    Version 2.0

    Use the following information to support your agency's implementation of the policy for responsible use of AI in government.

    The policy recognises that AI is used in many areas of the APS and everyday life. As adoption grows, staff at all levels will interact with AI and its outputs, directly or indirectly.

    The policy requires agencies to implement mandatory training for all staff on responsible and ethical AI use, regardless of their role. Agencies should also consider if it is appropriate for staff to complete annual refresher training. 

    Foundational AI training

    Foundational training focused on responsible and ethical AI use should address the following learning outcomes:

    Introduction to AI

    • identify different types of AI and their functions
    • recognise AI in everyday tools and services.

    Generative AI

    • understand how generative AI works
    • use clear prompts and verify outputs

    Foundations of safe and responsible AI use

    • recognise the importance of ethical AI use
    • understand how to remain accountable for AI use
    • appropriate management of security classified, personal and sensitive information
    • understand how AI ethics principles should be applied in practice, including how to manage AI for accuracy, bias, sensitivity and transparency

    The training must align with the Policy for the responsible use of AI in government, Australia’s AI Ethics Principles and expectations such as Australia’s Public Service Values and Code of Conduct. 

    AI fundamentals training

    The Digital Transformation Agency (DTA) has developed AI fundamentals training for agency use. This training meets the learning outcomes described above for foundational AI training and can be used to satisfy the policy requirement for staff training. 

    The training will be updated periodically to address changes in the AI landscape, such as evolutions in the technology and its use in government.

    As staff will likely interact with generative AI now and in the future, it is an area of focus for the training.

    The training is designed for all staff regardless of their experience using AI. It takes approximately 20 to 30 minutes to complete.

    It does not cover advanced topics such as model training, system development or instructions for specific technologies or platforms.

    Implement the training

    Agencies’ Learning and Development specialists can access the training for download through the APS Learning Bank. The training is provided in a format compatible with most e-learning platforms. Alternatively, agencies can choose to let their staff access the module directly through APSLearn.

    Agencies can use the training module as provided, or choose to modify it or incorporate it into an existing training program based on their specific context and requirements. 

    Agencies are encouraged to consider additional training for staff in consideration of their roles and responsibilities, such as those responsible for the procurement, development, training and deployment of AI systems.

    Report completion rates

    Where an agency implements the training, accountable officials should monitor completion rates and provide this information if requested by the DTA. This is in line with the activities to measure the implementation of the policy under the Standard for accountability. 

Connect with the digital community

Share, build or learn digital experience and skills with training and events, and collaborate with peers across government.