• If you are outside of Australia

    We have developed a version of this training that is intended to not be limited by location.

    However, as this version of the training module was originally developed for and by the Australian Government to support the implementation of the Policy for the responsible use of AI in government, it still refers to some Australian specific resources that may not be applicable in your jurisdiction. 

  • Evaluate: statements 26 - 30

  • This stage includes testing, verification, and validation of the whole AI system. It is assumed that agencies have existing capability on test management and on testing traditional software and systems. 

  • Testing should be done continuously at each component of the AI system lifecycle. The level of testing at each stage could be at unit, integration, system, or system-of-systems, depending on the scope of development at that stage and the test strategy.

    Testing can be divided into formal and informal phases. Formal testing is the phase when the system under test (SUT) is formally versioned, and the outcomes are evaluated for deciding whether to deploy to production or not. Statements within the standard apply to the formal testing phase. 

    Where an AI system or components have been procured, deployment and integration into the deployer’s environment may need to be done before starting formal testing.

    Testing of AI systems differs from testing traditional software as they can be probabilistic and non-deterministic in nature. While probabilistic systems do not give you an exact value, non-deterministic systems may have different outputs using the same inputs.

    Other key differences include your approach to regression testing. While making small changes to a non-AI system may have limited consequences, a step change in test data or in parameters can be significant for an AI system. This means you need to conduct more robust regression testing to mitigate the heightened risk of escaped defects.

    Note that an AI system which learns dynamically will change its behaviour without being formally updated. This means changes may occur on the same deployment version and without formal testing. This will require a more rigorous continuous monitoring post deployment. The development team or supplier should confirm whether your AI system has been designed to learn dynamically or statically.
     

  • Inclusive design embraces broad diversity to meet the varied needs and perspectives of a wide range of user groups.

  • Integrate: statements 31 - 32

  • The integrate stage of the AI lifecycle focuses on implementing and testing an AI system within an agency’s internal organisational environment, including with its systems and data.

  • Deploying a standalone or integrated AI system into existing IT infrastructure involves assessing compatibility and confirming data interoperability. This may require some reconfiguration to current systems.

    Agencies can achieve the best outcomes at this stage by adopting practices that closely align with those implemented at the test stage. This ensures the AI system has been thoroughly tested against its intended purpose prior to integration – a key measure before the AI system potentially contaminates the business environment. See Test section for a detailed discussion on testing methods.

    Following recommended practices for managing code integration workflows for AI systems will help agencies to maintain quality, security, and consistency.
     

  • Deploy: statements 33 - 36

  • The deploy stage involves introducing all the AI technical components, datasets, and related code into a production environment where it can start processing live data.

  • Deployment involves rigorous testing, governance, and security practices to ensure systems perform as intended in live environments. 

    Defining structures for secure deployment of an AI system is particularly crucial in a government setting. Deployment strategies should be incremental, consistent, and non-disruptive. 
     

  • Notes: 

    • The Australian Signals Directorate’s Australian Cyber Security Centre report on Deploying AI Systems Securely provides further advice on deploying AI systems securely.

  • Monitor: statements 37 - 39

  • The monitor stage of the AI lifecycle includes operating and maintaining the AI system. Monitoring is critical to ensuring the reliability, availability, performance, security, safety, and compliance of an AI system after it is deployed. 

  • Monitoring AI systems is critical because changes in the operating environment and inputs could result to degradation and potential harms. Effective monitoring includes continuous performance evaluation, anomaly detection, intervention, and proactive incident response. 

    The measures implemented at this stage helps identify if a system is generating outputs that misalign with its intended purpose and promptly remedy issues. 
     

  • Criterion 1 – Embrace diversity

  • Decommission: statements 40 - 42

Connect with the digital community

Share, build or learn digital experience and skills with training and events, and collaborate with peers across government.