• Design: statements 9 - 12

  • Designing AI systems that are effective, efficient, and ethical involves being clear on the problem, understanding the impacts of technical decisions, taking a design approach with humans at the centre and having a clear definition of success.

    In the design stage agencies consider how the AI system will operate with and impact existing processes, people, data, and technology. This includes considering potential malfunctions and harms.

    Without appropriate design an AI system could:

    • cause harm due to incorrect information,  caused by AI hallucinations, false positives, or false negatives
    • be used beyond their purpose
    • perpetuate existing injustices
    • be misused, misunderstood, or abused
    • be susceptible to malfunctions of another interacting system
    • experience behaviour and performance issues caused by other external factors.

    At the design stage agencies also determine the performance and reliability measures relevant to their AI system’s tasks. Considerations when selecting metrics include business, performance, safety, reliability, explainability, and transparency.
     

  • The design stage includes concept development, requirements engineering, and solution design.

  • Services not covered by the Digital Service Standard

    Agencies are recommended to apply the Digital Service Standard to existing staff facing services, though these services are not mandated.

    The Digital Service Standard does not apply to:

    • state, territory or local government services
    • personal ministerial websites that contain material on a minister’s political activities or views on issues not related to their ministerial role.

    State, territory or local government and third parties may choose to apply the Digital Service Standard to improve access and discoverability of their digital services.

    Some services may request an exemption from the Digital Service Standard. See the Exemptions section below.

  • Notes:



  • Data: statements 13 - 19

  • The data stage involves establishing the processes and responsibilities for managing data across the AI lifecycle. This stage includes data used in experimenting, training, testing, and operating AI systems.

  • Data used by an AI system can be classified into development and deployment data.

    Development data includes all inputs and outputs (and reference data for GenAI) used to develop the AI system. The dataset is made up of smaller datasets – train dataset, validation dataset, and test dataset.

    • Train dataset – this dataset is used to train the AI system. The AI system learns patterns in the train dataset. The train dataset is the largest subset of the modelling dataset. For GenAI, the train dataset may also include reference or contextual datasets such as retrieval-augmented generation (RAG) datasets and prompt datasets
    • Validation dataset – this dataset is used to evaluate the model's performance during model training. It is used to fine-tune and select the best-performing model, such as through cross validation
    • Test dataset – this dataset is used to evaluate the final model's performance on previously unseen data. This dataset helps provide unbiased evaluation of model performance.

    Deployment data includes AI system inputs such as live production data, user input data, configuration data, and AI system outputs such as predictions, recommendations, classifications, logs, and system health data. Deployment stage inputs are new and previously unseen by the AI system. 

    The performance of an AI system is dependent on robust management of data quality and the availability of data. 

    Key workstreams within this stage include: 

    • data orchestration – establishing central oversight of and planning the flow of data to an AI system from across datasets
    • data transformation – converting and optimising data for use by the AI system
    • feature engineering – methods to improve AI model training to better identify and learn patterns in the data
    • data quality – measuring dimensions of a dataset associated with greater performance and reliability
    • data validation – testing the consistency, accuracy, and reliability of the data to ensure it meets the requirements of the AI system
    • data integration and fusion – combining data from multiple sources to synchronise the flow of data to the AI system
    • data sharing – promoting reuse, reducing resources required for collection and analysis, and helping to build interoperability between systems and datasets
    • model dataset establishment – using real-world production data to build, refine, and contextualise a high-quality AI model.
       

  • Notes: 

  • Train: statements 20 - 25

  • The train stage covers the creation and selection of models and algorithms. The key activities in this stage include modelling, pre- and post-processing, model refinements, and fine-tuning. It also considers the use of pre-trained models and associated fine-tuning for the operational context.

  • Exemptions

    The DTA acknowledge that some agencies may be unable to meet one or more of the criteria set out by the Digital Service Standard due to a range of circumstances. These circumstances may include but are not limited to:

    • legacy technology barriers that the agency cannot reasonably overcome
    • substantial financial burden caused by changing a service to meet criteria.

    Exemptions may be granted for one or more of the criteria set out by the Digital Service Standard. This will be assessed on a case-by-case basis. Exemptions must be applied for through the DTA.

    Further information can be found in the Digital Experience Policy Exemption Guide.

    Note: Even if a service or website is not covered by the Digital Service Standard, or an exemption is received, obligations may still apply under relevant Australian legislation, for example accessibility requirements under the Disability Discrimination Act 1992.

    Off
  • AI training involves processing large amounts of data to enable AI models to recognise patterns, make predictions, draw inferences, and generate content. This process creates a mathematical model with parameters that can range from a few to trillions. Training an AI model might require adjustment of these parameters, entailing increased processing power and storage. 

    Training a model can be compute-heavy, relying on infrastructure that may be significantly expensive. The model architecture, including choice of the AI algorithm and learning strategy, together with the size of the model dataset, will influence the infrastructure requirements for the training environment.

    The AI Model encapsulates a complex mathematical relationship between input and output data that it derives from patterns in a modelling dataset. AI models can be chained together to provide more complex capabilities.

    Pre-processing and post-processing augment the capabilities of the AI model. Application, platform, and infrastructure components are shown here as well as they all contribute to the overall behaviour and performance of the whole AI system.

    Due to the number of mathematical computations involved and time taken to execute them, training can be a highly intensive stage of the AI lifecycle. This will depend on the infrastructure resources available, the algorithms used to train the AI model and the size of the training datasets.

    Key considerations during this stage include:

    • the model architecture, including the AI model and how components within the model interact, as well as the use of off-the-shelf or pre-trained models
    • selection and development of the algorithms and learning strategies used to train the AI model
    • an iterative process of implementing model architecture, setting hyperparameters, and training on model datasets
    • model validation tests, supplemented by human evaluation, which evaluate whether the model is fit-for-purpose and reliable
    • trained model selection assessments, which streamline development and enhance capabilities by comparing various models for the AI system
    • continuous improvement frameworks which set processes for measuring model outputs, business, and user feedback to manage model performance.

    If after multiple attempts of refinement, the model does not meet requirements or success criteria, a new model may need to be created, business requirements updated, or the model is retired.

    See the Design lifecycle stage for details on measuring model outputs, as well as business and user feedback, to manage AI model performance.

    See the Apply version control practices statement in the Whole of AI lifecycle section for detail on tracking changes to training models, trained models, algorithms, learning types, and hyperparameters.

  • Statement 20: Plan the model architecture

  • Lifecycle stages

    Introduces each lifecycle and it's accompanying statements noting which are required and which are recommended. 

  • A - E | F - P | Q - Z

  • Whole of AI lifecycle

Connect with the digital community

Share, build or learn digital experience and skills with training and events, and collaborate with peers across government.