-
-
-
Your responsibilities
To successfully meet this criterion, agencies need to:
- identify the most appropriate measure to monitor availability
- monitor the availability of the digital service based on expected user outcomes
- act to improve user outcomes.
When to apply
Apply Criterion 2 in the Live environment and consider it during Discovery. Collate metrics and monitor your digital service with a holistic approach. Report your results to build government’s view of its digital services landscape.
The Digital Performance Standard sets the expectation that at a minimum agencies should measure uptime to monitor the availability of the digital service, in line with the industry-standard approach. Where your agency has a mature monitoring framework in place, a more user-centric, comprehensive monitoring approach should be implemented as best practice.
Applying Criterion 2 should focus on the end-user experience to promote continuous improvement of your digital service.
Questions for consideration
- Can users access the digital service as intended?
- How does your agency currently monitor digital service availability?
- What story might the digital service availability metrics tell you?
- What service improvements are necessary?
How to apply criterion 2
-
Criterion 3 – Measure the success of your digital service
-
Monitor success to understand if your digital service is effective and working well.
-
Your responsibilities
To successfully meet this criterion, agencies need to:
- understand what success looks like for the digital service
- identify the most appropriate measure to monitor the success of the digital service
- regularly measure and monitor the effectiveness of the digital service and act to improve outcomes.
-
When to apply
Apply Criterion 3 in the Live environment and consider it during Discovery. Collate metrics and monitor your digital service with a holistic approach. Report your results to help government understand the impact and benefits of its ICT investments.
The Digital Performance Standard sets the expectation that, at a minimum, agencies should be monitoring how well they support users to finish the tasks they start in digital services. Where your agency has a mature monitoring framework in place, a more comprehensive monitoring approach that captures data along the user’s digital journey should be implemented as best practice.
Monitor Criterion 3 to evaluate the effectiveness of the digital services and identify areas for continuous improvement.
Off -
Questions for consideration
- What does success look like for the digital service?
- What rate of users complete their end-to-end transaction online?
- What data can be collected along the user journey?
- How can agencies support users to finish the online tasks they start?
-
Guidance and resources
- Digital Performance Standard Guidance
- Digital Service Standard
- GOV.UK – Measuring the success of your service
-
-
-
Your responsibilities
To successfully meet this criterion, agencies need to:
- understand what success looks like for the digital service
- identify the most appropriate measure to monitor the success of the digital service
- regularly measure and monitor the effectiveness of the digital service and act to improve outcomes.
When to apply
Apply Criterion 3 in the Live environment and consider it during Discovery. Collate metrics and monitor your digital service with a holistic approach. Report your results to help government understand the impact and benefits of its ICT investments.
The Digital Performance Standard sets the expectation that, at a minimum, agencies should be monitoring how well they support users to finish the tasks they start in digital services. Where your agency has a mature monitoring framework in place, a more comprehensive monitoring approach that captures data along the user’s digital journey should be implemented as best practice.
Monitor Criterion 3 to evaluate the effectiveness of the digital services and identify areas for continuous improvement.
Questions for consideration
- What does success look like for the digital service?
- What rate of users complete their end-to-end transaction online?
- What data can be collected along the user journey?
- How can agencies support users to finish the online tasks they start?
How to apply criterion 3
-
Criterion 4 – Measure if your digital service is meeting customer needs
-
Design and deliver digital services with a focus on customer satisfaction.
-
Your responsibilities
To successfully meet this criterion, agencies need to:
- identify the most appropriate measure to monitor the satisfaction rates of the digital service
- give users the ability to rate their satisfaction/dissatisfaction
- continuously monitor customer satisfaction of the digital service and act to improve outcomes.
-
When to apply
Apply Criterion 4 at all times in the Live environment and consider it during Discovery. Compile metrics and monitor the digital service with a holistic approach. Report the results to build government’s view of its digital services landscape.
Customer satisfaction is an industry-standard measure of digital service quality. At a minimum, customer feedback channels should be available on each page and at the end of a digital transaction.
A mature monitoring framework:
- adopts a best practice approach
- integrates feedback tools as users actively use digital services
- surpasses baseline requirements and provides in-depth insights on the user experience.
Apply Criterion 4 to encourage user-centred design and contribute to internal continuous improvement processes.
Off -
Questions for consideration
- Can users easily provide feedback within their digital experience?
- What can be learnt about the customer’s journey?
- What can agencies learn from digital services that have high customer satisfaction?
- Are user expectations aligned to their actual experience?
- What impacts factor into customer satisfaction?
-
Guidance and resources
- Digital Performance Standard Guidance
- GOV UK – Measuring user satisfaction
- Services NSW – Customer commitments
- Digital Service Standard
- The Australian Privacy Principles (oaic.gov.au)
-
-
-
Your responsibilities
To successfully meet this criterion, agencies need to:
- identify the most appropriate measure to monitor the satisfaction rates of the digital service
- give users the ability to rate their satisfaction/dissatisfaction
- continuously monitor customer satisfaction of the digital service and act to improve outcomes.
When to apply
Apply Criterion 4 at all times in the Live environment and consider it during Discovery. Compile metrics and monitor the digital service with a holistic approach. Report the results to build government’s view of its digital services landscape.
Customer satisfaction is an industry-standard measure of digital service quality. At a minimum, customer feedback channels should be available on each page and at the end of a digital transaction.
A mature monitoring framework:
- adopts a best practice approach
- integrates feedback tools as users actively use digital services
- surpasses baseline requirements and provides in-depth insights on the user experience.
Apply Criterion 4 to encourage user-centred design and contribute to internal continuous improvement processes.
Questions for consideration
- Can users easily provide feedback within their digital experience?
- What can be learnt about the customer’s journey?
- What can agencies learn from digital services that have high customer satisfaction?
- Are user expectations aligned to their actual experience?
- What impacts factor into customer satisfaction?
How to apply criterion 4
-
Criterion 5 – Analyse and report your digital performance
-
Reporting the performance of digital services across government enables transparency and accountability.
-
Your responsibilities
To successfully meet this criterion, agencies need to:
- establish internal processes to support performance data analysis and reporting
- report on progress during IOF states and post-implementation performance data, with key reporting requirements in the following states:
- Strategic planning and prioritisation: report on how you intend to implement a monitoring framework (Criterion 1) to your digital service
- Contestability: report (with evidence) that the Digital Performance Standard has been, or will be, applied to the digital service
- Assurance: report (with evidence) how you have applied the Digital Performance Standard to the digital service
- Operations: report (with evidence) on how the digital service continues to meet customer needs
- analyse performance results and act on any improvements to the agency’s digital service.
-
When to apply
Criterion 5 will apply throughout the design and implementation states of the IOF process and once the digital service is made available to the public in the Live environment.
Examples of analysis and reporting requirements will include but are not limited to:
- Strategic planning and prioritisation: outline the planned approach to applying the Digital Performance Standard
- Contestability: explain the plan and approach for meeting the Digital Performance Standard and when this will be implemented to support the investment proposal (Digital Capability Assessment Process)
- Assurance: demonstrate the progress of the ICT investment towards meeting the Digital Performance Standard and delivery milestones
- Operations: demonstrate that the ICT investment is continuing to meet customer needs.
As the monitoring framework matures, agencies should aim to achieve best practice, collecting and monitoring more meaningful user-centric metrics.
Off -
Questions for consideration
- Does the monitoring framework provide meaningful reporting data about ICT investment outcomes?
- Does the reporting data say how well we have delivered the benefits as stated in the ICT investment proposal?
- What outcomes are expected from the reported data?
- How can the reported data be used to improve outcomes for users?
- How can best-practice monitoring be applied to the digital services?
Connect with the digital community
Share, build or learn digital experience and skills with training and events, and collaborate with peers across government.