-
Criterion 8 – Innovate with purpose
-
Key recommendations
- Enable staff to use public generative AI tools for government information classified up to OFFICIAL while prioritising enterprise AI solutions that are security-assessed for handling classified information.
- Adopt a risk-based approach to public generative AI access with a default posture of enabling where appropriate, based on staff roles and security clearance levels, and using technical controls like upload blockers, splash screens and data loss prevention tools.
- Develop AI literacy through training and hands-on experience with approved tools, establishing competency frameworks and cross-agency knowledge sharing. AI capability is increasingly essential for all government employees, and ignoring this shift risks professional stagnation and undermines our ability to adapt and respond to future public service needs.
- Build on existing foundational training to reduce generative AI risks by ensuring staff have a strong understanding of privacy, security, professional conduct and integrity expectations. Key mandatory training programs include APS Foundations Security Awareness and Privacy Awareness.
- Integrate AI governance with existing frameworks including usage monitoring, incident response procedures, human validation requirements, and documentation of AI use.
- Manage access to public generative AI in accordance with the Protective Security Policy Framework, including Policy Advisory 001-2025 on OFFICIAL Information Use with Generative Artificial Intelligence.
PDF Poster:
How to use public generative AI tools responsibly -
Complementary frameworks
This guidance has been designed to complement and strengthen existing frameworks, legislation and practices, enabling agencies to harness AI benefits while maintaining their obligations.
Established governance frameworks provide a strong foundation for AI adoption when supported by appropriate human capability and training. These frameworks enable agencies to confidently expand AI use.
Related AI frameworks are available here digital.gov.au/policy/ai/related-frameworks
-
Appropriate public generative AI use examples
-
Research and information gatheringJordan, a policy analyst in an Australian Government department, needs to produce a comparative analysis of international approaches to a policy issue for a briefing. Jordan is considering using public generative AI tools to brainstorm, accelerate the research process and create first pass content that can be refined further.
What should Jordan do?
- Jordan uses a public generative AI tool to help identify publicly available research papers, reports, and case studies related to the policy issue without including any security classified information.
- Jordan can prompt the AI tool to generate targeted questions on international approaches, highlight key differences and create draft content for further development.
- Jordan carefully reviews the AI-generated content, cross-referencing it with authoritative sources to ensure it’s accurate and reliable. If the AI tool can cite its sources, Jordan can independently verify these to confirm or correct the information.
- If Jordan identifies incorrect, misleading or questionable information, they can clarify or reframe the prompt, compare outputs from different prompts or tools, verify claims against trusted sources and discard any unsupported or unverifiable claims.
- Before using any insights in briefings, Jordan consults with subject matter experts or experienced colleagues to validate complex or ambiguous content and check for accuracy and completeness.
-
Consider diverse user needs from the outset
Listen to and understand diverse user needsConsider diverse user needs from the outset to make sure your service caters to as many users as possible. Consider the different identities, characteristics, and perspectives of users to make your digital service
welcoming and inclusive for all.Conduct usability testing with diverse user groups
Do usability testing with individuals from diverse backgrounds, including those with different abilities, ages, and cultural
Off
contexts. Adopt inclusive prototyping techniques, to simulate the experiences of users with different abilities and identify potential challenges. Recognise that various aspects of a person’s identity, such as race, gender, and age all intersect to shape their digital experience. -
Tailored public communication materialsSofia is developing educational materials for diverse public audiences for a government program. She needs to develop a communications plan, web content and fact sheets tailored for communities across Australia, incorporating relevant local context. Sofia is considering using public generative AI tools to help develop initial drafts, research and organise local community information and improve readability for diverse audiences.
What should Sofia do?
- Sofia can ask public generative AI tools to suggest ways to present program information and tailor messaging for different community needs – without sharing personal information or information that’s security classified as OFFICIAL: Sensitive or above.
- Sofia employs the AI tool to help research local community information, rephrase bureaucratic terminology into plain language and organise content into clear, accessible formats.
- Sofia cross-checks AI-generated content against original sources, asks colleagues to fact-check drafts and contacts any external parties before including their information in public material.
- Sofia remains responsible for the final content, ensuring it accurately conveys current program guidelines and aligns with departmental requirements. Sofia uses public generative AI tools to suggest different ways to structure program information for broad community understanding and help rephrase complex program terminology into plain English.
-
Document analysis and interpretationAlex is reviewing a complex technical report to summarise its content and highlight strategic implications for the agency. Alex is considering using public generative AI platforms to assist with understanding and summarising the content.
What should Alex do?
- Alex can upload the report to a public generative AI tool to help break down complex concepts, explain unfamiliar terms and find further information on related topics.
- Alex asks the generative AI tool to help structure summaries and identify key themes while ensuring only OFFICIAL level government information is shared with the external platform.
- Alex validates AI-generated content by consulting subject matter experts to ensure technical accuracy and appropriate interpretation.
- Depending on team or agency expectations regarding disclosure of AI-influenced content, Alex may choose to include a statement advising readers how AI was used to develop the summary.
-
Community engagement planningChris is planning consultation activities for a community program. They need to develop engagement strategies and materials that will effectively reach diverse community groups and gather meaningful feedback, and project manage the consultation process to ensure it meets key timelines. Chris considers using public generative AI tools to brainstorm consultation approaches.
What should Chris do?
- Chris uses public generative AI tools to generate ideas for community engagement methods and consultation formats, taking care to only share OFFICIAL level information on the program and consultation objectives. Chris does not share any personal information about individuals connected to the consultations or the program itself with the AI platform.
- Chris asks the public AI tool to suggest project management strategies and draft project planning materials. For example, the tool can draft a consultation schedule, progress report templates and content and consultation session logistics checklists.
- Chris can also seek advice on developing survey questions and interview and focus group prompts and ways to track engagement and evaluate consultation outcomes, without entering any security classified or personal data.
- Chris asks the AI tool to suggest approaches for making consultations accessible to people with different language backgrounds and accessibility needs. Chris notes that public generative AI tools may not produce accurate translations from English into other languages. If translation is needed, Chris should follow agency guidelines for accessing professional translation services.
- Chris thoroughly checks and edits the AI-generated content and validates engagement strategies with colleagues, ensuring they align with government consultation standards and inclusive engagement principles.
- Chris reviews all draft consultation materials to ensure they meet plain language requirements and government communication guidelines.
-
Data analysis supportDinesh is conducting analysis on a dataset for a research project. The dataset contains OFFICIAL information and does not contain any personal or security classified information. Dinesh is wondering whether public generative AI tools can help identify patterns and generate insights from the data.
What should Dinesh do?
- Given the dataset contains only OFFICIAL information, with no personal or security classified information, Dinesh can use public generative AI tools to assist with data analysis and pattern identification.
- Dinesh provides only the necessary data elements to the public AI tool, avoiding any information security classified OFFICIAL: Sensitive or above.
- Noting that generative AI tools can produce biased or inaccurate outputs, Dinesh thoroughly validates all insights and analysis before using them in any official capacity.
- Before sharing findings more broadly, Dinesh asks a colleague to review the analysis to confirm it accurately reflects the data and meets government analytical standards.
-
-
-
Inappropriate public generative AI use examples
Connect with the digital community
Share, build or learn digital experience and skills with training and events, and collaborate with peers across government.