Statement 28: Test for safety, robustness, and reliability
Agencies must
- Criterion 101: Test the computational performance of the system. - This includes: - testing for response times, latency, and resource usage under various loads
- network and hardware load testing.
 
- Criterion 102: Test safety measures through negative testing methods, failure testing, and fault injection. - This includes: - testing for incorrect or harmful inputs.
 
- Criterion 103: Test reliability of the AI output, through stress testing over an extended period, simulating edge cases, and operating under extreme conditions. 
Agencies should
- Criterion 104: Undertake adversarial testing (red team testing), attempting to break security and privacy measures to identify weaknesses. - AI-specific attacks can be executed before, during, and after training. - Examples of attacks that can be made before and during training includes: - dataset poisoning
- algorithm poisoning
- model poisoning
- backdoor attacks.
 - Examples of attacks that can be made after training includes: - input attack and evasion
- reverse engineering the model and data.
 
 
 
              
  