How to Audit ISO 27001 Control 8.29: Security Testing in Development and Acceptance

Auditing ISO 27001 Annex A 8.29 Security Testing in Development and Acceptance is the technical verification of security validation integrated within the software engineering lifecycle. The Primary Implementation Requirement is automated security gates within CI/CD pipelines and formal UAT, providing the Business Benefit of preventing production-level exploits and ensuring management-authorized risk alignment.

ISO 27001 Annex A 8.29 Security Testing in Development and Acceptance Audit Checklist

This technical verification tool is designed for lead auditors to establish the efficacy of security testing within the software lifecycle. Use this checklist to validate compliance with ISO 27001 Annex A 8.29.

1. Security Testing Policy Formalisation Verified

Verification Criteria: A documented policy or technical standard exists defining the mandatory security testing activities required before software is promoted to production.

Required Evidence: Approved “Secure Testing Standard” or “SDLC Policy” specifying testing triggers, types, and required severity thresholds.

Pass/Fail Test: If the organisation cannot produce a formal document specifying the technical requirements for security testing, mark as Non-Compliant.

2. Static Application Security Testing (SAST) Integration Confirmed

Verification Criteria: Automated SAST tools are integrated into the development environment or CI/CD pipeline to scan source code for vulnerabilities during the build process.

Required Evidence: Build logs or CI/CD configuration files (e.g., Jenkinsfile, GitHub Actions) showing executed SAST scan stages.

Pass/Fail Test: If source code is committed to a production branch without an automated security scan being triggered, mark as Non-Compliant.

3. Dynamic Application Security Testing (DAST) Implementation Validated

Verification Criteria: Technical testing of the running application is performed in a staging environment to identify runtime vulnerabilities such as misconfigurations or broken access control.

Required Evidence: DAST tool reports (e.g., OWASP ZAP, Burp Suite, Veracode) dated within the current release cycle.

Pass/Fail Test: If the organisation only performs code scans but never tests the application in a running, functional state, mark as Non-Compliant.

4. Vulnerability Remediation Prior to Acceptance Verified

Verification Criteria: All identified security vulnerabilities are triaged and “Critical” or “High” risk findings are remediated before User Acceptance Testing (UAT) sign-off.

Required Evidence: Remediation logs or Jira tickets showing the closure of security findings cross-referenced with the production release date.

Pass/Fail Test: If a production release contains unmitigated “High” severity security vulnerabilities without an approved risk acceptance, mark as Non-Compliant.

5. User Acceptance Testing (UAT) Security Validation Confirmed

Verification Criteria: Acceptance testing includes specific scenarios to validate that security functional requirements (e.g., MFA, session timeouts) operate as designed.

Required Evidence: Signed UAT Test Plan and Results showing successful validation of security features and access control logic.

Pass/Fail Test: If UAT focuses solely on business functionality and ignores the validation of security controls, mark as Non-Compliant.

6. Independent Security Assessment of Major Changes Validated

Verification Criteria: Significant architectural changes or major releases undergo an independent security review or penetration test prior to deployment.

Required Evidence: Signed Penetration Test report or Independent Security Audit summary for the most recent major system update.

Pass/Fail Test: If a major system overhaul was promoted to production without an independent security assessment, mark as Non-Compliant.

7. Secure Configuration Testing of Testing Environments Verified

Verification Criteria: The environment used for security testing is hardened and configured to mirror the security posture of the production environment.

Required Evidence: Staging environment hardening checklists or IaC (Infrastructure as Code) templates showing parity between Staging and Production.

Pass/Fail Test: If security testing is performed in an environment with lower security settings than production (e.g., disabled firewalls), mark as Non-Compliant.

8. Automated Build Failure (Breaking the Build) Logic Verified

Verification Criteria: The CI/CD pipeline is configured to automatically stop the deployment if security testing identifies vulnerabilities exceeding a defined threshold.

Required Evidence: Pipeline configuration screenshots showing “Fail-on-High” or “Fail-on-Critical” security scan logic.

Pass/Fail Test: If security scans are “Information Only” and do not technically prevent the deployment of vulnerable code, mark as Non-Compliant.

9. Regression Testing of Security Controls Confirmed

Verification Criteria: Following any modification or patch, regression tests are performed to ensure that existing security controls remain functional and effective.

Required Evidence: Automated test suite logs showing successful execution of security regression test cases (e.g., auth, permissions, encryption).

Pass/Fail Test: If the organisation lacks automated or manual regression tests specifically for security functions after a code change, mark as Non-Compliant.

10. Management Sign-off and Acceptance Records Identified

Verification Criteria: A formalised record of management acceptance exists for every release, confirming that all required security testing has been completed and risks triaged.

Required Evidence: Signed Release Authorisation or “Go/No-Go” meeting minutes with a specific Security sign-off attribute.

Pass/Fail Test: If software is promoted to production via “silent deploy” without a formal record of management acceptance of the security status, mark as Non-Compliant.

ISO 27001 Annex A 8.29 SaaS / GRC Platform Failure Checklist
Control Requirement The ‘Checkbox Compliance’ Trap The Reality Check
Testing Policy Tool records “Policy.pdf” as ‘Uploaded’ and marks green. Verify Technical Thresholds. A policy is just words; the auditor must see the “Fail-Build” settings in the actual tool config.
SAST/DAST Coverage Platform identifies that a “Scanner” is connected via API. Check the Scan Scope. GRC tools often ignore that the scanner is only configured to scan 10% of the active repositories.
Vulnerability Triage SaaS tool reports “90% of findings closed” as a pass. Inspect the Risk Acceptances. Lazy teams “Accept” high risks to clear the GRC dashboard without actually fixing the code.
UAT Integrity Tool checks if “UAT Phase” is in the project plan. Review the Test Cases. If UAT doesn’t include “Negative Testing” (attempting unauthorised access), the ‘Security’ element is a lie.
Independent Review Platform assumes cloud providers handle all security tests. Demand the Bespoke Test. Shared responsibility means the organization must test its own code, not just the provider’s infrastructure.
Environment Parity Tool records “Staging Environment Exists”. Check Access Control. If Staging is wide open but Production is locked down, security tests in Staging are invalid.
Management Sign-off Tool identifies “Release Manager” checked a box. Verify Informed Consent. Did the manager see the open vulnerabilities before checking the box? Demand the risk summary report.
Fay Barker - High Table - ISO27001 Director

About the author

Stuart Barker
🎓 MSc Security 🛡️ Lead Auditor 30+ Years Exp 🏢 Ex-GE Leader

Stuart Barker

ISO 27001 Ninja

Stuart Barker is a veteran practitioner with over 30 years of experience in systems security and risk management. Holding an MSc in Software and Systems Security, he combines academic rigor with extensive operational experience, including a decade leading Data Governance for General Electric (GE).

As a qualified ISO 27001 Lead Auditor, Stuart possesses distinct insight into the specific evidence standards required by certification bodies. His toolkits represent an auditor-verified methodology designed to minimise operational friction while guaranteeing compliance.

Shopping Basket
Scroll to Top