ISO 27001:2022 Annex A 5.34 Privacy and protection of PII for AI Companies

ISO 27001 Annex A 5.34 for AI Companies

ISO 27001 Annex A 5.34 is a security control that mandates the identification and implementation of requirements for personal data protection. This preventive implementation requirement ensures legal compliance through rigorous data mapping, providing the business benefit of enhanced stakeholder trust and protection against membership inference attacks in AI models.

For an innovative AI company like yours, managing Personally Identifiable Information (PII) is not just a standard compliance task. It is a core component of building trust with your customers and partners, enabling you to innovate responsibly. The international standard for information security, ISO 27001, provides a clear framework for this crucial activity in ISO 27001 Annex A 5.34 Privacy and protection of PII.

The primary purpose of Control 5.34 is to ensure your organisation meets its legal, statutory, regulatory, and contractual requirements for the preservation, privacy, and protection of PII. It is a preventive control, designed to help you create clear guidelines and procedures that maintain risk at an acceptable level and safeguard the personal data you handle.

The “No-BS” Translation: Decoding the Requirement

Let’s strip away the consultant-speak. Annex A 5.34 is about ensuring you don’t accidentally train your model on your customer’s credit card numbers.

The Auditor’s View (ISO 27001)The AI Company View (Reality)
“The organisation shall identify and meet the requirements regarding the preservation of privacy and protection of PII according to applicable laws and regulations and contractual clauses.”Know where the PII is. Don’t just dump a CSV into S3 and forget about it. Scan it for emails, names, and phone numbers. If you find them, either delete them or encrypt them.
“Privacy impact assessments.”Check before you train. Before you start a new training run, ask: “Is there any personal data in this dataset?” If yes, anonymise it.

Shutterstock

PII data flow in AI pipelines

The Business Case: Why This Actually Matters for AI Companies

Why should a founder care about “PII Protection”? Because regurgitating PII is the fastest way to get sued.

The Sales Angle

Enterprise clients will ask: “How do you prevent our customer PII from leaking into your model responses?” If your answer is “We rely on the model’s safety filters,” you lose. If your answer is “We use a dedicated PII Redaction Layer (e.g., Presidio) before data hits the model, and we contractually guarantee zero-retention for inference,” you win the deal. A 5.34 is your privacy architecture.

The Risk Angle

The “Membership Inference” Attack: Researchers can query your model to find out if a specific person was in the training set (e.g., “Was John Smith treated for cancer?”). If your model reveals this, you have breached GDPR/HIPAA. A 5.34 forces you to de-identify data before training to prevent this.

DORA, NIS2 and AI Regulation: Privacy By Design

Regulators demand “Privacy by Design,” not just a privacy policy.

  • GDPR (Article 25): Requires “Data Protection by Design and Default.” You must implement technical measures (like pseudonymisation) to strip PII from training data.
  • EU AI Act: High-risk AI systems must have “data governance” that includes examination for possible biases. Often, bias correction requires processing sensitive data (e.g., race). You need a specific legal basis and strict controls (A 5.34) to do this lawfully.
  • DORA: Financial entities must protect client data. If your AI processes transaction data, you must treat it with the same security as a bank vault.

ISO 27001 Toolkit vs SaaS Platforms: The Privacy Trap

SaaS platforms help you generate a privacy policy, but they don’t help you engineer privacy. Here is why the ISO 27001 Toolkit is superior.

FeatureISO 27001 Toolkit (Hightable.io)Online SaaS Platform
The MethodDPIA Templates. A rigorous “Data Protection Impact Assessment” template that forces you to analyse AI specific risks.Checkbox Policy. Platforms give you a generic “Privacy Policy” for your website, but not the technical procedure for sanitising training data.
OwnershipYour Assessment. You keep the DPIA document. It proves you thought about the risk.Black Box. The platform says “Compliant” because you ticked a box, but you have no evidence of the actual risk analysis to show a regulator.
SimplicityData Mapping. A spreadsheet to map PII flows. “Data enters via API -> Redacted by Lambda -> Stored in S3.”Automated Noise. Data discovery tools often flag every string of 16 digits as a credit card, drowning you in false positives.
CostOne-off fee. Pay once. Protect privacy forever.Per-Data Source Cost. Privacy modules often charge by the number of databases connected.

The AI Magnifying Glass: Analysing Unique PII Risks in Your Workflow

While Control 5.34 applies to all businesses, your AI-driven workflows introduce unique risks.

Exposure of Sensitive Training Datasets

A breach of training data could lead to massive unauthorised disclosure. More advanced threats include model inversion attacks, where attackers reconstruct sensitive PII from a model’s outputs. Protecting the training set is as important as protecting the production database.

Disruption of Algorithmic Processes

If the integrity of PII is compromised (e.g., medical records altered), your diagnostic AI could make life-threatening errors. A failure to protect this data directly impacts business continuity and stakeholder trust.

Vulnerabilities in the AI Supply Chain

Your AI supply chain includes data labellers and cloud providers. You must ensure every partner handling PII complies with the same rigorous data protection standards. A weak link in the labelling team exposes your data.

Your Compliance Blueprint: Actionable Steps for PII Protection

Complying with Control 5.34 requires a structured approach.

Establish Clear Governance and Responsibility

Appoint a Privacy Officer (this can be a shared role in startups). Identify all applicable data protection laws (GDPR, CCPA) and document them in your Legal Register (A 5.31).

Develop Topic-Specific Policies and Procedures

Create a Privacy Policy that covers PII Classification, Secure Handling, and Data Retention. Specifically address PII in MLOps: how do you manage PII across development, testing, and production environments?

Implement Technical and Organisational Measures

Implement strong encryption (AES-256) for PII at rest. Use data masking or synthetic data in non-production environments. Never use live customer PII for testing code.

The Evidence Locker: What the Auditor Needs to See

When the audit comes, prepare these artifacts:

  • Data Processing Agreement (DPA): Signed agreements with your customers and your vendors.
  • Record of Processing Activities (ROPA): An Excel sheet listing what PII you have, why you have it, and where it goes.
  • DPIA Reports (PDF): Completed impact assessments for high-risk processing (e.g., training a new model on user data).
  • Subject Access Request Log: A log of user requests (e.g., “Delete my data”) and evidence that you fulfilled them within 30 days.

Common Pitfalls & Auditor Traps

Here are the top 3 ways AI companies fail this control:

  • The “Dev DB” Leak: Developers took a dump of the production database (with real PII) to test a new feature on their laptop. Instant GDPR violation. Use synthetic data.
  • The “Log File” Oversight: You encrypt the database, but your application logs print out the PII in plain text for debugging. The auditor checks the logs and finds email addresses. Fail.
  • The “Forever” Retention: You keep user data indefinitely “to improve the model.” You have no legal basis for this. You must delete data when the user leaves.

Handling Exceptions: The “Break Glass” Protocol

Sometimes you need to access raw PII to debug a critical error.

The PII Access Workflow:

  • Trigger: Production incident requiring inspection of specific user record.
  • Approval: DPO / Privacy Lead approves temporary access.
  • Access: Access granted via Just-In-Time (JIT) privilege for 1 hour.
  • Log: Access is logged and reviewed. Any data downloaded is deleted immediately after debugging.

The Process Layer: “The Standard Operating Procedure (SOP)”

How to operationalise A 5.34 using your existing stack (Linear, Excel).

  • Step 1: Mapping (Manual). Use the High Table ROPA Template to map data flows.
  • Step 2: Assessment (Manual). Before a new feature launch, Product Manager completes a “Privacy Checklist” ticket in Linear.
  • Step 3: Redaction (Automated). Configure libraries (e.g., Microsoft Presidio) to automatically redact PII from logs and training pipelines.
  • Step 4: Deletion (Automated). Set up cron jobs to delete user data 30 days after account termination.

By implementing a structured compliance programme, you can turn a complex obligation into a source of competitive advantage. Leveraging expert tools like the High Table ISO 27001 Toolkit provides the definitive path to achieving this.

ISO 27001:2022 Annex A 5.34 for AI Companies FAQ

What is ISO 27001 Annex A 5.34 for AI companies?

ISO 27001 Annex A 5.34 requires AI companies to identify and implement the privacy and protection of Personally Identifiable Information (PII) in accordance with applicable legislation and contractual requirements. For AI firms, this involves ensuring that 100% of PII within training datasets, fine-tuning scripts, and inference logs is protected against unauthorised processing or accidental exposure.

How do AI firms protect PII in training datasets for Annex A 5.34?

AI firms must implement technical measures to safeguard PII during the model development lifecycle. Compliance strategies typically include:

       
  • Pseudonymisation: Replacing sensitive identifiers with artificial identifiers to reduce re-identification risk by over 90%.
  •    
  • Differential Privacy: Adding mathematical “noise” to datasets to ensure individual records cannot be reverse-engineered from model outputs.
  •    
  • Automated Scrubbing: Using PII-detection algorithms to remove names, addresses, and emails before data enters the training pipeline.

Is a DPIA required for AI models?

Yes, under Annex A 5.34 and GDPR, a formal Data Protection Impact Assessment (DPIA) is mandatory for AI models involving high-risk processing of personal data. The assessment must document the “Right to be Forgotten” within AI weights—a complex technical requirement where 100% of a specific individual’s data must be effectively unlearned if requested.

How does ISO 27001 Annex A 5.34 intersect with GDPR for AI?

Annex A 5.34 acts as the technical bridge to GDPR compliance. While GDPR provides the legal framework, Annex A 5.34 provides the operational controls. Failure to align these can result in fines of up to €20 million or 4% of global annual turnover, making the integration of privacy-by-design into the AI architecture a critical business requirement.

What are the data residency requirements for AI GPU clusters under Annex A 5.34?

To meet privacy mandates, AI companies must ensure that PII processed by third-party GPU clusters or LLM APIs remains within legally approved jurisdictions. This requires documenting the physical location of servers to ensure 100% compliance with international data transfer laws, preventing the illegal export of personal data to non-equivalent regimes.

About the author

Stuart Barker
🎓 MSc Security 🛡️ Lead Auditor 30+ Years Exp 🏢 Ex-GE Leader

Stuart Barker

ISO 27001 Ninja

Stuart Barker is a veteran practitioner with over 30 years of experience in systems security and risk management. Holding an MSc in Software and Systems Security, he combines academic rigor with extensive operational experience, including a decade leading Data Governance for General Electric (GE).

As a qualified ISO 27001 Lead Auditor, Stuart possesses distinct insight into the specific evidence standards required by certification bodies. His toolkits represent an auditor-verified methodology designed to minimise operational friction while guaranteeing compliance.

Shopping Basket
Scroll to Top