Skip to content Skip to sidebar Skip to footer

What PIC/S Annex 22 Actually Means for Your Manufacturing Operations

Hands holding a glowing digital globe labeled “PIC/S Annex 22,” with connected icons representing AI, systems, and cloud controls.

Important Note: This article discusses a draft guidance document currently under consultation. PIC/S Annex 22 has not been finalised or implemented as of December 2025.

AI in pharmaceutical manufacturing is moving from experimental to regulated. In July 2025, the Pharmaceutical Inspection Co-operation Scheme (PIC/S), in partnership with the European Medicines Agency (EMA), released Annex 22 for stakeholder consultation—the first dedicated GMP framework proposal for artificial intelligence and machine learning in medicine production.

The draft Annex 22 sets proposed boundaries: which AI systems could work in critical manufacturing, how to validate them properly, and what controls are expected when the guidance finalises.

If you manufacture medicines under GMP and use AI in any area that affects product quality, patient safety, or data integrity, this proposed framework will likely apply to you once adopted.

Why Annex 22 Matters

Previous guidance tried forcing AI into frameworks built for traditional software. The original Annex 11 (Computerised Systems) was written before modern machine learning existed. It assumes deterministic systems—identical inputs always produce identical outputs.

AI doesn’t work that way.

Machine learning models find patterns in data and make predictions based on probabilities. They can produce different results from identical inputs depending on training data, hyperparameters, or random initialisation. Some systems learn continuously from new data, adapting behaviour during operation.

That unpredictability challenges GMP principles. When your automated system classifies a batch as acceptable or flags a critical deviation, regulators and patients need confidence that the decision is reproducible, explainable, and traceable.

The draft Annex 22 proposes to resolve this by drawing clear lines: some AI architectures could work in critical applications if properly controlled, while others would be prohibited.

Note: Concurrently with Annex 22, PIC/S and EMA released a revised Annex 11 that updates the broader computerised systems framework to complement the AI-specific guidance.

The Core Proposed Requirement: Static, Deterministic Models Only

Here’s the fundamental constraint proposed in the draft: only static AI models with deterministic outputs may be used in critical GMP applications.

What this means in practice:

Potentially Allowed:

  • Models where parameters freeze after training

  • Systems producing identical outputs from identical inputs every time

  • Supervised learning models that classify or predict based on validated data patterns

Proposed as Prohibited:

  • Dynamic models that continue learning during routine operation

  • Non-deterministic systems produce variable outputs from identical inputs

  • Generative AI and Large Language Models (LLMs) in critical contexts

  • Any system where behaviour changes without explicit validation

If you’ve deployed adaptive AI that evolves with new production data, be aware that the architecture is likely non-compliant with critical applications under the draft guidance. You may need a fundamental redesign before this guidance becomes mandatory.

Clarification: The prohibition targets non-deterministic outputs, not necessarily the underlying mathematical nature of the algorithm. The key is that identical inputs must produce identical, reproducible results.

Balance-scale diagram comparing “Allowed Models” (predictable outputs, static learning, supervised learning) versus “Prohibited Models” (variable outputs, dynamic learning, generative AI).
Model selection matters: predictable, static approaches are easier to justify for critical GMP applications.

What Counts as Critical Applications

The draft Annex 22 applies to computerised systems in medicinal product and active substance manufacturing, where AI models predict or classify data affecting:

  • Product quality: batch disposition, QC results, deviation classification

  • Patient safety: release decisions, stability predictions, contamination detection

  • Data integrity: automated review decisions, record classification, audit trail analysis

Non-critical applications—drafting SOPs, training materials, project planning—could use broader AI capabilities under the draft, but only with qualified Human-in-the-Loop (HITL) oversight.

The critical distinction: if the AI decision directly affects whether medicine reaches patients, it falls under the strictest proposed controls.

Five Key Requirement Areas in the Draft Guidance

Note: The following framework is an interpretive summary to help organisations understand the draft requirements. “Five Pillars” is not official PIC/S terminology.

1. Intended Use Definition—Before Anything Else

Every AI model would require comprehensive intended-use documentation approved by process Subject Matter Experts (SMEs) before acceptance testing begins.

This document must specify:

  • Complete description of what the model predicts or classifies

  • Full characterisation of input data, including expected variations

  • Identification of limitations, boundary conditions, and potential biases

  • Defined subgroups based on relevant characteristics (geography, equipment types, defect categories)

  • Clear operator responsibilities when used in HITL configurations

This isn’t checkbox documentation. It serves as the proposed foundation for the validation scope, test data requirements, and acceptance criteria. Without SME approval, the draft guidance indicates you cannot proceed to testing.

2. Test Data Independence—The Non-Negotiable Control

Here’s where most AI validation could fail under the draft: test data must be completely independent from training and validation datasets.

If you split-test data from a shared pool, you must prevent anyone with access to training data from viewing test data. The draft proposes:

  • Access restrictions with audit trails

  • Four-eyes principle enforcement

  • Version control with traceability

  • Documentation proving independence

This addresses the fundamental risk: if developers see test data, they may optimise models to pass tests rather than demonstrate real-world performance.

3. Representative Sampling and Verified Labelling

Test data must represent the complete input sample space—not just common scenarios.

Required sampling approach:

  • Stratified sampling covering all subgroups

  • Inclusion of rare variations and boundary conditions

  • Documented rationale for selection criteria

  • Coverage of potential error modes

Label verification requirements:
Test data labels must achieve high correctness through:

  • Independent expert verification

  • Validated equipment measurements

  • Laboratory testing results

  • Cross-validation by multiple SMEs

The draft emphasizes: you cannot validate an AI model with incorrectly labelled data. Garbage in, garbage out applies fully in GMP environments.

4. Explainability and Confidence Scoring

For critical applications, the draft proposes that AI must explain its decisions in ways humans can verify.

Feature attribution requirements:
Systems must capture and record which input features contributed to each classification. Techniques like SHAP values, LIME, or visual heat maps must highlight key factors driving outcomes.

This would allow auditors and SMEs to verify the model bases decisions on relevant, appropriate features—not spurious correlations.

Confidence scoring:
Models must log confidence scores for each prediction. When confidence falls below thresholds, the system should flag outcomes as “undecided” instead of making unreliable predictions. This creates clear escalation paths for human review when situations fall outside validated experience.

5. Performance Metrics and Acceptance Criteria

Before testing, SMEs must define performance metrics and acceptance criteria at least as stringent as the process being replaced.

Required metrics:

  • Confusion matrix (true positives, false positives, true negatives, false negatives)

  • Sensitivity (true positive rate)

  • Specificity (true negative rate)

  • Accuracy, precision, F1 score

  • Metrics specific to intended use

If you cannot quantify your current manual process performance, the draft guidance indicates you cannot validate the AI replacement. This forces an honest assessment of baseline capability before automation.

Proposed Operational Controls After Deployment

Validation wouldn’t end at go-live. The draft Annex 22 mandates continuous controls:

  • Change Control: Any change to the model, system, or manufacturing process would require evaluation for retesting. Decisions not to retest must be documented and justified.
  • Configuration Management: Models must be under configuration control with measures that detect unauthorised changes. Version control, access restrictions, and audit trails are mandatory.
  • Performance Monitoring: Continuous monitoring must detect model drift from defined metrics or changes in input data distribution. Automated alerts should trigger an investigation when performance degrades.
  • Input Space Monitoring: Regular checks ensure new operational data remains within the validated input sample space. Statistical measures must detect data drift before it affects model reliability.

What This Could Mean for Different Teams

Quality Assurance

Your audit scope may expand significantly. Inspectors could examine:

  • AI governance frameworks and policies

  • Multidisciplinary team composition and qualifications

  • Documentation traceability from intended use through operation

  • Training records for AI-specific risks and controls

Audit readiness would demand evidence that AI systems are “controllable, testable, and verifiable”—not just functional.

Data Science Teams

Technical architecture may need to change:

  • Dynamic learning capabilities must be disabled or removed for critical applications

  • Data governance requires technical controls ensuring test data independence

  • MLOps pipelines must incorporate GMP change control and configuration management

  • Explainability tools must be integrated into the model serving infrastructure

Manufacturing Operations

Process validation must address AI-specific risks:

  • Supplier qualification now includes AI development documentation requirements

  • Sensors feeding AI models must be qualified and calibrated

  • Deviation management expands to cover model performance issues

  • Batch records must consist of AI decision documentation

Australian Implementation Context

Important for Australian manufacturers: TGA has not formally adopted Annex 22 yet, but Australia is an active PIC/S member. The TGA participated in the consultation process and invited Australian stakeholders to provide feedback on the draft guidance.

Practical steps for Australian manufacturers:

  • Engage TGA proactively about AI systems in regulatory submissions

  • Map Annex 22 draft requirements to existing TGA guidance and your current QMS

  • Consider aligning with ISO 42001 (AI Management Systems) for comprehensive governance

  • Document rationale for any variations from the draft Annex 22 in your quality system

Final Status Update

As of December 2025, Annex 22 remains in draft consultation status. This consultation period remained open until October 7, 2025. Organisations should prepare for eventual implementation while recognising that requirements may evolve based on stakeholder feedback. The guidance represents a thoughtful approach to AI governance in GMP environments, but it is not yet an enforceable regulation.

Monitor the PIC/S publications page and EMA updates for the final version and official implementation timeline.

Infographic staircase titled “Achieving AI Compliance by 2026,” showing four steps: Foundation, Governance, Technical Controls, and Scale and Improvement.
A practical phased roadmap to build AI governance and controls for GMP use cases.

PIC/S Annex 22 & Pharma AI: FAQs

Q: What is PIC/S Annex 22 and its current status?

A: PIC/S Annex 22 is draft guidance released July 7, 2025, for stakeholder consultation, developed jointly by PIC/S and EMA. It proposes the first dedicated GMP framework for AI/ML in pharmaceutical manufacturing, focusing on patient safety, product quality, and data integrity. It is not yet finalised or enforceable.

Q: When will PIC/S Annex 22 become mandatory?

A: No implementation date has been set. Following the consultation period (July–October 2025), PIC/S and EMA will review stakeholder feedback before finalising the guidance. Organisations should monitor official PIC/S publications for the final version and implementation timeline.

Q: How does Annex 22 relate to the revised Annex 11?

A: Annex 22 complements the concurrently revised Annex 11 (Computerised Systems). While Annex 22 addresses AI-specific requirements, the revised Annex 11 provides updated foundational requirements for all computerised systems. Both were released together and should be implemented as a comprehensive framework.

Q: Does Annex 22 apply to Australia?

A: As a PIC/S member, Australian manufacturers will likely need to comply once finalised.

Q: What should we do during the draft phase?

A: Conduct gap assessments against draft requirements, inventory AI systems, classify critical vs. non-critical applications, and prepare cross-functional governance structures. Use this period to proactively prepare while awaiting final guidance.

Q: Are there any transitional provisions expected?

A: The draft does not specify transitional provisions. Organisations should prepare for implementation once finalised, but may need to seek clarification from PIC/S or national inspectorates about transition timelines for existing AI systems.