Maintenance Notice

Due to necessary scheduled maintenance, the JMIR Publications website will be unavailable from Wednesday, July 01, 2020 at 8:00 PM to 10:00 PM EST. We apologize in advance for any inconvenience this may cause you.

Who will be affected?

Accepted for/Published in: Journal of Medical Internet Research

Date Submitted: Nov 17, 2025
Open Peer Review Period: Nov 17, 2025 - Jan 12, 2026
Date Accepted: Dec 12, 2025
(closed for review but you can still tweet)

The final, peer-reviewed published version of this preprint can be found here:

Data Poisoning Vulnerabilities Across Health Care Artificial Intelligence Architectures: Analytical Security Framework and Defense Strategies

Abtahi F, Seoane F, Pau I, Vega M

Data Poisoning Vulnerabilities Across Health Care Artificial Intelligence Architectures: Analytical Security Framework and Defense Strategies

J Med Internet Res 2026;28:e87969

DOI: 10.2196/87969

PMID: 41575020

PMCID: 12881903

Data Poisoning Vulnerabilities Across Healthcare AI Architectures: Analytical Security Framework and Defense Strategies

  • Farhad Abtahi; 
  • Fernando Seoane; 
  • Ivan Pau; 
  • Mario Vega

ABSTRACT

Background:

Healthcare artificial intelligence (AI) systems are increasingly integrated into clinical workflows yet remain vulnerable to data poisoning attacks. A small number of manipulated training samples can compromise AI models used for diagnosis, documentation, and resource allocation. Existing privacy regulations, including HIPAA and GDPR, may inadvertently complicate anomaly detection and cross-institutional auditing, thereby limiting visibility into adversarial activity.

Objective:

This study provides a comprehensive threat analysis of data poisoning vulnerabilities across major healthcare AI architectures. The goals are to (1) identify attack surfaces in clinical AI systems, (2) evaluate the feasibility and detectability of poisoning attacks demonstrated in prior security research, and (3) propose a multi-layered defense framework appropriate for healthcare settings.

Methods:

We synthesized empirical findings from 41 key security studies published between 2019 and 2025 and integrated them into an analytical threat modeling framework specific to healthcare. We constructed eight hypothetical yet technically grounded attack scenarios across four categories: (A) architecture-specific attacks on CNNs, LLMs, and reinforcement learning agents; (B) infrastructure exploitation in federated learning and clinical documentation pipelines; (C) poisoning of critical resource allocation systems; and (D) supply chain attacks affecting commercial foundation models. Scenarios were aligned with realistic insider-access threat models and current clinical deployment practices.

Results:

Multiple empirical studies across diverse model sizes and datasets suggest that attackers with access to as few as 100–500 poisoned samples can compromise healthcare AI systems regardless of dataset scale, with attack success rates typically ≥60%. We estimate that detection delays commonly range from 6–12 months and may extend to years in distributed or privacy-constrained environments. Analytical scenarios highlight that (1) routine insider access creates numerous injection points across healthcare data infrastructure, (2) federated learning amplifies risks by obscuring attribution, and (3) supply chain compromises can simultaneously affect dozens to hundreds of institutions. Privacy regulations further complicate cross-patient correlation and model audit processes, substantially delaying the detection of subtle poisoning campaigns.

Conclusions:

Healthcare AI systems face significant security challenges that current regulatory frameworks and validation practices do not adequately address. We propose a multi-layered defense strategy that combines ensemble disagreement monitoring, adversarial testing, privacy-preserving yet auditable mechanisms, and strengthened governance requirements. Ensuring patient safety may require shifting from opaque, high-performance models toward more interpretable and constraint-driven architectures with verifiable robustness guarantees. Clinical Trial: This study did not involve human subjects or the collection, use, or analysis of personal or patient data. The research presents an analytical examination of data poisoning vulnerabilities in healthcare AI systems through theoretical attack scenarios and defense frameworks. As no human participants were involved and no personal data was accessed or processed, ethical approval was not required for this work. All examples and scenarios presented are hypothetical constructs designed to illustrate potential security vulnerabilities and do not reflect actual patient data, clinical records, or identifiable health information.


 Citation

Please cite as:

Abtahi F, Seoane F, Pau I, Vega M

Data Poisoning Vulnerabilities Across Health Care Artificial Intelligence Architectures: Analytical Security Framework and Defense Strategies

J Med Internet Res 2026;28:e87969

DOI: 10.2196/87969

PMID: 41575020

PMCID: 12881903

Download PDF


Request queued. Please wait while the file is being generated. It may take some time.

© The authors. All rights reserved. This is a privileged document currently under peer-review/community review (or an accepted/rejected manuscript). Authors have provided JMIR Publications with an exclusive license to publish this preprint on it's website for review and ahead-of-print citation purposes only. While the final peer-reviewed paper may be licensed under a cc-by license on publication, at this stage authors and publisher expressively prohibit redistribution of this draft paper other than for review purposes.