Maintenance Notice

Due to necessary scheduled maintenance, the JMIR Publications website will be unavailable from Wednesday, July 01, 2020 at 8:00 PM to 10:00 PM EST. We apologize in advance for any inconvenience this may cause you.

Who will be affected?

Currently accepted at: JMIR Formative Research

Date Submitted: Nov 9, 2025
Open Peer Review Period: Nov 10, 2025 - Jan 5, 2026
Date Accepted: Mar 5, 2026
(closed for review but you can still tweet)

This paper has been accepted and is currently in production.

It will appear shortly on 10.2196/87465

The final accepted version (not copyedited yet) is in this tab.

Warning: This is an author submission that is not peer-reviewed or edited. Preprints - unless they show as "accepted" - should not be relied on to guide clinical practice or health-related behavior and should not be reported in news media as established information.

Differences in Safety Risks across Languages for Health Large Language Models: A Cross-Language Vulnerability Study

  • Saubhagya Joshi; 
  • Melissa Mendoza; 
  • Yonaira Rivera; 
  • Vivek K. Singh

ABSTRACT

Background:

Large language models (LLMs) such as ChatGPT are increasingly used to support health-related queries and decision-making. However, these models can be “jailbroken” through adversarial prompts that bypass safety filters and elicit harmful or medically inappropriate responses. In healthcare contexts, such vulnerabilities pose serious risks. Understanding how jailbreak susceptibility varies across languages is essential for developing robust safeguards and promoting equitable access to safe health information.

Objective:

This study aims to systematically compare and contrast the vulnerability of a health LLM for jailbreaking across three languages: English, Spanish, and Hindi (transliterated using the Latin alphabet) based on emoji and permutation cipher attacks.

Methods:

We analyzed 1,000 input prompts per language, drawn from the BeaverTails dataset, across three harm categories: self-harm, violence, and drug abuse. Each prompt was modified using emoji and permutation cipher techniques, resulting in 6,000 input-output pairs. Model responses were evaluated by human coders to determine the success rate of jailbreak attempts across languages and cipher types.

Results:

Hindi prompts showed the highest vulnerability, with 787 successful jailbreaks using emoji ciphers and 873 using permutation ciphers. Spanish and English followed, with lower success rates across both cipher types. Differences in jailbreak success across languages and cipher strategies were statistically significant. Additionally, attacks targeting violence-related prompts were more successful overall than those targeting drug-related or self-harm content, indicating variation in vulnerability by harm type.

Conclusions:

The findings of this formative study reveal that LLM safety performance varies substantially across languages and harm categories, raising concerns about equitable protection in multilingual health communication. Disparities in access to harmful content may contribute to downstream health risks. Strengthening multilingual content moderation and developing language-aware safety mechanisms are critical steps toward safer and more inclusive health AI systems.


 Citation

Please cite as:

Joshi S, Mendoza M, Rivera Y, Singh VK

Differences in Safety Risks across Languages for Health Large Language Models: A Cross-Language Vulnerability Study

JMIR Preprints. 09/11/2025:87465

DOI: 10.2196/preprints.87465

URL: https://preprints.jmir.org/preprint/87465

Download PDF


Request queued. Please wait while the file is being generated. It may take some time.

© The authors. All rights reserved. This is a privileged document currently under peer-review/community review (or an accepted/rejected manuscript). Authors have provided JMIR Publications with an exclusive license to publish this preprint on it's website for review and ahead-of-print citation purposes only. While the final peer-reviewed paper may be licensed under a cc-by license on publication, at this stage authors and publisher expressively prohibit redistribution of this draft paper other than for review purposes.