Maintenance Notice

Due to necessary scheduled maintenance, the JMIR Publications website will be unavailable from Wednesday, July 01, 2020 at 8:00 PM to 10:00 PM EST. We apologize in advance for any inconvenience this may cause you.

Who will be affected?

Currently accepted at: JMIR Formative Research

Date Submitted: Sep 23, 2025
Date Accepted: Jan 22, 2026

This paper has been accepted and is currently in production.

It will appear shortly on 10.2196/84391

The final accepted version (not copyedited yet) is in this tab.

Warning: This is an author submission that is not peer-reviewed or edited. Preprints - unless they show as "accepted" - should not be relied on to guide clinical practice or health-related behavior and should not be reported in news media as established information.

Designing Psychologically-Grounded AI for Supporting Bystander-based Cyberaggression Intervention: A Mixed Methods Exploratory Study

  • Jinkyung Katie Park; 
  • Alina Yu; 
  • Vignesh Krishnan; 
  • Huaye Li; 
  • Linda Reddy; 
  • Vivek K Singh

ABSTRACT

Background:

Cyberaggression poses a growing threat to mental health, particularly among youth, contributing to increased distress, reduced self-esteem, and other adverse psychosocial outcomes. Although bystander intervention can mitigate the escalation and impact of cyberaggression, individuals often lack the confidence, strategies, or language to respond effectively in these high-stakes online interactions. Advances in generative artificial intelligence (AI) present a novel opportunity to facilitate digital behavior change by assisting bystanders with contextually appropriate, theory-informed intervention messages that promote safer online environments and support mental well-being.

Objective:

The current mixed-method design study explored the feasibility of using generative AI to support bystander intervention in cyberaggression on social media. Specifically, we examined whether AI can generate effective responses aligned with established intervention strategies and how these responses are perceived in terms of their potential to de-escalate online harm and foster behavior change.

Methods:

We collected 1,000 real-world cyberaggression examples from public social media datasets and generated bystander intervention responses using three distinct prompt strategies: a generic policy reminder, a baseline GPT prompt, and a theory-driven GPT prompt (AllyGPT). To evaluate the responses, we conducted computational linguistic analyses to assess their psycholinguistic features and carried out a mixed-methods evaluation. Three trained coders rated each message on favorability, conversational impact, and potential for behavior change, and later participated in semi-structured interviews to reflect on their evaluation process and perceptions of intervention effectiveness.

Results:

Linguistic analyses revealed that baseline GPT responses exhibited more emotionally positive and authentic language compared to AllyGPT responses, which showed a more analytical and assertive tone. Policy reminder messages were linguistically rigid and lacked emotional nuance. Human evaluation results showed that AllyGPT responses received the highest effective ratings for low-incivil cyberaggression cases in two dimensions (favorability and changing behavior); baselineGPT works better for mid and high levels for all effectiveness dimensions. For medium- and high-incivil aggressions, baseline GPT responses received the highest ratings across all three dimensions of effectiveness (favorability, discussion-shifting potential, and likelihood of changing bullying behavior), followed by AllyGPT, with policy reminders rated lowest. Qualitative feedback further emphasized that baseline GPT responses were perceived as natural and inclusive, while AllyGPT responses, although grounded in psychological theory, were sometimes viewed as overly direct. Policy reminders were considered clear but lacked persuasive impact.

Conclusions:

Our work showed that designing effective AI-generated bystander interventions requires more than just generating appropriate content; it demands a deep sensitivity to platform culture, social context, and user expectations. By combining psychological theory with adaptive, conversational design and ongoing feedback loops, future systems can better support bystanders, delivering interventions that are not only contextually appropriate but also socially resonant and behaviorally impactful. As such, this work serves as a foundation for scalable, human-centered AI systems that promote safer online spaces and users’ mental well-being.


 Citation

Please cite as:

Park JK, Yu A, Krishnan V, Li H, Reddy L, Singh VK

Designing Psychologically-Grounded AI for Supporting Bystander-based Cyberaggression Intervention: A Mixed Methods Exploratory Study

JMIR Preprints. 23/09/2025:84391

DOI: 10.2196/preprints.84391

URL: https://preprints.jmir.org/preprint/84391

Download PDF


Request queued. Please wait while the file is being generated. It may take some time.

© The authors. All rights reserved. This is a privileged document currently under peer-review/community review (or an accepted/rejected manuscript). Authors have provided JMIR Publications with an exclusive license to publish this preprint on it's website for review and ahead-of-print citation purposes only. While the final peer-reviewed paper may be licensed under a cc-by license on publication, at this stage authors and publisher expressively prohibit redistribution of this draft paper other than for review purposes.