Accepted for/Published in: JMIR Human Factors
Date Submitted: Nov 9, 2021
Date Accepted: Jun 14, 2022
Consumer perspectives on the use of Artificial Intelligence technology and automation in crisis support services: Mixed methods study
ABSTRACT
Background:
Emerging technologies, such as Artificial Intelligence (AI), have the potential to enhance service responsiveness and quality, improve reach to underserved groups, and help address lack of workforce capacity in health and mental health care. Yet, there has been little research on the acceptability of AI, particularly in mental health and crisis support, and how this may inform the development of responsible and responsive innovation in the area.
Objective:
This study aimed to explore: (1) the level of support for the use of technology and automation, such as AI, in Lifeline’s crisis support services in Australia, (2) the likelihood of service use if technology and automation were implemented, (3) the impact of demographic characteristics on the level of support and likelihood of service use, and (4) reasons for not using Lifeline’s crisis support services if technology and automation was implemented in the future.
Methods:
A mixed methods study involving a computer assisted telephone interview and an online survey was undertaken from 2019 to 2020 to explore expectations and anticipated outcomes of Lifeline’s crisis support services in a nationally representative community sample (n=1,300) and Lifeline help-seeker sample (n=553). Participants were aged from 18 to 93 years. Quantitative descriptive analysis, binary logistic regression models, and qualitative thematic analysis were undertaken to address the research aims.
Results:
One-third of community and help-seeker participants did not support the collection of information about service users through technology and automation, and about half reported that they would be less likely to use the service if automation was introduced. Significant demographic differences were found between the community and help-seeker samples. Of the demographics, only older age predicted being less likely to endorse technology and automation to tailor Lifeline’s crisis support service and use such services (OR=1.48-1.66, P<.01). The most common reason for reluctance, reported by both samples, was that respondents wanted to speak to a real person, with the assumption that human counsellors would be replaced with automated robots or machine services.
Conclusions:
Even though Lifeline plans to always have a real person providing its crisis support, help-seekers automatically fear this will not be the case if new technology and automation are introduced. Consequently, incorporating innovative use of technology to improve help-seeker outcomes in such services will require careful messaging and assurances that the human connection will continue.
Citation
Request queued. Please wait while the file is being generated. It may take some time.
Copyright
© The authors. All rights reserved. This is a privileged document currently under peer-review/community review (or an accepted/rejected manuscript). Authors have provided JMIR Publications with an exclusive license to publish this preprint on it's website for review and ahead-of-print citation purposes only. While the final peer-reviewed paper may be licensed under a cc-by license on publication, at this stage authors and publisher expressively prohibit redistribution of this draft paper other than for review purposes.