Accepted for/Published in: Journal of Medical Internet Research
Date Submitted: Jul 4, 2019
Open Peer Review Period: Jul 8, 2019 - Sep 2, 2019
Date Accepted: Jan 28, 2020
(closed for review but you can still tweet)
Transfer Learning for Risk Classification of Social Media Posts: Model Evaluation Study
ABSTRACT
Background:
Mental illness affects a significant portion of the worldwide population. Online mental health forums can provide a supportive environment for those afflicted and also generate a large amount of data which can be mined to predict mental health states using machine learning methods.
Objective:
We benchmark multiple methods of text feature representation for social media posts and compare their downstream use with automated machine learning (AutoML) tools. We test on datasets that contain posts labelled for perceived suicide risk or moderator attention in the context of self-harm. Specifically, we assess the ability of the methods to prioritize posts that a moderator would identify for immediate response.
Methods:
We used 1588 labeled posts from the CLPsych 2017 shared task collected from the Reachout.com forum (Milne et al., 2019). Posts were represented using lexicon based tools including VADER, Empath, LIWC and also used pre-trained artificial neural network models including DeepMoji, Universal Sentence Encoder, and GPT-1. We used TPOT and Auto-sklearn as AutoML tools to generate classifiers to triage the posts.
Results:
The top-performing system used features derived from the GPT-1 model, which was fine-tuned on over 150,000 unlabeled posts from Reachout.com. Our top system had a macro averaged F1 score of 0.572, providing a new state-of-the-art result on the CLPsych 2017 task. This was achieved without additional information from meta-data or preceding posts. Error analyses revealed that this top system often misses expressions of hopelessness. We additionally present visualizations that aid understanding of the learned classifiers.
Conclusions:
We show that transfer learning is an effective strategy for predicting risk with relatively little labeled data. We note that fine-tuning of pre-trained language models provides further gains when large amounts of unlabeled text is available.
Citation
Request queued. Please wait while the file is being generated. It may take some time.
Copyright
© The authors. All rights reserved. This is a privileged document currently under peer-review/community review (or an accepted/rejected manuscript). Authors have provided JMIR Publications with an exclusive license to publish this preprint on it's website for review and ahead-of-print citation purposes only. While the final peer-reviewed paper may be licensed under a cc-by license on publication, at this stage authors and publisher expressively prohibit redistribution of this draft paper other than for review purposes.