Authors
Xuhai Xu, Bingsheng Yao, Yuanzhe Dong, Saadia Gabriel, Hong Yu, James Hendler, Marzyeh Ghassemi, Anind K Dey, Dakuo Wang
Publication date
2024/3
Journal
Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies
Volume
8
Publisher
ACM New York, NY, USA
Description
Advances in large language models (LLMs) have empowered a variety of applications. However, there is still a significant gap in research when it comes to understanding and enhancing the capabilities of LLMs in the field of mental health. In this work, we present a comprehensive evaluation of multiple LLMs on various mental health prediction tasks via online text data, including Alpaca, Alpaca-LoRA, FLAN-T5, GPT-3.5, and GPT-4. We conduct a broad range of experiments, covering zero-shot prompting, few-shot prompting, and instruction fine-tuning. The results indicate a promising yet limited performance of LLMs with zero-shot and few-shot prompt designs for mental health tasks. More importantly, our experiments show that instruction finetuning can significantly boost the performance of LLMs for all tasks simultaneously. Our best-finetuned models, Mental-Alpaca and Mental-FLAN-T5, outperform the best …
Total citations
20222023202411355
Scholar articles
X Xu, B Yao, Y Dong, H Yu, J Hendler, AK Dey… - arXiv preprint arXiv:2307.14385, 2023
X Xu, B Yao, Y Dong, S Gabriel, H Yu, J Hendler… - Proceedings of the ACM on Interactive, Mobile …, 2024