The Chatbot Mental Health
The Chatbot Mental Health
Tripti Pal1, Adrija Saha1, Shounak Dutta1, Soumidatta Paul1, Rupsha Basu1,
Ranit De1, Triparna Mukherjee1, Devjyoti Ghosh1, Arunima Das1, Abhishek
Mukhopadhyay2*
1
Amity Institute of Biotechnology,
Amity University
Kolkata: 700135
2
Computer Science and Engineering
School of Engineering and Technology
Amity University
Kolkata: 700135
Email ID of the Corresponding author: amukhopadhyay@kol.amity.edu
ABSTRACT
Mental health disorders affect nearly a billion people worldwide, with anxiety and
depression as the leading contributors to this global crisis. Despite advancements in treatments
like Cognitive Behavioral Therapy (CBT) and psychotherapy, access to mental health care
remains a significant challenge due to financial constraints, geographical barriers, and a shortage
of trained professionals. Recent technological advancements, particularly in artificial intelligence
(AI), have introduced innovative solutions such as AI-powered chatbots, which offer scalable,
cost-effective, and accessible mental health support. This chapter explores the transformative
potential of AI chatbots in mental health care, focusing on their ability to simulate human
conversation, deliver therapeutic interventions, and provide 24/7 personalized support in a
stigma-free environment. By examining popular applications like Woebot, Wysa, MoodKit, and
Happify, this chapter highlights how AI-driven tools have effectively addressed symptoms of
anxiety, depression, and stress, particularly during the post-COVID-19 era. The chapter also
delves into the integration of cutting-edge technologies like Large Language Models (LLMs) to
enhance chatbot performance and user engagement. This chapter also outlines challenges such
as ethical concerns, long-term efficacy, and system integration, providing future research
directions for developing more holistic, inclusive, and impactful AI-driven mental health
interventions.
Artificial intelligence (AI)-based interventions such as chatbots and therapy bots and
specialized chat applications are able to overcome the above-mentioned shortcomings of
traditional psychotherapy. Technological influence on mental health has dramatically
accelerated social interaction and ways of communicating with others, as well as influencing
behavior and personality. Furthermore, Cyber Health Psychology has brought about a significant
transformation in the landscape of mental health support through the convergence of
psychology, health, and digital technology for better service delivery and more patient
participation. The emphasis here is on the psychological effects of technologies in healthcare
services that include mobile health applications, telemedicine, and online forums towards the
designing of digital interventions for healthy behavior and improvement of better mental health
outcomes [9]. Research findings indicate that online interventions have proved effective in
treating mental diseases like depression, anxiety, drug and substance use, and eating disorders
[10]. Additionally, the collaboration of technology with psychology, especially spatial computing
and AI, is promising to aid mental health [11]. AI psychotherapy can be delivered online through
smartphone applications and digital games.
Over the past decade, improvements in computational powers and the availability of
open-source technologies have provided AI-based chatbot programs with extensive popularity.
Improvements in AI and Natural Language Processing (NLP) ensure that it is easier to implement
chatbots to accommodate a variety of applications and to simulate human conversation with
more effectiveness. One of the most important technological innovations within mental health
care is the introduction of chatbots, or conversational agents. AI chatbots are advanced
conversational systems designed to simulate human interactions in a natural and intuitive
manner. These computer programs can think, learn, and complete tasks in combination with
humans or with independence, using big data, NLP and Machine Learning (ML) algorithms. These
chatbots utilize natural language processing to interpret user input and generate contextually
appropriate responses, enhancing their functionality to boost productivity while offering
conversation, guidance, and support [12] [13].
Given the increasing demand for mental health care on one hand and not enough mental
health care professionals available on the other hand, many AI-driven mental health apps and
chatbots have been developed to make the support more accessible, potential from therapy to
assistance. According to Moore & Caudill [14], a chatbot refers to a computer program that takes
the form of a human conversation, enabling it to respond to any question or message in text or
voice. It offers 24/7 confidential support aside from conventional, face-to-face therapy. The apps
provided examples like Happify, Moodkit, Woebot, and Wysa, wherein techniques of CBT more
than anything deliver it as mood tracking and daily mental health checks via conversational
interfaces. Such digital tools were helpful in managing both newly emerging mental health issues
and the aggravation of existing conditions.
AI chatbots can be considered a promising frontier in mental health care that has the
potential to empower the users, complement traditional therapy, and increase the quality of
mental health resources available on a global scale. These tools could play an indispensable role
in shaping the face of mental health care, making it more inclusive, accessible, adaptable, and
flexible to the ever-changing needs of individuals and communities all across the world.
This chapter also reviews the potential challenges and limitations of AI chatbots within mental
health care, including their ability to deliver scalable, round-the-clock support and overcome
stigma-related barriers by offering users privacy and anonymity. By analyzing various use cases,
therapeutic outcomes, and challenges in AI-based mental health interventions, this chapter aims
to present a comprehensive overview of the effectiveness and scalability of chatbot technology in
enhancing mental health outcomes, as well as its role in shaping the future of mental health care
delivery.
Language ambiguity is one of the most common problems that an AI chatbot needs to
tackle while detecting emotions from textual content. The users can use general expressions like
I am fine or everything is alright, which seem to denote a positive emotion that masks deeper
underlying negative emotions. To enable fine-tuned emotional detection, chatbots use deep
learning models that are pre-trained on massive datasets containing textual excerpts from social
media posts and transcripts of popular TV shows. Goemotions is a dataset developed by Google
AI that contains 58,000 comments from Reddit, labelled with 28 emotions [21].
Seq2Emo [22] is a model on which mental health chatbot Wysa and other chatbots are
based. At the core of this model is a seq-2-seq framework that uses encoder-decoder
architecture to process input texts and generate output texts. This is combined with a latent
variable chain model , which introduces a latent variable z (encodes global information about
emotions; the model can generalize unseen data, those not covered in the training set), which
conditions the model to better capture the subtle variations in the emotional context. The
encoder uses LSTM, Bi-LSTM or GRU to enable nuanced emotion detection over long texts. The
Seq2Emo model also makes use of pre-trained models like DeepMoji and Elmo embeddings to
enhance emotion detection. Elmo generates different representations of the same words based
on the emotional connotation of previous texts. GloVe is another embedding used that is not
context-specific, unlike Elmo and DeepMoji, to generate a general understanding of sentiments
and capture word association and semantic relationships. The vector representations of the
words (in the input text) generated are passed through a decoder, wherein attention mechanisms
are used to focus on the integral part of the text, and then it is classified as an emotion label.
BERT-CNN [20] is a most recent deep learning model; it is based on ISEAR and
SemEval. This model combines the functions of BERT and CNN as shown in figure 1. The
output of BERT (Bidirectional Encoder Representations Form Transformers) and then passes it
through the layers of Convoluted Neural Network (CNN). It is composed of three main
components.
o Preprocessing of the dataset
The BERT Base model processes text through 12 layers of self-attention to generate a
contextual vector representation.
A CNN, employed as a classifier.
Standard metrics like precision, recall, F1-score, and accuracy are used to evaluate the
model. It was observed that the BERT-CNN model could detect the happy emotion with 98%
precision, and its overall accuracy in detecting emotions is 94.7%.
With reference to figure 2, CBT-based chatbots make up the largest share of chatbots
dedicated to mental health. Cognitive restructuring, one of the core components of CBT, is how AI
chatbots support behavioral interventions, wherein chatbots respond with empathy and guide
users through technique to reframe their negative thoughts. With the help of another principle of
CBT, behavioral activation, wherein they suggest small tasks like taking a walk or engaging in a
hobby, chatbots help in breaking the cycle of inactivity that often accompanies depression.
Chatbots also make use of mindfulness training exercises.
Fig. 2. Distribution of Mental Health Chatbots
PTSD Coach [25] is another mental health chatbot made by the US Department of
Veteran Affairs for war veterans to help with their posttraumatic stress disorder. This app was
successful because of the incorporation of a distress scale consisting of rankings from 0 to 10.
Based on the ranking given by the user, the app will suggest CBT techniques to help them manage
their distress. PTSD Coach also provides personalized coping solutions tailored to the user distress
level, like journaling prompts, music, or art therapy. Ellie is another chatbot created for PTSD by
US DARPA (Defense Advanced Research Project Agency), which depends on its relationship-
building capabilities to gather personal information on the user and then use this information to
recognize the issues the user is going through without them ever explicitly stating so.
MISHA [27] was developed in collaboration with ETH Zurich. MISHA offers a chat-based
interface enriched with multimedia elements and regular notifications to engage users. It
incorporates evidence-based strategies from Cognitive Behavioral Therapy (CBT), mindfulness,
and psychoeducation to provide insights on stressors and effective coping techniques. The stress
management program integrates key CBT elements, including cognitive restructuring, and the
identification, evaluation, and modification of maladaptive thought patterns. Additionally,
techniques like Behavioral Activation (BA) and activity monitoring from CBT are applied to assist
participants in achieving their goals through a collaborative approach.
Youper, Melinda, XiaoE [28], and Aetna [29] are other chatbots based on CBT,
mindfulness, and the third wave CBT principle, Action and Commitment Therapy (ACT).
Youper allows chat/text-guided psychoeducation, SMART goal setting, as well as tracking
of progress. Aetna is a psychoeducational chatbot for students facing mental health
difficulties which is based on CBT, mindfulness training and positive psychology.
One of the prime concerns when an individual uses AI chatbots is the privacy and
confidentiality of the data. The AI systems on which these chatbots are based collect and analyze a
huge amount of personal data related to the user, and it is very important that the data is handled
in a secure manner and the privacy rights of the user are respected. The users should be able to
have transparent communication, and the AI model should be explainable, and it is of utmost
importance to ensure the consent of the user is received before deploying the AI interventions.
System vulnerabilities are identified by the stride model, which segregates the identified security
threats into spoofing, tampering, information disclosure, denial of service, etc. Data security falls
under 2 categories: the first category is the safe transfer of messages to the server where the
chatbot is hosted; the second category lies on the user end: how the messages and data are
stored, processed, or shared [30].
End-to-end encryption ensures that only the intended recipient can decrypt the message,
preventing unauthorized access by third parties. However, spoofing and tampering can still occur
if a third party gains access to the cryptographic keys. In public key encryption, the user's device
generates two key pairs: a private key and a public key. These key pairs are created using
protocols based on the RSA algorithm. Many chatbots utilize HTTPS, which secures data through
an encrypted connection via SSL (Secure Socket Layer) or TLS (Transport Layer Security). Unlike
end-to-end encryption, HTTPS provides point-to-point encryption, ensuring that data is encrypted
between the client and server but not necessarily end-to-end [31]. Woebot, Replika, and Anna use
this technique.
In many instances, sensitive Personally Identifiable Information (PII) is transmitted. Self-
destructing messages offer a practical solution, automatically erasing messages after a
predetermined period, without requiring user intervention. This process involves both the chatbot
and the user, ensuring that sensitive data is securely deleted within the set timeframe.
Article 5 (e) of the GDPR states that the personal data cannot be kept longer than
necessary for the purpose for which it is being processed. According to PHI, any information about
mental health status should be collected by a covered identity and not linked to a specific
individual. GDPR compliance requires an intent level of privacy. Article 32(a) of GDPR requires the
mental health chatbot companies to act to minimize and encrypt personal data. Open-domain
chatbots are a little more problematic than closed-domain chatbots. Closed-domain chatbots
working with PII require authentication and authorization, and users agree with terms and
conditions. On the contrary, no such precautions are taken by open-domain chat bots. A popular
mental health chatbot, Replika shares some user data with third-party tracking and analytics tools
like Google Analytics and cookies, which helps the company evaluate the type of customers. Wysa,
Youper, and Tess provide services to users for making complaints under HIPPA guidelines. The
new GDPR mandate requires that any technology handling users' personal data must implement
stringent security measures to prevent malicious encryption activities that could infringe on
individuals' personal rights [31].
There are great challenges in mental health care. Accessibility is one of the major barriers.
One of the major benefits of the AI chatbots is that they help enhance accessibility and provide
immediate support in mental health care. Traditional mental health services require
appointments most of the time, resulting in long waiting for periods for patients. AI chatbots, on
the other hand, if used, provide immediate support all round the clock, every day. This non-stop
availability would be essential for some people during times of crisis [32]. Regular monitoring of
sick patients is often considered fruitful as it optimizes their therapy. AI therapy plays an
important role in providing continuous support 24/7. New parents dealing with the ongoing stress
of a chronic illness, combined with the chaos that accompanies parental care, can just turn to the
AI therapists for advice and emotional support whenever needed. Some chronic diseases, such as
diabetes, can benefit from AI therapy, as they offer reminders for medications and lifestyle
modifications, thereby enhancing patient compliance to a great extent. Stressed career
professionals often get overwhelmed, and, due to time constraints, they often find it difficult to
move from their workspace to a traditional therapy session. AI therapists, on several platforms
that are available 24/7, highlight the conveniences of modern-day living by providing on-demand
solutions tailored to the needs of the individual. It takes out the pain of scheduling. With AI
therapy, one can seek help at any time. However late it is, we do not have to wait for office hours,
as help is available at the click of a button [33].
Youper [34] is a 24/7 available mental health support chatbot. It uses artificial intelligence
and machine learning to provide the user with personalized mental health support. The
interactions are based on various techniques involving NLP analysis and analytic data review to
track the progress of users and improve their recommendations. Youper holds real-time
conversations with the users, affording them the opportunity to receive support from wherever
they are and whenever they feel the need. Whenever the user chooses to check in, it provides a
reliable time and space for processing feelings and accessing resources. Youper combines this with
round-the-clock availability and smart personalization for flexibility and responsiveness on the
terms of the users.
Talktotherapist.app [35] offers 24/7 assistance through its AI-driven chatbot named
Sophia, designed to provide mental health support. Sophia is the world’s first chatbot designed
specifically for survivors of domestic violence. It creates a trusted space for survivors to connect
with individuals who can offer support, such as healthcare workers, police officers, and
psychologists. Developed by the Swiss human rights organization Spring ACT, Sophia is available
24/7, providing a listening ear and validation for survivors, regardless of their location. Accessible
from any internet-enabled device, it supports communication in English, French, German, and
Italian.
This 24/7 accessibility of AI chatbots ensures that we are never left alone during a crisis
and we don’t need to travel or wait for appointments by providing users with immediate
assistance and resources.
AI-powered mental health chatbots use sophisticated NLP and Deep Learning (DL) to
engage users in meaningful, personalized conversations that feel responsive and human-like. NLP
is what allows these chatbots to extract from text about the user intentions, emotional tones, and
other context-dependent meanings. The Seq2Seq deep learning model, based on Recurrent
Neural networks (RNNs), plays a pivotal role by remembering prior conversation parts, allowing
the chatbot to maintain conversational context. In practice, this model uses an encoder to
convert the input of the user into a vector summarizing their message and a decoder to generate
a response, building on the previous dialogue and providing users with relevant, coherent
answers. For example, if a user expresses anxiety, the chatbot will retain that context to deliver
appropriate support throughout the conversation.
The effectiveness of the chatbot depends a lot on the training data, which includes large
datasets of past mental health-related conversations, such as anonymized therapy session
transcripts or user interactions. These datasets teach the chatbot to recognize common patterns
in expressions of mental states like sadness or stress, allowing it to adjust its responses to be
empathetic and supportive. This training, however, limits the ability of the chatbot to handle
unfamiliar or complex issues outside the scope of the dataset, as it can only respond to topics it
has learned. Many companies gather and refine data to enhance chatbot responses in specific
areas, improving effectiveness over time. However, chatbots still perform best within certain
domains where they have substantial training data and may struggle with topics outside their
training.
They give real-time support and foster a secure, private space that will help users feel
comfortable enough to provide sensitive information in particular. Yet there are major ethical
issues around personal data and consent to be taken into consideration. The bots should be able
to identify situations that require a human, directing users to appropriate professional help when
necessary [36].
Woebot [37] provides users with personalized conversations by tailoring its responses to the
mental health needs of each individual, with a specific focus on alleviating symptoms of depression
and anxiety. By utilizing CBT principles and NLP, Woebot initiates with an assessment to understand
the user moods and current challenges. Then, based on user mental state and recent interaction, it
crafts the messages so users may know specific coping strategies, undertake exercises, and track
their moods. For example, if a user often brings up anxiety, then Woebot may suggest exercises
addressing anxious thoughts and thus build a custom dialogue in response to the unique experiences
of the individual over time. This allows the Woebot to grow alongside the user, where the system
provides instant support that only gets more individualized as the user continues their journey.
Wysa [37] also provides users with personalized conversations by integrating CBT and
mindfulness techniques into its interactions, designed to address stress, anxiety, and emotional well-
being. Beginning with a conversational assessment, Wysa identifies the main challenges of each user
and curates mindfulness exercises, guided journaling prompts, or other tools to support them. If a
user often expresses feeling overwhelmed, adjust the responses to emphasize stress management
resources and resilience-building exercises. This allows Wysa to create a customized support
experience that feels tailored to each individual, delivering time-to-time adaptive guidance that
evolves with the needs of the user over time. Through ongoing adaptation, Wysa provides
continuous relevant support, bridging gaps between formal therapy sessions and fostering a
sustained therapeutic connection.
Table 1: Summary of the above mentioned chatbots with the corresponding features they possess
4. EFFECTIVENESS AND IMPACT ON STUDENT MENTAL HEALTH
AI-based chatbots offer accessible mental health for anxiety and stress. They provide 24/7
interventions, helping overcome barriers like geography, cost, and stigma. However, challenges
persist, including a lack of empathy and privacy concerns. Depression and anxiety affect 280 million
and 301 million people, worsened by the COVID-19 pandemic [38]. While traditional treatments like
CBT are often inaccessible, AI chatbots like Woebot and Wysa utilize machine learning and natural
language processing to deliver effective, personalized support [39]. Research shows promise, but
further studies are needed to evaluate their long-term efficacy in treating mental health conditions
[33].
The COVID-19 pandemic has significantly affected the mental health of young people,
leading to increased levels of stress, anxiety, and depression. Studies [43] highlight AI chatbots
potential in addressing these issues found that XiaoE, a CBT-based chatbot for college students,
resulted in moderate reductions in depressive symptoms after one week, although smaller at one-
month follow-up. Participants valued its inner support but noted limitations like rigidity. Peuters et.
al., [44] evaluated the LIFEGOALS Health intervention, which included an AI chatbot, showing
positive effects on physical activity, sleep, and mood, especially during in-person schooling. While
engagement was challenging, users appreciated gamification and personalized information. These
findings suggest AI chatbots can support mental health during crises, but their long-term efficacy and
engagement strategies require further research.
A study by Yasukawa et al. [45] explored the potential of AI chatbots in assisting individuals
with specific health conditions. The researchers found that the AI chatbot significantly improved
adherence to internet-based cognitive-behavioral therapy (iCBT) among workers with subthreshold
depression, primarily boosting engagement rather than directly affecting clinical outcomes. Ogawa
et. al. [46] examined an AI chatbot for Parkinson’s disease patients, showing improvements in facial
expressivity and speech fluency, suggesting its role in remote symptom monitoring. Ulrich et. al., in
another study [47], developed the BalanceUP app, which uses a chatbot interface for migraine
management and to an improved mental wellbeing to a large extent, indicating its effectiveness as a
standalone digital intervention.
Bennion et. al. [55] studied AI chatbots for problem solving among older adults, comparing
MYLO and ELIZA. Both reduced problems like distress and anxiety to a large extent, with MYLO being
preferred. Usability ratings were below acceptable thresholds, indicating a need for improvement.
Methods: The trial was an unblinded randomized controlled study involving 83 university students
as participants. Participants were divided into either a test group receiving a brand new developed
intervention via a chatbot or a control group in the form of minimal level bibliotherapy, with a size of
n = 41 and n = 42, in the order given. A series of questionnaires were utilized to assess clinical
variables at baseline and every 4 weeks for 16 weeks. These included the Patient Health
Questionnaire-9 (PHQ-9), Generalized Anxiety Disorder Scale (GAD-7), and the Positive and Negative
Affect Scale (PANAS). To evaluate satisfaction and the therapeutic alliance, the Client Satisfaction
Questionnaire-8 (CSQ-8) and the Working Alliance Inventory-Short Revised (WAI-SR) were
administered following the intervention. Additionally, participants provided self-reported data on
adherence and feedback regarding the therapy chatbot.
Interventions: A therapy chatbot named XiaoNan was specifically developed for the trial and made
accessible through the WeChat Official Accounts Platform. Users can access the chatbot using the
smartphone app WeChat on Windows, macOS, Android, and iOS.
XiaoNan is designed to assist users in identifying and separating their emotions, thoughts,
reactions, and behaviors, akin to sessions with a CBT professional. Additionally, it allows users to
track their emotions daily or engage in casual conversations, offering empathetic interactions in line
with its therapeutic purpose.
Bibliotherapy intervention: Bibliotherapy is a key form of self-help therapy that utilizes literature to
address and alleviate a patient's issues. In this approach, patients engage in psychological
intervention by reading carefully selected literature, guided by the recommendations of
professionals.Previous studies have shown evidence of the effectiveness of bibliotherapy in the
treatment of moderate depression. A systematic review found that the long-term benefits of
bibliotherapy were confirmed through randomized clinical trials. Notably, even after a three-month
follow-up period, adult participants who engaged in cognitive bibliotherapy demonstrated sustained
improvement.
Statistical methods: The analysis was conducted using SPSS version 26, with the significance level set
at 0.05. Baseline variables, including age, gender, education, and initial clinical metrics, were
evaluated using ANOVA and chi-square tests to identify any significant differences between the
groups. This ensured the groups were comparable at the start of the study. Primary outcomes were
evaluated using univariable ANCOVA within the intention-to-treat (ITT) framework, accounting for
the univariate impact of group assignment on results while adjusting for baseline clinical variables.
Missing data, presumed to be missing at random, were managed through SPSS's multiple imputation
procedure. Cohen's d effect sizes were calculated to quantify effect magnitude. Independent t-tests
were conducted on the CSQ-8 and WAI-SR scores, along with self-reported adherence measures.
Additionally, responses to two open-ended questions were analyzed using a word frequency
analysis.
Result: The chatbots proved to be an effective medium for delivering self-help depression
treatment in real-world, pragmatic settings. The therapy chatbot led to a substantial reduction in
depression symptoms, as measured by PHQ-9, over a 16-week period. Additionally, it achieved a
modest reduction in anxiety, as assessed by GAD-7, within the initial 4 weeks. The chatbot
intervention demonstrated greater effectiveness compared to bibliotherapy. Both groups
demonstrated similar levels of client satisfaction and self-reported adherence. However, the chatbot
group experienced a downward trend in adherence rates, potentially attributable to technical and
content-related shortcomings in the chatbot design.Scores achieved in the WAI-SR test by users of
the chatbot were significantly higher compared to users of the chatbot, hence indicating that using
conversational AI can contribute to building a therapeutic alliance. Some of the previous studies
were in line with the present result, in which there were shorter periods of intervention, such as 2
and 4 weeks. An increase in self-disclosure was observed, as reflected in user feedback, where
participants expressed a greater tendency to share personal thoughts and feelings.Moreover, based
on the suggestions for XiaoNan, it indicates that more process factors are greater than content.
5. CHALLENGES
Mental health chatbots offer accessible support for individuals experiencing mental health
challenges.
Unlike human therapists, mental health chatbots often lack nuanced understanding and
contextual awareness, which can lead to inadequate responses in critical situations. For instance,
generative AI models sometimes respond in an insensitive way to distress signals, risking harm
rather than providing support. Such responses may exacerbate the emotional distress of a user, and
without immediate human intervention, these interactions can pose safety concerns [57].
Users express concerns about privacy and data security, which are given that chatbot
interactions are often recorded for machine learning improvements. In addition, the users may feel
uncertain about sharing personal issues with an AI, impacting the effectiveness of the chatbot as a
mental health tool. Transparent data use and improved security protocols could help alleviate these
concerns [58].
Many chatbots are based on pre-scripted responses, which can limit their ability to engage in
dynamic conversation or adapt to complex emotional states. This restriction can make interactions
feel less genuine and may reduce engagement over time. To handle this, advanced AI systems are in
development that would use NLP to identify emotions; they still fall significantly short, though, in the
expression of true empathy [60].
7. REFERENCES
1. “The Latest Mental Health Statistics: What the Numbers Say About the State of Our Minds in
2024.” Huntington Psychological Services, available online: ate-of-our-minds-in-2024.
Accessed 14 Nov. 2024.
2. "World Health Organization Mental Health." World Health Organization, available online:
https://www.who.int/health-topics/mental-health. Accessed 14 Nov. 2024.
3. Javaid, Syed Fahad, et al. "Epidemiology of anxiety disorders: global burden and
sociodemographic associations." Middle East Current Psychiatry, vol. 30, 2023, p. 44.
6. Patel, Vikram, et al. "Reducing the Treatment Gap for Mental Disorders: A WPA Survey."
World Psychiatry, vol. 9, no. 3, 2013.
7. “WHO Report Highlights Global Shortfall in Investment in Mental Health.” World Health
Organization, available online: th. Accessed 14 Nov. 2024.
8. Kapur, N., Gorman, L. S., Quinlivan, L., and Webb, R. T. "Mental Health Services: Quality,
Safety, and Suicide." BMJ Quality & Safety, vol. 31, no. 6, 2022, pp. 419–422.
9. Caponnetto, P., and Milazzo, M. "Cyber Health Psychology: The Use of New Technologies at
the Service of Psychological Well-Being and Health Empowerment." Health Psychology
Research, vol. 7, no. 2, 2019, p. 8559.
10. Taylor, C. B., Graham, A. K., Flatt, R. E., Waldherr, K., and Fitzsimmons-Craft, E. E. "Current
State of Scientific Evidence on Internet-Based Interventions for the Treatment of Depression,
Anxiety, Eating Disorders, and Substance Abuse: An Overview of Systematic Reviews and
Meta-Analyses." European Journal of Public Health, vol. 31, 2021, pp. I3–I10.
11. Spiegel, B. M. R., Liran, O., Clark, A., Samaan, J. S., Khalil, C., Chernoff, R., Reddy, K., and
Mehra, M. "Feasibility of Combining Spatial Computing and AI for Mental Health Support in
Anxiety and Depression." NPJ Digital Medicine, vol. 7, no. 1, 2024, p. 22.
12. Caldarini, G., Jaf, S., and McGarry, K. "A Literature Survey of Recent Advances in Chatbots."
Information, vol. 13, no. 1, 2022, p. 41.
13. Bala, Kumkum, Mukesh Kumar, Sayali Hulawale, and Sahil Pandita. "Chat-Bot for College
Management System Using AI." International Research Journal of Engineering and
Technology (IRJET), 2018, pp. 2395–0056.
14. Moore, J. R., and Caudill, R. "The Bot Will See You Now: A History and Review of Interactive
Computerized Mental Health Programs." The Psychiatric Clinics of North America, vol. 42, no.
4, 2019, pp. 627–634.
15. Abd-alrazaq, Alaa A., Mohannad Alajlani, Ali Abdallah Alalwan, Bridgette M. Bewick, Peter
Gardner, and Mowafa H. "An Overview of the Features of Chatbots in Mental Health: A
Scoping Review." International Journal of Medical Informatics, vol. 132, 2019, p. 103978.
16. Paulus, M. P., Huys, Q. J., and Maia, T. V. "A Roadmap for the Development of Applied
Computational Psychiatry." Biological Psychiatry: Cognitive Neuroscience and Neuroimaging,
vol. 1, no. 5, 2016, pp. 386–392.
17. Moilanen, J., van Berkel, N., Visuri, A., Gadiraju, U., van der Maden, W., and Hosio, S.
"Supporting Mental Health Self-Care Discovery Through a Chatbot." Frontiers in Digital
Health, vol. 5, 2023, p. 1034724.
18. Chakraborty, C., Pal, S., Bhattacharya, M., Dash, S., and Lee, S. S. "Overview of Chatbots with
Special Emphasis on Artificial Intelligence-Enabled ChatGPT in Medical Science." Frontiers in
Artificial Intelligence, vol. 6, 2023, p. 1237704.
19. COVID-19 Mental Disorders Collaborators. "Global Prevalence and Burden of Depressive and
Anxiety Disorders in 204 Countries and Territories in 2020 Due to the COVID-19 Pandemic."
The Lancet, vol. 398, no. 10312, 2021, pp. 1700–1712.
20. Abas, A. R., Elhenawy, I., Zidan, M., and Othman, M. "BERT-CNN: A Deep Learning Model for
Detecting Emotions from Text." Computers, Materials & Continua, vol. 71, no. 2, 2022, pp.
2943–2961, https://doi.org/10.32604/cmc.2022.021671.
21. CERC 2021: Collaborative European Research Conference, 9–10 Sept. 2021, Cork, Ireland.
22. Huang, C., Trabelsi, A., Qin, X., Farruque, N., and Zaiane, O. R. "Seq2Emo for Multi-Label
Emotion Classification Based on Latent Variable Chains Transformation." ArXiv,
abs/1911.02147, 2019.
23. Mitsea, E., Drigas, A., and Skianis, C. "Digitally Assisted Mindfulness in Training Self-
Regulation Skills for Sustainable Mental Health: A Systematic Review." Behavioral Sciences
(Basel, Switzerland), vol. 13, no. 12, 2023, p. 1008.
24. Greer, S., Ramo, D., Chang, Y. J., Fu, M., Moskowitz, J., and Haritatos, J. "Use of the Chatbot
'Vivibot' to Deliver Positive Psychology Skills and Promote Well-Being Among Young People
After Cancer Treatment: Randomized Controlled Feasibility Trial." JMIR mHealth and
uHealth, vol. 7, no. 10, 2019, p. e15018.
25. Gamble, A. "Artificial Intelligence and Mobile Apps for Mental Healthcare: A Social
Informatics Perspective." Aslib Journal of Information Management, vol. 72, no. 4, 2020, pp.
509–523.
26. Boucher, E. M., et al. "Artificially Intelligent Chatbots in Digital Mental Health Interventions:
A Review." Expert Review of Medical Devices, vol. 18, sup. 1, 2021, pp. 37–49.
27. Ulrich, S., Lienhard, N., Künzli, H., and Kowatsch, T. "A Chatbot-Delivered Stress
Management Coaching for Students (MISHA App): Pilot Randomized Controlled Trial." JMIR
mHealth and uHealth, vol. 12, 2024, p. e54945.
28. Casu, Mirko, Sergio Triscari, Sebastiano Battiato, Luca Guarnera, and Pasquale Caponnetto.
"AI Chatbots for Mental Health: A Scoping Review of Effectiveness, Feasibility, and
Applications." Applied Sciences, vol. 14, no. 13, 2024, p. 5889.
29. Ab Razak, N. I., Muhammad Yusoff, M. F., and Rahmat, R. W. O. K. "ChatGPT Review: A
Sophisticated Chatbot Model in Medical and Health-Related Teaching and Learning." BMSC,
2023.
30. Saeidnia, Hamid Reza, et al. "Ethical Considerations in Artificial Intelligence Interventions for
Mental Health and Well-Being: Ensuring Responsible Implementation and Impact." Social
Sciences, vol. 13, no. 7, 2024, p. 381.
31. Hasal, M., et al. "Chatbots: Security, Privacy, Data Protection, and Social Aspects."
Concurrency and Computation: Practice and Experience, vol. 33, no. 19, 2021.
32. Miner, Adam S., et al. "Talking to Machines About Personal Mental Health Problems." JAMA,
vol. 318, no. 13, 2017, pp. 1217–1218.
33. Vaidyam, Anmol N., et al. "Chatbots and Conversational Agents in Mental Health: A Review
of the Psychiatric Landscape." Canadian Journal of Psychiatry, vol. 64, no. 7, 2019, pp. 456–
464.
34. Fitzpatrick, Kathleen K., et al. "Delivering Cognitive Behavior Therapy to Young Adults with
Symptoms of Depression and Anxiety Using a Fully Automated Conversational Agent
(Woebot): A Randomized Controlled Trial." JMIR Mental Health, vol. 4, no. 2, 2017, p. e19.
35. Viduani, Anna, et al. "Chatbots in the Field of Mental Health: Challenges and Opportunities."
Advances in Experimental Medicine and Biology, 2023, doi:10.1007/978-3-031-10698-9_8.
36. Hill, Jennifer, et al. "Real Conversations with Artificial Intelligence: A Comparison Between
Human–Human Online Conversations and Human–Chatbot Conversations." Computers in
Human Behavior, vol. 49, 2015, pp. 245–250, doi:10.1016/j.chb.2015.02.026.
37. Daley, Kate, et al. "Preliminary Evaluation of the Engagement and Effectiveness of a Mental
Health Chatbot." Frontiers in Digital Health, vol. 2, 2020, p. 576361,
doi:10.3389/fdgth.2020.576361.
38. Xiong, Jiaqi, et al. "Impact of COVID-19 Pandemic on Mental Health in the General
Population: A Systematic Review." Journal of Affective Disorders, vol. 277, 2020, pp. 55–64.
39. Inkster, Becky, et al. "An Empathy-Driven, Conversational Artificial Intelligence Agent (Wysa)
for Digital Mental Wellbeing: Real-World Data Evaluation Mixed-Methods Study." JMIR
mHealth and uHealth, vol. 6, no. 11, 2018, p. e12106, doi:10.2196/12106.
40. Fulmer, Rachel, et al. "Using Psychological Artificial Intelligence (Tess) to Relieve Symptoms
of Depression and Anxiety: Randomized Controlled Trial." JMIR Mental Health, vol. 5, no. 4,
2018, p. e64, doi:10.2196/mental.9782.
41. Bickmore, Timothy W., et al. "Automated Interventions for Multiple Health Behaviors Using
Conversational Agents." Patient Education and Counseling, vol. 92, no. 2, 2013, pp. 142–148,
doi:10.1016/j.pec.2013.05.011.
42. Abd-Alrazaq, Anas A., et al. "Effectiveness and Safety of Using Chatbots to Improve Mental
Health: Systematic Review and Meta-Analysis." Journal of Medical Internet Research, vol. 22,
no. 7, 2020, p. e16021.
43. He, Liang, et al. "Effectiveness of an AI Chatbot Based on Cognitive Behavioral Therapy for
Mental Health Support Among College Students During the COVID-19 Pandemic." Journal of
Affective Disorders, vol. 289, 2021, pp. 123–131.
44. Peuters, H., et al. "Impact of an mHealth Intervention with a Chatbot Component on
Lifestyle Factors Among Adolescents During COVID-19: The LIFEGOALS Study." Journal of
Medical Internet Research, vol. 24, no. 3, 2022, p. e29155.
46. Ogawa, T., et al. "Evaluation of an AI-Driven Chatbot for Remote Symptom Monitoring in
Patients with Parkinson’s Disease: Effects on Facial Expressivity and Speech Fluency." Journal
of Parkinson’s Disease, vol. 12, no. 3, 2022, pp. 745–754.
47. Ulrich, M., et al. "Effectiveness of the BalanceUP App, an AI-Based Chatbot for Mental
Wellbeing Improvement in Migraine Patients: A Pilot Study." Digital Health, vol. 8, 2022, p.
20552076221084191.
48. Vereschagin, V., et al. "Evaluating the Minder App: An AI-Based Chatbot for Cognitive–
Behavioral Therapy in Substance Use Reduction Among University Students." Journal of
Substance Abuse Treatment, vol. 125, 2021, p. 108311.
49. Prochaska, J. J., et al. "Effectiveness of the Woebot Chatbot for Substance Use and Craving
Reduction in an 8-Week Program: A Randomized Controlled Trial." Addiction, vol. 117, no. 1,
2022, pp. 180–190.
50. So, M., et al. "A Comparison of Guided Versus Unguided AI Chatbot Interventions for
Problem Gambling: A Randomized Controlled Study." Journal of Gambling Studies, vol. 38,
no. 2, 2022, pp. 563–580.
51. Olano-Espinosa, E., et al. "Efficacy of the Dejal@bot AI Chatbot Versus Usual Care for
Smoking Cessation in Primary Care: A Randomized Controlled Trial." Tobacco Control, vol.
30, no. 5, 2021, pp. 560–566.
54. Oh, S., et al. "Effectiveness of a Mobile App-Based Interactive Cognitive Behavioral Therapy
Chatbot Compared to a Paperback Book in Managing Panic Disorder: A Randomized
Controlled Trial." Journal of Anxiety Disorders, vol. 85, 2022, p. 102524.
55. Bennion, M. R., et al. "The Role of Artificial Intelligence Chatbots in Problem Solving for
Older Adults: A Comparative Study of MYLO and ELIZA." Aging & Mental Health, vol. 25, no.
8, 2021, pp. 1440–1448.
56. Thunström, L., et al. "Usability Comparison of a Digital Human (BETSY) Versus a Text-Based
Chatbot: A Pilot Study on Enhancing User Experience and Engagement." Computers in
Human Behavior, vol. 127, 2022, p. 107058.
57. Van Dyke, Thomas, et al. "The Effect of Consumer Privacy Empowerment on Trust and
Privacy Concerns in E-Commerce." Electronic Markets, vol. 17, no. 1, 2007, pp. 68–81.
58. Coghlan, Simon, et al. "To Chat or Bot to Chat: Ethical Issues with Using Chatbots in Mental
Health." Digital Health, vol. 9, Jan. 2023, p. 20552076231183542.
60. Diaz, Anthony, and Daehan Kwak. "Leveraging Chatbots for Mental Health Enhancement."
2023 International Conference on Computational Science and Computational Intelligence
(CSCI), IEEE, 2023, pp. 137–143.
61. Tian, Wensheng, et al. "A Privacy-Preserving Framework for Mental Health Chatbots Based
on Confidential Computing." 2022 IEEE Smartworld, Ubiquitous Intelligence & Computing,
Scalable Computing & Communications, Digital Twin, Privacy Computing, Metaverse,
Autonomous & Trusted Vehicles (SmartWorld/UIC/ScalCom/DigitalTwin/PriComp/Meta),
IEEE, 2022, pp. 1119–1124.
62. Agbih, S. "Housing Facilities for Asylum Seekers in Germany: Ethical Concerns Regarding
Social Exclusion, Othering and Negative Effects on Health." European Journal of Public
Health, vol. 29, no. Supplement_4, Nov. 2019, p. ckz185.174.
63. Knightbridge, Stephen M., et al. "Using Participatory Action Research in a Community-Based
Initiative Addressing Complex Mental Health Needs." Australian & New Zealand Journal of
Psychiatry, vol. 40, no. 4, Apr. 2006, pp. 325–332.
64. Griffiths, P. D., et al. "Post-Mortem Confirmation of Fetal Brain Abnormalities: Challenges
Highlighted by the MERIDIAN Cohort Study." BJOG: An International Journal of Obstetrics &
Gynaecology, vol. 128, no. 7, June 2021, pp. 1174–1182.
65. Beatty, C., et al. "Evaluating the Therapeutic Alliance with a Free-Text CBT Conversational
Agent (Wysa): A Mixed-Methods Study." Frontiers in Digital Health, vol. 4, 2022, p. 847991.
66. Darcy, A., et al. "Evidence of Human-Level Bonds Established with a Digital Conversational
Agent: Cross-Sectional, Retrospective Observational Study." JMIR Formative Research, vol. 5,
2021, p. e27868.
67. Ahmed, A., et al. "Thematic Analysis on User Reviews for Depression and Anxiety Chatbot
Apps: Machine Learning Approach." JMIR Formative Research, vol. 6, 2022, p. e27654.
68. Malik, T., et al. "Evaluating User Feedback for an Artificial Intelligence–Enabled, Cognitive
Behavioral Therapy–Based Mental Health App (Wysa): Qualitative Thematic Analysis." JMIR
Human Factors, vol. 9, 2022, p. e35668.
69. Ta, V., et al. "User Experiences of Social Support from Companion Chatbots in Everyday
Contexts: Thematic Analysis." Journal of Medical Internet Research, vol. 22, 2020, p. e16235.
70. Haque, M. R., and S. Rubya. "An Overview of Chatbot-Based Mobile Mental Health Apps:
Insights from App Description and User Reviews." JMIR mHealth and uHealth, vol. 11, 2023,
p. e44838.
71. Kettle, L., and Y.-C. Lee. "User Experiences of Well-Being Chatbots." Human Factors: The
Journal of the Human Factors and Ergonomics Society, vol. 66, 2024, pp. 1703–1723.
72. Zhang, M., and J. Li. "A Commentary of GPT-3 in MIT Technology Review 2021."
Fundamental Research, vol. 1, 2021, pp. 831–833.
73. Vaswani, A., et al. "Attention Is All You Need." Preprint at arXiv, 2023,
http://arxiv.org/abs/1706.03762.
74. Bubeck, S., et al. "Sparks of Artificial General Intelligence: Early Experiments with GPT-4."
Preprint at arXiv, 2023, http://arxiv.org/abs/2303.12712.
75. Tu, T., et al. "Towards Conversational Diagnostic AI." Preprint at arXiv, 2024,
http://arxiv.org/abs/2401.05654.
76. Salvi, F., Ribeiro, M. H., Gallotti, R., and West, R. "On the Conversational Persuasiveness of
Large Language Models: A Randomized Controlled Trial." Preprint at arXiv, 2024,
2http://arxiv.org/abs/2403.14380.
77. Strachan, J. W. A., et al. "Testing Theory of Mind in Large Language Models and Humans."
Nature Human Behaviour, forthcoming.
78. Yin, Y., Jia, N., and Wakslak, C. J. "AI Can Help People Feel Heard, but an AI Label Diminishes
This Impact." Proceedings of the National Academy of Sciences of the USA, vol. 121, 2024,
e2319112121.
79. Young, J., et al. "The Role of AI in Peer Support for Young People: A Study of Preferences for
Human- and AI-Generated Responses." Proceedings of the CHI Conference on Human Factors
in Computing Systems, ACM, Honolulu, HI, USA, 2024, pp. 1–18.
80. Li, J. Z., Herderich, A., and Goldenberg, A. "Skill but Not Effort Drive GPT Overperformance
over Humans in Cognitive Reframing of Negative Scenarios." Preprint.
81. Malik, A. "OpenAI’s ChatGPT Now Has 100 Million Weekly Active Users." TechCrunch AI, 6
Nov. 2023, https://techcrunch.com/2023/11/06/openais-chatgpt-now-has-100-million-
weekly-active-users/.
82. Salesforce. "Top Generative AI Statistics for 2024." Salesforce News & Insights, 2023,
https://www.salesforce.com/news/stories/generative-ai-statistics/.
83. Pandya. "The Age of Generative AI: Over Half of Americans Have Used Generative AI and
Most Believe It Will Help Them Be More Creative." Adobe Blog, 22 Apr. 2024,
https://blog.adobe.com/en/publish/2024/04/22/age-generative-ai-over-half-americans-
have-used-generative-ai-most-believe-will-help-them-be-more-creative.
84. Li, H., Zhang, R., Lee, Y.-C., Kraut, R. E., and Mohr, D. C. "Systematic Review and Meta-
Analysis of AI-Based Conversational Agents for Promoting Mental Health and Well-Being."
NPJ Digital Medicine, vol. 6, 2023, p. 236.