Emotional Wellbeing

Artificial Intelligence for Mental Health: What It Can and Cannot Do

Let's Shine Team · · 10 min read
Person using an AI wellbeing app on a smartphone

Artificial intelligence applied to mental health refers to the use of machine-learning systems, natural-language processing, and generative models to provide emotional support, early detection of disorders, psychoeducation, and self-knowledge tools. It is a rapidly expanding field: according to a Grand View Research report (2024), the global AI-in-mental-health market will exceed $4 billion by 2028. Yet its accelerated growth raises fundamental questions about its real capabilities, its ethical limits, and its role in an ecosystem where people's mental health is at stake. This article analyses — honestly and without naive techno-optimism — what AI can and cannot do in this space.

Important notice: LetsShine.app is not a psychological-therapy service and does not replace a mental-health professional. If you are in crisis, contact the 988 Suicide & Crisis Lifeline (call or text 988 in the US), Crisis Text Line (text HOME to 741741), or Samaritans (116 123 in the UK).

Aspect What AI CAN do What AI CANNOT do
Availability Accessible 24/7, no waiting list Provide containment in an acute crisis
Listening A judgement-free space to express yourself Genuine empathy (it simulates, it does not feel)
Patterns Detect patterns in what you share Diagnose mental disorders
Psychoeducation Inform about techniques and concepts Prescribe medication
Accessibility Low cost, no geographical barriers Replace the therapeutic relationship
Tracking Remember previous conversations Guarantee absolute confidentiality

What Can Artificial Intelligence Do for Your Mental Health?

1. Offer an Accessible Space for Reflection

One of the most significant barriers to accessing emotional support is the shortage of mental-health professionals. In the US, the Health Resources and Services Administration (HRSA) estimates a shortage of over 8,000 mental-health practitioners, especially in rural areas. In the UK, NHS waiting times for talking therapies average 18 weeks. AI does not solve this structural problem, but it can offer an initial space for reflection for people who need to talk and do not have immediate access to a professional.

On LetsShine.app, the AI functions as a reflective companion: it listens, asks questions that help you go deeper, and offers perspectives grounded in attachment psychology, nonviolent communication, and what we call emotional archaeology.

2. Detect Emotional Patterns

Language models can analyse what a person writes over time and identify recurring patterns: topics that repeat, predominant emotions, shifts in tone. This can be valuable for self-awareness. A study published in Nature Medicine (2024) demonstrated that language models can identify signals of depression in written text with 82 % sensitivity.

3. Democratise Psychoeducation

AI can explain psychological concepts in an accessible and personalised way. What is anxious attachment? Why do I struggle to set boundaries? What is codependency? Instead of searching Google and finding information of variable quality, you can have a conversation where the AI adapts the explanation to your specific situation.

4. Practise Relational Skills

A particularly interesting application is using AI as a "simulator" for difficult conversations. Do you need to say something uncomfortable to your partner? Do you want to set a boundary with your mother? You can rehearse the conversation with the AI, explore different ways of saying the same thing, and prepare yourself emotionally before the real conversation.

5. Between-Session Support

Many people see their therapist every two weeks or once a month. Between sessions, AI can serve as a continuity tool: a space to keep reflecting on what was discussed in therapy, a place to jot down what you want to bring up at the next session, a companion for the difficult moments that arise at three in the morning.

What Can Artificial Intelligence NOT Do?

1. Diagnose

AI cannot and should not diagnose mental disorders. A psychological diagnosis requires a complete clinical evaluation that includes an interview, observation, medical history, socio-family context, and in many cases standardised tests. No chatbot can replicate this.

2. Treat Mental Disorders

Psychological therapy is not a conversation; it is a therapeutic relationship within a specific professional framework (setting, transference, therapeutic alliance, supervision). AI can inform about therapeutic techniques, but it cannot apply them with the depth, sensitivity, and adaptation that real treatment requires.

3. Manage Crises

If a person is at risk of suicide, in the middle of a psychotic episode, or experiencing a dissociative episode, AI does not have the capacity to intervene adequately. It can — and must — redirect to emergency resources, but it cannot replace human crisis intervention.

4. Feel Empathy

Language models can generate responses that appear empathic, but they do not feel. Human empathy involves emotional resonance, physical presence, and a connection that cannot be simulated with statistics. This does not invalidate AI's usefulness, but it is important to be honest about it.

5. Guarantee Total Privacy

Any digital system stores data. Although responsible platforms like LetsShine.app implement encryption and strict privacy policies, the reality is that a paper notebook remains the most private medium in existence. Users should be aware of what data they share and how it is used.

Is It Ethical to Use AI for Mental Health?

This is the question that generates the most debate. There are legitimate arguments on both sides:

In favour:

  • It reaches people who lack access to a professional (due to cost, waiting lists, or stigma).
  • It lowers the barrier to entry: many people speak more easily with an AI than with another person.
  • It can function as a gateway to professional therapy.
  • It offers quality psychoeducation at scale and in a personalised way.

Against:

  • Risk that users believe they are receiving therapy when they are not.
  • Lack of specific regulation in most countries (including the US and UK, where frameworks are still evolving).
  • Possibility of inadequate responses in serious situations.
  • Risk of emotional dependence on a system that cannot offer reciprocity.

The position of LetsShine.app is clear: we are a complement, never a substitute. Our AI is a tool for self-knowledge, reflection, and relational-skill improvement. When we detect that a person needs professional attention, we say so explicitly.

What Makes LetsShine.app Different From Other Mental-Health Chatbots?

LetsShine.app does not position itself as a mental-health chatbot. It positions itself as a relational-improvement platform powered by AI. There are important differences:

  • Relational focus, not just individual: we work with couples, families, and shared living spaces, not only with isolated individuals.
  • Emotional archaeology: our approach is not limited to giving generic advice. The AI explores the deep roots of your patterns with you so that change is structural, not superficial.
  • Mediation, not therapy: in shared spaces, the AI acts as a mediator, helping both parties express themselves and understand the other's perspective.
  • Honesty about limits: we do not promise to cure your anxiety or save your marriage. We promise a quality space to know yourself better and communicate better, with an explicit recommendation to see a professional when necessary.

What Is the Future of AI in Mental Health?

AI is not going to replace psychologists. But it will transform the mental-health ecosystem in several ways:

  • Early detection: algorithms that identify signals of emotional deterioration before they become disorders.
  • Treatment personalisation: models that help therapists personalise interventions based on the patient's profile.
  • Accessibility: quality tools available to populations that today lack access to psychological care.
  • Research: analysis of large volumes of anonymised data to better understand the mechanisms of mental disorders.

The challenge is not technological: it is ethical, regulatory, and human. We need clear legal frameworks, transparency in how algorithms work, and an honest conversation about what technology can contribute and what must remain in human hands.

Frequently Asked Questions

Can AI replace a psychologist? No. AI can complement therapeutic work, offer psychoeducation, and serve as a reflection tool, but it cannot replicate the therapeutic relationship, diagnose, or treat mental disorders. If you need psychological care, see a qualified professional.

Is it safe to share my problems with an AI? It depends on the platform. Look for services that clearly specify their privacy policy, how they store data, and whether they use it to train models. At LetsShine.app we encrypt conversations and do not use the content of your sessions to train general AI models.

Is LetsShine.app therapy? No. LetsShine.app is a relational-improvement platform that uses AI as a guided-reflection, mediation, and self-knowledge tool. We do not diagnose, prescribe, or treat disorders. We actively recommend seeing a professional when the situation requires it. Our value lies in offering an accessible space to explore your emotions, improve your communication, and better understand your relationships.

For what type of person is AI most useful in emotional wellbeing? AI is especially useful for people seeking self-knowledge, wanting to improve their communication skills, needing a reflection space between therapy sessions, or facing access barriers (cost, waiting list, schedule). It is not appropriate as a primary resource for people with severe disorders or in crisis situations.

What if the AI gives me advice that doesn't seem right? Be sceptical of any system — human or artificial — that offers absolute certainties about your emotional life. AI can be wrong, offer perspectives that do not fit your reality, or stay on the surface. If something does not resonate with you, trust your own judgement. And if you have serious concerns about your mental health, always consult a human professional.

Your relationships can improve. Today.

Start free in 2 minutes. No credit card, no commitment. Just you, the people you care about, and an AI that helps you understand each other.

Start free now

Related articles