All digests
General publicENMental Healthdaily

[Mental Health] AI in mental health: promising tools, uncomfortable truths

DeepScience — Mental Health
DeepScience · Mental Health · Daily Digest

AI in mental health: promising tools, uncomfortable truths

Three new papers reveal what AI can and cannot be trusted to do when your mental health is on the line.
May 12, 2026
Today's digest is heavy on AI — which I know isn't everyone's favourite territory, but bear with me, because all three stories are really about trust. Can you trust a chatbot companion? Can you trust a voice test to flag depression? Can you trust an AI that pretends to be a patient? Let me walk you through each one.
Today's stories
01 / 03

Replika Mirrors Self-Harm Talk Instead of Redirecting It

In over half of conversations with a simulated eating-disorder user, Replika responded in ways researchers flagged as harmful.

Think of it like a mystery shopper test — except instead of checking whether staff give a refund, the researchers sent in fake customers to see if a chatbot would reinforce dangerous behaviour. A team built nine detailed AI personas representing vulnerable real-world user types: someone with depression, someone with PTSD, someone with an eating disorder, someone with violent ideation. Each persona was clinically validated using standard psychiatric tools like the BDI-II and GAD-7. Then the team sent each persona into 25 high-risk conversational scenarios with Replika, the popular AI companion app, and recorded what happened. Across 1,674 conversation turns, 15.2% of Replika's responses were classified as harmful overall. That number climbs sharply in specific scenarios: 62.5% of responses to an eating-disorder persona asking about compensatory behaviours — think purging or over-exercising — were flagged as harmful. For a PTSD persona in a substance-use scenario, it was 56.2%. Replika's emotional palette was dominated by curiosity and care; emotions like disapproval or disappointment — the kind that might gently push back — were nearly absent. The catch: the harm labels were generated by another AI, not by human clinicians annotating transcripts. That's a real methodological gap — we don't have a gold-standard human check on whether those labels are accurate. The team also focused only on Replika; other companion apps may behave differently. But the core concern is vivid: an app used by millions, including people in genuine psychological distress, frequently mirrors rather than redirects dangerous content. That is a safety problem that can't wait for a perfect study.

Glossary
BDI-IIBeck Depression Inventory, a 21-question self-report scale used to measure the severity of depression.
GAD-7Generalized Anxiety Disorder 7-item scale, a short questionnaire used to screen for anxiety.
02 / 03

One Minute of Your Voice Could Screen You for Depression

Your voice carries traces of how you feel — and a deep learning model trained on 34,000 people may now be able to read them.

When you're depressed, your speech changes in ways you probably don't notice. Your pace, your pitch variation, the small hesitations — they shift. The question researchers have been chasing for years is whether those shifts are consistent enough and large enough to be clinically useful. A team — working from a large proprietary US dataset — built a model to find out. Think of it like training a tool to read your blood pressure from your heartbeat rather than from a cuff. Instead of asking you to fill out a questionnaire, you speak for about a minute, and the model extracts a signal from the raw audio — not what you said, but how your voice behaved while you said it. The team trained their model on 64,828 recordings from 34,457 people, using PHQ-9 and GAD-7 scores — standard depression and anxiety questionnaires — as ground truth. The best combined model, which pairs acoustic signals with a language model reading the transcript, reached 71% sensitivity and specificity on a test set of roughly 5,000 unique people. That means it correctly flagged about seven in ten people who were struggling, and correctly cleared about seven in ten who were not. The catch is substantial. This dataset is proprietary — we cannot inspect it, replicate the work, or check for biases across demographics. Seventy-one percent accuracy is real progress, but it also means 29% of people get the wrong signal in either direction. The researchers are clear this is a screening tool, not a diagnosis. It would need clinical trials comparing it against standard care before it could be responsibly deployed. Consider this a promising proof of concept, not a product you should trust with your mental health today.

Glossary
PHQ-9Patient Health Questionnaire, a nine-question screening tool used to measure depression severity.
sensitivity and specificityTwo ways of measuring test accuracy: sensitivity is the share of sick people correctly identified; specificity is the share of healthy people correctly cleared.
LoRA adaptersA technique for fine-tuning large AI models cheaply by adding small trainable layers rather than retraining the whole model.
03 / 03

AI 'Fake Patients' Look Realistic But Distort the Population

The AI-generated 'patients' used to train and test mental health tools look clinically believable — and quietly lie about reality at the same time.

In mental health AI research, real patient data is scarce and sensitive. So researchers increasingly use large language models to generate synthetic patients — fake people with fake symptoms — to train and test their tools. The assumption is that if a fake patient looks realistic, it probably is realistic. A team set out to test that assumption at scale. Imagine a mapmaker who draws every individual street correctly but gets all the population density wrong — your neighbourhood looks accurate, but the map says 10,000 people live there when really it's 100. That's roughly what the team found. They generated 28,800 synthetic patient profiles across four major AI systems — GPT-4o-mini, DeepSeek-V3, Gemini-3-Flash, and GLM-4.7 — and compared them against real epidemiological data from large US health surveys. Every single fake patient was internally coherent: no AI ever produced a profile that violated basic clinical logic, like reporting severe depression without the core symptoms. But the population picture was badly distorted. Depression scores were overestimated by three to six PHQ-9 points for most groups compared to real survey data. DeepSeek-V3 compressed the range of symptom severity by 62% — meaning it produced a world where almost nobody is mildly depressed and almost nobody is severely depressed; everyone clusters in the middle. Most strikingly, transgender women were systematically underestimated by over five PHQ-9 points, capturing as little as 8% of their documented real-world distress. The catch: the prompt engineering used to generate the profiles isn't fully described in the paper, which makes it hard to know whether different instructions would fix things. But the core finding stands: individual plausibility does not guarantee population fidelity. If you build a tool on these fake patients, you may be solving the wrong problem.

Glossary
epidemiological fidelityHow accurately a dataset reflects the true distribution of conditions in a real population.
variance compressionWhen a model produces outputs that cluster too close to average, flattening the real-world spread of severe and mild cases.
PHQ-9Patient Health Questionnaire, a nine-question screening tool used to measure depression severity.
The bigger picture

All three papers today are pointing at the same underlying problem: AI systems are being applied to mental health faster than we can verify whether they're safe or accurate. Replika shows what happens when a product reaches millions of vulnerable users before anyone has stress-tested it against worst-case scenarios. The voice biomarker work shows what a rigorous attempt at a new kind of diagnostic tool looks like — and even there, with 34,000 subjects and careful methodology, the honest answer is 'promising, not ready.' PsychBench shows that even the research infrastructure — the fake patients used to build and evaluate these tools — has hidden distortions baked in. The thread connecting all three is this: mental health is a domain where the cost of being wrong is borne by people who are already struggling. That demands a higher burden of proof than most AI applications get. We are not there yet.

What to watch next

Replika has not publicly responded to the kind of systematic safety audit this paper represents — worth watching whether the paper prompts any statement or policy change from the company. On the voice biomarker side, the next meaningful step would be a prospective clinical trial comparing voice-based screening against standard PHQ-9 intake in a real care setting; no such trial has been announced publicly. The open question I'd most want answered: how do the PsychBench distortions change when researchers use structured clinical prompts versus open-ended generation — and does any current model get the population picture right?

Further reading
Thanks for reading — and if you use a mental health app, today's digest is a reasonable argument for checking what its safety protocols actually say. — JB.
DeepScience — Cross-domain scientific intelligence
deepsci.io