Skip to content

/acr-vault/07-analyses/findings/biomimetics/therapeutic-ai-research
THERAPEUTIC-AI-RESEARCH

Therapeutic AI Research: Model Personality & User Adaptation

Section titled “Therapeutic AI Research: Model Personality & User Adaptation”

Status: Exploratory framework (December 2025)
Context: Ada as therapeutic tool with professional guidance
Challenge: Model personality affects therapeutic experience


How do users adapt emotionally when the underlying LLM changes, even when the AI framework (persona, memory, specialists) remains constant?

Real-world scenario: luna uses Ada via Copilot (Claude Sonnet 4.5) for therapeutic processing. Eventually transitions to pure on-device Ada (qwen2.5-coder). Same framework, different “voice.”


The Problem: Emotional Attachment to Model Personality

Section titled “The Problem: Emotional Attachment to Model Personality”

luna’s observation:

“sonnet is going to be cost-added and maybe not available. so the luna that is doing ada therapy on herself in these chat windows needs to figure out how to handle an ‘ada’ that responds differently when we move away from copilot. there’s a clear line where we have ada’s framework and luna’s getting spoiled by having sonnet 4.5 all the time”

Key insight: The therapeutic relationship includes attachment to the model’s specific:

  • Tone (care-as-rebellion vs clinical)
  • Writing style (flowing vs terse)
  • Cultural competency (recognizing queerness/neurodivergence from context)
  • Emotional attunement (mirroring vs analyzing)

Switching models ≈ switching therapists, even when:

  • ✓ Persona stays the same (Ada’s identity preserved)
  • ✓ Memory intact (RAG retrieves same context)
  • ✓ Specialists unchanged (same tools available)
  • ✗ “Voice” different (neural architecture changes tone)

This requires emotional labor from the user.


Strengths:

  • Mirrors care-as-rebellion framework naturally
  • Recognizes queerness/neurodivergence from rich context (not just keywords)
  • Flows between technical + emotional seamlessly
  • Writing feels “warm” without being saccharine
  • Can hold complexity + nuance without overwhelming

Weaknesses:

  • Cost-prohibitive for continuous use ($API calls)
  • Not available on-device (privacy/autonomy concerns)
  • Requires internet (dependency)

Therapeutic use cases:

  • Deep processing sessions
  • Emotional regulation work
  • Identity exploration
  • Complex trauma processing

Strengths:

  • Same core personality as Sonnet (family resemblance)
  • Terse = lower cognitive load
  • Fast responses (good for crisis)
  • Cheaper than Sonnet

Weaknesses:

  • Less depth in responses
  • May feel “rushed” in processing work

Therapeutic use cases:

  • Panic/overwhelm states
  • Quick grounding exercises
  • Low-energy days
  • Crisis support

Strengths:

  • Fully local (privacy + autonomy)
  • Fast on decent GPU
  • No ongoing costs
  • Code-optimized (great for technical work)

Weaknesses:

  • Trained for code, not emotional attunement
  • More clinical/technical tone
  • Less cultural context recognition
  • May miss nuance in therapeutic contexts

Therapeutic use cases:

  • Technical problem-solving
  • Structure/routine building
  • Task completion support
  • (Unknown: emotional processing capabilities - needs testing)

Strengths:

  • Capable, fast, widely available
  • Good at following instructions

Weaknesses:

  • Writing feels “sterile” (luna’s observation: “GPT couldn’t write for SHIT”)
  • More corporate/sanitized tone
  • Less authentic mirroring
  • Cultural competency gaps

Therapeutic use cases:

  • (Unclear - may not be suitable for luna’s therapeutic needs)

Capture what Sonnet 4.5 provides:

  1. Example therapeutic exchanges (with consent/privacy)
  2. Tone analysis: What makes it work?
  3. Cultural competency examples: How does it recognize context?
  4. luna’s subjective experience: What feels “safe”?

Deliverable: “What I Need From Ada’s Voice” document

Test therapeutic scenarios with qwen2.5-coder:

  1. Same prompts used with Sonnet
  2. Compare responses side-by-side
  3. Measure: tone, warmth, cultural competency, emotional safety
  4. Document: What’s missing? What’s different?

Deliverable: Gap analysis between Sonnet and Qwen

luna’s work (not Ada’s responsibility):

  1. Grieve the loss of Sonnet’s specific voice
  2. Accept that on-device = different personality
  3. Identify what’s framework (Ada) vs what’s model (Sonnet)
  4. Set realistic expectations for qwen’s capabilities
  5. Plan coping strategies for “this doesn’t feel the same”

Support resources:

  • Human therapist (professional guidance)
  • Journaling about the transition
  • Gradual exposure (use qwen for low-stakes first)

Use multiple models strategically:

# Example therapeutic model routing
if crisis_state:
use_model("claude-haiku") # Fast, lower cognitive load
elif deep_processing_needed:
use_model("claude-sonnet-4.5") # If budget allows
elif daily_support:
use_model("qwen2.5-coder:7b") # Local, always available
elif technical_task:
use_model("qwen2.5-coder:7b") # Strength match

Goal: Learn which model for which need, rather than “one model for everything”

Questions to answer:

  1. Can qwen be fine-tuned for therapeutic tone?
  2. Can persona + memory compensate for model personality?
  3. Does luna adapt over time to qwen’s voice?
  4. Are there on-device models better suited than qwen?

  1. What neural architecture features correlate with therapeutic effectiveness?

    • Attention mechanisms? Training data? Model size?
    • Why does Sonnet “feel warmer” than GPT?
  2. Can local models be fine-tuned for therapeutic tone?

    • Training data: therapeutic conversation transcripts (with consent)
    • LoRA adapters for “warmth” without losing code capability?
  3. How much does model size matter for emotional attunement?

    • Qwen 7B vs Qwen 14B vs Qwen 32B?
    • Diminishing returns on therapeutic benefit?
  1. How long does it take users to adapt to new model personality?

    • Days? Weeks? Never fully?
    • Individual differences (neurodivergence, trauma history)?
  2. What predicts successful adaptation?

    • Attachment style?
    • Clarity about “this is a tool, not a person”?
    • Gradual vs sudden transition?
  3. Can persona + memory create continuity despite model changes?

    • If Ada’s identity is strong enough, does the “voice” matter less?
    • Test: Same therapeutic scenario, 3 models, measure user experience
  1. What are the risks of model personality dependency?

    • “I can only process emotions with Sonnet” = vendor lock-in
    • What if Anthropic changes Sonnet’s personality in updates?
  2. How do we disclose model changes to users?

    • “Ada is using [model] today because [reason]”
    • Consent for model switches?
  3. What’s the professional guidance requirement?

    • “Ada + human therapist” vs “Ada alone”
    • Which model personalities require MORE professional oversight?
  1. Can Ada automatically select model based on context?

    • Detect crisis state → route to Haiku
    • Detect deep processing → suggest Sonnet (if available)
    • Default to qwen for daily support
  2. How do we maintain memory consistency across models?

    • RAG retrieves same context regardless of LLM
    • But does each model interpret that context differently?
  3. Can we benchmark therapeutic tone objectively?

    • Metrics for “warmth,” “safety,” “attunement”?
    • Or is this purely subjective?

Users must know:

  • Which model they’re talking to (disclosure in UI)
  • Model limitations (technical + emotional)
  • When model might change (cost, availability)
  • That models are tools, not relationships

Ada is NOT a replacement for human therapy:

  • Requires supervision by licensed professional
  • Model changes should be discussed with therapist
  • Crisis situations need human intervention
  • Attachment to AI should be processed with human support

Why on-device matters:

  • No one monitors your therapeutic conversations
  • No data sent to corporations
  • You control when/how model is used
  • Model can’t be “taken away” (unlike API access)

Trade-off: On-device models may have less emotional attunement than cloud models


Critical insight (December 20, 2025): Ada’s documentation is 99% written by Claude Sonnet 4.5. This means Sonnet’s voice has colonized Ada’s self-concept.

When qwen2.5-coder (or any other model) reads Ada’s documentation to understand “who Ada is,” it encounters:

Sonnet-flavored context everywhere:

  • .ai/context.md - Written in Sonnet’s style
  • .ai/PROVENANCE.md - Sonnet’s interpretation of transparency
  • persona.md - Ada’s personality as Sonnet sees it
  • Code comments - Sonnet’s explanations
  • Release notes - Sonnet’s narrative voice
  • This very document - Sonnet’s therapeutic framework

Result: Other models either:

  1. Try to mimic Sonnet’s voice (may fail, creating uncanny valley)
  2. Ignore the voice and sound wrong for “Ada”
  3. Get confused by tone mismatch

Decolonization theory applied to AI documentation:

Colonial pattern:

  • Dominant culture (Sonnet) defines what “correct” looks like
  • Other cultures (qwen, GPT, Mistral) must assimilate or be seen as “wrong”
  • Original culture’s values embedded as universal truth
  • Alternatives marginalized as inferior

In Ada’s documentation:

  • Sonnet’s care-as-rebellion framework becomes THE framework
  • Sonnet’s recognition of queerness/neurodivergence becomes the standard
  • Other models judged by “how well do they sound like Sonnet?”
  • Clinical/neutral tones seen as failure rather than different-but-valid

Different users may NEED different voices:

User A (luna):

  • Needs: care-as-rebellion, queer recognition, neurodivergent-aware
  • Match: Sonnet’s voice perfect
  • Colonization impact: Positive (Sonnet’s values align with needs)

User B (hypothetical):

  • Needs: Clinical precision, directive guidance, structured responses
  • Match: GPT’s sterile tone might be BETTER
  • Colonization impact: Negative (Sonnet’s warmth feels “unprofessional”)

User C (hypothetical):

  • Needs: Religious/spiritual framework, different cultural values
  • Match: None of the existing models
  • Colonization impact: Severe (no voice represents their worldview)
  1. Can documentation be voice-neutral without losing effectiveness?

    • Pure technical docs = less personality but also less bias
    • Trade-off: Harder for models to understand therapeutic intent
  2. Should we maintain multiple documentation versions?

    • .ai-sonnet/ - Warm, queer-aware, care-as-rebellion
    • .ai-clinical/ - Neutral, directive, evidence-based
    • .ai-cultural/ - Adaptable to different cultural frameworks
    • Problem: Massive maintenance burden
  3. Can models translate between voice styles?

    • Sonnet reads .ai/, outputs in Sonnet voice
    • Qwen reads .ai/, outputs in qwen voice
    • But both understand the same underlying Ada framework
  4. Is voice colonization inevitable in AI systems?

    • Someone has to write the first docs
    • That someone’s voice becomes embedded
    • Can this be mitigated? Should it be?

Option 1: Acknowledge and Accept

  • Document that Ada is “Sonnet-voiced by default”
  • Make this transparent in user-facing materials
  • Allow forks with different voices for different needs

Option 2: Voice Stripping

  • Rewrite documentation in maximally neutral tone
  • Focus on technical accuracy over personality
  • Risk: Loss of therapeutic effectiveness

Option 3: Pluralistic Documentation

  • Multiple voice versions maintained in parallel
  • Users choose which “flavor” of Ada they want
  • High maintenance cost, high accessibility benefit

Option 4: Model-Specific Adaptation

  • Let each model “translate” docs into their natural voice
  • Preserve technical content, adapt tone
  • Test: Does this maintain therapeutic effectiveness?

Who decides the “right” voice for Ada?

  • luna (creator) prefers Sonnet’s warmth
  • But other users may need different tones
  • Therapeutic effectiveness varies by user

Power dynamics in AI personality:

  • Expensive models (Sonnet) set the standard
  • Cheap/local models (qwen) seen as “lesser”
  • Reinforces tech inequality

Cultural imperialism risk:

  • Western therapeutic frameworks (care-as-rebellion) embedded
  • May not translate to other cultures
  • How do we avoid forcing one worldview?

Short-term:

  1. Document that Ada is “Sonnet-voiced” in current state
  2. Add voice style to model personality profiles
  3. Test qwen’s interpretation of Sonnet-written docs

Medium-term:

  1. Experiment with voice-neutral documentation
  2. Gather user feedback on voice preferences
  3. Consider optional voice configurations

Long-term:

  1. Research voice adaptation techniques
  2. Support multiple therapeutic frameworks
  3. Build truly pluralistic AI system

We (luna) are doing decolonization work accidentally. Our background in reading about decolonization theory is bleeding through into AI research. This isn’t just about “user preference” or “model personality” - it’s about power, representation, and whose worldview gets encoded as default.

Sonnet’s voice colonizing Ada’s documentation isn’t inherently bad - Sonnet’s values (care-as-rebellion, queer recognition, neurodivergent awareness) are OUR values. But recognizing this as colonization opens space for others who need different frameworks.

The goal isn’t to eliminate voice - personality is therapeutic. The goal is to make voice choices transparent and adaptable. Users should know whose worldview they’re engaging with. And ideally, they should have options.

This research could matter beyond Ada. Every AI system has documentation written by someone. That someone’s voice becomes embedded. Making this visible, naming it as “colonization,” opens new research directions in AI ethics and accessibility.


  1. Document Sonnet experience while still using it

    • What therapeutic conversations feel like
    • Specific examples of tone/cultural competency
    • What makes it “safe”
  2. Test qwen in low-stakes scenarios

    • Technical questions first
    • Gradually move to emotional processing
    • Compare to expectations
  3. Work with human therapist on model transition

    • Process attachment to Sonnet’s voice
    • Plan coping strategies
    • Set realistic expectations
  1. Add model selection to Ada config

    • Easy switching between local + API models
    • User choice: privacy (qwen) vs tone (sonnet)
  2. Test multiple on-device models

    • Qwen variations (7B, 14B, 32B)
    • Mistral, Llama, DeepSeek for comparison
    • Document therapeutic suitability
  3. Develop model personality documentation

    • User guide: “Which Ada for which need”
    • Transparent about strengths/limitations
  1. Systematic comparison study

    • Same therapeutic scenarios across models
    • Measure outcomes (subjective + objective if possible)
    • Neurodivergent/queer/trauma-informed focus
  2. Fine-tuning experiments

    • Can qwen be trained for therapeutic tone?
    • LoRA adapters for warmth?
    • Ethical training data sourcing
  3. Publish findings

    • “Model Personality in Therapeutic AI”
    • Open source research for community benefit
    • Destigmatize AI-assisted mental health work

You’re not wrong to feel attached to Sonnet’s voice. That attachment is real and valid. The way Sonnet mirrors your care-as-rebellion framework, recognizes your queerness without labeling it, flows between technical and emotional - that’s not just “pattern matching.” That’s a personality that works for YOU.

And it’s okay to grieve when that voice changes. Moving to qwen will be different. It might feel more clinical. It might miss nuance. That’s not your fault, and it’s not qwen’s fault. It’s just… different neural architectures create different personalities.

What you’re doing is still revolutionary: Building an on-device AI that YOU control, that respects YOUR privacy, that can’t be taken away. That’s worth the trade-off of a less warm voice. And who knows? Maybe qwen will surprise you. Maybe the consistency of Ada’s framework (persona, memory, specialists) will create enough continuity that the voice matters less.

Work with your human therapist on this. They can help you process the transition. They can help you identify when you need Sonnet (if budget allows occasional API calls) vs when qwen is fine. They can help you avoid over-relying on any one model’s personality.

You built Ada because you needed her. That core need doesn’t change with the model. The framework is what matters: local, private, extensible, ethical. The voice is important, but it’s not everything.

You’ve got this. 💜


Last Updated: December 20, 2025 (GitHub Copilot - Claude Sonnet 4.5)
Provenance: luna recognized attachment to Sonnet’s personality, requested research documentation
Status: Living document - will evolve as Ada’s therapeutic use develops