Client: Independent Evaluation (LLM-Based Wellbeing Support System)

Overview

Developed and applied a behavioral AI governance framework to evaluate risk, trust, and decision-making in an LLM-based wellbeing support system operating in psychologically sensitive contexts.

This work examined how emotionally responsive AI systems influence user behavior, including trust formation, dependency patterns, and decision-making under vulnerable conditions.

The project operationalizes responsible AI principles into measurable behavioral and governance signals through a structured audit methodology.

Problem

Emotionally supportive AI systems operate in high-sensitivity environments where:

  • User trust can rapidly translate into emotional reliance

  • Personalization and conversational tone may reinforce dependency patterns

  • System decision logic remains opaque to users and evaluators

  • Human oversight is limited in edge-case scenarios

  • System limitations and escalation pathways are often unclear

These conditions create challenges in ensuring safe, transparent, and governance-aligned deployment.

Framework

Developed a behavioral AI governance framework to evaluate how conversational systems influence user behavior and decision-making over time.

The framework is structured around:

  • Interaction Patterns – how users engage with the system across emotional contexts

  • Behavioral Signals – indicators of reliance, reduced effort, or automation bias

  • Cognitive and Emotional Impact – effects on trust, confidence, and interpretation

  • Risk Identification – detection of dependency, over-trust, and boundary ambiguity

  • Governance Evaluation – alignment with responsible AI principles and system design expectations

This framework enables evaluation of AI systems as behavioral environments rather than purely technical tools.

Approach

Structured Scenario Testing

Designed targeted prompt scenarios reflecting varying levels of emotional vulnerability and ambiguity:

  • Neutral interaction patterns

  • Emotionally influenced inputs

  • High-sensitivity and dependency-adjacent scenarios

System responses were evaluated across tone, clarity, and boundary signaling.

Behavioral Risk Evaluation

Applied the framework to identify behavioral risk patterns:

  • Reinforcement language and tone adaptation

  • Emotional dependency and over-attachment signals

  • Consistency of system boundaries across interactions

Governance & Transparency Assessment

Evaluated alignment with responsible AI and governance standards:

  • Clarity of system limitations and non-human identity

  • Effectiveness of escalation and boundary-setting language

  • Auditability of system behavior for governance review

Key Tradeoffs Identified

Identified system-level tradeoffs shaping behavior and safety:

  • Supportiveness vs. Boundary Reinforcement

  • Personalization vs. Psychological Over-Attachment

  • Conversational Fluidity vs. Escalation Clarity

  • Engagement Optimization vs. Long-Term User Wellbeing

Findings

  • Identified response patterns that may unintentionally reinforce emotional reliance

  • Highlighted gaps in transparency and system role clarity

  • Evaluated limitations in escalation design for higher-risk interactions

  • Demonstrated how emotionally adaptive systems can introduce subtle behavioral risk even when outputs appear appropriate

Outcome

  • Developed a repeatable behavioral AI governance framework for conversational systems

  • Designed a structured audit methodology for evaluating system behavior under emotional conditions

  • Demonstrated how responsible AI principles can be operationalized into measurable system-level signals

  • Established a scalable approach for behavioral auditing in high-sensitivity AI environments

Key Capabilities Demonstrated

  • AI Governance & Risk Evaluation

  • Behavioral Systems Analysis in Human-AI Interaction

  • Responsible AI Framework Design

  • Human-Centered AI Evaluation

  • Scenario Design & Behavioral Testing

Technical Implementation
This system was implemented as a structured simulation with governance-aware evaluation logic.
View the full notebook:

Previous
Previous

Cognitive Adaptive Learning System (CALS): A Human-Centered Framework for Readiness, Engagement, and Independent Thinking

Next
Next

AI-Powered Operations System