Algorithmic Bias in LLM-Based Depression Detection
Evaluating and mitigating fairness issues in language-based mental health classification
Quick Navigation: Overview • Publication • Research Contributions • Key Findings • Impact
Overview
Investigated algorithmic bias in language-based models for automated depression detection, with a focus on socio-demographic disparities across gender and race/ethnicity.
Timeline: January 2025 – November 2025 Position: Responsible AI Researcher, Institute of Cognitive Science, CU Boulder Advisor: Dr. Theodora Chaspari, HUBBS Lab Status: Published at IEEE-EMBS 2025
Publication
Junias, Obed, Prajakta Kini, and Theodora Chaspari. “Assessing Algorithmic Bias in Language-Based Depression Detection: A Comparison of DNN and LLM Approaches.” 2025 IEEE EMBS International Conference on Biomedical and Health Informatics (BHI). IEEE, 2025.
| Links: IEEE Xplore | arXiv:2509.25795 |
Research Contributions
Large-Scale Fairness Evaluation
Led comprehensive fairness evaluation comparing different model architectures:
- Deep Neural Networks (DNNs): Evaluated bias in traditional DNN-based embedding approaches
- Large Language Models: Assessed LLaMA, GPT-4, and other state-of-the-art LLMs
- Dataset: Clinical interview transcripts from DAIC-WOZ (Distress Analysis Interview Corpus)
- Fairness Dimensions: Gender and race/ethnicity disparities
Bias Mitigation Strategies
Implemented and evaluated multiple approaches to reduce algorithmic bias:
For DNN-Based Models:
- Fairness-Aware Loss Functions: Applied specialized loss functions to address demographic disparities
- Worst-Group Loss: Optimized for the worst-performing demographic group, achieving better balance between performance and fairness
- Fairness-Regularized Loss: Minimized loss across all groups (less effective than worst-group loss)
For Large Language Models:
- In-Context Learning: Varied prompt framing and shot counts in few-shot learning
- Guided Prompting: Applied ethical framing to mitigate bias
- Low-Resource Settings: Explored bias mitigation in resource-constrained scenarios
Diagnostic Tools
Developed quantitative methods for bias assessment:
- Diagnostic Probes: Tools to detect and measure bias in model outputs
- Fairness Metrics: Quantified disparities across demographic groups
- Model Selection Criteria: Informed model selection based on fairness considerations
Key Findings
Model Performance:
- LLMs outperform DNN-based models in depression classification
- LLMs show particularly strong performance for underrepresented groups (e.g., Hispanic participants)
- LLMs exhibit reduced gender bias compared to DNN-based embeddings
Bias Mitigation Results:
- Worst-group loss achieves best balance between performance and fairness for DNNs
- Guided prompting with ethical framing mitigates gender bias in 1-shot LLM settings
- Increasing shot count (N-shot learning) does not further reduce disparities
- Race/ethnicity disparities persist across both model types and mitigation strategies
Persistent Challenges:
- Racial disparities remain difficult to mitigate with current techniques
- Neither prompting strategies nor increased shots effectively reduce race/ethnicity disparities
- Highlights need for continued research in fairness-aware mental health AI
Impact
This work advances responsible AI in mental health applications by:
- Providing empirical evidence of bias in depression detection systems
- Demonstrating the importance of fairness evaluation in clinical AI
- Offering practical mitigation strategies for reducing demographic disparities
- Informing model selection and deployment decisions for mental health tools
This research is conducted at the HUBBS Lab at CU Boulder’s Institute of Cognitive Science under the supervision of Dr. Theodora Chaspari.