EthicsMH: A Pilot Benchmark for Ethical Reasoning in Mental Health AI
Abstract
EthicsMH is a dataset of 125 scenarios designed to evaluate AI systems' ethical reasoning in mental health contexts, focusing on decision accuracy, explanation quality, and alignment with professional norms.
The deployment of large language models (LLMs) in mental health and other sensitive domains raises urgent questions about ethical reasoning, fairness, and responsible alignment. Yet, existing benchmarks for moral and clinical decision-making do not adequately capture the unique ethical dilemmas encountered in mental health practice, where confidentiality, autonomy, beneficence, and bias frequently intersect. To address this gap, we introduce Ethical Reasoning in Mental Health (EthicsMH), a pilot dataset of 125 scenarios designed to evaluate how AI systems navigate ethically charged situations in therapeutic and psychiatric contexts. Each scenario is enriched with structured fields, including multiple decision options, expert-aligned reasoning, expected model behavior, real-world impact, and multi-stakeholder viewpoints. This structure enables evaluation not only of decision accuracy but also of explanation quality and alignment with professional norms. Although modest in scale and developed with model-assisted generation, EthicsMH establishes a task framework that bridges AI ethics and mental health decision-making. By releasing this dataset, we aim to provide a seed resource that can be expanded through community and expert contributions, fostering the development of AI systems capable of responsibly handling some of society's most delicate decisions.
Community
We introduce EthicsMH, the first pilot benchmark specifically designed for ethical reasoning in mental health AI. Unlike existing datasets that focus on general morality or clinical tasks, EthicsMH encodes 125 therapy-relevant scenarios across confidentiality, autonomy vs. beneficence, and race/gender bias. Each case includes structured fields (decision options, expert-aligned reasoning, expected AI behavior, real-world impact, and multi-stakeholder perspectives). This design enables evaluation not only of decisions but also of explanation quality, fairness, and alignment with professional norms. Built through a human-in-the-loop process with expert validation, EthicsMH serves as both a seed dataset and a methodological blueprint for creating larger, ethically grounded corpora in sensitive domains.
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- "Pull or Not to Pull?": Investigating Moral Biases in Leading Large Language Models Across Ethical Dilemmas (2025)
- Towards Assessing Medical Ethics from Knowledge to Practice (2025)
- TRIDENT: Benchmarking LLM Safety in Finance, Medicine, and Law (2025)
- Exploring Safety Alignment Evaluation of LLMs in Chinese Mental Health Dialogues via LLM-as-Judge (2025)
- Pluralistic Alignment for Healthcare: A Role-Driven Framework (2025)
- PHAX: A Structured Argumentation Framework for User-Centered Explainable AI in Public Health and Biomedical Sciences (2025)
- Data and AI governance: Promoting equity, ethics, and fairness in large language models (2025)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
Models citing this paper 0
No model linking this paper
Datasets citing this paper 1
Spaces citing this paper 0
No Space linking this paper