EMNLP 2025

November 06, 2025

Suzhou, China

Would you like to see your presentation here, made available to a global audience of researchers?
Add your own presentation or have us affordably record your next conference.

Large Language Models (LLMs) often exhibit gender bias, resulting in unequal treatment of male and female subjects across different contexts. To address this issue, we propose a novel data generation framework that fosters exploratory thinking in LLMs. Our approach prompts models to generate story pairs featuring male and female protagonists in structurally identical, morally ambiguous scenarios, then elicits and compares their moral judgments. When inconsistencies arise, the model is guided to produce balanced, gender-neutral judgments. These story-judgment pairs are used to fine-tune or optimize the models via Direct Preference Optimization (DPO). Experimental results show that our method significantly reduces gender bias while preserving or even enhancing general model capabilities. We will release the code and generated data.

Downloads

SlidesPaperTranscript English (automatic)

Next from EMNLP 2025

AutoDCWorkflow: LLM-based Data Cleaning Workflow Auto-Generation and Benchmark
poster

AutoDCWorkflow: LLM-based Data Cleaning Workflow Auto-Generation and Benchmark

EMNLP 2025

+1Lan LiLiri Fang
Liri Fang and 3 other authors

06 November 2025