Papers
1–3 of 3Research Paper·Jan 29, 2026
KnowBias: Mitigating Social Bias in LLMs via Know-Bias Neuron Enhancement
Large language models (LLMs) exhibit social biases that reinforce harmful stereotypes, limiting their safe deployment. Most existing debiasing methods adopt a suppressive paradigm by modifying paramet...
8.0 viability
Research Paper·Feb 10, 2026
Evaluating Social Bias in RAG Systems: When External Context Helps and Reasoning Hurts
Social biases inherent in large language models (LLMs) raise significant fairness concerns. Retrieval-Augmented Generation (RAG) architectures, which retrieve external knowledge sources to enhance the...
5.0 viability
Research Paper·Jan 21, 2026
Self-Blinding and Counterfactual Self-Simulation Mitigate Biases and Sycophancy in Large Language Models
Fair decisions require ignoring irrelevant, potentially biasing, information. To achieve this, decision-makers need to approximate what decision they would have made had they not known certain facts, ...
2.0 viability