Mitigating Gender Bias in Distilled Language Models via Counterfactual Role Reversal Paper • 2203.12574 • Published Mar 23, 2022
Operationalizing a Threat Model for Red-Teaming Large Language Models (LLMs) Paper • 2407.14937 • Published Jul 20, 2024 • 1
AlexaTM 20B: Few-Shot Learning Using a Large-Scale Multilingual Seq2Seq Model Paper • 2208.01448 • Published Aug 2, 2022
Measuring Fairness of Text Classifiers via Prediction Sensitivity Paper • 2203.08670 • Published Mar 16, 2022