Repository logo

SoftAdaClip: A Smooth Clipping Strategy for Fair and Private Model Training

Loading...
Thumbnail Image

Journal Title

Journal ISSN

Volume Title

Publisher

Abstract

Differential privacy (DP) provides strong protection for sensitive data, but often reduces model performance and fairness, especially for underrepresented groups. One major reason is gradient clipping in DP-SGD, which can disproportionately suppress learning signals for minority subpopulations. Although adaptive clipping can enhance utility, it still relies on uniform hard clipping, which may restrict fairness. To address this, we introduce SoftAdaClip, a differentially private training method that replaces hard clipping with a smooth, tanh-based transformation to preserve relative gradient magnitudes while bounding sensitivity. We evaluate SoftAdaClip on various datasets, including MIMIC-III (clinical text), GOSSIS-eICU (structured healthcare), and Adult Income (tabular data). Our results show that SoftAdaClip reduces subgroup disparities by up to 87% compared to DP-SGD and up to 48% compared to Adaptive-DPSGD, and these reductions in subgroup disparities are statistically significant. These findings underscore the importance of integrating smooth transformations with adaptive mechanisms to achieve fair and private model training.

Description

This thesis introduces SoftAdaClip, a novel differentially private training strategy that replaces traditional hard gradient clipping with a smooth, tanh-based transformation. The method aims to improve both model utility and fairness by preserving informative gradients while maintaining strong privacy guarantees. Through extensive experiments on healthcare and tabular datasets, SoftAdaClip demonstrates significant improvements in accuracy and subgroup fairness compared to standard DP-SGD and Adaptive-DPSGD.

Keywords

Machine Learning, Privacy, Fairness, Deep Learning, Differential Privay

Citation