UP | HOME

Data Forging Is Harder Than You Think

Abstract

Recent research has introduced data forging attacks, which involve replacing mini-batches used in training with different ones that yield nearly identical model parameters. These attacks pose serious privacy concerns, as they can undermine membership inference predictions and falsely suggest machine unlearning without actual unlearning. Given such critical privacy implications, this paper aims to scrutinize existing attacks and understand the notion of data forging. First, we argue that state-of-the-art data forging attacks have key limitations, which make them unrealistic and easily detectable. Through experimentation on multiple hardware platforms, we demonstrate that approximation errors that existing attacks report are orders-of-magnitude higher than benign errors caused by numerical deviations. Next, we formulate data forging as an optimisation problem and show that solving it via simple gradient-based methods also results in high approximation errors. Finally, we theoretically analyse data forging for logistic regression. Our theoretical results suggest, even for logistic regression, it is difficult to efficiently find forged batches. In conclusion, our findings call for a reevaluation of existing attacks and highlight that data forging is still an intriguing open problem.

Details

The full paper can be found here. The paper has been accepted at the Privacy Regulation and Protection in Machine Learning Workshop at ICLR 2024.