Skip to yearly menu bar Skip to main content


Poster
in
Workshop: The Impact of Memorization on Trustworthy Foundation Models

Evaluating Memorization in Parameter-Efficient Fine-tuning

Sanghyun Hong · Nicholas Carlini · Alexey Kurakin

[ ] [ Project Page ]
Sat 19 Jul 8:30 a.m. PDT — 9:30 a.m. PDT
 
presentation: The Impact of Memorization on Trustworthy Foundation Models
Sat 19 Jul 8:25 a.m. PDT — 5 p.m. PDT

Abstract:

We study the impact of an emerging fine-tuning paradigm, parameter-efficient fine-tuning (PEFT), on privacy. We use an off-the-shelf data extraction attack as a vehicle to comprehensively evaluate memorization on three language models fine-tuned on two datasets, repeated 3–5 times with different random seeds. Our main findings are: (1) for practitioners employing PEFT to construct personalized models, the fine-tuned models have lower privacy risks while maintaining reasonable utility; (2) for developers designing new PEFT algorithms, while safer than standard fine-tuning, certain design choices in the algorithms increases memorization unexpectedly; and (3) for researchers auditing the privacy of fine-tuned models, employing weak differential privacy is sufficient to mitigate existing data extraction risks without significantly compromising model utility.

Chat is not available.