Computer Science 5th Year Master of Science Thesis Presentation

— 2:00pm

Location:
In Person - Gates Hillman 9115

Speaker:
MIHIR DHAMANKAR , Master's Student, Computer Science Department, Carnegie Mellon University
https://mdkar.github.io/

Extraction of Training Data from Fine-Tuned Large Language Models

Large Language Models have been shown to perform well on natural language tasks, even those they were not explicitly trained to perform. Fine tuning these models on smaller datasets has become a popular technique to achieve high performance on specific tasks. However, fine tuning can lead to the memorization of training data, which may be a privacy concern. In this work, I investigated the extraction of training data from fine-tuned large language models. I conducted a series of experiments to determine how easily private training data can be extracted from fine-tuned models using different data extraction techniques. I also investigated how the amount of training data used for fine-tuning, the number of epochs, the length and content of each training sample, and the fine-tuning technique and parameters used affect the ease of data extraction. I found that data extraction is simple with direct access to the model if the model is not trained only on prompt completions. I also found that the proportion of data that can be extracted increased with the amount of data used for fine-tuning (for a constant number of epochs). This work has implications for the privacy of individuals whose data is used for fine-tuning, as well as for businesses or groups that use fine-tuned models in public facing software. 

Thesis Committee:

Matt Fredrikson (Chair)
Yuvraj Agarwal

Additional Information


Add event to Google
Add event to iCal