Current Natural language processing (NLP) techniques are firmly ingrained in Big Data to manage a wide range of text-based jobs. However, using these methods may require handling private or sensitive data. As a result, privacy in natural language processing has come under more attention. A prominent and extensively studied strategy is Differential Privacy (DP), drawing the focus of many researchers on how to integrate this privacy enhancement tool into NLP models. Nevertheless, an aspect requiring further research is the assessment of word-level DP mechanisms against adversary NLP models. Consequently, this thesis delves into the investigation of unveil privacy, with the objective of revealing the original text inputs from the privatized, obfuscated text outputs generated by the word-level DP mechanisms. Our study focuses on identifying the prerequisites essential for attackers to successfully make correct inferences from privatized data which involves the examination of training methodologies and architectural features essential for the success of an adversary model. We also address suitable metrics to comprehensively assess the success of these adversary models and consequently, our research contributes to the field by developing a benchmark to evaluate and compare the resilience of word-level DP mechanisms in their resistance against an attacker's attempts to infer the original text inputs.
Name | Type | Size | Last Modification | Last Editor |
---|---|---|---|---|
240205MT_FurkanYilmaz_KickOffPresentation.pptx | 1,51 MB | 17.06.2024 Versions | ||
240615 MasterThesis_FurkanYilmaz.pdf | 5,91 MB | 17.06.2024 | ||
240617 MT_FurkanYilmaz_FinalPresentation.pptx | 6,93 MB | 17.06.2024 |