Skip to content

A research repository exploring potential data leakage vulnerabilities in Large Language Models (LLMs). This work analyzes existing literature, methodologies, and privacy implications in modern LLM architectures, providing comprehensive summaries and insights from various research papers.

Notifications You must be signed in to change notification settings

Saketh1702/Data-Leakage-Detection-in-LLMs

Repository files navigation

Data Leakage Detection in LLMs

A framework for detecting data leakage and bias in LLMs (e.g., Llama-2, Mistral) using n-gram metrics and one-shot prompting. BLEURT and ROUGE-L models are used to evaluate similarity between reference and model outputs for guided and general prompts. The framework analyzes model behavior on MMLU and TruthfulQA benchmarks to identify training data memorization and gender stereotyping patterns.

About

A research repository exploring potential data leakage vulnerabilities in Large Language Models (LLMs). This work analyzes existing literature, methodologies, and privacy implications in modern LLM architectures, providing comprehensive summaries and insights from various research papers.

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Contributors 5