Data Contamination and Leakage Detection

Research on identifying, preventing, and mitigating data contamination and leakage in training and evaluation of large language models

Hero image

Data Contamination and Leakage Detection

Research on Large Language Models in Data Contamination and Leakage Detection

Exposing the Blind Spot in Multimodal LLMs

Exposing the Blind Spot in Multimodal LLMs

A Framework to Detect Data Contamination Across Text & Images

Finding Hidden Memories in Large Language Models

Finding Hidden Memories in Large Language Models

Automated Detection of Privacy Vulnerabilities at Scale

Quantifying Data Extraction Risks in LLMs

Quantifying Data Extraction Risks in LLMs

A sequence-level probability approach to measuring training data leakage

Security Risks in Code Language Models

Security Risks in Code Language Models

Investigating Data Extraction Vulnerabilities Before and After Fine-tuning

Uncovering Dataset Contamination in LLMs

Uncovering Dataset Contamination in LLMs

A new metric for measuring training data leakage into evaluation sets

The Hidden Danger in AI Evaluation

The Hidden Danger in AI Evaluation

How LLMs judging other LLMs creates security vulnerabilities

Securing LLMs from Toxic Training Data

Securing LLMs from Toxic Training Data

A Data Attribution Approach to Finding & Filtering Unsafe Content

Privacy at Risk: Stealing Personal Data from LLMs

Privacy at Risk: Stealing Personal Data from LLMs

New technique extracts personally identifiable information from language models

Rethinking Data Poisoning in LLMs

Rethinking Data Poisoning in LLMs

From Security Vulnerabilities to Development Opportunities

Safeguarding AI Evaluation Integrity

Safeguarding AI Evaluation Integrity

Detecting benchmark contamination with innovative watermarking techniques

Evolving LLM Benchmarks

Evolving LLM Benchmarks

From Static to Dynamic Evaluation: Combating Data Contamination

Poison Pills in LLMs: Hidden Vulnerabilities

Poison Pills in LLMs: Hidden Vulnerabilities

How targeted data poisoning compromises AI security

Detecting LLM Training Data Exposure

Detecting LLM Training Data Exposure

New Attack Method Requires Only Generated Outputs

Defending Against Dead Code Poisoning

Defending Against Dead Code Poisoning

Novel detection techniques to secure code generation models

The Hidden Dangers of Private LLM Evaluations

The Hidden Dangers of Private LLM Evaluations

Security risks in closed-door model assessments

Unveiling the Black Box of LLM Training Data

Unveiling the Black Box of LLM Training Data

A novel approach to detect data imprints in proprietary models

The IP Protection Dilemma in LLM Fine-Tuning

The IP Protection Dilemma in LLM Fine-Tuning

Balancing utility and intellectual property protection for hardware design

The Illusion of LLM Benchmark Success

The Illusion of LLM Benchmark Success

Revealing the failures of contamination mitigation strategies

The LLM Memorization Challenge

The LLM Memorization Challenge

How language models can complete texts they weren't explicitly trained on

Defending Against Data Poisoning

Defending Against Data Poisoning

Understanding threats to deep learning security

Security Risks in Code-Generating LLMs

Security Risks in Code-Generating LLMs

Uncovering sensitive information disclosure vulnerabilities

Password Vulnerabilities in Fine-tuned LLMs

Password Vulnerabilities in Fine-tuned LLMs

How sensitive data can leak through model parameters

Rethinking Data Markets for LLMs

Rethinking Data Markets for LLMs

Game theory reveals flaws in current data valuation methods

Efficient Detection of AI Memory Leaks

Efficient Detection of AI Memory Leaks

A streamlined approach to measuring training data memorization in AI models

Key Takeaways

Summary of Research on Data Contamination and Leakage Detection