
Targeted Prompt Injection Attacks Against Code LLMs
New security vulnerabilities in AI code generation tools
This research introduces TPIA (Target-specific Prompt Injection Attack), a novel attack paradigm that can manipulate Code LLMs to generate malicious code with specific harmful functionality.
- Successfully demonstrated on major models including ChatGPT, Claude, and Gemini
- Achieves high attack success rates while maintaining code functionality
- Functions through carefully crafted prompts that bypass security measures
- Exposes significant security vulnerabilities in widely-used code generation tools
Implications for Security: Organizations using AI-powered coding assistants need to implement enhanced security protocols and monitoring systems, as these attacks can be executed without access to model parameters or training data.
TPIA: Towards Target-specific Prompt Injection Attack against Code-oriented Large Language Models