Prompt injection attacks and defenses in LLM-integrated applications
-
Updated
Jan 22, 2024 - Python
Prompt injection attacks and defenses in LLM-integrated applications
The Prompt Injection Testing Tool is a Python script designed to assess the security of your AI system's prompt handling against a predefined list of user prompts commonly used for injection attacks. This tool utilizes the OpenAI GPT-3.5 model to generate responses to system-user prompt pairs and outputs the results to a CSV file for analysis.
Image Prompt Injection is a Python script that demonstrates how to embed a secret prompt within an image using steganography techniques. This hidden prompt can be later extracted by an AI system for analysis, enabling covert communication with AI models through images.
ASCII Art Prompt Injection is a novel approach to hacking AI assistants using ASCII art. This project leverages the distracting nature of ASCII art to bypass security measures and inject prompts into large language models, such as GPT-4, leading them to provide unintended or harmful responses.
LLM Security Project with Llama Guard
Client SDK to send LLM interactions to Vibranium Dome
Prompt Engineering Tool for AI Models with cli prompt or api usage
Add a description, image, and links to the prompt-injection-tool topic page so that developers can more easily learn about it.
To associate your repository with the prompt-injection-tool topic, visit your repo's landing page and select "manage topics."