Whispers in the Machine: Confidentiality in LLM-integrated Systems
-
Updated
Jul 9, 2024 - Python
Whispers in the Machine: Confidentiality in LLM-integrated Systems
The Security Toolkit for LLM Interactions
GitHub repository for a tool that detects and filters malicious prompts before they are entered into a Retrieval-Augmented Generation (RAG) database, ensuring data integrity and security.
π€― AI Security EXPOSED! Live Demos Showing Hidden Risks of π€ Agentic AI Flows: πPrompt Injection, β£οΈ Data Poisoning. Watch the recorded session:
LMpi (Language Model Prompt Injector) is a tool designed to test and analyze various language models, including both API-based models and local models like those from Hugging Face.
πΌ another CV template for your job application, yet powered by Typst and more
MER is a software that identifies and highlights manipulative communication in text from human conversations and AI-generated responses. MER benchmarks language models for manipulative expressions, fostering development of transparency and safety in AI. It also supports manipulation victims by detecting manipulative patterns in human communication.
Curated + custom prompt injections.
A prompt defence is a multi-layer defence that can be used to protect your applications against prompt injection attacks.
ChatGPT Jailbreaks, GPT Assistants Prompt Leaks, GPTs Prompt Injection, LLM Prompt Security, Super Prompts, Prompt Hack, Prompt Security, Ai Prompt Engineering, Adversarial Machine Learning.
Short list of indirect prompt injection attacks for OpenAI-based models.
Advanced Code and Text Manipulation Prompts for Various LLMs. Suitable for Siri, GPT-4o, Claude, Llama3, Gemini, and other high-performance open-source LLMs.
A benchmark for prompt injection detection systems.
Detecting malicious prompts used to exploit large language models (LLMs) by leveraging supervised machine learning classifiers
This repository provides implementation to formalize and benchmark Prompt Injection attacks and defenses
π LangKit: An open-source toolkit for monitoring Large Language Models (LLMs). π Extracts signals from prompts & responses, ensuring safety & security. π‘οΈ Features include text quality, relevance metrics, & sentiment analysis. π A comprehensive tool for LLM observability. π
Every practical and proposed defense against prompt injection.
PromptyAPI, people's LLM-based applications security layer
Add a description, image, and links to the prompt-injection topic page so that developers can more easily learn about it.
To associate your repository with the prompt-injection topic, visit your repo's landing page and select "manage topics."