Home / Companies / WhyLabs / Blog / Post Details
Content Deep Dive

Navigating Threats: Detecting LLM Prompt Injections and Jailbreaks

Blog post from WhyLabs

Post Details
Company
Date Published
Author
Felipe Adachi
Word Count
1,978
Language
English
Hacker News Points
-
Summary

This blog post discusses the issue of malicious attacks on language models (LLMs) such as jailbreak attacks and prompt injections. It presents two methods of detecting these attacks using LangKit, an open-source package for feature extraction for LLM and NLP applications. The first method involves comparing incoming user prompts to a set of known jailbreak/prompt injection attacks, while the second method is based on the assumption that under a prompt injection attack, the original prompt will not be followed by the model. Both methods have limitations, but they can help mitigate the issues associated with malicious LLM attacks.