The Lakera LLM Security Playbook provides a comprehensive resource for understanding AI security, focusing on the distinction between prompt attacks and non-prompt attacks in generative AI. It offers a detailed analysis of vulnerabilities in large language models (LLMs) and includes practical advice on data sanitization and personal identifiable information detection. The guide introduces Gandalf, an online game for learning AI security, and highlights the Lakera Guard security solution to counter AI threats. Featuring a database with nearly 30 million LLM attack data points, the playbook uses real-world examples to clarify common misconceptions and offers actionable guidelines for mitigating vulnerabilities, making it a valuable tool for professionals and enthusiasts aiming to secure AI systems effectively.