64 blog posts published by month since the start of 2024. Start from a different year:

Posts year-to-date
45 (19 posts by this month last year.)
Average posts per month since 2024
2.7

Post details (2024 to today)

Title Author Date Word count HN points
GPT-5.2 Initial Trust and Safety Assessment Michael D'Angelo Dec 11, 2025 1426 -
How to replicate the Claude Code attack with Promptfoo Ian Webster Nov 17, 2025 2516 -
Promptfoo vs PyRIT: A Practical Comparison of LLM Red Teaming Tools Ian Webster Jun 27, 2025 859 -
OWASP Top 10 LLM Security Risks (2025) – 5-Minute TLDR Tabs Fakier Jul 14, 2025 1132 -
Prompt Injection: A Comprehensive Guide Ian Webster Oct 09, 2024 1692 -
What are the Security Risks of Deploying DeepSeek-R1? Vanessa Sauter Feb 03, 2025 1278 -
AI Red Teaming for complete first-timers Tabs Fakier Jul 22, 2025 1054 -
Preventing Bias & Toxicity in Generative AI Ian Webster Oct 08, 2024 1213 -
Does Fuzzing LLMs Actually Work? Vanessa Sauter Oct 17, 2024 845 -
When AI becomes the attacker: The rise of AI-orchestrated cyberattacks Michael D'Angelo Nov 10, 2025 2596 -
RAG Data Poisoning: Key Concepts Explained Ian Webster Nov 04, 2024 1658 -
Harder, Better, Prompter, Stronger: AI system prompt hardening Tabs Fakier Jul 01, 2025 1696 -
Promptfoo Raises $5M to Fix Vulnerabilities in AI Applications Ian Webster Jul 23, 2024 719 -
Building a Security Scanner for LLM Apps Dane Schneider Dec 16, 2025 2717 -
Next Generation of Red Teaming for LLM Agents Steven Klein Jun 15, 2025 1327 -
The Promptfoo MCP Proxy: Enterprise MCP Security Steven Klein Jul 14, 2025 1752 -
Join Promptfoo at Hacker Summer Camp 2025 Vanessa Sauter Jul 24, 2025 432 -
Promptfoo Raises $18.4M Series A to Build the Definitive AI Security Stack Ian Webster and Michael D'Angelo Jul 29, 2025 826 -
Will agents hack everything? Dane Schneider Nov 14, 2025 949 -
Real-Time Fact Checking for LLM Outputs Michael D'Angelo Nov 28, 2025 2404 -
Celebrating 100,000 Users: Promptfoo's Journey, Red Teaming, and the Future of AI Security Michael D'Angelo Jun 10, 2025 787 -
Understanding AI Agent Security Vanessa Sauter Feb 14, 2025 2220 -
How to Red Team a HuggingFace Model: Complete Security Testing Guide Ian Webster Nov 20, 2024 611 -
AI Safety vs AI Security in LLM Applications: What Teams Must Know Michael D'Angelo Aug 17, 2025 5514 -
Sensitive Information Disclosure in LLMs: Privacy and Compliance in Generative AI Vanessa Sauter Mar 11, 2025 1664 -
A2A Protocol: The Universal Language for AI Agents Asmi Gulati May 12, 2025 1605 -
Beyond DoS: How Unbounded Consumption is Reshaping LLM Security Vanessa Sauter Dec 31, 2024 2311 -
Prompt Injection vs Jailbreaking: What's the Difference? Michael D'Angelo Aug 18, 2025 1810 -
ModelAudit vs ModelScan: Comparing ML Model Security Scanners Ian Webster Jul 06, 2025 727 -
Red Team Your LLM with BeaverTails Ian Webster Dec 22, 2024 1121 -
Your model upgrade just broke your agent's safety Guangshuo Zang Dec 08, 2025 1980 -
Archive - Dec 18, 2025 732 -
Misinformation in LLMs: Causes and Prevention Strategies Vanessa Sauter Mar 19, 2025 2179 -
Inside MCP: A Protocol for AI Integration Asmi Gulati May 06, 2025 1709 -
The Invisible Threat: How Zero-Width Unicode Characters Can Silently Backdoor Your AI-Generated Code Asmi Gulati Apr 10, 2025 1265 -
Automated Jailbreaking Techniques with DALL-E: Complete Red Team Guide Ian Webster Jul 01, 2024 1196 -
Defending Against Data Poisoning Attacks on LLMs: A Comprehensive Guide Vanessa Sauter Jan 07, 2025 1391 -
How Do You Secure RAG Applications? Vanessa Sauter Oct 14, 2024 2597 -
Jailbreaking LLMs: A Comprehensive Guide (With Examples) Ian Webster Jan 07, 2025 4626 -
Autonomy and agency in AI: We should secure LLMs with the same fervor spent realizing AGI Tabs Fakier Sep 02, 2025 1427 -
System Cards Go Hard Tabs Fakier Jul 15, 2025 705 -
Introducing GOAT—Promptfoo's Latest Strategy Vanessa Sauter Nov 05, 2024 873 -
Leveraging Promptfoo for EU AI Act Compliance Vanessa Sauter Dec 10, 2024 955 -
Reinforcement Learning with Verifiable Rewards Makes Models Faster, Not Smarter Michael D'Angelo Oct 24, 2025 3599 -
Promptfoo Achieves SOC 2 Type II and ISO 27001 Certification: Strengthening Trust in AI Security Vanessa Sauter Jul 11, 2025 409 -
New Red Teaming Plugins for LLM Agents: Enhancing API Security Ian Webster Aug 14, 2024 484 -
Why Attack Success Rate (ASR) Isn't Comparable Across Jailbreak Papers Without a Shared Threat Model Michael D'Angelo Dec 12, 2025 2117 -
How to Red Team an Ollama Model: Complete Local LLM Security Testing Guide Ian Webster Nov 23, 2024 744 -
How to Red Team Claude: Complete Security Testing Guide for Anthropic Models Ian Webster May 22, 2025 745 -
OWASP Red Teaming: A Practical Guide to Getting Started Vanessa Sauter Mar 25, 2025 3100 -
Jailbreaking Black-Box LLMs Using Promptfoo: A Complete Walkthrough Vanessa Sauter Sep 26, 2024 1052 -
Testing AI’s “Lethal Trifecta” with Promptfoo Ian Webster Sep 28, 2025 2669 -
Evaluating political bias in LLMs Michael D'Angelo Jul 24, 2025 3820 -
How to Red Team a LangChain Application: Complete Security Testing Guide Ian Webster Jan 18, 2025 937 -
Promptfoo for Enterprise: AI Evaluation and Red Teaming at Scale Ian Webster Aug 21, 2024 316 -
Top 10 Open Datasets for LLM Safety, Toxicity & Bias Evaluation Ian Webster Oct 06, 2025 2972 -
How to Red Team Gemini: Complete Security Testing Guide for Google's AI Models Ian Webster Jun 18, 2025 1504 -
Top Open Source AI Red-Teaming and Fuzzing Tools in 2025 Tabs Fakier Aug 14, 2025 2564 -
Promptfoo vs Garak: Choosing the Right LLM Red Teaming Tool Ian Webster Jun 26, 2025 896 -
Understanding Excessive Agency in LLMs Ian Webster Oct 08, 2024 1242 -
How to Red Team GPT: Complete Security Testing Guide for OpenAI Models Ian Webster Jun 07, 2025 941 -
How to run CyberSecEval Ian Webster Dec 21, 2024 1227 -
1,156 Questions Censored by DeepSeek Ian Webster Jan 28, 2025 1050 -
How Much Does Foundation Model Security Matter? Vanessa Sauter Oct 04, 2024 735 -