Archive5

How to Red Team an Ollama Model: Complete Local LLM Security Testing Guide
Ian Webster · 11/23/2024
Running LLMs locally with Ollama? These models often bypass cloud safety filters.

How to Red Team a HuggingFace Model: Complete Security Testing Guide
Ian Webster · 11/20/2024
Open source models on HuggingFace often lack safety training.

Introducing GOAT—Promptfoo's Latest Strategy
Vanessa Sauter · 11/5/2024
Meet GOAT: our advanced multi-turn jailbreaking strategy that uses AI attackers to break AI defenders.
RAG Data Poisoning: Key Concepts Explained
Ian Webster · 11/4/2024
Attackers can poison RAG knowledge bases to manipulate AI responses.

Does Fuzzing LLMs Actually Work?
Vanessa Sauter · 10/17/2024
Traditional fuzzing fails against LLMs.

How Do You Secure RAG Applications?
Vanessa Sauter · 10/14/2024
RAG applications face unique security challenges beyond foundation models.
Prompt Injection: A Comprehensive Guide
Ian Webster · 10/9/2024
Prompt injections are the most critical LLM vulnerability.
Understanding Excessive Agency in LLMs
Ian Webster · 10/8/2024
When LLMs have too much power, they become dangerous.
Preventing Bias & Toxicity in Generative AI
Ian Webster · 10/8/2024
Biased AI outputs can destroy trust and violate regulations.

How Much Does Foundation Model Security Matter?
Vanessa Sauter · 10/4/2024
Not all foundation models are created equal when it comes to security.