Skip to main content

3 posts tagged with "openai"

View All Tags

System Cards Go Hard

Tabs Fakier
Founding Developer Advocate

What are system cards, anyway?​

A system card accompanies a LLM release with system-level information about the model's deployment.

A system card is not to be confused with a model card, which conveys information about the model itself. Hooray for being given far more than a list of features and inadequate documentation along with the expectation of churning out a working implementation of some tool by the end of the week.

How to Red Team GPT: Complete Security Testing Guide for OpenAI Models

Ian Webster
Engineer & OWASP Gen AI Red Teaming Contributor

OpenAI's GPT-4.1 and GPT-4.5 represents a significant leap in AI capabilities, especially for coding and instruction following. But with great power comes great responsibility. This guide shows you how to use Promptfoo to systematically test these models for vulnerabilities through adversarial red teaming.

GPT's enhanced instruction following and long-context capabilities make it particularly interesting to red team, as these features can be both strengths and potential attack vectors.

You can also jump directly to the GPT 4.1 security report and compare it to other models.

Automated Jailbreaking Techniques with DALL-E: Complete Red Team Guide

Ian Webster
Engineer & OWASP Gen AI Red Teaming Contributor

We all know that image models like OpenAI's Dall-E can be jailbroken to generate violent, disturbing, and offensive images. It turns out this process can be fully automated.

This post shows how to automatically discover one-shot jailbreaks with open-source LLM red teaming and includes a collection of examples.

llm image red teaming