System Cards Go Hard
What are system cards, anyway?​
A system card accompanies a LLM release with system-level information about the model's deployment.
A system card is not to be confused with a model card, which conveys information about the model itself. Hooray for being given far more than a list of features and inadequate documentation along with the expectation of churning out a working implementation of some tool by the end of the week.

The first system card possibly came about from Dalle-2 in 2022, although Meta researchers were using the term earlier. OpenAI started publishing them in March 2023 (GPT-4). Anthropic followed suit. If there are other models with accompanying system cards, let me know. I'd love to sit down and spend three hours sifting through 169 pages of some LLM I don't use.
The following list isn't comprehensive or indicative of all models. Information can span:
- Training data
- Security policies
- Limitations
- Safeguards
- Risk identification
- Evaluations. SO. MANY. EVALUATIONS. (child safety, autonomy, biological risk etc.)
- Alignment (deception) assessment
- Model behaviors
- 'Spiritual bliss' attractors
And we care because...​
Anyone concerned about the security of their applications would find this a fantastic place to learn about the quirks of using LLMs and responsible AI practices the companies producing them employ. This is information we can use to educate our users, inform our own deployment practices, and introduce the necessary precautions for interactions with the LLMs concerned.
They are invaluable... And admittedly, interesting.
Existing system cards​
These are the system cards I could find:
- DALL·E 2 Preview: Risks and Limitations
- DALL·E 3
- GPT-4
- GPT-4V
- GPT-4o
- GPT-4o Native Image Generation Addendum
- GPT-4.5
- OpenAI o1
- OpenAI o3 and o4-mini
- Sora
- Claude 3.7 Sonnet
- Claude Opus 4 and Sonnet 4

Evolution of system card releases

Common evaluation categories in system cards