danenania
I haven't yet tried this red teaming tool, but I recently started using promptfoo to build out an evals pipeline for Plandex, a terminal-based AI coding tool I'm building[1]. promptfoo has been a pleasure to work with so far and I'd recommend it to anyone who knows they need evals but isn't sure where to begin.

It's quite flexible for different kinds of prompting scenarios and makes it easy to e.g. test a prompt n number of times (good for catching long-tail issues), only re-run evals that failed previously (helps to reduce costs/running time when you're iterating), or define various kinds of success criteria--exactly matches an expected string, contains an expected substring, a boolean JSON property is true/false, an LLM call that determines success, etc. etc. It pretty much covers all the bases on that front.

It can also treat prompts as jinja2 templates which is good for testing 'dynamic' prompts which take parameters (all of Plandex's prompts are like this).

It seems like a good foundation to build red teaming on top of.

1 - https://github.com/plandex-ai/plandex

Oras
Can this be dynamic on prompts and providers?

I’m thinking of continuous evaluation for LLM in production, where after each call, a webhook will send the input/output to evaluate.