OpenPipe is a flexible playground for comparing and optimizing LLM prompts. It lets you quickly generate, test and compare candidate prompts with realistic sample data.
These are simple experiments users have created that show how OpenPipe works. Feel free to fork them and start experimenting yourself.
You can use our hosted version of OpenPipe at https://openpipe.ai. You can also clone this repository and run it locally.
Visualize Responses
Inspect prompt completions side-by-side.
Test Many Inputs
OpenPipe lets you template a prompt. Use the templating feature to run the prompts you're testing against many potential inputs for broad coverage of your problem space.
Translate between Model APIs
Write your prompt in one format and automatically convert it to work with any other model.

Refine your prompts automatically
Use a growing database of best-practice refinements to improve your prompts automatically.

🪄 Auto-generate Test Scenarios
OpenPipe includes a tool to generate new test scenarios based on your existing prompts and scenarios. Just click "Autogenerate Scenario" to try it out!
- All models available through the OpenAI chat completion API
- Llama2 7b chat, 13b chat, 70b chat.
- Anthropic's Claude 1 Instant and Claude 2
- Install Postgresql.
- Install NodeJS 20 (earlier versions will very likely work but aren't tested).
- Install
pnpm
:npm i -g pnpm
- Clone this repository:
git clone https://github.com/openpipe/openpipe
- Install the dependencies:
cd openpipe && pnpm install
- Create a
.env
file (cp .env.example .env
) and enter yourOPENAI_API_KEY
. - Update
DATABASE_URL
if necessary to point to your Postgres instance and runpnpm prisma db push
to create the database. - Create a GitHub OAuth App and update the
GITHUB_CLIENT_ID
andGITHUB_CLIENT_SECRET
values. (Note: a PR to make auth optional when running locally would be a great contribution!) - Start the app:
pnpm dev
. - Navigate to http://localhost:3000