An innovative tool designed to revolutionize the testing, evaluation, and security of LLM applications. This versatile tool supports a test-driven development approach, allowing developers to optimize prompts, models, and APIs efficiently.
Whether you’re using CLI, integrating into CI/CD, or seeking robust security through automated red teaming, promptfoo offers a comprehensive solution to enhance the reliability and security of your LLM apps.
promptfoo
is a tool for testing, evaluating, and red-teaming LLM apps.
With promptfoo, you can:
The goal: test-driven LLM development instead of trial-and-error.
npx promptfoo@latest init
There are many different ways to evaluate prompts. Here are some reasons to consider promptfoo:
Start by establishing a handful of test cases – core use cases and failure cases that you want to ensure your prompt can handle.
As you explore modifications to the prompt, use promptfoo eval
to rate all outputs. This ensures the prompt is actually improving overall.
As you collect more examples and establish a user feedback loop, continue to build the pool of test cases.
To get started, run this command:
npx promptfoo@latest init
This will create a promptfooconfig.yaml
placeholder in your current directory.
After editing the prompts and variables to your liking, run the eval command to kick off an evaluation:
npx promptfoo@latest eval
For more information click here.
shadow-rs is a Windows kernel rootkit written in Rust, demonstrating advanced techniques for kernel manipulation…
Extract and execute a PE embedded within a PNG file using an LNK file. The…
Embark on the journey of becoming a certified Red Team professional with our definitive guide.…
This repository contains proof of concept exploits for CVE-2024-5836 and CVE-2024-6778, which are vulnerabilities within…
This took me like 4 days (+2 days for an update), but I got it…
MaLDAPtive is a framework for LDAP SearchFilter parsing, obfuscation, deobfuscation and detection. Its foundation is…