As Large Language Models (LLMs) become increasingly integral to various applications, ensuring their security is paramount.
Open-source LLM scanners play a crucial role in identifying vulnerabilities and mitigating risks associated with these models. Here’s an overview of some key open-source tools available on GitHub:
1. Vigil
- Function: Vigil is a Python library and REST API designed to detect and mitigate security threats in LLM prompts and responses. It identifies issues such as prompt injections and jailbreak attempts.
- Features: Modular and extensible scanners, canary tokens, and support for custom detection signatures.
- Stars: 200+.
2. Garak
- Function: A command-line tool for vulnerability scanning of LLMs, focusing on threats like prompt injections, hallucinations, and data leakage.
- Features: Supports multiple LLM platforms, heuristic and LLM-based detection methods.
- Stars: 1,000+.
3. LLMFuzzer
- Function: An open-source fuzzing framework for testing LLMs and their integrations via APIs.
- Features: Modular architecture, various fuzzing strategies, and API integration testing.
- Stars: 200+.
4. Agentic Security
- Function: A vulnerability scanner for Agent Workflows and LLMs, protecting against jailbreaks, fuzzing, and multimodal attacks.
- Features: Comprehensive fuzzing, API integration, and reinforcement learning-based attacks.
- Stars: Not specified.
5. Promptmap
- Function: A tool for testing prompt injection attacks against generative AI applications.
- Features: Automated tests for direct prompt injection, prompt leaking, and P2SQL injection.
- Stars: Not specified.
6. BurpGPT
- Function: An extension for Burp Suite that integrates LLMs to enhance web application security testing.
- Features: AI-enhanced vulnerability scanning, web traffic analysis, and custom-trained LLM support.
- Stars: 2,000+.
7. Purple Llama
- Function: Focuses on enhancing LLM security through tools like Llama Guard and Code Shield.
- Features: Benchmarks and models for mitigating LLM risks.
- Stars: Significant community interest, exact number not specified.
These tools contribute significantly to the security landscape of LLMs by providing open-source solutions for vulnerability detection and mitigation.
They enable developers and security professionals to proactively address potential threats and ensure more robust AI deployments.