21 skills found
promptfoo / PromptfooTest your prompts, agents, and RAGs. Red teaming/pentesting/vulnerability scanning for AI. Compare performance of GPT, Claude, Gemini, Llama, and more. Simple declarative configs with command line and CI/CD integration. Used by OpenAI and Anthropic.
AgentEvalHQ / AgentEvalAgentEval is the comprehensive .NET toolkit for AI agent evaluation—tool usage validation, RAG quality metrics, stochastic evaluation, and model comparison—built first for Microsoft Agent Framework (MAF) and Microsoft.Extensions.AI. What RAGAS, PromptFoo and DeepEval do for Python, AgentEval does for .NET
promptfoo / Promptfoo ActionThe GitHub Action for Promptfoo. Test your prompts, agents, and RAGs. AI Red teaming, pentesting, and vulnerability scanning for LLMs. Compare performance of GPT, Claude, Gemini, Llama, and more. Simple declarative configs with command line and CI/CD integration.
rsfl / Splunk MCP Llm SiemulatorA Docker lab integrating Splunk SIEM with Ollama LLM via MCP for AI security operations. Features Promptfoo OWASP evaluation, TA-ollama and TA-mcp-jsonrpc add-ons, dual bind-mount log ingestion, and real-time HEC streaming across six indexes for MITRE ATLAS TTP detection.
perzeuss / Dify PromptfooEvaluate Dify assistants with promptfoo!
rsfl / Splunk MCP Llm Siemulator LinuxLinux version of Splunk MCP LLM MCP SIEMulator . A Docker lab integrating Splunk SIEM with Ollama LLM via Model Context Protocol for AI-powered security operations. Features Promptfoo evaluation, OpenWebUI chat interface, Splunk UF and Raw HEC logging for real-time event ingestion and LLM-assisted incident response testing.
eon01 / LLMPromptEngineeringForDevelopersFilesThis repository contains the code snippets used in "LLM Prompt Engineering For Developers"
kpavlov / Koog Spring Boot AssistantKotlin + SpringBoot + Koog + Promptfoo example
TomasHer / Prompting BlueprintsYour guide to the Agentic AI evolution. **Prompting Blueprints** offers a curated collection of concepts and tactics for building autonomous AI workflows. Master tool-specific playbooks, backed by structured prompt packs and rigorous evaluations for the latest AI models.
openclay-ai / OpenclayRuntime-secured AI tooling framework for production-grade LLM applications, protecting against prompt injection, jailbreaks, and adversarial attacks.
promptfoo / Mini FooMini promptfoo used for interviews
yukinagae / Genkitx PromptfooCommunity Plugin for Genkit to use Promptfoo
GenAIGator / AI RedTeaming With PromptFooA collection of AI red teaming tests built with Promptfoo to simulate adversarial prompts, detect prompt injection and jailbreak vulnerabilities, and evaluate the security of LLM applications and agents.
promptfoo / MCP Agent ProviderA promptfoo custom provider to test MCP servers with our evil mcp server
christhesoul / Minitest PromptfooA simple Ruby wrapper for testing your LLM prompts with Promptfoo
syamsasi99 / Prompt Evaluatorprompt-evaluator is an open-source toolkit for evaluating, testing, and comparing LLM prompts. It provides a GUI-driven workflow for running prompt tests, tracking token usage, visualizing results, and ensuring reliability across models like OpenAI, Claude, and Gemini.
promptfoo / Redscan LitePromptfoo Interview Exercise
docker / Docker Model Runner And MCP With PromptfooExamples of how to use Docker Model Runner, Docker MCP Toolkit, and Promptfoo together to evaluate models, agents, and MCP servers
stephenc222 / Example PromptfooExample project demonstrating Promptfoo
blanky0230 / Promptfoo LabNotes, thoughts, and practical examples about just one single tool; that might proof useful in navingating through newly announced seas.