SkillAgentSearch skills...

Verity

Perplexity style AI answer engine for AI PCs with CPU,GPU and NPU support

Install / Use

/learn @rupeshs/Verity
About this skill

Quality Score

0/100

Supported Platforms

Claude Code
Cursor

README

🔍Verity

Verity is a Perplexity-style AI search and answer engine that runs fully locally on AI PCs.
It combines SearXNG-powered search, retrieval, and local LLM reasoning to generate grounded, verifiable answers — without relying on cloud-based LLM providers.

Verity screenshot

✨ Key Features

  • Fully Local, AI PC Ready - Optimized for Intel AI PCs using OpenVINO (CPU / iGPU / NPU), Ollama (CPU / CUDA / Metal)
  • No need of any paid APIs
  • Privacy by Design - Search and inference can be fully self-hosted
  • SearXNG-Powered Search - Self-hosted, privacy-friendly meta search engine
  • Designed for fact-grounded, concise answers
  • OpenVINO,Ollama models,llama.cpp server or any OpenAI compatible LLM servers supported
  • Modular architecture
  • CLI and WebUI support
  • API server support
  • Powered by Jan-nano 4B model,or configure any model

Supported LLM runtimes:

  • OpenVINO - CPU / iGPU / NPU
  • Ollama - CPU / CUDA / Apple Metal
  • llama.cpp server or any OpenAI compatible LLM server

Dependencies

  • Python 3.10 or higher
  • uv - fast Python package and project manager
  • Node.js

How to Install and Run

Follow the steps to install and run on Windows(Automated).

  • Step 1: Install SearXNG by following this guide
  • Step 2: Clone/Download verity repo
  • Step 3: Double click install.bat to install

To run CLI interactive app double click start.bat. To run WebUI start the webserver by double clicking start-webserver.bat and start frontend by double clicking start-webui.bat. OpenVINO models will be downloaded in the first run.

Manual installation

Ensure that you have Python 3.x and Node.js installed.

  • Install SearXNG by following this guide
  • Clone Verity repo, enter into verity dirctory
  • Create a virtual environment and activate it
  • Run the below commands :
pip install -r requirements.txt
playwright install
cd src/frontend
npm install

For backend configurations create a .env file refer the .env.sample file.

To run interactive CLI app run the below python src/app.py

Run WebUI

First run the API server python src/api_server.py

Start the WebUI

cd src/frontend
npm run dev

Models

We have tested the following models.

| LLM Provider | Recommended Model | Processor | | --------------------- | ----------------------------------|------------ | OpenVINO | rupeshs/jan-nano-int4-ov | CPU/GPU | | OpenVINO | rupeshs/jan-nano-int4-npu-ov | NPU (Intel)| | Ollama | mannix/jan-nano:latest | CPU/CUDA | | llama.cpp server | Jan-nano-Q4_K_M.gguf | CPU/CUDA/Metal |

Tested using Intel AI PC with Intel Core Ultra Series 1 processor with CPU/GPU/NPU.

How to use LLama.cpp server with Verity

Run the llama.cpp server:

llama-server.exe -m Jan-nano-Q4_K_M.gguf -c 4096 --port 9000

If you are changing port configure in the .env file

OPENAI_LLM_BASE_URL=http://localhost:8000

:exclamation: You can use any OpenAI compatible LLM servers with Verity.

How to use Ollama models with Verity

Use the below config in .env file. update LLM provider(LLM_PROVIDER) as ollama and use model(LLM_MODEL_PATH) as mannix/jan-nano:latest.

Run the below command to pull the model.

ollama pull mannix/jan-nano:latest

Docker (Without OpenVINO Support)

To build and start the services: Configurations in .env_docker file

docker compose up --build

To open verity webui http://localhost:5000/ You can use llama.cpp server, ollama or any OpenAI compatible LLM server.

API server - Streaming API (SSE)

Verity API server exposes one API endpoint,it streams model responses using Server-Sent Events (SSE). http://localhost:8000/

Endpoint

GET /ask?question=<query>

Response Type

text/event-stream

Events

| Event Name | Description | |------------|------------| | search | Search phase started | | read | Reading and processing search results | | think | Generating final answer | | token | Partial answer token | | done | Stream completed | | error | Eroor message |

CLI (go)

Verity CLI screenshot

Follow the steps to run Verity CLI

  • Step 1: Install SearXNG by following this guide

  • Step 2: Build and run verity CLI tool,in the terminal enter into verity/src/cli folder and run the below command :

    go build

  • Step 3:Install and run the Verity Web server by double clicking install.bat & followed by start-webserver.bat.

  • Step 4: Verity CLI tool can be used as : verity.exe "your question"

Verity MCP server

Verity MCP server running with LM Studio

To use MCP server

  • Install and run start-mcpserver.bat

Add the following configuration :

{
  "mcpServers": {
    "verity": {
      "command": "npx",
      "args": [
        "mcp-remote",
        "http://127.0.0.1:8000/mcp"
      ],
      "timeout": 120000
    }
  }
}

Related Skills

Hook Development

96.8k

This skill should be used when the user asks to "create a hook", "add a PreToolUse/PostToolUse/Stop hook", "validate tool use", "implement prompt-based hooks", "use ${CLAUDE_PLUGIN_ROOT}", "set up event-driven automation", "block dangerous commands", or mentions hook events (PreToolUse, PostToolUse, Stop, SubagentStop, SessionStart, SessionEnd, UserPromptSubmit, PreCompact, Notification). Provides comprehensive guidance for creating and implementing Claude Code plugin hooks with focus on advanced prompt-based hooks API.

MCP Integration

96.8k

This skill should be used when the user asks to "add MCP server", "integrate MCP", "configure MCP in plugin", "use .mcp.json", "set up Model Context Protocol", "connect external service", mentions "${CLAUDE_PLUGIN_ROOT} with MCP", or discusses MCP server types (SSE, stdio, HTTP, WebSocket). Provides comprehensive guidance for integrating Model Context Protocol servers into Claude Code plugins for external tool and service integration.

Plugin Structure

96.8k

This skill should be used when the user asks to "create a plugin", "scaffold a plugin", "understand plugin structure", "organize plugin components", "set up plugin.json", "use ${CLAUDE_PLUGIN_ROOT}", "add commands/agents/skills/hooks", "configure auto-discovery", or needs guidance on plugin directory layout, manifest configuration, component organization, file naming conventions, or Claude Code plugin architecture best practices.

Skill Development

96.8k

This skill should be used when the user wants to "create a skill", "add a skill to plugin", "write a new skill", "improve skill description", "organize skill content", or needs guidance on skill structure, progressive disclosure, or skill development best practices for Claude Code plugins.

View on GitHub
GitHub Stars46
CategoryCustomer
Updated8d ago
Forks4

Languages

Python

Security Score

80/100

Audited on Mar 24, 2026

No findings