Verity
Perplexity style AI answer engine for AI PCs with CPU,GPU and NPU support
Install / Use
/learn @rupeshs/VerityREADME
🔍Verity
Verity is a Perplexity-style AI search and answer engine that runs fully locally on AI PCs.
It combines SearXNG-powered search, retrieval, and local LLM reasoning to generate grounded, verifiable answers — without relying on cloud-based LLM providers.

✨ Key Features
- Fully Local, AI PC Ready - Optimized for Intel AI PCs using OpenVINO (CPU / iGPU / NPU), Ollama (CPU / CUDA / Metal)
- No need of any paid APIs
- Privacy by Design - Search and inference can be fully self-hosted
- SearXNG-Powered Search - Self-hosted, privacy-friendly meta search engine
- Designed for fact-grounded, concise answers
- OpenVINO,Ollama models,llama.cpp server or any OpenAI compatible LLM servers supported
- Modular architecture
- CLI and WebUI support
- API server support
- Powered by Jan-nano 4B model,or configure any model
Supported LLM runtimes:
- OpenVINO - CPU / iGPU / NPU
- Ollama - CPU / CUDA / Apple Metal
- llama.cpp server or any OpenAI compatible LLM server
Dependencies
- Python 3.10 or higher
- uv - fast Python package and project manager
- Node.js
How to Install and Run
Follow the steps to install and run on Windows(Automated).
- Step 1: Install SearXNG by following this guide
- Step 2: Clone/Download verity repo
- Step 3: Double click
install.batto install
To run CLI interactive app double click start.bat.
To run WebUI start the webserver by double clicking start-webserver.bat and start frontend by double clicking start-webui.bat.
OpenVINO models will be downloaded in the first run.
Manual installation
Ensure that you have Python 3.x and Node.js installed.
- Install SearXNG by following this guide
- Clone Verity repo, enter into verity dirctory
- Create a virtual environment and activate it
- Run the below commands :
pip install -r requirements.txt
playwright install
cd src/frontend
npm install
For backend configurations create a .env file refer the .env.sample file.
To run interactive CLI app run the below
python src/app.py
Run WebUI
First run the API server
python src/api_server.py
Start the WebUI
cd src/frontend
npm run dev
Models
We have tested the following models.
| LLM Provider | Recommended Model | Processor | | --------------------- | ----------------------------------|------------ | OpenVINO | rupeshs/jan-nano-int4-ov | CPU/GPU | | OpenVINO | rupeshs/jan-nano-int4-npu-ov | NPU (Intel)| | Ollama | mannix/jan-nano:latest | CPU/CUDA | | llama.cpp server | Jan-nano-Q4_K_M.gguf | CPU/CUDA/Metal |
Tested using Intel AI PC with Intel Core Ultra Series 1 processor with CPU/GPU/NPU.
How to use LLama.cpp server with Verity
Run the llama.cpp server:
llama-server.exe -m Jan-nano-Q4_K_M.gguf -c 4096 --port 9000
If you are changing port configure in the .env file
OPENAI_LLM_BASE_URL=http://localhost:8000
:exclamation: You can use any OpenAI compatible LLM servers with Verity.
How to use Ollama models with Verity
Use the below config in .env file. update LLM provider(LLM_PROVIDER) as ollama and use model(LLM_MODEL_PATH) as mannix/jan-nano:latest.
Run the below command to pull the model.
ollama pull mannix/jan-nano:latest
Docker (Without OpenVINO Support)
To build and start the services:
Configurations in .env_docker file
docker compose up --build
To open verity webui http://localhost:5000/
You can use llama.cpp server, ollama or any OpenAI compatible LLM server.
API server - Streaming API (SSE)
Verity API server exposes one API endpoint,it streams model responses using Server-Sent Events (SSE).
http://localhost:8000/
Endpoint
GET /ask?question=<query>
Response Type
text/event-stream
Events
| Event Name | Description |
|------------|------------|
| search | Search phase started |
| read | Reading and processing search results |
| think | Generating final answer |
| token | Partial answer token |
| done | Stream completed |
| error | Eroor message |
CLI (go)

Follow the steps to run Verity CLI
-
Step 1: Install SearXNG by following this guide
-
Step 2: Build and run verity CLI tool,in the terminal enter into
verity/src/clifolder and run the below command :go build -
Step 3:Install and run the Verity Web server by double clicking
install.bat& followed bystart-webserver.bat. -
Step 4: Verity CLI tool can be used as :
verity.exe "your question"
Verity MCP server

To use MCP server
- Install and run
start-mcpserver.bat
Add the following configuration :
{
"mcpServers": {
"verity": {
"command": "npx",
"args": [
"mcp-remote",
"http://127.0.0.1:8000/mcp"
],
"timeout": 120000
}
}
}
Related Skills
Hook Development
96.8kThis skill should be used when the user asks to "create a hook", "add a PreToolUse/PostToolUse/Stop hook", "validate tool use", "implement prompt-based hooks", "use ${CLAUDE_PLUGIN_ROOT}", "set up event-driven automation", "block dangerous commands", or mentions hook events (PreToolUse, PostToolUse, Stop, SubagentStop, SessionStart, SessionEnd, UserPromptSubmit, PreCompact, Notification). Provides comprehensive guidance for creating and implementing Claude Code plugin hooks with focus on advanced prompt-based hooks API.
MCP Integration
96.8kThis skill should be used when the user asks to "add MCP server", "integrate MCP", "configure MCP in plugin", "use .mcp.json", "set up Model Context Protocol", "connect external service", mentions "${CLAUDE_PLUGIN_ROOT} with MCP", or discusses MCP server types (SSE, stdio, HTTP, WebSocket). Provides comprehensive guidance for integrating Model Context Protocol servers into Claude Code plugins for external tool and service integration.
Plugin Structure
96.8kThis skill should be used when the user asks to "create a plugin", "scaffold a plugin", "understand plugin structure", "organize plugin components", "set up plugin.json", "use ${CLAUDE_PLUGIN_ROOT}", "add commands/agents/skills/hooks", "configure auto-discovery", or needs guidance on plugin directory layout, manifest configuration, component organization, file naming conventions, or Claude Code plugin architecture best practices.
Skill Development
96.8kThis skill should be used when the user wants to "create a skill", "add a skill to plugin", "write a new skill", "improve skill description", "organize skill content", or needs guidance on skill structure, progressive disclosure, or skill development best practices for Claude Code plugins.
