SkillAgentSearch skills...

Dingo

Dingo: A Comprehensive AI Data, Model and Application Quality Evaluation Tool

Install / Use

/learn @MigoXLab/Dingo

README

<!-- SEO Meta Information and Structured Data --> <div itemscope itemtype="https://schema.org/SoftwareApplication" align="center" xmlns="http://www.w3.org/1999/html"> <meta itemprop="name" content="Dingo: A Comprehensive AI Data Quality Evaluation Tool"> <meta itemprop="description" content="Comprehensive AI-powered data quality assessment platform for machine learning datasets, LLM training data validation, hallucination detection, and RAG system evaluation"> <meta itemprop="applicationCategory" content="Data Quality Software"> <meta itemprop="operatingSystem" content="Cross-platform"> <meta itemprop="programmingLanguage" content="Python"> <meta itemprop="url" content="https://github.com/MigoXLab/dingo"> <meta itemprop="downloadUrl" content="https://pypi.org/project/dingo-python/"> <meta itemprop="softwareVersion" content="latest"> <meta itemprop="license" content="Apache-2.0"> <!-- logo --> <p align="center"> <img src="docs/assets/dingo-logo.png" width="300px" style="vertical-align:middle;" alt="Dingo AI Data Quality Evaluation Tool Logo"> </p> <!-- badges --> <p align="center"> <a href="https://github.com/pre-commit/pre-commit"><img src="https://img.shields.io/badge/pre--commit-enabled-brightgreen?logo=pre-commit&logoColor=white" alt="pre-commit"></a> <a href="https://pypi.org/project/dingo-python/"><img src="https://img.shields.io/pypi/v/dingo-python.svg" alt="PyPI version"></a> <a href="https://pypi.org/project/dingo-python/"><img src="https://img.shields.io/pypi/pyversions/dingo-python.svg" alt="Python versions"></a> <a href="https://github.com/MigoXLab/dingo/blob/main/LICENSE"><img src="https://img.shields.io/github/license/DataEval/dingo" alt="License"></a> <a href="https://github.com/MigoXLab/dingo/stargazers"><img src="https://img.shields.io/github/stars/DataEval/dingo" alt="GitHub stars"></a> <a href="https://github.com/MigoXLab/dingo/network/members"><img src="https://img.shields.io/github/forks/DataEval/dingo" alt="GitHub forks"></a> <a href="https://github.com/MigoXLab/dingo/issues"><img src="https://img.shields.io/github/issues/DataEval/dingo" alt="GitHub issues"></a> <a href="https://mseep.ai/app/dataeval-dingo"><img src="https://mseep.net/pr/dataeval-dingo-badge.png" alt="MseeP.ai Security Assessment Badge" height="20"></a> <a href="https://deepwiki.com/MigoXLab/dingo"><img src="https://deepwiki.com/badge.svg" alt="Ask DeepWiki"></a> <a href="https://archestra.ai/mcp-catalog/dataeval__dingo"><img src="https://archestra.ai/mcp-catalog/api/badge/quality/DataEval/dingo" alt="Trust Score"></a> <a href="https://clawhub.ai/e06084/dingo"><img src="https://img.shields.io/badge/ClawHub-Skill-orange?logo=data:image/svg+xml;base64,PHN2ZyB4bWxucz0iaHR0cDovL3d3dy53My5vcmcvMjAwMC9zdmciIHZpZXdCb3g9IjAgMCAyNCAyNCI+PHRleHQgeT0iMTgiIGZvbnQtc2l6ZT0iMTYiPvCfpp48L3RleHQ+PC9zdmc+" alt="ClawHub Skill"></a> </p> </div> <div align="center">

English · 简体中文 · 日本語

</div> <!-- join us --> <p align="center"> 👋 join us on <a href="https://discord.gg/Jhgb2eKWh8" target="_blank">Discord</a> and <a href="./docs/assets/wechat.jpg" target="_blank">WeChat</a> </p> <p align="center"> If you like Dingo, please give us a ⭐ on GitHub! <br/> <a href="https://github.com/MigoXLab/dingo/stargazers" target="_blank"> <img src="docs/assets/clickstar_2.gif" alt="Click Star" width="480"> </a> </p>

Introduction

Dingo is A Comprehensive AI Data, Model and Application Quality Evaluation Tool, designed for ML practitioners, data engineers, and AI researchers. It helps you systematically assess and improve the quality of training data, fine-tuning datasets, and production AI systems.


🚀 Enterprise Dingo SaaS Version

Need a production-grade data quality platform? Try Dingo SaaS Enterprise Edition!

✨ Compared to the open-source version, SaaS provides:

  • 🌐 Web UI - Visual evaluation interface, no coding required
  • 🔐 Access Control - JWT + Google OAuth 2.0
  • 📊 Visual Reports - Interactive charts, trend analysis, export features
  • 🔌 RESTful API - Seamless integration with existing systems

📝 How to Get Free SaaS Code

👉 Apply for Dingo SaaS Repository Access

Review time: 1-5 business days | Suitable for enterprise data governance, team collaboration


Why Dingo?

🎯 Production-Grade Quality Checks - From pre-training datasets to RAG systems, ensure your AI gets high-quality data

🗄️ Multi-Source Data Integration - Seamlessly connect to Local files, SQL databases (PostgreSQL/MySQL/SQLite), HuggingFace datasets, and S3 storage

🔍 Multi-Field Evaluation - Apply different quality rules to different fields in parallel (e.g., ISBN validation for isbn, text quality for title)

🤖 RAG System Assessment - Comprehensive evaluation of retrieval and generation quality with 5 academic-backed metrics

🧠 LLM & Rule & Agent Hybrid - Combine fast heuristic rules (30+ built-in) with LLM-based deep assessment

🚀 Flexible Execution - Run locally for rapid iteration or scale with Spark for billion-scale datasets

📊 Rich Reporting - Detailed quality reports with GUI visualization and field-level insights

Architecture Diagram

Architecture of dingo

Quick Start

Installation

# Core package (includes rule evaluation, LLM evaluation, MCP server, datasource support)
pip install dingo-python

# With HHEM hallucination detection model (requires transformers + torch)
pip install "dingo-python[hhem]"

# With all features (HHEM + Agent)
pip install "dingo-python[all]"

Example Use Cases of Dingo

1. Evaluate LLM chat data

from dingo.config.input_args import EvaluatorLLMArgs
from dingo.io.input import Data
from dingo.model.llm.text_quality.llm_text_quality_v4 import LLMTextQualityV4
from dingo.model.rule.rule_common import RuleSpecialCharacter

data = Data(
    data_id='123',
    prompt="hello, introduce the world",
    content="�I am 8 years old. ^I love apple because:"
)


def llm():
    LLMTextQualityV4.dynamic_config = EvaluatorLLMArgs(
        key='YOUR_API_KEY',
        api_url='https://api.openai.com/v1/chat/completions',
        model='gpt-4o',
    )
    res = LLMTextQualityV4.eval(data)
    print(res)

def rule():
    res = RuleSpecialCharacter().eval(data)
    print(res)

rule()

2. Evaluate Dataset

from dingo.config import InputArgs
from dingo.exec import Executor

# Evaluate a dataset from Hugging Face
if __name__ == '__main__':
    input_data = {
        "input_path": "tatsu-lab/alpaca",  # Dataset from Hugging Face
        "dataset": {
            "source": "hugging_face",
            "format": "plaintext"  # Format: plaintext
        },
        "executor": {
            "result_save": {
                "bad": True  # Save evaluation results
            }
        },
        "evaluator": [
            {
                "evals": [
                    {"name": "RuleColonEnd"},
                    {"name": "RuleSpecialCharacter"}
                ]
            }
        ]
    }

    input_args = InputArgs(**input_data)
    executor = Executor.exec_map["local"](input_args)
    result = executor.execute()
    print(result)

Command Line Interface

Evaluate with Rule Sets

dingo eval --input .github/env/local_plaintext.json

Evaluate with LLM (e.g., GPT-4o)

dingo eval --input .github/env/local_json.json

MCP Server

Dingo includes a built-in Model Context Protocol (MCP) server for AI agent integration:

# Start MCP server (SSE transport, default port 8000)
dingo serve

# Custom port
dingo serve --port 9000

# stdio transport (for Claude Desktop)
dingo serve --transport stdio

For detailed setup and Cursor/Claude Desktop integration, see the dedicated documentation:

English · 简体中文 · 日本語

Video Demonstration

To help you get started quickly with Dingo MCP, we've created a video walkthrough:

https://github.com/user-attachments/assets/aca26f4c-3f2e-445e-9ef9-9331c4d7a37b

This video demonstrates step-by-step how to use Dingo MCP server with Cursor.


📚 Data Quality Metrics

Dingo provides 100+ evaluation metrics across multiple dimensions, combining rule-based speed with LLM-based depth.

Metric Categories

| Category | Examples | Use Case | |----------|----------|----------| | Pretrain Text Quality | Completeness, Effectiveness, Similarity, Security | LLM pre-training data filtering | | SFT Data Quality | Honest, Helpful, Harmless (3H) | Instruction fine-tuning data | | RAG Evaluation | Faithfulness, Context Precision, Answer Relevancy | RAG system assessment | | Hallucination Detection | HHEM-2.1-Open, Factuality Check | Production AI reliability | | Classification | Topic categorization, Content labeling | Data organization | | Multimodal | Image-text relevance, VLM quality, OCR visual evaluation | Vision-language data | | Security | PII detection, Perspective API toxicity | Privacy and safety |

📊 View Complete Metrics Documentation →
📖 RAG Evaluation Guide → | 中文版
🔍 Hallucination Detection Guide → | 中文版
Factuality Assessment Guide → | 中文版
👁️ VLM Render Judge Guide → | 中文版

Most metrics are backed by academic research to ensure scientific rigor.

Quick Metric Usage

llm_config = {
    "model": "gpt-4o",
    "key": "YOUR_API_KEY",
    "api_
View on GitHub
GitHub Stars671
CategoryDevelopment
Updated1d ago
Forks68

Languages

Python

Security Score

100/100

Audited on Mar 31, 2026

No findings