SkillAgentSearch skills...

Nanowakeword

A lightweight, open-source, and intelligent wake word detection engine. Train custom, high-accuracy models with minimal effort.

Install / Use

/learn @arcosoph/Nanowakeword

README

<p align="center"> <img src="https://raw.githubusercontent.com/arcosoph/nanowakeword/main/assets/logo/logo_0.png" alt="Logo" width="290"> </p> <p align="center"> <a href="https://colab.research.google.com/github/arcosoph/nanowakeword/blob/main/notebooks/Train_Your_First_Wake_Word_Model.ipynb"><img alt="Open In Colab" src="https://img.shields.io/badge/Open%20in%20Colab-FFB000?logo=googlecolab&logoColor=white"></a> <a href="https://discord.gg/rYfShVvacB"><img alt="Join the Discord" src="https://img.shields.io/badge/Join%20the%20Discord-5865F2?logo=discord&logoColor=white"></a> <a href="https://pypi.org/project/nanowakeword/"><img alt="PyPI" src="https://img.shields.io/pypi/v/nanowakeword.svg?color=6C63FF&logo=pypi&logoColor=white"></a> <a href="https://pypi.org/project/nanowakeword/"><img alt="Python" src="https://img.shields.io/pypi/pyversions/nanowakeword.svg?color=3776AB&logo=python&logoColor=white"></a> <a href="https://pepy.tech/projects/nanowakeword"><img alt="PyPI Downloads" src="https://static.pepy.tech/personalized-badge/nanowakeword?period=total&units=INTERNATIONAL_SYSTEM&left_color=GRAY&right_color=BLACK&left_text=downloads"></a> <a href="https://pypi.org/project/nanowakeword/"><img alt="License" src="https://img.shields.io/pypi/l/nanowakeword?color=white&logo=apache&logoColor=black"></a> </p>

NanoWakeWord is a next-generation, adaptive framework designed to build high-performance, custom wake word models. More than just a tool, it’s an intelligent engine that understands your data and optimizes the entire training process to deliver exceptional accuracy and efficiency.

Quick Access

Choose Your Architecture, Build Your Pro Model

NanoWakeWord is a versatile framework offering a rich library of neural network architectures. Each is optimized for different scenarios, allowing you to build the perfect model for your specific needs. This Colab notebook lets you experiment with any of them.

| Architecture | Recommended Use Case | Performance Profile | Start Training | | :--- | :--- | :--- | :--- | | DNN | General use on resource-constrained devices (e.g., MCUs). | Fastest Training, Low Memory | ▶️ Launch | | RNN | Baseline experiments or educational purposes. | Better than DNN | ▶️ Launch | | CNN | Short, sharp, and explosive wake words. | Efficient Feature Extraction | ▶️ Launch | | LSTM | Noisy environments or complex, multi-syllable phrases. | Best-in-Class Noise Robustness | ▶️ Launch | | GRU | A faster, lighter alternative to LSTM with similar high performance. | Balanced: Speed & Robustness | ▶️ Launch | | CRNN | Challenging audio requiring both feature and context analysis. | Hybrid Power: CNN + RNN | ▶️ Launch | | TCN | Modern, high-speed sequential processing. | Faster than RNN (Parallel) | ▶️ Launch | | QuartzNet| Top accuracy with a small footprint on edge devices. | Parameter-Efficient & Accurate | ▶️ Launch | | Transformer| Deep Contextual Understanding via Self-Attention mechanism. | SOTA Performance & Flexibility | ▶️ Launch | | Conformer | State-of-the-art hybrid for ultimate real-world performance. | SOTA: Global + Local Features | ▶️ Launch | | E-Branchformer| Bleeding-edge research for potentially the highest accuracy. | Peak Accuracy Potential | ▶️ Launch |


[!NOTE] Nanowakeword is under active development. For important updates, version-specific notes, and the latest stability status of all features, please refer to our official status document.

➡️ View Latest Release Notes & Project Status

State-of-the-Art Features and Architecture

Nanowakeword is not merely a tool; it's a holistic, end-to-end ecosystem engineered to democratize the creation of state-of-the-art, custom wake word models. It moves beyond simple scripting by integrating a series of automated, production-grade systems that orchestrate the entire lifecycle—from data analysis and feature engineering to advanced training and deployment-optimized inference.

<details> <summary><strong>1. Automated ML Engineering for Peak Performance</strong></summary>

The cornerstone and "brain" of the framework is its data-driven configuration engine. This system performs a holistic analysis of your unique dataset and hardware environment to replace hours of manual, error-prone hyper-parameter tuning with a single, intelligent process. It crafts a powerful, optimized training baseline by synergistically determining:

  • Adaptive Architectural Scaling: It doesn't just use a fixed architecture; it sculpts one for you. The engine dynamically scales the model's complexity—tuning its depth, width, and regularization (e.g., layers, neurons, dropout) to perfectly match the volume and complexity of your training data. This core function is critical for preventing both underfitting on small datasets and overfitting on large ones.

  • Optimized Training & Convergence Strategy: Based on data characteristics, it formulates a multi-stage, dynamic learning rate schedule and determines the precise training duration required to reach optimal convergence. This ensures the model is trained to its full potential without wasting computational resources on diminishing returns.

  • Hardware-Aware Performance Tuning: The engine profiles your entire hardware stack (CPU cores, system RAM, and GPU VRAM) to maximize throughput at every stage. It calculates the maximum efficient batch sizes for data generation, augmentation, and model training, ensuring that your hardware's full potential is unlocked.

  • Automatic Pre-processing: Just drop your raw audio files (.mp3, .flac, .pcm, etc.) into the data folders — NanoWakeWord automatically handles resampling, channel conversion, and format standardization.

  • Data-Driven Augmentation Policy: Rather than applying a generic augmentation strategy, the engine crafts a custom augmentation policy. It analyzes the statistical properties of your provided noise and reverberation files to tailor the intensity, probability, and type of on-the-fly augmentations, creating a training environment that mirrors real-world challenges.

While this engine provides a state-of-the-art baseline, it does not sacrifice flexibility. Advanced users retain full, granular control and can override any of the dozens of automatically generated parameters by simply specifying their desired value in the .yaml file.

</details> <details> <summary><strong>2. The Production-Grade Data Pipeline: From Raw Audio to Optimized Features</strong></summary>

Recognizing that data is the bedrock of any great model, Nanowakeword automates the entire data engineering lifecycle with a pipeline designed for scale and quality:

  • Phonetic Adversarial Negative Generation: This is a key differentiator. The system moves beyond generic noise and random words by performing a phonetic analysis of your wake word. It then synthesizes acoustically confusing counter-examples—phrases that sound similar but are semantically different. This forces the model to learn fine-grained phonetic boundaries, dramatically reducing the false positive rate in real-world use.

  • Dynamic On-the-Fly Augmentation: During training, a powerful augmentation engine injects a rich tapestry of real-world acoustic scenarios in real-time. This includes applying background noise at varying SNR levels, convolving clips with room impulse responses (RIR) for realistic reverberation, and applying a suite of other transformations like pitch shifting and filtering.

  • Seamless Large-Scale Data Handling (mmap): The framework shatters the

View on GitHub
GitHub Stars51
CategoryEducation
Updated3d ago
Forks7

Languages

Python

Security Score

100/100

Audited on Mar 23, 2026

No findings