180 skills found · Page 1 of 6
aidlearning / AidLearning FrameWork🔥🔥🔥AidLearning is a powerful AIOT development platform, AidLearning builds a linux env supporting GUI, deep learning and visual IDE on Android...Now Aid supports CPU+GPU+NPU for inference with high performance acceleration...Linux on Android or HarmonyOS
neuralmagic / DeepsparseSparsity-aware deep learning inference runtime for CPUs
xororz / Local DreamRun Stable Diffusion on Android Devices with Snapdragon NPU acceleration. Also supports CPU/GPU inference.
ELS-RD / Transformer DeployEfficient, scalable and enterprise-grade CPU/GPU inference server for 🤗 Hugging Face transformer models 🚀
RQLuo / MixTeX Latex OCRMixTeX multimodal LaTeX, ZhEn, and, Table OCR. It performs efficient CPU-based inference in a local offline on Windows.
RWKV / Rwkv.cppINT4/INT5/INT8 and FP16 inference on CPU for RWKV language model
Tencent / TurboTransformersa fast and user-friendly runtime for transformer inference (Bert, Albert, GPT2, Decoders, etc) on CPU and GPU.
pnnbao97 / VieNeu TTSVietnamese TTS with instant voice cloning • On-device • Real-time CPU inference • 24kHz audio quality • Chuyển văn bản thành giọng nói tiếng Việt • Text to speech tiếng Việt • TTS tiếng Việt
kennethleungty / Llama 2 Open Source LLM CPU InferenceRunning Llama 2 and other Open-Source LLMs on CPU Inference Locally for Document Q&A
microsoft / T MACLow-bit LLM inference on CPU/NPU with lookup table
intel / Auto RoundSOTA rounding-based quantization for high-accuracy low-bit LLM inference, seamlessly optimized for CPU/XPU/CUDA, with multi-datatype support and full compatibility with vLLM, SGLang, and Transformers.
Geekgineer / YOLOs CPPCross-Platform Production-ready C++ inference engine for YOLO models (v5-v12, YOLO26). Unified API for detection, segmentation, pose estimation, OBB, and classification. Built on ONNX Runtime and OpenCV. Optimized for CPU/GPU with quantization support.
cunjian / Pytorch Face LandmarkFast and accurate face landmark detection library using PyTorch; Support 68-point semi-frontal and 39-point profile landmark detection; Support both coordinate-based and heatmap-based inference; Up to 100 FPS landmark inference speed with SOTA face detector on CPU.
Daniil-Osokin / Lightweight Human Pose Estimation 3d Demo.pytorchReal-time 3D multi-person pose estimation demo in PyTorch. OpenVINO backend can be used for fast inference on CPU.
abacaj / Mpt 30B InferenceRun inference on MPT-30B using CPU
Maknee / Minigpt4.cppPort of MiniGPT4 in C++ (4bit, 5bit, 6bit, 8bit, 16bit CPU inference with GGML)
PaddlePaddle / AnakinHigh performance Cross-platform Inference-engine, you could run Anakin on x86-cpu,arm, nv-gpu, amd-gpu,bitmain and cambricon devices.
brontoguana / KrasisKrasis is a Hybrid LLM runtime which focuses on efficient running of larger models on consumer grade VRAM limited hardware
rlouf / McxExpress & compile probabilistic programs for performant inference on CPU & GPU. Powered by JAX.
andrewkchan / Deepseek.cppCPU inference for the DeepSeek family of large language models in C++