2,338 skills found · Page 2 of 78
tensorflow / Tflite MicroInfrastructure to enable deployment of ML models to low-power resource-constrained embedded targets (including microcontrollers and digital signal processors).
coqui-ai / STT🐸STT - The deep learning toolkit for Speech-to-Text. Training and deploying STT models has never been so easy.
666DZY666 / Micronetmicronet, a model compression and deploy lib. compression: 1、quantization: quantization-aware-training(QAT), High-Bit(>2b)(DoReFa/Quantization and Training of Neural Networks for Efficient Integer-Arithmetic-Only Inference)、Low-Bit(≤2b)/Ternary and Binary(TWN/BNN/XNOR-Net); post-training-quantization(PTQ), 8-bit(tensorrt); 2、 pruning: normal、regular and group convolutional channel pruning; 3、 group convolution structure; 4、batch-normalization fuse for quantization. deploy: tensorrt, fp32/fp16/int8(ptq-calibration)、op-adapt(upsample)、dynamic_shape
NVIDIA / Model OptimizerA unified library of SOTA model optimization techniques like quantization, pruning, distillation, speculative decoding, etc. It compresses deep learning models for downstream deployment frameworks like TensorRT-LLM, TensorRT, vLLM, etc. to optimize inference speed.
aws / Sagemaker Python SDKA library for training and deploying machine learning models on Amazon SageMaker
neulab / Prompt2modelprompt2model - Generate Deployable Models from Natural Language Instructions
u14app / Gemini Next ChatDeploy your private Gemini application for free with one click, supporting Gemini 1.5, Gemini 2.0 models.
tensorflow / Model OptimizationA toolkit to optimize ML models for deployment for Keras and TensorFlow, including quantization and pruning.
intentee / PaddlerOpen-source LLM load balancer and serving platform for self-hosting LLMs at scale 🏓🦙 Alternative to projects like llm-d, Docker Model Runner, etc but with less moving parts and simple deployments built around ggml ecosystem. Runs on CPU and GPU.
modelfoxdotdev / ModelfoxModelFox makes it easy to train, deploy, and monitor machine learning models.
soeaver / Caffe ModelCaffe models (including classification, detection and segmentation) and deploy files for famouse networks
NVlabs / GR00T WholeBodyControlWelcome to GR00T Whole-Body Control (WBC)! This is a unified platform for developing and deploying advanced humanoid controllers. This includes: Decoupled WBC models used in NVIDIA Isaac-Gr00t, Gr00t N1.5 and N1.6 and GEAR-SONIC
The-AI-Summer / Deep Learning In ProductionBuild, train, deploy, scale and maintain deep learning models. Understand ML infrastructure and MLOps using hands-on examples.
ShannonAI / Service StreamerBoosting your Web Services of Deep Learning Applications.
open-edge-platform / Training ExtensionsTrain, Evaluate, Optimize, Deploy Computer Vision Models via OpenVINO™
imfing / Keras Flask Deploy Webapp:smiley_cat: Pretty & simple image classifier app template. Deploy your own trained model or pre-trained model (VGG, ResNet, Densenet) to a web app using Flask in 10 minutes.
yeyupiaoling / Whisper FinetuneFine-tune the Whisper speech recognition model to support training without timestamp data, training with timestamp data, and training without speech data. Accelerate inference and support Web deployment, Windows desktop deployment, and Android deployment
rocketride-org / Rocketride ServerHigh-performance AI pipeline engine with a C++ core and 50+ Python-extensible nodes. Build, debug, and scale LLM workflows with 13+ model providers, 8+ vector databases, and agent orchestration, all from your IDE. Includes VS Code extension, TypeScript/Python SDKs, and Docker deployment.
tensorflow / Tfjs NodeTensorFlow powered JavaScript library for training and deploying ML models on Node.js.
qualcomm / AI Hub ModelsQualcomm® AI Hub Models is our collection of state-of-the-art machine learning models optimized for performance (latency, memory etc.) and ready to deploy on Qualcomm® devices.