319 skills found · Page 1 of 11
deepspeedai / DeepSpeedDeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
gvergnaud / Ts Pattern🎨 The exhaustive Pattern Matching library for TypeScript, with smart type inference.
mistralai / Mistral InferenceOfficial inference library for Mistral models
py-why / DowhyDoWhy is a Python library for causal inference that supports explicit modeling and testing of causal assumptions. DoWhy is based on a unified language for causal inference, combining causal graphical models and potential outcomes frameworks.
Trusted-AI / Adversarial Robustness ToolboxAdversarial Robustness Toolbox (ART) - Python Library for Machine Learning Security - Evasion, Poisoning, Extraction, Inference - Red and Blue Teams
NVIDIA / DALIA GPU-accelerated library containing highly optimized building blocks and an execution engine for data processing to accelerate deep learning training and inference applications.
open-edge-platform / AnomalibAn anomaly detection library comprising state-of-the-art algorithms and features such as experiment management, hyper-parameter optimization, and edge inference.
huggingface / Parler TtsInference and training library for high-quality TTS models.
turboderp-org / Exllamav2A fast inference library for running LLMs locally on modern consumer-class GPUs
zjhellofss / KuiperInfer校招、秋招、春招、实习好项目!带你从零实现一个高性能的深度学习推理库,支持大模型 llama2 、Unet、Yolov5、Resnet等模型的推理。Implement a high-performance deep learning inference library step by step
alexzhang13 / RlmGeneral plug-and-play inference library for Recursive Language Models (RLMs), supporting various sandboxes.
NVIDIA / TransformerEngineA library for accelerating Transformer models on NVIDIA GPUs, including using 8-bit and 4-bit floating point (FP8 and FP4) precision on Hopper, Ada and Blackwell GPUs, to provide better performance with lower memory utilization in both training and inference.
NVIDIA / Model OptimizerA unified library of SOTA model optimization techniques like quantization, pruning, distillation, speculative decoding, etc. It compresses deep learning models for downstream deployment frameworks like TensorRT-LLM, TensorRT, vLLM, etc. to optimize inference speed.
vitoplantamura / OnnxStreamLightweight inference library for ONNX files, written in C++. It can run Stable Diffusion XL 1.0 on a RPI Zero 2 (or in 298MB of RAM) but also Mistral 7B on desktops and servers. ARM, x86, WASM, RISC-V supported. Accelerated by XNNPACK. Python, C# and JS(WASM) bindings available.
uTensor / UTensorTinyML AI inference library
sdcb / PaddleSharp.NET/C# binding for Baidu paddle inference library and PaddleOCR
VoltaML / VoltaML⚡VoltaML is a lightweight library to convert and run your ML/DL deep learning models in high performance inference runtimes like TensorRT, TorchScript, ONNX and TVM.
matteocourthoud / Awesome Causal InferenceA curated list of causal inference libraries, resources, and applications.
sipeed / TinyMaixTinyMaix is a tiny inference library for microcontrollers (TinyML).
blackjax-devs / BlackjaxBlackJAX is a Bayesian Inference library designed for ease of use, speed and modularity.