2,180 skills found · Page 1 of 73
huggingface / Transformers🤗 Transformers: the model-definition framework for state-of-the-art machine learning models in text, vision, audio, and multimodal models, for both inference and training.
meta-llama / LlamaInference code for Llama models
facebookresearch / Segment AnythingThe repository provides code for running inference with the SegmentAnything Model (SAM), links for downloading the trained model checkpoints, and example notebooks that show how to use the model.
black-forest-labs / FluxOfficial inference repo for FLUX.1 models
FunAudioLLM / CosyVoiceMulti-lingual large voice generation model, providing inference, training and deployment full-stack ability.
facebookresearch / Sam2The repository provides code for running inference with the Meta Segment Anything Model 2 (SAM 2), links for downloading the trained model checkpoints, and example notebooks that show how to use the model.
meta-llama / Llama CookbookWelcome to the Llama Cookbook! This is your go to guide for Building with Llama: Getting started with Inference, Fine-Tuning, RAG. We also show you how to solve end to end problems using Llama model family and using them on various provider services
meta-llama / CodellamaInference code for CodeLlama models
NVIDIA / TensorRT LLMTensorRT LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and supports state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. TensorRT LLM also contains components to create Python and C++ runtimes that orchestrate the inference execution in a performant way.
bentoml / OpenLLMRun any open-source LLMs, such as DeepSeek and Llama, as OpenAI compatible API endpoint in the cloud.
huggingface / Text Generation InferenceLarge Language Model Text Generation Inference
mistralai / Mistral InferenceOfficial inference library for Mistral models
Const-me / WhisperHigh-performance GPGPU inference of OpenAI's Whisper automatic speech recognition (ASR) model
xorbitsai / InferenceSwap GPT for any LLM by changing a single line of code. Xinference lets you run open-source, speech, and multimodal models on cloud, on-prem, or your laptop — all through one unified, production-ready inference API.
facebookresearch / Sam3The repository provides code for running inference and finetuning with the Meta Segment Anything Model 3 (SAM 3), links for downloading the trained model checkpoints, and example notebooks that show how to use the model.
bentoml / BentoMLThe easiest way to serve AI apps and models - Build Model Inference APIs, Job queues, LLM apps, Multi-model pipelines, and more!
OptimalScale / LMFlowAn Extensible Toolkit for Finetuning and Inference of Large Foundation Models. Large Models for All.
py-why / DowhyDoWhy is a Python library for causal inference that supports explicit modeling and testing of causal assumptions. DoWhy is based on a unified language for causal inference, combining causal graphical models and potential outcomes frameworks.
google / Gemma.cpplightweight, standalone C++ inference engine for Google's Gemma models.
allenai / OLMoModeling, training, eval, and inference code for OLMo