SkillAgentSearch skills...

SDK

AI video generation SDK — JSX for videos. One API for Kling, Flux, ElevenLabs, Sora. Built on Vercel AI SDK.

Install / Use

/learn @vargHQ/SDK

README

<p align="center"> <h1 align="center">varg — AI Video Generation SDK</h1> <p align="center">Create AI videos with JSX. One SDK for Kling, Flux, ElevenLabs, Sora and more. Built on Vercel AI SDK.</p> </p> <p align="center"> <a href="https://www.npmjs.com/package/vargai"><img src="https://img.shields.io/npm/v/vargai.svg" alt="npm version"></a> <a href="https://www.npmjs.com/package/vargai"><img src="https://img.shields.io/npm/dm/vargai.svg" alt="npm downloads"></a> <a href="https://github.com/vargHQ/sdk/stargazers"><img src="https://img.shields.io/github/stars/vargHQ/sdk" alt="GitHub stars"></a> <a href="https://github.com/vargHQ/sdk/blob/main/LICENSE.md"><img src="https://img.shields.io/badge/license-Apache--2.0-blue" alt="License"></a> </p> <p align="center"> <a href="https://docs.varg.ai">Docs</a> &middot; <a href="https://app.varg.ai">Dashboard</a> &middot; <a href="https://docs.varg.ai/quickstart">Quickstart</a> &middot; <a href="https://docs.varg.ai/sdk/models">Models</a> &middot; <a href="https://discord.gg/varg">Discord</a> </p>

varg is an open-source TypeScript SDK for AI video generation. One API key, one gateway — generate images, video, speech, music, lipsync, and captions through varg.* providers. Write videos as JSX components (like React), render locally or in the cloud.

Get started

For AI agents (recommended)

Install the varg skill into Claude Code, Cursor, Windsurf, or any agent that supports skills. Zero code — just prompt.

# 1. Install the varg skill
npx -y skills add vargHQ/skills --all --copy -y

# 2. Set your API key (get one at app.varg.ai)
export VARG_API_KEY=varg_live_xxx

# 3. Create your first video
claude "create a 10-second product video for white sneakers, 9:16, UGC style, with captions and background music"

The agent writes declarative JSX, varg handles AI generation + caching + rendering.

For developers

# Install with bun (recommended)
bun install vargai ai

# Or with npm
npm install vargai ai

# Set up project (auth, skills, hello.tsx, cache dirs)
bunx vargai init

vargai init handles everything: signs you in, installs the agent skill, creates a starter template, and sets up your project structure.

Then render the starter template:

bunx vargai render hello.tsx

Or ask your AI agent to create something from scratch.

How it works

Your prompt / JSX code
        |
   varg gateway (api.varg.ai)
   /     |      \        \
 Kling  Flux  ElevenLabs  Wan ...   (AI providers)
   \     |      /        /
    varg render engine
        |
   output.mp4
  • One API key (VARG_API_KEY) routes to all providers through the varg gateway
  • Declarative JSX — compose videos like React components with <Clip>, <Video>, <Music>, <Captions>
  • Automatic caching — same props = instant cache hit at $0. Re-render without re-generating
  • Local or cloud — render with bunx vargai render locally, or submit via the Cloud Render API

Quick examples

Image to video

import { Render, Clip, Image, Video } from "vargai/react";
import { varg } from "vargai/ai";

const character = Image({
  prompt: "cute kawaii orange cat, round body, big eyes, Pixar style",
  model: varg.imageModel("nano-banana-pro"),
  aspectRatio: "9:16",
});

export default (
  <Render width={1080} height={1920}>
    <Clip duration={5}>
      <Video
        prompt={{ text: "cat waves hello, bounces happily", images: [character] }}
        model={varg.videoModel("kling-v3")}
      />
    </Clip>
  </Render>
);
bunx vargai render hello.tsx

With music and captions

import { Render, Clip, Image, Video, Speech, Captions, Music } from "vargai/react";
import { varg } from "vargai/ai";

const character = Image({
  model: varg.imageModel("nano-banana-pro"),
  prompt: "friendly robot, blue metallic, expressive eyes",
  aspectRatio: "9:16",
});

const voiceover = Speech({
  model: varg.speechModel("eleven_v3"),
  voice: "adam",
  children: "Hello! I'm your AI assistant. Let me show you something cool!",
});

export default (
  <Render width={1080} height={1920}>
    <Music prompt="upbeat electronic, cheerful" model={varg.musicModel()} volume={0.15} />
    <Clip duration={5}>
      <Video
        prompt={{ text: "robot talking, subtle head movements", images: [character] }}
        model={varg.videoModel("kling-v3")}
      />
    </Clip>
    <Captions src={voiceover} style="tiktok" color="#ffffff" withAudio />
  </Render>
);

Talking head with lipsync

import { Render, Clip, Image, Video, Speech, Captions, Music } from "vargai/react";
import { varg } from "vargai/ai";

const voiceover = Speech({
  model: varg.speechModel("eleven_v3"),
  voice: "josh",
  children: "With varg, you can create any videos at scale!",
});

const baseCharacter = Image({
  prompt: "woman, sleek black bob hair, fitted black t-shirt, natural look",
  model: varg.imageModel("nano-banana-pro"),
  aspectRatio: "9:16",
});

const animatedCharacter = Video({
  prompt: {
    text: "woman speaking naturally, subtle head movements, friendly expression",
    images: [baseCharacter],
  },
  model: varg.videoModel("kling-v3"),
});

export default (
  <Render width={1080} height={1920}>
    <Music prompt="modern tech ambient, subtle electronic" model={varg.musicModel()} volume={0.1} />
    <Clip duration={5}>
      <Video
        prompt={{ video: animatedCharacter, audio: voiceover }}
        model={varg.videoModel("sync-v2-pro")}
      />
    </Clip>
    <Captions src={voiceover} style="tiktok" color="#ffffff" withAudio />
  </Render>
);

Components

| Component | Purpose | Key props | |-----------|---------|-----------| | <Render> | Root container | width, height, fps | | <Clip> | Time segment | duration, transition, cutFrom, cutTo | | <Image> | AI or static image | prompt, src, model, zoom, aspectRatio, resize | | <Video> | AI or source video | prompt, src, model, volume, cutFrom, cutTo | | <Speech> | Text-to-speech | voice, model, volume, children | | <Music> | Background music | prompt, src, model, volume, loop, ducking | | <Title> | Text overlay | position, color, start, end | | <Subtitle> | Subtitle text | backgroundColor | | <Captions> | Auto-generated subs | src, srt, style, color, activeColor, withAudio | | <Overlay> | Positioned layer | left, top, width, height, keepAudio | | <Split> | Side-by-side | direction | | <Slider> | Before/after reveal | direction | | <Swipe> | Tinder-style cards | direction, interval | | <TalkingHead> | Animated character | character, src, voice, model, lipsyncModel | | <Packshot> | End card with CTA | background, logo, cta, blinkCta |

Caption styles

<Captions src={voiceover} style="tiktok" />     // word-by-word highlight
<Captions src={voiceover} style="karaoke" />    // fill left-to-right
<Captions src={voiceover} style="bounce" />     // words bounce in
<Captions src={voiceover} style="typewriter" /> // typing effect

Transitions

67 GL transitions available:

<Clip transition={{ name: "fade", duration: 0.5 }}>
<Clip transition={{ name: "crossfade", duration: 0.5 }}>
<Clip transition={{ name: "wipeleft", duration: 0.5 }}>
<Clip transition={{ name: "cube", duration: 0.8 }}>

Models

All models are accessed through varg.* — one API key, one provider.

import { varg } from "vargai/ai";

Video

| Model | Use case | Credits (5s) | |-------|----------|-------------| | varg.videoModel("kling-v3") | Best quality, latest | 150 | | varg.videoModel("kling-v3-standard") | Good quality, cheaper | 50 | | varg.videoModel("kling-v2.5") | Previous gen, reliable | 50 | | varg.videoModel("wan-2.5") | Good for characters | 50 | | varg.videoModel("minimax") | Alternative | 50 | | varg.videoModel("sync-v2-pro") | Lipsync (video + audio) | 50 |

Image

| Model | Use case | Credits | |-------|----------|---------| | varg.imageModel("nano-banana-pro") | Versatile, fast | 5 | | varg.imageModel("nano-banana-pro/edit") | Image-to-image editing | 5 | | varg.imageModel("flux-schnell") | Fast generation | 5 | | varg.imageModel("flux-pro") | High quality | 25 | | varg.imageModel("recraft-v3") | Alternative | 10 |

Audio

| Model | Use case | Credits | |-------|----------|---------| | varg.speechModel("eleven_v3") | Text-to-speech | 25 | | varg.speechModel("eleven_multilingual_v2") | Multilingual TTS | 25 | | varg.musicModel() | Music generation | 25 | | varg.transcriptionModel("whisper") | Speech-to-text | 5 |

1 credit = $0.01. Cache hits are always free.

CLI

bunx vargai login                              # sign in (email OTP or API key)
bunx vargai init                               # set up project (auth + skills + template)
bunx vargai render video.tsx                   # render a video
bunx vargai render video.tsx --preview         # free preview with placeholders
bunx vargai render video.tsx --verbose         # render with detailed output
bunx vargai balance                            # check credit balance
bunx vargai topup                              # add credits
bunx vargai run image --prompt "sunset"        # generate a single image
bunx vargai run video --prompt "ocean waves"   # generate a single video
bunx vargai list                               # list available models and actions
bunx vargai studio                             # open visual editor

Environment

# Required — one key for everything
VARG_API_KEY=varg_live_xxx

Get your API key at app.varg.ai. Bun auto-loads .env files.

<details> <summary>Bring your own keys (optional)</summary>

You can use provider keys directly if you prefer:

FAL_API_KEY=fal_xxx                # fal.ai direct
ELEVENLABS_API_KEY=xxx             # Elev
View on GitHub
GitHub Stars249
CategoryDevelopment
Updated1h ago
Forks15

Languages

TypeScript

Security Score

100/100

Audited on Mar 31, 2026

No findings