SkillAgentSearch skills...

Pallaidium

PALLAIDIUM — a generative AI movie studio, seamlessly integrated into the Blender Video Editor (VSE), enabling end-to-end production from script to screen and back.

Install / Use

/learn @tin2tin/Pallaidium

README

<h1 align="center">PALLAIDIUM</h1> <p align="center">A free generative AI movie studio integrated into the Blender Video Editor.</p> <p align="center"> <a href="https://discord.gg/HMYpnPzbTm"><img src="https://img.shields.io/badge/Chat%20with%20us%20on%20Discord--blue?style=social&logo=discord" alt="Chat with us" title="Chat with us"></a> <a href="https://twitter.com/tintwotin"><img src="https://img.shields.io/twitter/follow/tintwotin" alt="Follow us on X" title="Follow us on X"></a> <p> <hr>

Workflow examples with Pallaidium, Blender Screenwriter, and GPT4Blender

https://github.com/user-attachments/assets/81d30bc1-01f6-4b52-8ce7-abf53d53e854

AI-generate video, image, and audio from text prompts or video, image, or text strips.

Generation Matrix

This matrix provides a quick overview of the core generative capabilities, mapping input types to possible outputs.

| Input | Image | Video | Text | Audio | Music | Speech | | :--- | :---: | :---: | :---: | :---: | :---: | :---: | | Text | ✅ | ✅ | ✅ | ✅ | ✅ | ✅ | | Image| ✅ | ✅ | ✅ | ✅ | | | | Video| ✅ | ✅ | ✅ | ✅ | | |

Features

| | | | |----------------------------------------------------|-----------------------------------------------------|-----------------------------------------------------| | Text to video | Text to image | Text to text | | Text to speech | Text to audio | Text to music | | Image to image | Image to video | Image to text | | Video to video | Video to Image | Video to text | | ControlNet | OpenPose | Canny | | ADetailer | IP Adapter Face | IP Adapter Style | | Multiple LoRAs | LoRA Weight | Style selector | | Seed | Quality steps | Strip power | | Frames (Duration) | Word power | Model card selector | | Batch conversion | Batch refinement of images. | Prompt batching | | Batch upscale & refinement of movies. | Render-to-path selector. | Render finished notification. | | User-defined file path for generated files. | Seed and prompt added to strip name. | One-click install and uninstall dependencies. |

Requirements

  • Windows. Limited support for Linux and MacOS
  • Blender 4.5.3
  • A CUDA-supported Nvidia card with at least 6 GB VRAM or MPS.
  • CUDA: 12.4
  • 20+ GB HDD. (Each model is 6+ GB).

For Mac and Linux, we'll have to rely on contributor support. So, post your issues here for Mac: https://github.com/tin2tin/Pallaidium/issues/106 and here for Linux: https://github.com/tin2tin/Pallaidium/issues/105, and hope some contributor wants to help you out.

How to install

  • First, download and install git (must be on PATH): https://git-scm.com/downloads

  • Download Blender 4.5.X (not 5.0) https://builder.blender.org/download/daily/ and unzip it into the Documents folder.

  • Download the add-on: https://github.com/tin2tin/text_to_video/archive/refs/heads/main.zip

  • On Windows, right-click on the Blender(blender.exe) icon and "Run Blender as Administrator"(or you'll get write permission errors).

  • Install the add-on as usual: Preferences > Add-ons > Install > select file > enable the add-on.

  • In the Generative AI add-on preferences, hit the "Uninstall Dependencies" button (to clear out any incompatible libs).

  • Restart Blender via "Run as Administrator".

  • In the Generative AI add-on preferences, hit the "Install Dependencies" button.

  • Restart the computer and run Blender via "Run as Administrator".

  • Open the add-on UI in the Sequencer > Sidebar > Generative AI.

  • 5-10 GB must be downloaded first the first time any model is executed.

  • When you have Pallaidium installed, reach out on Discord: https://discord.gg/HMYpnPzbTm or leave a note on how it is working for you. It means the world to me to know someone is using it!

If any Python modules are missing, use this add-on to install them manually: | :------------- | https://github.com/tin2tin/blender_pip | If "WARNING: Failed to find MSVC", install "Tools for Visual Studio": https://aka.ms/vs/17/release/vs_BuildTools.exe If error: "Missing DLL", install Microsoft Visual C++ Redistributable: https://learn.microsoft.com/en-us/cpp/windows/latest-supported-vc-redist?view=msvc-170

Uninstall

Hugging Face Diffusers models are downloaded from the hub and saved to a local cache directory. Delete the folder manually:

On Linux and macOS: ~/.cache/huggingface/hub

On Windows: %userprofile%\.cache\huggingface\hub

Usage

The addon panel is located in the Video Sequence Editor > Sidebar (N-Panel) > Generative AI tab.

Basic Workflow

  1. Choose Output: In the "Output" section of the panel, select whether you want to generate an Image, Video, Audio, or Text.
  2. Select Model: Based on your output choice, select a specific AI model from the dropdown list.
  3. Set Input:
    • For Prompts: Leave the Input dropdown on Prompts.
    • For Strips: Select one or more strips in the VSE timeline and set the Input dropdown to Strips.
  4. Configure: Enter your text prompts and adjust parameters like resolution, quality steps, and guidance scale.
  5. Generate:
    • If using Prompts, click the Generate button.
    • If using Strips, click the Generate from Strips button.

The generated media will be saved to the directory specified in the addon preferences and automatically added to your VSE timeline on a new channel.

Change Log

2026-02-26: Add: Ltx-2 Multi Input - Custom LoRA + detailer LoRA in last step.

2026-02-22: Add: LTX-2 Video as Distilled or Multi-Input (via. Meta Strips)

2026-01-23: Add: Support for BFL-FLUX.2 Klein 4B and 9B

2025-11-28: Add: FLUX2 & Z-Image & Optimized model-loading

2025-10-27: Remove 30 dead models.

2025-10-1: Fix: Deps. Add: Qwen Multi-image Edit.

2025-7-05: Add: FLUX Kontext Relight.

<details> <summary><strong>Older Changes</strong></summary>

2025-6-26: Add: FLUX.1 Dev Kontext. Update Diffusers by installing this: "git+https://github.com/huggingface/diffusers.git" with the Python Module Manager add-on (link below), and restart Blender.

2025-6-22: Add: Long string parsing for Chatterbox (for Audiobooks). Use Blender 5.0 Alpha.

2025-6-14: Add: Chroma txt2vid

2025-6-12: Add: ZuluVision/MoviiGen1.1_Prompt_Rewriter

2025-6-01: Add: Chatterbox with zero shot text to speech and speech to speech voice cloning.

2025_5_29: Add: Flux depth+canny+redux Fix: neg Flux input + py libs

2025-5-12: Add: F5-TTS Voice Cloning

2025-5-09: Update: LTX 0.9.7 w. img, txt & vid input

2025-5-08: FramePack (Img/+Last Img), Img+txt for MMAudio, Flux De-distilled

2025-4-11: Improved LTX 0.95, MetaData, StripPicker, Wan t2i

2025-2-25: Add: MMAudio for Video to Sync Audio

2025-2-21: Support for Skywork/SkyReels-V1-Hunyuan-T2V/I2V. Need a full update of dependencies! (Thx newgenai79 for int4 transformer)

2025-2-15: Add: LoRA support for HunyuanVideo + better preset

2025-2-12: Add multi-media prompting via: OmniGen

2025-2-10: Update: a-r-r-o-w/LTX-Video-0.9.1-diffusers ZhengPeng7/BiRefNet_HR MiaoshouAI/Florence-2-large-PromptGen-v2.0 New: ostris/Flex.1-alpha Alpha-VLLM/Lumina-Image-2.0 Efficient-Large-Model/Sana_1600M_1024px_diffusers Fix: Frame by frame (SD XL) Remove: Corcelio/mobius

2025-1-26: Add: MiniMax Cloud txt/img/subject to video (insert your MiniMax API key in MiniMax_API.txt) and fast FLUX LoRA

2025-1-15: FLUX: faster img2img and inpaint

2024-11-2: Add: Image Background Removal, Stable Diffusion 3.5 Medium, Fast Flux(t2i)

2024-9-19: Add: Image to Video for CogVideoX

2024-9-15: Add: LoRA import for Flux

2024-9-14: Add: Flux Inpaint & Img2img.

2024-9-4: Add: Florence 2 (Image Caption), AudioLDM2-Large, CogVideox-2b, flash_attn on Win.

2024-9-2: Add: Vid2vid for CogVideoX-5b and Parler TTS

2024-8-28: Make CogVideox-5b run on 6 GB VRAM & Flux on 2 GB VRAM

2024-8-27: Add: CogVideoX-5b Remove: Low-quality models

2024-8-5: Add: Flux Dev - NB. needs update of dependencies and 24 GB VRAM

2024-8-2: Add: Flux Schnell - NB. needs update of dependencies and 24 GB VRAM

2024-7-12: Add: Kwai/Kolors (txt2img & img2img)

2024-6-13: Add: SD3 - A "Read" token from HuggingFace must be entered, it's free (img2img). Fix: Installation of Dependencies

2024-6-6: Add: Stable Audio Open, Frame:-1 will inherit duration.

2024-6-1: IP Adapter(When using SDXL): Face (Image or folder), Style (image or folder) New image models: Mobius, OpenVision, Juggernaut X Hyper

2024-4-29: Add: PixArt Sigma 2k, PixArt 10

View on GitHub
GitHub Stars1.4k
CategoryContent
Updated6h ago
Forks123

Languages

Python

Security Score

100/100

Audited on Mar 24, 2026

No findings