Pallaidium
PALLAIDIUM — a generative AI movie studio, seamlessly integrated into the Blender Video Editor (VSE), enabling end-to-end production from script to screen and back.
Install / Use
/learn @tin2tin/PallaidiumREADME
Workflow examples with Pallaidium, Blender Screenwriter, and GPT4Blender
https://github.com/user-attachments/assets/81d30bc1-01f6-4b52-8ce7-abf53d53e854
AI-generate video, image, and audio from text prompts or video, image, or text strips.
Generation Matrix
This matrix provides a quick overview of the core generative capabilities, mapping input types to possible outputs.
| Input | Image | Video | Text | Audio | Music | Speech | | :--- | :---: | :---: | :---: | :---: | :---: | :---: | | Text | ✅ | ✅ | ✅ | ✅ | ✅ | ✅ | | Image| ✅ | ✅ | ✅ | ✅ | | | | Video| ✅ | ✅ | ✅ | ✅ | | |
Features
| | | | |----------------------------------------------------|-----------------------------------------------------|-----------------------------------------------------| | Text to video | Text to image | Text to text | | Text to speech | Text to audio | Text to music | | Image to image | Image to video | Image to text | | Video to video | Video to Image | Video to text | | ControlNet | OpenPose | Canny | | ADetailer | IP Adapter Face | IP Adapter Style | | Multiple LoRAs | LoRA Weight | Style selector | | Seed | Quality steps | Strip power | | Frames (Duration) | Word power | Model card selector | | Batch conversion | Batch refinement of images. | Prompt batching | | Batch upscale & refinement of movies. | Render-to-path selector. | Render finished notification. | | User-defined file path for generated files. | Seed and prompt added to strip name. | One-click install and uninstall dependencies. |
Requirements
- Windows. Limited support for Linux and MacOS
- Blender 4.5.3
- A CUDA-supported Nvidia card with at least 6 GB VRAM or MPS.
- CUDA: 12.4
- 20+ GB HDD. (Each model is 6+ GB).
For Mac and Linux, we'll have to rely on contributor support. So, post your issues here for Mac: https://github.com/tin2tin/Pallaidium/issues/106 and here for Linux: https://github.com/tin2tin/Pallaidium/issues/105, and hope some contributor wants to help you out.
How to install
-
First, download and install git (must be on PATH): https://git-scm.com/downloads
-
Download Blender 4.5.X (not 5.0) https://builder.blender.org/download/daily/ and unzip it into the Documents folder.
-
Download the add-on: https://github.com/tin2tin/text_to_video/archive/refs/heads/main.zip
-
On Windows, right-click on the Blender(blender.exe) icon and "Run Blender as Administrator"(or you'll get write permission errors).
-
Install the add-on as usual: Preferences > Add-ons > Install > select file > enable the add-on.
-
In the Generative AI add-on preferences, hit the "Uninstall Dependencies" button (to clear out any incompatible libs).
-
Restart Blender via "Run as Administrator".
-
In the Generative AI add-on preferences, hit the "Install Dependencies" button.
-
Restart the computer and run Blender via "Run as Administrator".
-
Open the add-on UI in the Sequencer > Sidebar > Generative AI.
-
5-10 GB must be downloaded first the first time any model is executed.
-
When you have Pallaidium installed, reach out on Discord: https://discord.gg/HMYpnPzbTm or leave a note on how it is working for you. It means the world to me to know someone is using it!
If any Python modules are missing, use this add-on to install them manually: | :------------- | https://github.com/tin2tin/blender_pip | If "WARNING: Failed to find MSVC", install "Tools for Visual Studio": https://aka.ms/vs/17/release/vs_BuildTools.exe If error: "Missing DLL", install Microsoft Visual C++ Redistributable: https://learn.microsoft.com/en-us/cpp/windows/latest-supported-vc-redist?view=msvc-170
Uninstall
Hugging Face Diffusers models are downloaded from the hub and saved to a local cache directory. Delete the folder manually:
On Linux and macOS: ~/.cache/huggingface/hub
On Windows: %userprofile%\.cache\huggingface\hub
Usage
The addon panel is located in the Video Sequence Editor > Sidebar (N-Panel) > Generative AI tab.
Basic Workflow
- Choose Output: In the "Output" section of the panel, select whether you want to generate an
Image,Video,Audio, orText. - Select Model: Based on your output choice, select a specific AI model from the dropdown list.
- Set Input:
- For Prompts: Leave the
Inputdropdown onPrompts. - For Strips: Select one or more strips in the VSE timeline and set the
Inputdropdown toStrips.
- For Prompts: Leave the
- Configure: Enter your text prompts and adjust parameters like resolution, quality steps, and guidance scale.
- Generate:
- If using
Prompts, click theGeneratebutton. - If using
Strips, click theGenerate from Stripsbutton.
- If using
The generated media will be saved to the directory specified in the addon preferences and automatically added to your VSE timeline on a new channel.
Change Log
2026-02-26: Add: Ltx-2 Multi Input - Custom LoRA + detailer LoRA in last step.
2026-02-22: Add: LTX-2 Video as Distilled or Multi-Input (via. Meta Strips)
2026-01-23: Add: Support for BFL-FLUX.2 Klein 4B and 9B
2025-11-28: Add: FLUX2 & Z-Image & Optimized model-loading
2025-10-27: Remove 30 dead models.
2025-10-1: Fix: Deps. Add: Qwen Multi-image Edit.
2025-7-05: Add: FLUX Kontext Relight.
<details> <summary><strong>Older Changes</strong></summary>2025-6-26: Add: FLUX.1 Dev Kontext. Update Diffusers by installing this: "git+https://github.com/huggingface/diffusers.git" with the Python Module Manager add-on (link below), and restart Blender.
2025-6-22: Add: Long string parsing for Chatterbox (for Audiobooks). Use Blender 5.0 Alpha.
2025-6-14: Add: Chroma txt2vid
2025-6-12: Add: ZuluVision/MoviiGen1.1_Prompt_Rewriter
2025-6-01: Add: Chatterbox with zero shot text to speech and speech to speech voice cloning.
2025_5_29: Add: Flux depth+canny+redux Fix: neg Flux input + py libs
2025-5-12: Add: F5-TTS Voice Cloning
2025-5-09: Update: LTX 0.9.7 w. img, txt & vid input
2025-5-08: FramePack (Img/+Last Img), Img+txt for MMAudio, Flux De-distilled
2025-4-11: Improved LTX 0.95, MetaData, StripPicker, Wan t2i
2025-2-25: Add: MMAudio for Video to Sync Audio
2025-2-21: Support for Skywork/SkyReels-V1-Hunyuan-T2V/I2V. Need a full update of dependencies! (Thx newgenai79 for int4 transformer)
2025-2-15: Add: LoRA support for HunyuanVideo + better preset
2025-2-12: Add multi-media prompting via: OmniGen
2025-2-10: Update: a-r-r-o-w/LTX-Video-0.9.1-diffusers ZhengPeng7/BiRefNet_HR MiaoshouAI/Florence-2-large-PromptGen-v2.0 New: ostris/Flex.1-alpha Alpha-VLLM/Lumina-Image-2.0 Efficient-Large-Model/Sana_1600M_1024px_diffusers Fix: Frame by frame (SD XL) Remove: Corcelio/mobius
2025-1-26: Add: MiniMax Cloud txt/img/subject to video (insert your MiniMax API key in MiniMax_API.txt) and fast FLUX LoRA
2025-1-15: FLUX: faster img2img and inpaint
2024-11-2: Add: Image Background Removal, Stable Diffusion 3.5 Medium, Fast Flux(t2i)
2024-9-19: Add: Image to Video for CogVideoX
2024-9-15: Add: LoRA import for Flux
2024-9-14: Add: Flux Inpaint & Img2img.
2024-9-4: Add: Florence 2 (Image Caption), AudioLDM2-Large, CogVideox-2b, flash_attn on Win.
2024-9-2: Add: Vid2vid for CogVideoX-5b and Parler TTS
2024-8-28: Make CogVideox-5b run on 6 GB VRAM & Flux on 2 GB VRAM
2024-8-27: Add: CogVideoX-5b Remove: Low-quality models
2024-8-5: Add: Flux Dev - NB. needs update of dependencies and 24 GB VRAM
2024-8-2: Add: Flux Schnell - NB. needs update of dependencies and 24 GB VRAM
2024-7-12: Add: Kwai/Kolors (txt2img & img2img)
2024-6-13: Add: SD3 - A "Read" token from HuggingFace must be entered, it's free (img2img). Fix: Installation of Dependencies
2024-6-6: Add: Stable Audio Open, Frame:-1 will inherit duration.
2024-6-1: IP Adapter(When using SDXL): Face (Image or folder), Style (image or folder) New image models: Mobius, OpenVision, Juggernaut X Hyper
2024-4-29: Add: PixArt Sigma 2k, PixArt 10
