MVIMP
Mixed Video and Image Manipulation Program
Install / Use
/learn @CyFeng16/MVIMPREADME
Welcome to MVIMP 👋
The name MVIMP (Mixed Video and Image Manipulation Program) was inspired by the name GIMP (GNU Image Manipulation Program), which hope it can help more people.
I realize that training a good-performance AI model is kind of just one side of the story, make it easy to use for others is the other thing. Thus, this repository built to embrace out-of-the-box AI ability to manipulate multimedia. Last but not least, wish you have fun!
| Model | Input | Output | Parallel | Colab Link |
|:------------------------------------------------------:|:------:|:------:|:----------------------:|:---------------------------------------------------------------------------------------------------------------------:|
| AnimeGAN | Images | Images | True | |
| AnimeGANv2 | Images | Images | True |
|
| DAIN | Video | Video | False |
|
| DeOldify | Images | Images | True |
|
| Photo3D | Images | Videos | True(not recommmended) |
|
| Waifu2x | Images | Images | True |
|
You are welcomed to discuss future features in this issue.
AnimeGANv2

Original repository: TachibanaYoshino/AnimeGANv2
The improved version of AnimeGAN, which converts landscape photos/videos(todo) to anime. The improvement directions of AnimeGANv2 mainly include the following 4 points:
- Solve the problem of high-frequency artifacts in the generated image.
- It is easy to train and directly achieve the effects in the paper.
- Further, reduce the number of parameters of the generator network. (generator size: 8.17 Mb), The lite version has a smaller generator model.
- Use new high-quality style data, which come from BD movies as much as possible.
| Dependency | Version | |:------------:|:----------------------------------:| | TensorFLow | 1.15.2 | | CUDA Toolkit | 10.0(tested locally) / 10.1(colab) | | Python | 3.6.8(3.6+) |
Usage:
-
ColabYou can open our jupyter notebook through colab link.
-
Local# Step 1: Prepare git clone https://github.com/CyFeng16/MVIMP.git cd MVIMP python3 preparation.py # Step 2: Put your photos into ./Data/Input/ # Step 3: Infernece python3 inference_animeganv2.py -s {The_Style_You_Choose} -
Description of Parameters
| params | abbr. | Default | Description | |- |- |- |- | | --style | -s | Hayao | The anime style you want to get. |
| Style name | Anime style | |- |- | | Hayao | Miyazaki Hayao | | Shinkai | Makoto Shinkai | | Paprika | Kon Satoshi |
AnimeGAN

Original repository: TachibanaYoshino/AnimeGAN
This is the Open source of the paper <AnimeGAN: a novel lightweight GAN for photo animation>, which uses the GAN framwork to transform real-world photos into anime images.
| Dependency | Version | |:------------:|:----------------------------------:| | TensorFLow | 1.15.2 | | CUDA Toolkit | 10.0(tested locally) / 10.1(colab) | | Python | 3.6.8(3.6+) |
Usage:
-
ColabYou can open our jupyter notebook through colab link.
-
Local# Step 1: Prepare git clone https://github.com/CyFeng16/MVIMP.git cd MVIMP python3 preparation.py -f animegan # Step 2: Put your photos into ./Data/Input/ # Step 3: Infernece python3 inference_animegan.py
DAIN

Original repository: baowenbo/DAIN
Depth-Aware video frame INterpolation (DAIN) model explicitly detect the occlusion by exploring the depth cue. We develop a depth-aware flow projection layer to synthesize intermediate flows that preferably sample closer objects than farther ones.
This method achieves SOTA performance on the Middlebury dataset. Video are provided here.
The current version of DAIN (in this repo) can smoothly run 1080p video frame insertion even on GTX-1080 GPU card, as long as you turn -hr on (see Description of Parameters below).
| Dependency | Version | |:------------:|:-----------------------------------------------------:| | PyTroch | 1.0.0 | | CUDA Toolkit | 9.0(colab tested) | | Python | 3.6.8(3.6+) | | GCC | 4.9(Compiling PyTorch 1.0.0 extension files (.c/.cu)) |
P.S. Make sure your virtual env has torch-1.0.0 and torchvision-0.2.1 with CUDA-9.0 . ~~You can use the following command:~~ You can find out dependencies issue at #5 and #16 .
Usage:
-
ColabYou can open our jupyter notebook through colab link.
-
Local# Step 1: Prepare git clone https://github.com/CyFeng16/MVIMP.git cd MVIMP python3 preparation.py -f dain # Step 2: Put a single video file into ./Data/Input/ # Step 3: Infernece python3 inference_dain.py -input your_input.mp4 -ts 0.5 -hr -
Description of Parameters
| params | abbr. | Default | Description | |-------------------|--------|------------|---------------------------------------------------------------------------------------------------------------------------------------------| | --input_video | -input | / | The input video name. | | --time_step | -ts | 0.5 | Set the frame multiplier.<br>0.5 corresponds to 2X;<br>0.25 corresponds to 4X;<br>0.125 corresponds to 8X. | | --high_resolution | -hr | store_true | Default is False(action:store_true).<br>Turn it on when you handling FHD videos,<br>A frame-splitting process will reduce GPU memory usage. |
DeOldify

Original repository: jantic/DeOldify
DeOldify is a Deep Learning based project for colorizing and restoring old images and video!
We are now integrating the inference capabilities of the DeOldify model (both Artistic and Stable, no Video) with our MVIMP repository, and keeping the input and output interfaces consistent.
| Dependency | Version | |:------------:|:--------------------------:| | PyTroch | 1.5.0 | | CUDA Toolkit | 10.1(tested locally/colab) | | Python | 3.6.8(3.6+) |
Other Python dependencies listed in colab_requirements.txt, and will be auto installed while running preparation.py.
Usage:
-
ColabYou can open our jupyter notebook through colab link.
-
Local# Step 1: Prepare git clone https://github.com/CyFeng16/MVIMP.git cd MVIMP python3 preparation.py -f deoldify # Step 2: Infernece python3 -W ignore inference_deoldify.py -art
Related Skills
qqbot-channel
345.9kQQ 频道管理技能。查询频道列表、子频道、成员、发帖、公告、日程等操作。使用 qqbot_channel_api 工具代理 QQ 开放平台 HTTP 接口,自动处理 Token 鉴权。当用户需要查看频道、管理子频道、查询成员、发布帖子/公告/日程时使用。
docs-writer
100.0k`docs-writer` skill instructions As an expert technical writer and editor for the Gemini CLI project, you produce accurate, clear, and consistent documentation. When asked to write, edit, or revie
model-usage
345.9kUse CodexBar CLI local cost usage to summarize per-model usage for Codex or Claude, including the current (most recent) model or a full model breakdown. Trigger when asked for model-level usage/cost data from codexbar, or when you need a scriptable per-model summary from codexbar cost JSON.
arscontexta
2.9kClaude Code plugin that generates individualized knowledge systems from conversation. You describe how you think and work, have a conversation and get a complete second brain as markdown files you own.
