SkillAgentSearch skills...

DreamDance

DreamDance: Personalized Text-to-video Generation by Combining Text-to-Image Synthesis and Motion Transfer

Install / Use

/learn @robo-alex/DreamDance
About this skill

Quality Score

0/100

Supported Platforms

Zed

README

DreamDance: Personalized Text-to-video Generation by Combining Text-to-Image Synthesis and Motion Transfer

Results of Pipeline 1

dance_1

dance_2

orange_justice_1

orange_justice_2

The motion transfer is quite successful, even if the the character in the reference video performs large motion, like dancing and rotating.

Note that limited by the computing resources, we only generated the imitation videos of low-resolution. The performance of motion imitation is good.

Results of Pipeline 2

Input images of prompt: miguel playing guitar on the street, pixar, cartoon, high quality, full body, single person

input_guitar

Output video

output_guitar

Input images of prompt: miguel running in a forest, pixar, cartoon, green eyes, red hat, high quality, standing, full body, single person

input_running

Output video

output_running

Input images of prompt: miguel in a forest, pixar, cartoon, green eyes, red hat, high quality, standing, full body, single person

input_2

Output video

output_2

Input images with prompt: miguel, pixar, cartoon, playing guitar, high quality, full body, single person

input_guitar_2

Output video

output_guitar_2

We noticed that if the changes are even larger, the interpolation still handled the video synthesis pretty well. Although the are some artifacts in the mid-frames, our limitations are mainly from the input image generation side. If future text-to-image synthesis models have the capability of generating more promising images with high consistency of all the factors above, frame interpolation will be a powerful method of text-to-video generation.

View on GitHub
GitHub Stars14
CategoryContent
Updated2y ago
Forks1

Languages

Jupyter Notebook

Security Score

75/100

Audited on Jan 4, 2024

No findings