SoraReview
The official GitHub page for the review paper "Sora: A Review on Background, Technology, Limitations, and Opportunities of Large Vision Models".
Install / Use
/learn @lichao-sun/SoraReviewREADME
Sora: A Review on Background, Technology, Limitations, and Opportunities of Large Vision Models
🔍 See our paper: "Sora: A Review on Background, Technology, Limitations, and Opportunities of Large Vision Models"
🔍 See our newest Video Generation paper: "Mora: Enabling Generalist Video Generation via A Multi-Agent Framework"
![]()
)
📧 Please let us know if you find a mistake or have any suggestions by e-mail: lis221@lehigh.edu
Table of Contents
About
Sora is a text-to-video generative AI model, released by OpenAI in February 2024. The model is trained to generate videos of realistic or imaginative scenes from text instructions and shows potential in simulating the physical world. Based on public technical reports and reverse engineering, this paper presents a comprehensive review of the model's background, related technologies, applications, remaining challenges, and future directions of text-to-video AI models. We first trace Sora's development and investigate the underlying technologies used to build this "world simulator". Then, we describe in detail the applications and potential impact of Sora in multiple industries ranging from filmmaking and education to marketing. We discuss the main challenges and limitations that need to be addressed to widely deploy Sora, such as ensuring safe and unbiased video generation. Lastly, we discuss the future development of Sora and video generation models in general, and how advancements in the field could enable new ways of human-AI interaction, boosting productivity and creativity of video generation.
<div align="center"> <img src="https://raw.githubusercontent.com/lichao-sun/SoraReview/main/image/sora_framework.png" width="85%"></div>Updates
- 📄 [28/02/2024] Our paper has been uploaded to arXiv and was selected as the Daily Paper by Hugging Face.
History of Generative AI in the Vision Domain
<div align="center"> <img src="https://raw.githubusercontent.com/lichao-sun/SoraReview/main/image/history.png" width="85%"></div>Paper List
<div align="center"> <img src="https://raw.githubusercontent.com/lichao-sun/SoraReview/main/image/paper_list_structure.png" width="70%"></div>Technology
Data Pre-processing
- (NeurIPS'23) Patch n’Pack: Navit, A Vision Transformer for Any Aspect Ratio and Resolution [paper]
- (ICLR'21) An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale [paper][code]
- (arXiv 2013.12) Auto-Encoding Variational Bayes [paper]
- (ICCV'21) Vivit: A Video Vision Transformer [paper][code]
- (ICML'21) Is Space-Time Attention All You Need for Video Understanding? [paper][code]
- (NeurIPS'17) Neural Discrete Representation Learning [paper][code]
- (CVPR'22) High-Resolution Image Synthesis with Latent Diffusion Models [paper][code]
Modeling
- (JMLR'22) Cascaded Diffusion Models for High Fidelity Image Generation [paper]
- (ICLR'22) Progressive Distillation for Fast Sampling of Diffusion Models [paper][code]
- Imagen Video: High Definition Video Generation with Diffusion Models [paper]
- (CVPR'23) Align Your Latents: High-Resolution Video Synthesis with Latent Diffusion Models [paper]
- (ICCV'23) Scalable Diffusion Models with Transformers [paper]
- (CVPR'23) All Are Worth Words: A ViT Backbone for Diffusion Models [paper][code]
- (ICCV'23) Masked Diffusion Transformer Is a Strong Image Synthesizer [paper][code]
- (arXiv 2023.12) DiffiT: Diffusion Vision Transformers for Image Generation [paper][code]
- (CVPR'24) GenTron: Delving Deep into Diffusion Transformers for Image and Video Generation [paper]
- (arXiv 2023.09) LAVIE: High-Quality Video Generation with Cascaded Latent Diffusion Models [paper][code]
- (arXiv 2024.01) Latte: Latent Diffusion Transformer for Video Generation [paper][code]
- (arXiv 2024.03) Scaling Rectified Flow Transformers for High-Resolution Image Synthesis [paper]
Language Instruction Following
- Improving Image Generation with Better Captions [paper]
- (arXiv 2022.05) CoCa: Contrastive Captioners are Image-Text Foundation Models [paper][code]
- (arXiv 2022.12) VideoCoCa: Video-Text Modeling with Zero-Shot Transfer from Contrastive Captioners [paper]
- (CVPR'23) InstructPix2Pix: Learning to Follow Image Editing Instructions [paper][code]
- (NeurlPS'23) Visual Instruction Tuning [paper][code]
- (ICML'23) mPLUG-2: A Modularized Multi-modal Foundation Model Across Text, Image, and Video [paper][code]
- (arXiv 2022.05) GIT: A Generative Image-to-text Transformer for Vision and Language [paper][code]
- (CVPR'23) Vid2Seq: Large-Scale Pretraining of a Visual Language Model for Dense Video Captioning [paper][code]
Prompt Engineering
- (arXiv 2023.10) Unleashing the Potential of Prompt Engineering in Large Language Models: A Comprehensive Review [paper]
- (arXiv 2023.04) Boosted Prompt Ensembles for Large Language Models [paper][code]
- (NeurIPS'23) Optimizing Prompts for Text-to-Image Generation [paper][code]
- (CVPR'23) VoP: Text-Video Co-operative Prompt Tuning for Cross-Modal Retrieval [paper][code]
- (ICCV'23) Tune-a-Video: One-Shot Tuning of Image Diffusion Models for Text-to-Video Generation [paper][code]
- (CVPR'22) Image Segmentation Using Text and Image Prompts [paper][code]
- (*ACM
Security Score
Audited on Jan 15, 2026
