SkillAgentSearch skills...

Commavq

commaVQ is a dataset of compressed driving video

Install / Use

/learn @commaai/Commavq
About this skill

Quality Score

0/100

Supported Platforms

Universal

README

<div align="center"> <h1>commaVQ challenge</h1> <h3> <a href="https://comma.ai/leaderboard">Leaderboard</a> <span> · </span> <a href="https://comma.ai/jobs">comma.ai/jobs</a> <span> · </span> <a href="https://discord.comma.ai">Discord</a> <span> · </span> <a href="https://x.com/comma_ai">X</a> </h3> </div>

| Source Video | Compressed Video | Future Prediction | | --------------- | ---------------- |------------------ | | <video src="https://github.com/commaai/commavq/assets/29985433/91894bf7-592b-4204-b3f2-3e805984045c"> | <video src="https://github.com/commaai/commavq/assets/29985433/3a799ac8-781e-461c-bf14-c15cea42b985"> | <video src="https://github.com/commaai/commavq/assets/29985433/f6f7699b-b6cb-4f9c-80c9-8e00d75fbfae"> |

A world model is a model that can predict the next state of the world given the observed previous states and actions.

World models are essential to training all kinds of intelligent agents, especially self-driving models.

commaVQ contains:

  • encoder/decoder models used to heavily compress driving scenes
  • a world model trained on 3,000,000 minutes of driving videos
  • a dataset of 100,000 minutes of compressed driving videos

Task

Lossless compression challenge: make me smaller! $500 challenge

Losslessly compress 5,000 minutes of driving video "tokens". Go to ./compression/ to start

Prize: highest compression rate on 5,000 minutes of driving video (~915MB) - Challenge ended July, 1st 2024 11:59pm AOE

Submit a single zip file containing the compressed data and a python script to decompress it into its original form using this form. Top solutions are listed on comma's official leaderboard.

<!-- TABLE-START --> <table class="ranked"> <thead> <tr> <th> </th> <th> score </th> <th> name </th> <th> method </th> </tr> </thead> <tbody> <tr> <td> </td> <td> 3.4 </td> <td> <a href="https://github.com/szabolcs-cs"> szabolcs-cs </a> </td> <td> self-compressing neural network </td> </tr> <tr> <td> </td> <td> 3.0 </td> <td> <a href="https://github.com/SAT-oO"> SAT-oO </a> </td> <td> arithmetic coding with GPT </td> </tr> <tr> <td> </td> <td> 2.9 </td> <td> <a href="https://github.com/BradyWynn"> BradyWynn </a> </td> <td> arithmetic coding with GPT </td> </tr> <tr> <td> </td> <td> 2.7 </td> <td> <a href="https://github.com/ylevental"> ylevental </a> </td> <td> arithmetic coding with GPT </td> </tr> <tr> <td> </td> <td> 2.7 </td> <td> <a href="https://github.com/ksd3"> ksd3 </a> </td> <td> arithmetic coding with GPT </td> </tr> <tr> <td> </td> <td> 2.6 </td> <td> <a href="https://github.com/pkourouklidis"> pkourouklidis </a> 👑 </td> <td> arithmetic coding with GPT </td> </tr> <tr> <td> </td> <td> 2.3 </td> <td> anonymous </td> <td> zpaq </td> </tr> <tr> <td> </td> <td> 2.3 </td> <td> <a href="https://github.com/rostislav"> rostislav </a> </td> <td> zpaq </td> </tr> <tr> <td> </td> <td> 2.2 </td> <td> anonymous </td> <td> zpaq </td> </tr> <tr> <td> </td> <td> 2.2 </td> <td> anonymous </td> <td> zpaq </td> </tr> <tr> <td> </td> <td> 2.2 </td> <td> <a href="https://github.com/0x41head"> 0x41head </a> </td> <td> zpaq </td> </tr> <tr> <td> </td> <td> 2.2 </td> <td> <a href="https://github.com/tillinf"> tillinf </a> </td> <td> zpaq </td> </tr> <tr> <td> </td> <td> 2.2 </td> <td> <a href="https://github.com/ylevental"> ylevental </a> </td> <td> zpaq </td> </tr> <tr> <td> </td> <td> 2.2 </td> <td> <a href="https://github.com/nuniesmith"> nuniesmith </a> </td> <td> zpaq </td> </tr> <tr> <td> </td> <td> 1.6 </td> <td> baseline </td> <td> lzma </td> </tr> </tbody> </table> <!-- TABLE-END -->

Overview

A VQ-VAE [1,2] was used to heavily compress each video frame into 128 "tokens" of 10 bits each. Each entry of the dataset is a "segment" of compressed driving video, i.e. 1min of frames at 20 FPS. Each file is of shape 1200x8x16 and saved as int16.

A world model [3] was trained to predict the next token given a context of past tokens. This world model is a Generative Pre-trained Transformer (GPT) [4] trained on 3,000,000 minutes of driving videos following a similar recipe to [5].

Examples

./notebooks/encode.ipynb and ./notebooks/decode.ipynb for an example of how to visualize the dataset using a segment of driving video from comma's drive to Taco Bell

./notebooks/gpt.ipynb for an example of how to use the world model to imagine future frames.

./compression/compress.py for an example of how to compress the tokens using lzma

Download the dataset

  • Using huggingface datasets
import numpy as np
from datasets import load_dataset
# load the first shard
data_files = {'train': ['data-0000.tar.gz']}
ds = load_dataset('commaai/commavq', data_files=data_files)
tokens = np.array(ds['train'][0]['token.npy'])
poses = np.array(ds['train'][0]['pose.npy'])
  • Manually download from huggingface datasets repository: https://huggingface.co/datasets/commaai/commavq

References

[1] Van Den Oord, Aaron, and Oriol Vinyals. "Neural discrete representation learning." Advances in neural information processing systems 30 (2017).

[2] Esser, Patrick, Robin Rombach, and Bjorn Ommer. "Taming transformers for high-resolution image synthesis." Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. 2021.

[3] https://worldmodels.github.io/

[4] Vaswani, Ashish, et al. "Attention is all you need." Advances in neural information processing systems 30 (2017).

[5] Micheli, Vincent, Eloi Alonso, and François Fleuret. "Transformers are Sample-Efficient World Models." The Eleventh International Conference on Learning Representations. 2022.

View on GitHub
GitHub Stars367
CategoryContent
Updated1d ago
Forks70

Languages

Jupyter Notebook

Security Score

95/100

Audited on Apr 5, 2026

No findings