Livebot
LiveBot: Generating Live Video Comments Based on Visual and Textual Contexts (AAAI 2019)
Install / Use
/learn @lancopku/LivebotREADME
LiveBot
This is the codes and datasets for the papers: LiveBot: Generating Live Video Comments Based on Visual and Textual Contexts.
What is Live Video Comments?
Live video commenting, which is also called ''video barrage'' (''弹幕'' in Chinese or ''Danmaku'' in Japanese), is an emerging feature on online video sites that allows real-time comments from viewers to fly across the screen like bullets or roll at the right side of the screen.
Requirements
- Ubuntu 16.0.4
- Python 3.5
- Pytorch 0.4.1
- Sklearn >= 0.19.1
Datasets
-
Processed dataset can be directly used for our codes to reproduce the results reported in the paper. It should be downloaded from Google Drive or Baidu Pan, and put in the folder /data.
-
Raw dataset consists of the videos and the corresponding live comments that directly downloaded from the Bilibili video websites. It can be found at Google Drive or Baidu Pan. After processed with the scripts in the folder /data, it can be transformed into the processed datasets above.
Livebot Model
-
Step 1: Download the processed dataset above
-
Step 2: Train a model
python3 codes/transformer.py -mode train -dir CKPT_DIR -
Step 3: Restore the checkpoint and evaluate the model
python3 codes/transformer.py -mode test -restore CKPT_DIR/checkpoint.pt -dir CKPT_DIR
Process a raw dataset (Optional)
- Step 1: Extract the frames from the videos and the comments from the .ass files.
python3 data/extract.py - Step 2: Convert the extracted images and text into the format required by our model.
python3 data/preprocess.py - Step 3: Construct the candidate set for the evaluation of the model.
python3 data/add_candidate.py
Note
-
More details regarding the model and the dataset can be found in our paper.
-
The code is currently non-deterministic due to various GPU ops, so you are likely to end up with a slightly better or worse evaluation.
Citation
Hopefully the codes and the datasets are useful for the future research. If you use the above codes or datasets for your research, please kindly cite our paper:
@inproceedings{livebot,
author = {Shuming Ma and
Lei Cui and
Damai Dai and
Furu Wei and
Xu Sun},
title = {LiveBot: Generating Live Video Comments Based on Visual and Textual Contexts},
booktitle = {{AAAI} 2019},
year = {2019}
}
Related Skills
qqbot-channel
353.1kQQ 频道管理技能。查询频道列表、子频道、成员、发帖、公告、日程等操作。使用 qqbot_channel_api 工具代理 QQ 开放平台 HTTP 接口,自动处理 Token 鉴权。当用户需要查看频道、管理子频道、查询成员、发布帖子/公告/日程时使用。
docs-writer
100.7k`docs-writer` skill instructions As an expert technical writer and editor for the Gemini CLI project, you produce accurate, clear, and consistent documentation. When asked to write, edit, or revie
model-usage
353.1kUse CodexBar CLI local cost usage to summarize per-model usage for Codex or Claude, including the current (most recent) model or a full model breakdown. Trigger when asked for model-level usage/cost data from codexbar, or when you need a scriptable per-model summary from codexbar cost JSON.
arscontexta
3.1kClaude Code plugin that generates individualized knowledge systems from conversation. You describe how you think and work, have a conversation and get a complete second brain as markdown files you own.
