This is the official implementation for VideoTree
Authors: Ziyang Wang*, Shoubin Yu*, Elias Stengel-Eskin*, Jaehong Yoon, Feng Cheng, Gedas Bertasius, Mohit Bansal
We introduce VideoTree, a query-adaptive and hierarchical framework for long-video understanding with LLMs. Specifically, VideoTree dynamically extracts query-related information from the input video and builds a tree-based video representation for LLM reasoning.
Install environment.
Python 3.8 or above is required.
git clone https://github.com/Ziyang412/VideoTree.git
cd VideoTree
python3 -m venv videetree_env
source activate videetree_env/bin/activate
pip install openai
pip3 install torch torchvision torchaudio --index-url https://download.pytorch.org/whl/cu118
pip install pandas
pip install transformers==4.28.1
pip install accelerate
Download dataset annotations and extracted captions.
Download data.zip from the File LLoVi provided.
unzip data.zip
You could extract captions for EgoSchema at ./data
. It also contains dataset annotations.
Specifically, LaViLa base model is leveraged to extract EgoSchema captions at 1 FPS.
Download EgoSchema Videos.
Please follow EgoSchema to download the orginal EgoSchema videos. After downloading, please extract the videos into 1 FPS video frames (save in image format for faster loading speed). Please save in the format of ./data/egoschema_frames/{video_id}/{frame_id}.jpg
. Then, to further speed up the tree building process, we extract the visual features for each frame using EVA-CLIP-8B and save the features in ./data/egoschema_features/{video_id}.pt
.
python data_extraction/extract_images.py
python data_extraction/extract_features.py
Since the orginal Kmeans-pytorch package doesn't set a iteration limit and will cause perpetual loop issue, we update the init file of the original kmeans-pytorch package.
git clone https://github.com/subhadarship/kmeans_pytorch
cd kmeans_pytorch
Please replace the init file in "kmeans_pytorch" folder with the file we provide in "./kmeans_pytorch" folder (this repo). And run the following command.
pip install --editable .
Due to the limit of time, we are still updating the codebase. We will also incorporate the scipts/captions for NeXT-QA and IntentQA in the future.
Please update the feature, asgs (in util.py) and output path before running the code.
sh scripts/breath_expansion.sh
Please update the feature, the output of last step (the relevance output path and first level cluster information) and output path before running the code.
python depth_expansion.py
Please update the tree node index file (output of last step), data files and output path before running the code.
sh scripts/egoschema_qa.sh
--save_info: save more information, e.g. token usage, detailed prompts, etc.
--num_examples_to_run: how many examples to run. -1 (default) to run all.
--start_from_scratch: ignore existing output files. Start from scratch.
We thank the developers of LLoVi, LifelongMemory, EVA-CLIP, Kmeans-pytorch and SKlearn Clustering for their public code release. We also thank the authors of VideoAgent for the helpful discussion.
Please cite our paper if you use our models in your works:
@article{wang2024videotree,
title={VideoTree: Adaptive Tree-based Video Representation for LLM Reasoning on Long Videos},
author={Wang, Ziyang and Yu, Shoubin and Stengel-Eskin, Elias and Yoon, Jaehong and Cheng, Feng and Bertasius, Gedas and Bansal, Mohit},
journal={arXiv preprint arXiv:2405.19209},
year={2024}
}