VELoRA: A Low-Rank Adaptation Approach for Efficient RGB-Event based Recognition, Lan Chen, Haoxiang Yang, Pengpeng Shao, Haoyu Song, Xiao Wang*, Zhicheng Zhao, Yaowei Wang, Yonghong Tian [Paper]
Pattern recognition leveraging both RGB and Event cameras can significantly enhance performance by deploying deep neural networks that utilize a fine-tuning strategy. Inspired by the successful application of large models, the introduction of such large models can also be considered to further enhance the performance of multi-modal tasks. However, fully fine-tuning these models leads to inefficiency and lightweight fine-tuning methods such as LoRA and Adapter have been proposed to achieve a better balance between efficiency and performance. To our knowledge, there is currently no work that has conducted parameter-efficient fine-tuning (PEFT) for RGB-Event recognition based on pre-trained foundation models. To address this issue, this paper proposes a novel PEFT strategy to adapt the pre-trained foundation vision models for the RGB-Event-based classification. Specifically, given the RGB frames and event streams, we extract the RGB and event features based on the vision foundation model ViT with a modality-specific LoRA tuning strategy. The frame difference of the dual modalities is also considered to capture the motion cues via the frame difference backbone network. These features are concatenated and fed into high-level Transformer layers for efficient multi-modal feature learning via modality-shared LoRA tuning. Finally, we concatenate these features and feed them into a classification head to achieve efficient fine-tuning.
- clone this repository
git clone https://github.com/Event-AHU/VELoRA.git
- Environment Setting
Python 3.9
torch 2.2.1
easydict 1.12
ftfy 6.1.3
Jinja2 3.1.3
scipy 1.12.0
tqdm 4.66.2
numpy 1.23.0
Pillow 10.2.0
torchvision 0.17.1
sentence-transformers 2.4.0
peft 0.10.0
timm 0.9.16
- PokerEvent Dataset
BaiduYun (178GB): 链接:https://pan.baidu.com/s/1vQnHZUqQ1o58SajvtE-uHw?pwd=AHUE 提取码:AHUE
- HARDVS Dataset
[Event Images] 链接:https://pan.baidu.com/s/1OhlhOBHY91W2SwE6oWjDwA?pwd=1234 提取码:1234
train
CUDA_VISIBLE_DEVICES=1 python -m torch.distributed.launch --nproc_per_node=1 --master_port=20000 train.py MARS
testing
CUDA_VISIBLE_DEVICES=1 python -m torch.distributed.launch --nproc_per_node=1 --master_port=20000 eval.py MARS
Model | File Size | Update Date | Results on PokerEvent | Download Link |
---|---|---|---|---|
ViT-B/16 | 1.88GB | Dec 29, 2024 | 57.99 | https://pan.baidu.com/s/1kCWwksfFJ_NoLpm9zFzPPQ?pwd=6ac4 |
If you find this work helps your research, please star this GitHub and cite the following papers:
@article{chen2024velora,
title={VELoRA: A Low-Rank Adaptation Approach for Efficient RGB-Event based Recognition},
author={Chen, Lan and Yang, Haoxiang and Shao, Pengpeng and Song, Haoyu and Wang, Xiao and Zhao, Zhicheng and Wang, Yaowei and Tian, Yonghong},
journal={arXiv preprint arXiv:2412.20064},
year={2024}
}
If you have any questions about these works, please feel free to leave an issue.