Skip to content

UpChen/2024_FusionNet

Repository files navigation

Fusion of Short-term and Long-term Attention for Video Mirror Detection (ICME Oral 2024)

Mingchen Xu, Yukun Lai, Ze Ji, Jing Wu

School of Computer Science and Informatics, Cardiff University

Requirement

cuda==11.1
cudnn==8.0
torch==1.9.0
timm==0.9.6
transformers==4.30.2
pytorch-lightning==1.5.10
medpy==0.4.0
einops==0.6.1

Usage Instructions

  1. Setup

    Clone the repository and navigate to its directory:

    git clone https://github.com/UpChen/2024_FusionNet.git
    cd 2024_FusionNet
  2. Dataset Preparation

    Download and unzip (ViMirr dataset).

  3. Training

    To train the model, execute:

    python train.py
  4. Testing

    After training, update the checkpoint file path in the infer.py script. Then, test the trained model using:

    python infer.py
  5. Evaluating

    After testing, update the results file path in the eval.py script. Then, evaluate the results using:

    python eval.py

Citation

If you use this code or the associated paper in your work, please cite:

@INPROCEEDINGS {10688367,
author = {M. Xu and J. Wu and Y. Lai and Z. Ji},
booktitle = {2024 IEEE International Conference on Multimedia and Expo (ICME)},
title = {Fusion of Short-term and Long-term Attention for Video Mirror Detection},
year = {2024},
volume = {},
issn = {},
pages = {1-9},
abstract = {Techniques for detecting mirrors from static images have witnessed rapid growth in recent years. However, these methods detect mirrors from single input images. Detecting mirrors from video requires further consideration of temporal consistency between frames. We observe that humans can recognize mirror candidates, from just one or two frames, based on their appearance (e.g. shape, color). However, to ensure that the candidate is indeed a mirror (not a picture or a window), we often need to observe more frames for a global view. This observation motivates us to detect mirrors by fusing appearance features extracted from a short-term attention module and context information extracted from a long-term attention module. To evaluate the performance, we build a challenging benchmark dataset of 19,255 frames from 281 videos. Experimental results demonstrate that our method achieves state-of-the-art performance on the benchmark dataset.},
keywords = {shape;image color analysis;streaming media;benchmark testing;feature extraction;transformers;real-time systems;mirrors;data mining},
doi = {10.1109/ICME57554.2024.10688367},
url = {https://doi.ieeecomputersociety.org/10.1109/ICME57554.2024.10688367},
publisher = {IEEE Computer Society},
address = {Los Alamitos, CA, USA},
month = {jul}
}

Contact

If you have any questions, please feel free to contact me via [email protected].

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published