Skip to content

Visually explaining 3D-CNN predictions for video classification with an adaptive occlusion sensitivity analysis

Notifications You must be signed in to change notification settings

uchiyama33/AOSA

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

20 Commits
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Adaptive occlusion sensitivity analysis

This repository contains the implementation of the paper "Visually explaining 3D-CNN predictions for video classification with an adaptive occlusion sensitivity analysis".

This repository includes the work that is distributed in the Apache License 2.0.

Requirements

  • docker > 20.
  • docker-compose

Instalation

Download model parameters

  1. Download the parameters of R3D fine-tuned on UCF-101 from here
  2. Place the downloaded file (save_200.pth) into data/r3d_models/finetuning/ucf101/r3d50_K_fc/

Create docker container

  1. $cd .devcontainer
  2. $docker-compose up -d
  3. $docker attach aosa

Example

Please refer to occlusion_sensitivity_analysis.ipynb.

If there is no enough GPU memory, please try to small "batchsize" in the example codes.

Models and dataset utils

We use the code from the following repository for 3D-CNN models and dataset utilities. To download datasets and other resources, please refer to this repository.

kenshohara/3D-ResNets-PyTorch: 3D ResNets for Action Recognition (CVPR 2018)

Citation

@article{uchiyama2022visually,
  title = {{Visually explaining 3D-CNN predictions for video classification with an adaptive occlusion sensitivity analysis}},
  author = {Uchiyama, Tomoki and Sogi, Naoya and Niinuma, Koichiro and Fukui, Kazuhiro},
  journal={arXiv preprint arXiv:2207.12859},
  year = {2022}
}

About

Visually explaining 3D-CNN predictions for video classification with an adaptive occlusion sensitivity analysis

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published