S3D Network is reported in the ECCV 2018 paper Rethinking Spatiotemporal Feature Learning: Speed-Accuracy Trade-offs in Video Classification. It has been shown by Xie that replacing standard 3D convolutions with spatial and temporal separable 3D convolutions 1) reduces the total number of parameters, 2) is more computationally efficient, and even 3) improves the performance in terms of accuracy.
In this repository, we release the pretrained S3D network in PyTorch. We pretrained S3D on Kinetics-400 dataset and it achieves 72.08% top1 accuracy (top5: 90.35%) on the validation set of the dataset.
Network | Top 1 (%) | Top 5 (%) |
---|---|---|
I3D | 71.1 | 89.3 |
S3D (reported by author) | 72.2 | 90.6 |
S3D (our implementation) | 72.1 | 90.4 |
First, clone this repository and download this weight file. Then, just run the code using
$ python main.py
This will output the top 5 Kinetics classes predicted by the model with corresponding probability.
Top 5 classes ... with probability
riding a bike ... 0.9937429
biking through snow ... 0.0041600233
riding mountain bike ... 0.0010456557
riding unicycle ... 0.00088055385
motorcycling ... 0.00014455811
We implement and release this repository together with our TASED-Net project (ICCV 2019). If you find it useful, you might want to consider citing our work.
@inproceedings{min2019tased,
title={TASED-Net: Temporally-Aggregating Spatial Encoder-Decoder Network for Video Saliency Detection},
author={Min, Kyle and Corso, Jason J},
booktitle={Proceedings of the IEEE International Conference on Computer Vision},
pages={2394--2403},
year={2019}
}