MovieNet: A Holistic Dataset for Movie Understanding
CUHK-SenseTime Joint Lab, The Chinese University of Hong Kong
European Conference on Computer Vision (ECCV) 2020, Glasgow, Scotland, United Kingdom

Overview

Recent years have seen remarkable advances in visual understanding. However, how to understand a story-based long video with artistic styles, e.g. movie, remains challenging. In this paper, we introduce MovieNet -- a holistic dataset for movie understanding. MovieNet contains 1,100 movies with a large amount of multi-modal data, e.g. trailers, photos, plot descriptions, etc. Besides, different aspects of manual annotations are provided in MovieNet, including 1.1M characters with bounding boxes and identities, 42K scene boundaries, 2.5K aligned description sentences, 65K tags of place and action, and 92K tags of cinematic style. To the best of our knowledge, MovieNet is the largest dataset with richest annotations for comprehensive movie understanding. Based on MovieNet, we set up several benchmarks for movie understanding from different angles. Extensive experiments are executed on these benchmarks to show the immeasurable value of MovieNet and the gap of current approaches towards comprehensive movie understanding. We believe that such a holistic dataset would promote the researches on story-based long video understanding and beyond.

Dataset

Statistics

Materials


Citation

@inproceedings{huang2020movie,
    title={MovieNet: A Holistic Dataset for Movie Understanding},
    author={Huang, Qingqiu and Xiong, Yu and Rao, Anyi and Wang, Jiaze and Lin, Dahua},
    booktitle = {The European Conference on Computer Vision (ECCV)}, 
    year={2020}
}

Contact

Qingqiu Huang: hq016 [AT] ie.cuhk.edu.hk