BriefGPT.xyz
Mar, 2022
基于遮挡自编码器的视频无监督预训练技术
VideoMAE: Masked Autoencoders are Data-Efficient Learners for Self-Supervised Video Pre-Training
HTML
PDF
Zhan Tong, Yibing Song, Jue Wang, Limin Wang
TL;DR
本文研究了使用VideoMAE进行自主监督视频预训练(SSVP)的数据高效性问题,并通过适当的视频屏蔽达到了良好的表现,进而证明数据质量对SSVP更加重要。
Abstract
Pre-training video transformers on extra large-scale datasets is generally required to achieve
premier performance
on relatively small datasets. In this paper, we show that
video masked autoencoders
(VideoMAE) ar
→