BriefGPT.xyz
Sep, 2020
通过解耦场景和运动来增强无监督视频表示学习
Enhancing Unsupervised Video Representation Learning by Decoupling the Scene and the Motion
HTML
PDF
Jinpeng Wang, Yuting Gao, Ke Li, Xinyang Jiang, Xiaowei Guo...
TL;DR
提出了一种解耦场景和物体运动信息的DSM方法,通过构造正负剪辑来加强模型对物体运动信息的关注,减少场景信息的影响,并在两项任务上进行实验,发现在UCF101和HMDB51数据集上动作识别任务的准确率分别提高了8.1%和8.8%。
Abstract
One significant factor we expect the
video representation learning
to capture, especially in contrast with the image representation learning, is the
object motion
. However, we found that in the current mainstream
→