TL;DR提出了两个大规模视频领域自适应数据集,并研究了不同的视频 DA 集成方法,展示了同时对齐和学习时间动态可实现有效对齐的效果,最后提出了一个基于时间注意力的对抗适应网络 TA3N,实现更有效的领域对齐,达到了四个视频 DA 数据集的最优性能。
Abstract
Although various image-based domain adaptation (DA) techniques have been proposed in recent years, domain shift in videos is still not well-explored. Most previous works only evaluate performance on small-scale datasets which are saturated. Therefore, we first propose two largescale video DA datasets with much larger domain discrepancy: UCF-HMDB_full and Kin