BriefGPT.xyz
Jan, 2022
使用统一条件模型的自训练视觉语言BERTs
Self-Training Vision Language BERTs with a Unified Conditional Model
HTML
PDF
Xiaofeng Yang, Fengmao Lv, Fayao Liu, Guosheng Lin
TL;DR
提出了一种自我训练的方法,用于从未标注的图像数据中训练VL-BERT模型,模型采用统一的有条件模型,能够执行零样本条件生成,通过该方法使用仅300k个未标注的额外数据,可以获得与训练了300万个图像数据的相似模型大小的模型相媲美或甚至更好的性能
Abstract
Natural language BERTs are trained with language corpus in a self-supervised manner. Unlike natural language BERTs, vision language BERTs need paired data to train, which restricts the scale of VL-
bert
pretraining. We propose a
→