BriefGPT.xyz
Jun, 2023
弃用预训练:将视觉-语言模型应用于未知语言
Stop Pre-Training: Adapt Visual-Language Models to Unseen Languages
HTML
PDF
Yasmine Karoui, Rémi Lebret, Negar Foroutan, Karl Aberer
TL;DR
本文提出一种简单而有效的方法,利用多语言预训练语言模型(MPLM)和跨语言上下文化词嵌入对 VLP 进行了调整和适应,以在未见过的语言上具有出色的表现,而无需大量平行语料库。
Abstract
vision-language pre-training
(VLP) has advanced the performance of many vision-language tasks, such as
image-text retrieval
, visual entailment, and visual reasoning. The pre-training mostly utilizes lexical datab
→