BriefGPT.xyz
Oct, 2023
ArchBERT: 神经架构和自然语言的双模理解
ArchBERT: Bi-Modal Understanding of Neural Architectures and Natural Languages
HTML
PDF
Mohammad Akbari, Saeed Ranjbar Alvar, Behnam Kamranian, Amin Banitalebi-Dehkordi, Yong Zhang
TL;DR
提出了一种用于联合学习和理解神经架构和自然语言的双模态模型ArchBERT,引入了被称为MAM的预训练策略,提供了两个新的双模态数据集进行方法的训练和验证,通过一系列下游任务的实验验证了ArchBERT的性能。
Abstract
Building
multi-modal language models
has been a trend in the recent years, where additional modalities such as image, video, speech, etc. are jointly learned along with natural languages (i.e., textual information). Despite the success of these
→