TL;DR该研究实现了扩展 MAE 预训练策略 RetroMAE 的统一框架以支持多样化的句子表示任务,并证明其在零样本检索和其他下游任务中的有效性,为未来的句子表示预训练的设计提供了实证建议。
Abstract
Despite the progresses on pre-trained language models, there is a lack of
unified frameworks for pre-trained sentence representation. As such, it calls
for different pre-training methods for specific scenarios, a