BriefGPT.xyz
Feb, 2024
高效的语言自适应预训练:为波兰语扩展最先进的大型语言模型
Efficient Language Adaptive Pre-training: Extending State-of-the-Art Large Language Models for Polish
HTML
PDF
Szymon Ruciński
TL;DR
使用英语大型语言模型(LLMs)通过微调,成功地生成了波兰文文本,并证明了该方法对于为现有LLMs添加新语言是可行的。
Abstract
This study explores the potential of
fine-tuning
foundational English Large Language Models (LLMs) for
generating polish text
. The first step involves Language Adaptive Pre-training (LAPT) on a high-quality datas
→