DataFunTalk
Jun 21, 2023 · Artificial Intelligence
Low‑Resource NLP Pretraining: Methodology, Experiments, and Zero‑Shot Applications
This article presents a low‑resource NLP pretraining approach that combines transformer‑based language modeling with contrastive vector learning, details the unsupervised sample‑pair construction, introduces a camel‑shaped masking distribution, and demonstrates through extensive experiments that the resulting model achieves strong zero‑shot NLU, NLG, and retrieval performance while requiring minimal compute and data.
NLPZero-shotcontrastive learning
0 likes · 10 min read