Toggle navigation
Home
People
Publications
Resources
Reading Group
Talks
self-supervision
Rethinking Self-Supervision Objectives for Generalizable Coherence Modeling
We show empirically that increasing the density of negative samples improves the basic model, and using a global negative queue further improves and stabilizes the model while training with hard negative samples.
UXLA: A Robust Unsupervised Data Augmentation Framework for Zero-Resouce Cross-Lingual NLP
We propose UXLA, a novel data augmentation framework for self-supervised learning in zero-resource transfer learning scenarios.
×
Cite