Can You Label Less by Using Out-of-Domain Data? Active & Transfer Learning with Few-shot Instructions
Abstract
Labeling social-media data for custom dimensions of toxicity and social bias is challenging and labor-intensive. Existing transfer and active learning approaches meant to reduce annotation effort require fine-tuning, which suffers from over-fitting to noise and can cause domain shift with small sample sizes. In this work, we propose a novel Active Transfer Few-shot Instructions (ATF) approach which requires no fine-tuning. ATF leverages the internal linguistic knowledge of pre-trained language models (PLMs) to facilitate the transfer of information from existing pre-labeled datasets (source-domain task) with minimum labeling effort on unlabeled target data (target-domain task). Our strategy can yield positive transfer achieving a mean AUC gain of 10.5% compared to no transfer with a large 22b parameter PLM. We further show that annotation of just a few target-domain samples via active learning can be beneficial for transfer, but the impact diminishes with more annotation effort (26% drop in gain between 100 and 2000 annotated examples). Finally, we find that not all transfer scenarios yield a positive gain, which seems related to the PLMs initial performance on the target-domain task.
Additional Information
Attribution 4.0 International (CC BY 4.0). We would like to thank the Caltech SURF program for contributing to the funding of this project and especially the named donor Carolyn Ash. This material is based upon work supported by the National Science Foundation under Grant # 2030859 to the Computing Research Association for the CIFellows Project. Anima Anandkumar is partially supported by Bren Named Chair Professorship at Caltech and is a paid employee of Nvidia. Sara Kangaslahti was a paid part-time intern at Nvidia during this project.Attached Files
Accepted Version - 2211.11798.pdf
Files
Name | Size | Download all |
---|---|---|
md5:0153e4052f17c09504512dca8aab5331
|
2.2 MB | Preview Download |
Additional details
- Eprint ID
- 118560
- Resolver ID
- CaltechAUTHORS:20221221-004733367
- Caltech Summer Undergraduate Research Fellowship (SURF)
- NSF
- CCF-2030859
- Bren Professor of Computing and Mathematical Sciences
- NVIDIA Corporation
- Created
-
2022-12-22Created from EPrint's datestamp field
- Updated
-
2023-06-02Created from EPrint's last_modified field