Repository logo

Infoscience

  • English
  • French
Log In
Logo EPFL, École polytechnique fédérale de Lausanne

Infoscience

  • English
  • French
Log In
  1. Home
  2. Academic and Research Output
  3. Conferences, Workshops, Symposiums, and Seminars
  4. PERFECT: Prompt-free and Efficient Few-shot Learning with Language Models
 
conference paper

PERFECT: Prompt-free and Efficient Few-shot Learning with Language Models

Mahabadi, Rabeeh Karimi  
•
Zettlemoyer, Luke
•
Henderson, James
Show more
January 1, 2022
Proceedings Of The 60Th Annual Meeting Of The Association For Computational Linguistics (Acl 2022), Vol 1: (Long Papers)
60th Annual Meeting of the Association-for-Computational-Linguistics (ACL)

Current methods for few-shot fine-tuning of pretrained masked language models (PLMs) require carefully engineered prompts and verbalizers for each new task to convert examples into a cloze-format that the PLM can score. In this work, we propose PERFECT, a simple and efficient method for few-shot fine-tuning of PLMs without relying on any such handcrafting, which is highly effective given as few as 32 data points. PERFECT makes two key design choices: First, we show that manually engineered task prompts can be replaced with task-specific adapters that enable sample-efficient fine-tuning and reduce memory and storage costs by roughly factors of 5 and 100, respectively. Second, instead of using handcrafted verbalizers, we learn new multi-token label embeddings during fine-tuning, which are not tied to the model vocabulary and which allow us to avoid complex auto-regressive decoding. These embeddings are not only learnable from limited data but also enable nearly 100x faster training and inference. Experiments on a wide range of few shot NLP tasks demonstrate that PERFECT, while being simple and efficient, also outperforms existing state-of-the-art few-shot learning methods. Our code is publicly available at https://github.com/facebookresearch/perfect.git.

  • Details
  • Metrics
Type
conference paper
Web of Science ID

WOS:000828702303047

Author(s)
Mahabadi, Rabeeh Karimi  
Zettlemoyer, Luke
Henderson, James
Saeidi, Marzieh
Mathias, Lambert
Stoyanov, Veselin
Yazdani, Majid
Date Issued

2022-01-01

Publisher

ASSOC COMPUTATIONAL LINGUISTICS-ACL

Publisher place

Stroudsburg

Published in
Proceedings Of The 60Th Annual Meeting Of The Association For Computational Linguistics (Acl 2022), Vol 1: (Long Papers)
ISBN of the book

978-1-955917-21-6

Start page

3638

End page

3652

Subjects

Computer Science, Artificial Intelligence

•

Computer Science, Interdisciplinary Applications

•

Linguistics

•

Computer Science

•

Linguistics

Editorial or Peer reviewed

REVIEWED

Written at

EPFL

Event nameEvent placeEvent date
60th Annual Meeting of the Association-for-Computational-Linguistics (ACL)

Dublin, IRELAND

May 22-27, 2022

Available on Infoscience
September 26, 2022
Use this identifier to reference this record
https://infoscience.epfl.ch/handle/20.500.14299/191017
Logo EPFL, École polytechnique fédérale de Lausanne
  • Contact
  • infoscience@epfl.ch

  • Follow us on Facebook
  • Follow us on Instagram
  • Follow us on LinkedIn
  • Follow us on X
  • Follow us on Youtube
AccessibilityLegal noticePrivacy policyCookie settingsEnd User AgreementGet helpFeedback

Infoscience is a service managed and provided by the Library and IT Services of EPFL. © EPFL, tous droits réservés