Repository logo

Infoscience

  • English
  • French
Log In
Logo EPFL, École polytechnique fédérale de Lausanne

Infoscience

  • English
  • French
Log In
  1. Home
  2. Academic and Research Output
  3. Conferences, Workshops, Symposiums, and Seminars
  4. Direction is what you need: Improving Word Embedding Compression in Large Language Models
 
conference paper

Direction is what you need: Improving Word Embedding Compression in Large Language Models

Balazy, Klaudia
•
Banaei, Mohammadreza  
•
Lebret, Remi  
Show more
January 1, 2021
Repl4Nlp 2021: Proceedings Of The 6Th Workshop On Representation Learning For Nlp
Joint Conference of 59th Annual Meeting of the Association-for-Computational-Linguistics (ACL) / 11th International Joint Conference on Natural Language Processing (IJCNLP) / 6th Workshop on Representation Learning for NLP (RepL4NLP)

The adoption of Transformer-based models in natural language processing (NLP) has led to great success using a massive number of parameters. However, due to deployment constraints in edge devices, there has been a rising interest in the compression of these models to improve their inference time and memory footprint. This paper presents a novel loss objective to compress token embeddings in the Transformer-based models by leveraging an AutoEncoder architecture. More specifically, we emphasize the importance of the direction of compressed embeddings with respect to original uncompressed embeddings. The proposed method is task-agnostic and does not require further language modeling pre-training. Our method significantly outperforms the commonly used SVD-based matrix-factorization approach in terms of initial language model Perplexity. Moreover, we evaluate our proposed approach over SQuAD v1.1 dataset and several downstream tasks from the GLUE benchmark, where we also outperform the baseline in most scenarios. Our code is public.(1).

  • Details
  • Metrics
Type
conference paper
DOI
10.18653/v1/2021.repl4nlp-1.32
Web of Science ID

WOS:000694699300032

Author(s)
Balazy, Klaudia
Banaei, Mohammadreza  
Lebret, Remi  
Tabor, Jacek
Aberer, Karl  
Date Issued

2021-01-01

Publisher

ASSOC COMPUTATIONAL LINGUISTICS-ACL

Publisher place

Stroudsburg

Published in
Repl4Nlp 2021: Proceedings Of The 6Th Workshop On Representation Learning For Nlp
ISBN of the book

978-1-954085-72-5

Start page

322

End page

330

Editorial or Peer reviewed

REVIEWED

Written at

EPFL

EPFL units
LSIR  
Event nameEvent placeEvent date
Joint Conference of 59th Annual Meeting of the Association-for-Computational-Linguistics (ACL) / 11th International Joint Conference on Natural Language Processing (IJCNLP) / 6th Workshop on Representation Learning for NLP (RepL4NLP)

ELECTR NETWORK

Aug 01-06, 2021

Available on Infoscience
September 25, 2021
Use this identifier to reference this record
https://infoscience.epfl.ch/handle/20.500.14299/181652
Logo EPFL, École polytechnique fédérale de Lausanne
  • Contact
  • infoscience@epfl.ch

  • Follow us on Facebook
  • Follow us on Instagram
  • Follow us on LinkedIn
  • Follow us on X
  • Follow us on Youtube
AccessibilityLegal noticePrivacy policyCookie settingsEnd User AgreementGet helpFeedback

Infoscience is a service managed and provided by the Library and IT Services of EPFL. © EPFL, tous droits réservés