Repository logo

Infoscience

  • English
  • French
Log In
Logo EPFL, École polytechnique fédérale de Lausanne

Infoscience

  • English
  • French
Log In
  1. Home
  2. Academic and Research Output
  3. Conferences, Workshops, Symposiums, and Seminars
  4. Revisiting Offline Compression: Going Beyond Factorization-based Methods for Transformer Language Models
 
conference paper

Revisiting Offline Compression: Going Beyond Factorization-based Methods for Transformer Language Models

Banaei, Mohammadreza  
•
Balazy, Klaudia
•
Kasymov, Artur
Show more
Augenstein, I
•
Vlachos, A
January 1, 2023
17Th Conference Of The European Chapter Of The Association For Computational Linguistics, Eacl 2023
17th Conference of the European-Chapter of the Association-for-Computational-Linguistics (EACL)

Recent transformer language models achieve outstanding results in many natural language processing (NLP) tasks. However, their enormous size often makes them impractical on memory-constrained devices, requiring practitioners to compress them to smaller networks. In this paper, we explore offline compression methods, meaning computationally-cheap approaches that do not require further finetuning of the compressed model. We challenge the classical matrix factorization methods by proposing a novel, better-performing autoencoder-based framework. We perform a comprehensive ablation study of our approach, examining its different aspects over a diverse set of evaluation settings. Moreover, we show that enabling collaboration between modules across layers by compressing certain modules together positively impacts the final model performance. Experiments on various NLP tasks demonstrate that our approach significantly outperforms commonly used factorization-based offline compression methods.

  • Details
  • Metrics
Type
conference paper
DOI
10.18653/v1/2023.findings-eacl.133
Web of Science ID

WOS:001181085100131

Author(s)
Banaei, Mohammadreza  
Balazy, Klaudia
Kasymov, Artur
Lebret, Remi  
Tabor, Jacek
Aberer, Karl  
Editors
Augenstein, I
•
Vlachos, A
Date Issued

2023-01-01

Publisher

Assoc Computational Linguistics-Acl

Publisher place

Stroudsburg

Published in
17Th Conference Of The European Chapter Of The Association For Computational Linguistics, Eacl 2023
ISBN of the book

978-1-959429-47-0

Start page

1788

End page

1805

Subjects

Technology

Editorial or Peer reviewed

REVIEWED

Written at

EPFL

EPFL units
LSIR  
Event nameEvent placeEvent date
17th Conference of the European-Chapter of the Association-for-Computational-Linguistics (EACL)

Dubrovnik, CROATIA

MAY 02-06, 2023

FunderGrant Number

National Centre of Science (Poland)

2019/33/B/ST6/00894

Natural Sciences at the Jagiellonian University

POIR.04.04.00-00-14DE/18-00

Foundation for Polish Science - European Union under the European Regional Development Fund

Available on Infoscience
May 1, 2024
Use this identifier to reference this record
https://infoscience.epfl.ch/handle/20.500.14299/207595
Logo EPFL, École polytechnique fédérale de Lausanne
  • Contact
  • infoscience@epfl.ch

  • Follow us on Facebook
  • Follow us on Instagram
  • Follow us on LinkedIn
  • Follow us on X
  • Follow us on Youtube
AccessibilityLegal noticePrivacy policyCookie settingsEnd User AgreementGet helpFeedback

Infoscience is a service managed and provided by the Library and IT Services of EPFL. © EPFL, tous droits réservés