Repository logo

Infoscience

  • English
  • French
Log In
Logo EPFL, École polytechnique fédérale de Lausanne

Infoscience

  • English
  • French
Log In
  1. Home
  2. Academic and Research Output
  3. Conferences, Workshops, Symposiums, and Seminars
  4. Interpreting Language Models Through Knowledge Graph Extraction
 
conference paper not in proceedings

Interpreting Language Models Through Knowledge Graph Extraction

Swamy, Vinitra  
•
Romanou, Angelika  
•
Jaggi, Martin  
December 14, 2021
35th Conference on Neural Information Processing Systems (NeurIPS 2021)

Transformer-based language models trained on large text corpora have enjoyed immense popularity in the natural language processing community and are commonly used as a starting point for downstream tasks. While these models are undeniably useful, it is a challenge to quantify their performance beyond traditional accuracy metrics. In this paper, we compare BERT-based language models through snapshots of acquired knowledge at sequential stages of the training process. Structured relationships from training corpora may be uncovered through querying a masked language model with probing tasks. We present a methodology to unveil a knowledge acquisition timeline by generating knowledge graph extracts from cloze "fill-in-the-blank" statements at various stages of RoBERTa's early training. We extend this analysis to a comparison of pretrained variations of BERT models (DistilBERT, BERT-base, RoBERTa). This work proposes a quantitative framework to compare language models through knowledge graph extraction (GED, Graph2Vec) and showcases a part-of-speech analysis (POSOR) to identify the linguistic strengths of each model variant. Using these metrics, machine learning practitioners can compare models, diagnose their models' behavioral strengths and weaknesses, and identify new targeted datasets to improve model performance.

  • Files
  • Details
  • Metrics
Type
conference paper not in proceedings
ArXiv ID

2111.08546

Author(s)
Swamy, Vinitra  
Romanou, Angelika  
Jaggi, Martin  
Date Issued

2021-12-14

Total of pages

13

Editorial or Peer reviewed

REVIEWED

Written at

EPFL

EPFL units
MLO  
Event nameEvent placeEvent date
35th Conference on Neural Information Processing Systems (NeurIPS 2021)

Online

December 6-14, 2021

Available on Infoscience
March 22, 2022
Use this identifier to reference this record
https://infoscience.epfl.ch/handle/20.500.14299/186528
Logo EPFL, École polytechnique fédérale de Lausanne
  • Contact
  • infoscience@epfl.ch

  • Follow us on Facebook
  • Follow us on Instagram
  • Follow us on LinkedIn
  • Follow us on X
  • Follow us on Youtube
AccessibilityLegal noticePrivacy policyCookie settingsEnd User AgreementGet helpFeedback

Infoscience is a service managed and provided by the Library and IT Services of EPFL. © EPFL, tous droits réservés