Repository logo

Infoscience

  • English
  • French
Log In
Logo EPFL, École polytechnique fédérale de Lausanne

Infoscience

  • English
  • French
Log In
  1. Home
  2. Academic and Research Output
  3. Conferences, Workshops, Symposiums, and Seminars
  4. Breaking the Language Barrier: Improving Cross-Lingual Reasoning with Structured Self-Attention
 
conference paper

Breaking the Language Barrier: Improving Cross-Lingual Reasoning with Structured Self-Attention

Foroutan, Negar  
•
Banaei, Mohammadreza  
•
Aberer, Karl  
Show more
December 1, 2023
Findings of the Association for Computational Linguistics: EMNLP 2023
Conference on Empirical Methods in Natural Language Processing (EMNLP 2023)

In this work, we study whether multilingual language models (MultiLMs) can transfer logical reasoning abilities to other languages when they are fine-tuned for reasoning in a different language. We evaluate the cross-lingual reasoning abilities of MultiLMs in two schemes: (1) where the language of the context and the question remain the same in the new languages that are tested (i.e., the reasoning is still monolingual, but the model must transfer the learned reasoning ability across languages), and (2) where the language of the context and the question is different (which we term code-switched reasoning). On two logical reasoning datasets, RuleTaker and LeapOfThought, we demonstrate that although MultiLMs can transfer reasoning ability across languages in a monolingual setting, they struggle to transfer reasoning abilities in a code-switched setting. Following this observation, we propose a novel attention mechanism that uses a dedicated set of parameters to encourage cross-lingual attention in code-switched sequences, which improves the reasoning performance by up to 14% and 4% on the RuleTaker and LeapOfThought datasets, respectively. 1 * Equal contribution 1 Our code is available at https://github.com/negarforoutan/multilingual-code-switched-reasoning.

  • Files
  • Details
  • Metrics
Type
conference paper
DOI
10.18653/v1/2023.findings-emnlp.632
Author(s)
Foroutan, Negar  

EPFL

Banaei, Mohammadreza  

EPFL

Aberer, Karl  

EPFL

Bosselut, Antoine  

EPFL

Date Issued

2023-12-01

Publisher

Association for Computational Linguistics

Published in
Findings of the Association for Computational Linguistics: EMNLP 2023
Start page

9422

End page

9442

Editorial or Peer reviewed

REVIEWED

Written at

EPFL

EPFL units
NLP  
LSIR  
Event nameEvent acronymEvent placeEvent date
Conference on Empirical Methods in Natural Language Processing (EMNLP 2023)

EMNLP 2023

Singapore

2022-12-06 - 2022-12-10

Available on Infoscience
March 10, 2025
Use this identifier to reference this record
https://infoscience.epfl.ch/handle/20.500.14299/247675
Logo EPFL, École polytechnique fédérale de Lausanne
  • Contact
  • infoscience@epfl.ch

  • Follow us on Facebook
  • Follow us on Instagram
  • Follow us on LinkedIn
  • Follow us on X
  • Follow us on Youtube
AccessibilityLegal noticePrivacy policyCookie settingsEnd User AgreementGet helpFeedback

Infoscience is a service managed and provided by the Library and IT Services of EPFL. © EPFL, tous droits réservés