Repository logo

Infoscience

  • English
  • French
Log In
Logo EPFL, École polytechnique fédérale de Lausanne

Infoscience

  • English
  • French
Log In
  1. Home
  2. Academic and Research Output
  3. Conferences, Workshops, Symposiums, and Seminars
  4. CAR: Conceptualization-Augmented Reasoner for Zero-Shot Commonsense Question Answering
 
conference paper

CAR: Conceptualization-Augmented Reasoner for Zero-Shot Commonsense Question Answering

Wang, Weiqi
•
Fang, Tianqing
•
Ding, Wenxuan
Show more
2023
Findings of the Association for Computational Linguistics: EMNLP 2023
The 2023 Conference on Empirical Methods in Natural Language Processing

The task of zero-shot commonsense question answering evaluates models on their capacity to reason about general scenarios beyond those presented in specific datasets. Existing approaches for tackling this task leverage external knowledge from CommonSense Knowledge Bases (CSKBs) by pre-training the model on synthetic QA pairs constructed from CSKBs. In these approaches, negative examples (distractors) are formulated by randomly sampling from CSKBs using fairly primitive keyword constraints. However, two bottlenecks limit these approaches: the inherent incompleteness of CSKBs limits the semantic coverage of synthetic QA pairs, and the lack of human annotations makes the sampled negative examples potentially uninformative and contradictory. To tackle these limitations above, we propose Conceptualization-Augmented Reasoner (CAR), a zero-shot commonsense question-answering framework that fully leverages the power of conceptualization. Specifically, CAR abstracts a commonsense knowledge triple to many higher-level instances, which increases the coverage of the CSKB and expands the ground-truth answer space, reducing the likelihood of selecting false-negative distractors. Extensive experiments demonstrate that CAR more robustly generalizes to answering questions about zero-shot commonsense scenarios than existing methods, including large language models, such as GPT3.5 and ChatGPT. Our code, data, and model checkpoints are available at https://github.com/HKUSTKnowComp/CAR.

  • Files
  • Details
  • Metrics
Type
conference paper
DOI
10.18653/v1/2023.findings-emnlp.902
Scopus ID

2-s2.0-85183304088

Author(s)
Wang, Weiqi

Hong Kong University of Science and Technology

Fang, Tianqing

Hong Kong University of Science and Technology

Ding, Wenxuan

Hong Kong University of Science and Technology

Xu, Baixuan

Hong Kong University of Science and Technology

Liu, Xin

Hong Kong University of Science and Technology

Song, Yangqiu

Hong Kong University of Science and Technology

Bosselut, Antoine  

École Polytechnique Fédérale de Lausanne

Date Issued

2023

Publisher

Association for Computational Linguistics (ACL)

Published in
Findings of the Association for Computational Linguistics: EMNLP 2023
ISBN of the book

9798891760615

Start page

13520

End page

13545

Editorial or Peer reviewed

REVIEWED

Written at

EPFL

EPFL units
NLP  
Event nameEvent acronymEvent placeEvent date
The 2023 Conference on Empirical Methods in Natural Language Processing

EMNLP 2023

Singapore

2023-12-06 - 2023-12-10

FunderFunding(s)Grant NumberGrant URL

EPFL Center for Imaging

EPFL Science Seed Fund

NSFC Fund

U20B2053

Show more
Available on Infoscience
January 8, 2026
Use this identifier to reference this record
https://infoscience.epfl.ch/handle/20.500.14299/257687
Logo EPFL, École polytechnique fédérale de Lausanne
  • Contact
  • infoscience@epfl.ch

  • Follow us on Facebook
  • Follow us on Instagram
  • Follow us on LinkedIn
  • Follow us on X
  • Follow us on Youtube
AccessibilityLegal noticePrivacy policyCookie settingsEnd User AgreementGet helpFeedback

Infoscience is a service managed and provided by the Library and IT Services of EPFL. © EPFL, tous droits réservés