QASem Parsing: Text-to-text Modeling of QA-based Semantics


Journal article


Ayal Klein, Eran Hirsch, Ron Eliav, Valentina Pyatkin, Avi Caciularu, Ido Dagan
EMNLP, 2022

View PDF Paper Data Poster
Cite

Cite

APA   Click to copy
Klein, A., Hirsch, E., Eliav, R., Pyatkin, V., Caciularu, A., & Dagan, I. (2022). QASem Parsing: Text-to-text Modeling of QA-based Semantics. EMNLP.


Chicago/Turabian   Click to copy
Klein, Ayal, Eran Hirsch, Ron Eliav, Valentina Pyatkin, Avi Caciularu, and Ido Dagan. “QASem Parsing: Text-to-Text Modeling of QA-Based Semantics.” EMNLP (2022).


MLA   Click to copy
Klein, Ayal, et al. “QASem Parsing: Text-to-Text Modeling of QA-Based Semantics.” EMNLP, 2022.


BibTeX   Click to copy

@article{ayal2022a,
  title = {QASem Parsing: Text-to-text Modeling of QA-based Semantics},
  year = {2022},
  journal = {EMNLP},
  author = {Klein, Ayal and Hirsch, Eran and Eliav, Ron and Pyatkin, Valentina and Caciularu, Avi and Dagan, Ido}
}

Abstract

Various works suggest the appeals of incorporating explicit semantic representations when addressing challenging realistic NLP scenarios. Common approaches offer either comprehensive linguistically-based formalisms, like AMR, or alternatively Open-IE, which provides a shallow and partial representation. More recently, an appealing trend introduces semi-structured natural-language structures as an intermediate meaning-capturing representation, often in the form of questions and answers.In this work, we further promote this line of research by considering three prior QA-based semantic representations. These cover verbal, nominalized and discourse-based predications, regarded as jointly providing a comprehensive representation of textual information — termed QASem. To facilitate this perspective, we investigate how to best utilize pre-trained sequence-to-sequence language models, which seem particularly promising for generating representations that consist of natural language expressions (questions and answers). In particular, we examine and analyze input and output linearization strategies, as well as data augmentation and multitask learning for a scarce training data setup. Consequently, we release the first unified QASem parsing tool, easily applicable for downstream tasks that can benefit from an explicit semi-structured account of information units in text.