Journal article
EMNLP, 2022
APA
Click to copy
Klein, A., Hirsch, E., Eliav, R., Pyatkin, V., Caciularu, A., & Dagan, I. (2022). QASem Parsing: Text-to-text Modeling of QA-based Semantics. EMNLP.
Chicago/Turabian
Click to copy
Klein, Ayal, Eran Hirsch, Ron Eliav, Valentina Pyatkin, Avi Caciularu, and Ido Dagan. “QASem Parsing: Text-to-Text Modeling of QA-Based Semantics.” EMNLP (2022).
MLA
Click to copy
Klein, Ayal, et al. “QASem Parsing: Text-to-Text Modeling of QA-Based Semantics.” EMNLP, 2022.
BibTeX Click to copy
@article{ayal2022a,
title = {QASem Parsing: Text-to-text Modeling of QA-based Semantics},
year = {2022},
journal = {EMNLP},
author = {Klein, Ayal and Hirsch, Eran and Eliav, Ron and Pyatkin, Valentina and Caciularu, Avi and Dagan, Ido}
}
Various works suggest the appeals of incorporating explicit semantic representations when addressing challenging realistic NLP scenarios. Common approaches offer either comprehensive linguistically-based formalisms, like AMR, or alternatively Open-IE, which provides a shallow and partial representation. More recently, an appealing trend introduces semi-structured natural-language structures as an intermediate meaning-capturing representation, often in the form of questions and answers.In this work, we further promote this line of research by considering three prior QA-based semantic representations. These cover verbal, nominalized and discourse-based predications, regarded as jointly providing a comprehensive representation of textual information — termed QASem. To facilitate this perspective, we investigate how to best utilize pre-trained sequence-to-sequence language models, which seem particularly promising for generating representations that consist of natural language expressions (questions and answers). In particular, we examine and analyze input and output linearization strategies, as well as data augmentation and multitask learning for a scarce training data setup. Consequently, we release the first unified QASem parsing tool, easily applicable for downstream tasks that can benefit from an explicit semi-structured account of information units in text.