@article{laskar-etal-2022-domain,
title = "Domain Adaptation with Pre-trained Transformers for Query-Focused Abstractive Text Summarization",
author = "Laskar, Md Tahmid Rahman and
Hoque, Enamul and
Huang, Jimmy Xiangji",
journal = "Computational Linguistics",
volume = "48",
number = "2",
month = jun,
year = "2022",
address = "Cambridge, MA",
publisher = "MIT Press",
url = "https://aclanthology.org/2022.cl-2.2",
doi = "10.1162/coli_a_00434",
pages = "279--320",
abstract = "The Query-Focused Text Summarization (QFTS) task aims at building systems that generate the summary of the text document(s) based on the given query. A key challenge in addressing this task is the lack of large labeled data for training the summarization model. In this article, we address this challenge by exploring a series of domain adaptation techniques. Given the recent success of pre-trained transformer models in a wide range of natural language processing tasks, we utilize such models to generate abstractive summaries for the QFTS task for both single-document and multi-document scenarios. For domain adaptation, we apply a variety of techniques using pre-trained transformer-based summarization models including transfer learning, weakly supervised learning, and distant supervision. Extensive experiments on six datasets show that our proposed approach is very effective in generating abstractive summaries for the QFTS task while setting a new state-of-the-art result in several datasets across a set of automatic and human evaluation metrics.",
}
<?xml version="1.0" encoding="UTF-8"?>
<modsCollection xmlns="http://www.loc.gov/mods/v3">
<mods ID="laskar-etal-2022-domain">
<titleInfo>
<title>Domain Adaptation with Pre-trained Transformers for Query-Focused Abstractive Text Summarization</title>
</titleInfo>
<name type="personal">
<namePart type="given">Md</namePart>
<namePart type="given">Tahmid</namePart>
<namePart type="given">Rahman</namePart>
<namePart type="family">Laskar</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Enamul</namePart>
<namePart type="family">Hoque</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Jimmy</namePart>
<namePart type="given">Xiangji</namePart>
<namePart type="family">Huang</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<originInfo>
<dateIssued>2022-06</dateIssued>
</originInfo>
<typeOfResource>text</typeOfResource>
<genre authority="bibutilsgt">journal article</genre>
<relatedItem type="host">
<titleInfo>
<title>Computational Linguistics</title>
</titleInfo>
<originInfo>
<issuance>continuing</issuance>
<publisher>MIT Press</publisher>
<place>
<placeTerm type="text">Cambridge, MA</placeTerm>
</place>
</originInfo>
<genre authority="marcgt">periodical</genre>
<genre authority="bibutilsgt">academic journal</genre>
</relatedItem>
<abstract>The Query-Focused Text Summarization (QFTS) task aims at building systems that generate the summary of the text document(s) based on the given query. A key challenge in addressing this task is the lack of large labeled data for training the summarization model. In this article, we address this challenge by exploring a series of domain adaptation techniques. Given the recent success of pre-trained transformer models in a wide range of natural language processing tasks, we utilize such models to generate abstractive summaries for the QFTS task for both single-document and multi-document scenarios. For domain adaptation, we apply a variety of techniques using pre-trained transformer-based summarization models including transfer learning, weakly supervised learning, and distant supervision. Extensive experiments on six datasets show that our proposed approach is very effective in generating abstractive summaries for the QFTS task while setting a new state-of-the-art result in several datasets across a set of automatic and human evaluation metrics.</abstract>
<identifier type="citekey">laskar-etal-2022-domain</identifier>
<identifier type="doi">10.1162/coli_a_00434</identifier>
<location>
<url>https://aclanthology.org/2022.cl-2.2</url>
</location>
<part>
<date>2022-06</date>
<detail type="volume"><number>48</number></detail>
<detail type="issue"><number>2</number></detail>
<extent unit="page">
<start>279</start>
<end>320</end>
</extent>
</part>
</mods>
</modsCollection>
%0 Journal Article
%T Domain Adaptation with Pre-trained Transformers for Query-Focused Abstractive Text Summarization
%A Laskar, Md Tahmid Rahman
%A Hoque, Enamul
%A Huang, Jimmy Xiangji
%J Computational Linguistics
%D 2022
%8 June
%V 48
%N 2
%I MIT Press
%C Cambridge, MA
%F laskar-etal-2022-domain
%X The Query-Focused Text Summarization (QFTS) task aims at building systems that generate the summary of the text document(s) based on the given query. A key challenge in addressing this task is the lack of large labeled data for training the summarization model. In this article, we address this challenge by exploring a series of domain adaptation techniques. Given the recent success of pre-trained transformer models in a wide range of natural language processing tasks, we utilize such models to generate abstractive summaries for the QFTS task for both single-document and multi-document scenarios. For domain adaptation, we apply a variety of techniques using pre-trained transformer-based summarization models including transfer learning, weakly supervised learning, and distant supervision. Extensive experiments on six datasets show that our proposed approach is very effective in generating abstractive summaries for the QFTS task while setting a new state-of-the-art result in several datasets across a set of automatic and human evaluation metrics.
%R 10.1162/coli_a_00434
%U https://aclanthology.org/2022.cl-2.2
%U https://doi.org/10.1162/coli_a_00434
%P 279-320
Markdown (Informal)
[Domain Adaptation with Pre-trained Transformers for Query-Focused Abstractive Text Summarization](https://aclanthology.org/2022.cl-2.2) (Laskar et al., CL 2022)
ACL