SLIDE 12 Proposed Approach (cont’d)
(a) Pre-train the BERTSUM model on a generic abstractive summarization corpus.
[CLS] SentQ [SEP] [CLS] Sent1 [SEP] … [CLS] SentN [SEP] [CLS] Sent1 [SEP] [CLS] Sent2 [SEP] … [CLS] SentN [SEP]
Input: Document{Sent1, Sent2, ... SentN} Input: Query{SentQ}, Document{Sent1 ... SentN} Pre-train the BERTSUM model on a large generic abstractive summarization dataset. Incorporate query relevance into the pre-trained BERTSUM and fine-tune on the target domain. Transfer Learning
Query: What is the benefit of reality shows? Document: Even if reality shows were not enlightening, they generate massive revenues that can be used for funding more sophisticated programs. Take BBC for example, it
entertaining reality shows such as total wipeout as well as brilliant documentaries. Summary: Reality show generates revenues. Document: The argument that too evil can be prevented by assassination is highly questionable. The figurehead of an evil government is not necessarily the lynchpin that hold it together. Therefore, if Hitler had been assassinated, it is pure supposition that the Nazi would have acted any differently to how they did act. Summary: The idea that assassinations can prevent injustice is questionable.
Transformer Decoder BERT Encoder (b) Fine-tune the pre-trained model for the QFAS task on the target domain. Transformer Decoder BERT Encoder
12