Lecture image placeholder

Premium content

Access to this content requires a subscription. You must be a premium user to view this content.

Monthly subscription - $9.99Pay per view - $4.99Access through your institutionLogin with Underline account
Need help?
Contact us
Lecture placeholder background

ACL-IJCNLP 2021

August 02, 2021

Thailand

Would you like to see your presentation here, made available to a global audience of researchers?
Add your own presentation or have us affordably record your next conference.

keywords:

long-document, retrospective, recurrence mechanism

Transformers are not suited for processing long documents, due to their quadratically increasing memory and time consumption. Simply truncating a long document or applying the sparse attention mechanism will incur the context fragmentation problem or lead to an inferior modeling capability against comparable model sizes. In this paper, we propose ERNIE-Doc, a document-level language pretraining model based on Recurrence Transformers. Two well-designed techniques, namely the retrospective feed mechanism and the enhanced recurrence mechanism, enable ERNIE-Doc, which has a much longer effective context length, to capture the contextual information of a complete document. We pretrain ERNIE-Doc to explicitly learn the relationships among segments with an additional document-aware segment-reordering objective. Various experiments were conducted on both English and Chinese document-level tasks. ERNIE-Doc improved the state-of-the-art language modeling result of perplexity to 16.8 on WikiText-103. Moreover, it outperformed competitive pretraining models by a large margin on most language understanding tasks, such as text classification and question answering.

Downloads

SlidesPaper

Next from ACL-IJCNLP 2021

OntoED: Low-resource Event Detection with Ontology Embedding
technical paper

OntoED: Low-resource Event Detection with Ontology Embedding

ACL-IJCNLP 2021

Ningyu ZhangShumin Deng
Shumin Deng and 1 other author

02 August 2021

Similar lecture

Parameter Selection: Why We Should Pay More Attention to It
poster

Parameter Selection: Why We Should Pay More Attention to It

ACL-IJCNLP 2021

+1Si-An Chen
Si-An Chen and 3 other authors

02 August 2021