ReadOnce Transformers: Reusable Representations of Text for Transformers

Shih-Ting Lin
Shih-Ting Lin
Cited by: 0|Bibtex|Views4
Other Links: arxiv.org

Abstract:

While large-scale language models are extremely effective when directly fine-tuned on many end-tasks, such models learn to extract information and solve the task simultaneously from end-task supervision. This is wasteful, as the general problem of gathering information from a document is mostly task-independent and need not be re-learne...More

Code:

Data:

Full Text
Your rating :
0

 

Tags
Comments