×
REPT: Bridging Language Models and Machine Reading Comprehension via Retrieval-Based Pre-training · Fangkai Jiao, Yangyang Guo, Yilin Niu, Feng Ji, ...
May 10, 2021 · We present REPT, a REtrieval-based Pre-Training approach. In particular, we introduce two self-supervised tasks to strengthen evidence extraction during pre- ...
This work introduces two self-supervised tasks to strengthen evidence extraction during pre-training, which is further inherited by downstream MRC tasks ...
Aug 1, 2021 · In this paper, we present a novel pre-training ap- proach, REPT, to bridge the gap between pre- trained language models and machine reading com-.
This is the repo for the paper REPT: Bridging Language Models and Machine Reading Comprehension via Retrieval-Based Pre-training (ACL-IJCNLP 2021: Findings).
The proposed model can be applied to other machine reading comprehension models through pretraining. To enhance the performance of the proposed machine reading ...
May 31, 2021 · Bibliographic details on REPT: Bridging Language Models and Machine Reading Comprehension via Retrieval-Based Pre-training.
REPT: Bridging Language Models and Machine Reading Comprehension via Retrieval-Based Pre-training. Fangkai Jiao, Yangyang Guo, Yilin Niu, Feng Ji, Feng-Lin Li ...
This project focuses on enhancing open-source large language models through instruction-tuning and providing comprehensive evaluations of their performance.
Co-authors ; REPT: Bridging language models and machine reading comprehension via retrieval-based pre-training. F Jiao, Y Guo, Y Niu, F Ji, FL Li, L Nie.