all AI news
Drop your Decoder: Pre-training with Bag-of-Word Prediction for Dense Passage Retrieval
April 23, 2024, 4:50 a.m. | Guangyuan Ma, Xing Wu, Zijia Lin, Songlin Hu
cs.CL updates on arXiv.org arxiv.org
Abstract: Masked auto-encoder pre-training has emerged as a prevalent technique for initializing and enhancing dense retrieval systems. It generally utilizes additional Transformer decoder blocks to provide sustainable supervision signals and compress contextual information into dense representations. However, the underlying reasons for the effectiveness of such a pre-training technique remain unclear. The usage of additional Transformer-based decoders also incurs significant computational costs. In this study, we aim to shed light on this issue by revealing that masked …
abstract arxiv auto bag cs.cl cs.ir decoder encoder however information prediction pre-training retrieval supervision sustainable systems training transformer transformer decoder type word
More from arxiv.org / cs.CL updates on arXiv.org
Jobs in AI, ML, Big Data
Artificial Intelligence – Bioinformatic Expert
@ University of Texas Medical Branch | Galveston, TX
Lead Developer (AI)
@ Cere Network | San Francisco, US
Research Engineer
@ Allora Labs | Remote
Ecosystem Manager
@ Allora Labs | Remote
Founding AI Engineer, Agents
@ Occam AI | New York
AI Engineer Intern, Agents
@ Occam AI | US