all AI news
Drop your Decoder: Pre-training with Bag-of-Word Prediction for Dense Passage Retrieval
April 23, 2024, 4:50 a.m. | Guangyuan Ma, Xing Wu, Zijia Lin, Songlin Hu
cs.CL updates on arXiv.org arxiv.org
Abstract: Masked auto-encoder pre-training has emerged as a prevalent technique for initializing and enhancing dense retrieval systems. It generally utilizes additional Transformer decoder blocks to provide sustainable supervision signals and compress contextual information into dense representations. However, the underlying reasons for the effectiveness of such a pre-training technique remain unclear. The usage of additional Transformer-based decoders also incurs significant computational costs. In this study, we aim to shed light on this issue by revealing that masked …
abstract arxiv auto bag cs.cl cs.ir decoder encoder however information prediction pre-training retrieval supervision sustainable systems training transformer transformer decoder type word
More from arxiv.org / cs.CL updates on arXiv.org
Jobs in AI, ML, Big Data
AI Research Scientist
@ Vara | Berlin, Germany and Remote
Data Architect
@ University of Texas at Austin | Austin, TX
Data ETL Engineer
@ University of Texas at Austin | Austin, TX
Lead GNSS Data Scientist
@ Lurra Systems | Melbourne
Lead Data Scientist, Commercial Analytics
@ Checkout.com | London, United Kingdom
Data Engineer I
@ Love's Travel Stops | Oklahoma City, OK, US, 73120