all AI news
Dwell in the Beginning: How Language Models Embed Long Documents for Dense Retrieval
April 8, 2024, 4:46 a.m. | Jo\~ao Coelho, Bruno Martins, Jo\~ao Magalh\~aes, Jamie Callan, Chenyan Xiong
cs.CL updates on arXiv.org arxiv.org
Abstract: This study investigates the existence of positional biases in Transformer-based models for text representation learning, particularly in the context of web document retrieval. We build on previous research that demonstrated loss of information in the middle of input sequences for causal language models, extending it to the domain of representation learning. We examine positional biases at various stages of training for an encoder-decoder model, including language model pre-training, contrastive pre-training, and contrastive fine-tuning. Experiments with …
abstract arxiv biases build causal context cs.cl cs.ir document documents embed information language language models loss representation representation learning research retrieval study text transformer transformer-based models type web
More from arxiv.org / cs.CL updates on arXiv.org
Jobs in AI, ML, Big Data
Senior Machine Learning Engineer
@ GPTZero | Toronto, Canada
ML/AI Engineer / NLP Expert - Custom LLM Development (x/f/m)
@ HelloBetter | Remote
Doctoral Researcher (m/f/div) in Automated Processing of Bioimages
@ Leibniz Institute for Natural Product Research and Infection Biology (Leibniz-HKI) | Jena
Seeking Developers and Engineers for AI T-Shirt Generator Project
@ Chevon Hicks | Remote
Data Analyst (Salesforce)
@ Lisinski Law Firm | Latin America
Data Analyst
@ Fusemachines | India - Remote