all AI news
Hyper-CL: Conditioning Sentence Representations with Hypernetworks
March 15, 2024, 4:48 a.m. | Young Hyun Yoo, Jii Cha, Changhyeon Kim, Taeuk Kim
cs.CL updates on arXiv.org arxiv.org
Abstract: While the introduction of contrastive learning frameworks in sentence representation learning has significantly contributed to advancements in the field, it still remains unclear whether state-of-the-art sentence embeddings can capture the fine-grained semantics of sentences, particularly when conditioned on specific perspectives. In this paper, we introduce Hyper-CL, an efficient methodology that integrates hypernetworks with contrastive learning to compute conditioned sentence representations. In our proposed approach, the hypernetwork is responsible for transforming pre-computed condition embeddings into corresponding …
abstract art arxiv contributed cs.cl embeddings fine-grained frameworks introduction methodology paper perspectives representation representation learning semantics state type
More from arxiv.org / cs.CL updates on arXiv.org
Jobs in AI, ML, Big Data
Lead Developer (AI)
@ Cere Network | San Francisco, US
Research Engineer
@ Allora Labs | Remote
Ecosystem Manager
@ Allora Labs | Remote
Founding AI Engineer, Agents
@ Occam AI | New York
AI Engineer Intern, Agents
@ Occam AI | US
AI Research Scientist
@ Vara | Berlin, Germany and Remote