Dec. 7, 2023, 1:12 p.m. | /u/dancingnightly

Natural Language Processing www.reddit.com

I am working on an application where we would like to improve our multilingual support. Right now, we use the \`multi-qa-MiniLM-L6-cos-v1 \` sometimes, and at other times, thinks like mcontriever, at other times, embeddings from large language models (e.g. GPT-J/20B/other more recent models).

It just seems though that since mcontriever (which is a bit too unwieldy) and the older multi-qa models, there aren't any fast and good models, unlike say the BGE / E5 models (which as far as I …

application embeddings generative gpt gpt-j language language models languagetechnology large language large language models multilingual sbert self-hosted sota support

Founding AI Engineer, Agents

@ Occam AI | New York

AI Engineer Intern, Agents

@ Occam AI | US

AI Research Scientist

@ Vara | Berlin, Germany and Remote

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne