April 12, 2024, 6:45 p.m. | /u/wantondevious

Machine Learning www.reddit.com

Hi,

I'm wondering if anyone has a survey of the easiest way to handle classification tasks where the input token space is >> 512 (or whatever the single GPU models are limited to).

I'm working in a complex space. I'm looking at a ranking (actually may even be simply binary, with a class imbalance) type problem, so not generative, but where the text's contents is important, not just some pooled version of the embeddings, and so I'd like to make …

bert classification documents gpu machinelearning ranking space style survey tasks token tokens

Artificial Intelligence – Bioinformatic Expert

@ University of Texas Medical Branch | Galveston, TX

Lead Developer (AI)

@ Cere Network | San Francisco, US

Research Engineer

@ Allora Labs | Remote

Ecosystem Manager

@ Allora Labs | Remote

Founding AI Engineer, Agents

@ Occam AI | New York

AI Engineer Intern, Agents

@ Occam AI | US