April 12, 2024, 6:45 p.m. | /u/wantondevious

Machine Learning www.reddit.com

Hi,

I'm wondering if anyone has a survey of the easiest way to handle classification tasks where the input token space is >> 512 (or whatever the single GPU models are limited to).

I'm working in a complex space. I'm looking at a ranking (actually may even be simply binary, with a class imbalance) type problem, so not generative, but where the text's contents is important, not just some pooled version of the embeddings, and so I'd like to make …

bert classification documents gpu machinelearning ranking space style survey tasks token tokens

AI Research Scientist

@ Vara | Berlin, Germany and Remote

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne

Senior Machine Learning Engineer (MLOps)

@ Promaton | Remote, Europe

Senior Software Engineer, Generative AI (C++)

@ SoundHound Inc. | Toronto, Canada