Feb. 21, 2022, 11:36 a.m. | Tim Schopf

Towards Data Science - Medium towardsdatascience.com

An introduction to multiprocessing predictions of large machine learning and deep learning models

Image by Slejven Djurakovic on Unsplash

The current trend in AI research is moving towards the development of increasingly larger Deep Learning models, which constantly surpass each other in terms of performance. Recent examples in Natural Language Processing (NLP) are GPT-3, XLNet or the classical BERT transformer models. While the ever-improving results are inspiring researchers enthusiastically to research even larger models, this development also has a …

hugging face nlp parallel-computing text classification transformers

Lead Developer (AI)

@ Cere Network | San Francisco, US

Research Engineer

@ Allora Labs | Remote

Ecosystem Manager

@ Allora Labs | Remote

Founding AI Engineer, Agents

@ Occam AI | New York

AI Engineer Intern, Agents

@ Occam AI | US

AI Research Scientist

@ Vara | Berlin, Germany and Remote