Feb. 10, 2024, 4:42 p.m. | Vyacheslav Efimov

Towards Data Science - Medium towardsdatascience.com

Large Language Models, GPT-2 — Language Models Are Unsupervised Multitask Learners

Acing GPT capabilities by turning it into a powerful multitask zero-shot model

Introduction

GPT is a well-known series of models whose last versions are currently dominating in various NLP tasks. The first GPT version was a significant milestone: being trained on enormous 120M parameters, this model demonstrated state-of-the-art performance on top benchmarks. Starting from this point, researchers tried to improve the base version.

In 2019, researchers from OpenAI officially …

capabilities gpt gpt-2 language language models large language large language models machine learning nlp series tasks transformers unsupervised versions zero-shot

Doctoral Researcher (m/f/div) in Automated Processing of Bioimages

@ Leibniz Institute for Natural Product Research and Infection Biology (Leibniz-HKI) | Jena

Research Scholar (Technical Research)

@ Centre for the Governance of AI | Hybrid; Oxford, UK

Lead Software Engineer, Machine Learning

@ Monarch Money | Remote (US)

Investigator, Data Science

@ GSK | Stevenage

Alternance - Assistant.e Chef de Projet Data Business Intelligence (H/F)

@ Pernod Ricard | FR - Paris - The Island

Expert produit Big Data & Data Science - Services Publics - Nantes

@ Sopra Steria | Nantes, France