April 15, 2024, 8 p.m. | code_your_own_AI

code_your_own_AI www.youtube.com

New LLM from Microsoft: WizardLM-2 8x22B, based on Mixtral 8x22B (by Mistral AI), but further fine-tuned and stage-DPO aligned plus RLEIF (instruction quality reward model (IRM) combined with the process supervision reward model (PRM)). Open-source?

HuggingFace link to WizardLM-2 8x22B LLM (fine-tuned and aligned):
https://huggingface.co/microsoft/WizardLM-2-8x22B

#airesearch
#airesearch
#microsoft

huggingface llm microsoft mistral mistral ai mixtral mixtral 8x22b process process supervision quality reward model stage supervision

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne

Senior Machine Learning Engineer (MLOps)

@ Promaton | Remote, Europe

Research Scientist (Computer Science)

@ Nanyang Technological University | NTU Main Campus, Singapore

Intern - Sales Data Management

@ Deliveroo | Dubai, UAE (Main Office)