Dec. 7, 2023, 1 a.m. | Adnan Hassan

MarkTechPost www.marktechpost.com

Researchers developed the CoDi-2 Multimodal Large Language Model (MLLM) from UC Berkeley, Microsoft Azure AI, Zoom, and UNC-Chapel Hill to address the problem of generating and understanding complex multimodal instructions, as well as excelling in subject-driven image generation, vision transformation, and audio editing tasks. This model represents a significant breakthrough in establishing a comprehensive multimodal […]


The post This AI Research Introduces CoDi-2: A Groundbreaking Multimodal Large Language Model Transforming the Landscape of Interleaved Instruction Processing and Multimodal Output Generation …

ai research ai shorts applications artificial intelligence azure azure ai berkeley computer vision editors pick groundbreaking hill image image generation landscape language language model large language large language model machine learning microsoft microsoft azure microsoft azure ai mllm multimodal multimodal large language model processing research researchers staff tech news technology uc berkeley understanding zoom

More from www.marktechpost.com / MarkTechPost

Founding AI Engineer, Agents

@ Occam AI | New York

AI Engineer Intern, Agents

@ Occam AI | US

AI Research Scientist

@ Vara | Berlin, Germany and Remote

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Consultant Senior Power BI & Azure - CDI - H/F

@ Talan | Lyon, France