Aug. 3, 2023, noon | code_your_own_AI

code_your_own_AI www.youtube.com

A novel approach to "embodied AI" by integrating a large language model (LLM) with a Vision Transformer (ViT) and various sensor modalities, creating a single general-purpose multimodal language model: PaLM-E.

Opening up a new direction for embodied language models, highlighting the potential of these models in real-world applications and their ability to learn from experience and adapt to their environment.

PaLM-E: An Embodied Multimodal Language Model
https://palm-e.github.io/assets/palm-e.pdf
(all rights with authors)

#ai
#vision
#largelanguagemodels
#palm-e

applications embodied embodied ai general highlighting language language model language models large language large language model llm multimodal novel palm palm-e robotics robotics & ai sensor transformer vision vit world

Founding AI Engineer, Agents

@ Occam AI | New York

AI Engineer Intern, Agents

@ Occam AI | US

AI Research Scientist

@ Vara | Berlin, Germany and Remote

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne