all AI news
[R] Visualization-of-Thought Elicits Spatial Reasoning in Large Language Models
April 14, 2024, 6:14 p.m. | /u/SeawaterFlows
Machine Learning www.reddit.com
**Abstract**:
>Large language models (LLMs) have exhibited impressive performance in language comprehension and various reasoning tasks. However, their abilities in spatial reasoning, a crucial aspect of human cognition, remain relatively unexplored. Human possess a remarkable ability to create mental images of unseen objects and actions through a process known as **the Mind's Eye**, enabling the imagination of the unseen world. Inspired by this cognitive capacity, we propose **Visualization-of-Thought** (**VoT**) prompting. VoT aims to elicit spatial reasoning of LLMs …
abstract cognition create enabling however human images imagination language language models large language large language models llms machinelearning mind objects performance process reasoning spatial tasks through world
More from www.reddit.com / Machine Learning
Jobs in AI, ML, Big Data
Data Architect
@ University of Texas at Austin | Austin, TX
Data ETL Engineer
@ University of Texas at Austin | Austin, TX
Lead GNSS Data Scientist
@ Lurra Systems | Melbourne
Senior Machine Learning Engineer (MLOps)
@ Promaton | Remote, Europe
Software Engineer, Data Tools - Full Stack
@ DoorDash | Pune, India
Senior Data Analyst
@ Artsy | New York City