Feb. 12, 2024, 5:43 a.m. | Xidong Feng Ziyu Wan Muning Wen Stephen Marcus McAleer Ying Wen Weinan Zhang Jun Wang

cs.LG updates on arXiv.org arxiv.org

Recent works like Tree-of-Thought (ToT) and Reasoning via Planning (RAP) aim to augment the reasoning capabilities of LLMs by using tree-search algorithms to guide multi-step reasoning. These methods rely on prompting a pre-trained model to serve as a value function and focus on problems with low search depth. As a result, these methods will not work in domains where the pre-trained LLM does not have enough knowledge to serve as an effective value function or in domains that require long-horizon …

aim algorithms alphazero capabilities cs.ai cs.cl cs.lg decoding focus function guide language language model large language large language model llms low planning prompting rap reasoning search search algorithms serve thought training tree value via

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne

Senior Machine Learning Engineer (MLOps)

@ Promaton | Remote, Europe

Business Data Scientist, gTech Ads

@ Google | Mexico City, CDMX, Mexico

Lead, Data Analytics Operations

@ Zocdoc | Pune, Maharashtra, India