all AI news
Towards A Unified View of Answer Calibration for Multi-Step Reasoning
Feb. 27, 2024, 5:44 a.m. | Shumin Deng, Ningyu Zhang, Nay Oo, Bryan Hooi
cs.LG updates on arXiv.org arxiv.org
Abstract: Large Language Models (LLMs) employing Chain-of-Thought (CoT) prompting have broadened the scope for improving multi-step reasoning capabilities. We generally divide multi-step reasoning into two phases: path generation to generate the reasoning path(s); and answer calibration post-processing the reasoning path(s) to obtain a final answer. However, the existing literature lacks systematic analysis on different answer calibration approaches. In this paper, we summarize the taxonomy of recent answer calibration techniques and break them down into step-level and …
abstract arxiv capabilities cs.ai cs.cl cs.ir cs.lg generate language language models large language large language models llms path post-processing processing prompting reasoning thought type view
More from arxiv.org / cs.LG updates on arXiv.org
Testing the Segment Anything Model on radiology data
2 days, 9 hours ago |
arxiv.org
Calorimeter shower superresolution
2 days, 9 hours ago |
arxiv.org
Jobs in AI, ML, Big Data
Software Engineer for AI Training Data (School Specific)
@ G2i Inc | Remote
Software Engineer for AI Training Data (Python)
@ G2i Inc | Remote
Software Engineer for AI Training Data (Tier 2)
@ G2i Inc | Remote
Data Engineer
@ Lemon.io | Remote: Europe, LATAM, Canada, UK, Asia, Oceania
Artificial Intelligence – Bioinformatic Expert
@ University of Texas Medical Branch | Galveston, TX
Lead Developer (AI)
@ Cere Network | San Francisco, US