Abstract
Numerous applications of large language models (LLMs) rely on their ability to perform step-by-step reasoning. However, the reasoning behavior of LLMs remains poorly understood, posing challenges to research, development, and safety. To address this gap, we introduce landscape of thoughts-the first visualization tool for users to inspect the reasoning paths of chain-of-thought and its derivatives on any multi-choice dataset. Specifically, we represent the states in a reasoning path as feature vectors that quantify their distances to all answer choices. These features are then visualized in two-dimensional plots using t-SNE. Qualitative analysis shows that the landscape of thoughts effectively distinguishes between strong and weak models, correct and incorrect answers, as well as different reasoning tasks. It also uncovers undesirable reasoning patterns, such as low consistency and high uncertainty. Additionally, users can adapt our tool to a neural model that predicts any property they observe. We showcase this advantage by adapting our tool to a lightweight verifier, which significantly improves reasoning by evaluating the correctness of reasoning paths.
Original language | English |
---|---|
Title of host publication | ICLR 2025 Workshop on Reasoning and Planning for Large Language Models |
Publisher | International Conference on Learning Representations |
Pages | 1-22 |
Number of pages | 22 |
Publication status | Published - Apr 2025 |
Event | ICLR 2025 Workshop on Reasoning and Planning for Large Language Models - , Singapore Duration: 27 Apr 2025 → 27 Apr 2025 https://openreview.net/group?id=ICLR.cc/2025/Workshop/LLM_Reason_and_Plan#tab-accept |
Workshop
Workshop | ICLR 2025 Workshop on Reasoning and Planning for Large Language Models |
---|---|
Country/Territory | Singapore |
Period | 27/04/25 → 27/04/25 |
Internet address |