LLM Reasoning 相关度: 7/10

Eliciting Numerical Predictive Distributions of LLMs Without Autoregression

Julianna Piskorz, Katarzyna Kobalczyk, Mihaela van der Schaar
arXiv: 2603.02913v1 发布: 2026-03-03 更新: 2026-03-03

AI 摘要

论文探索了无需自回归生成即可从LLM内部表征中提取数值预测分布的方法。

主要贡献

  • 提出利用探针从LLM内部表征预测数值分布统计量
  • 验证了LLM嵌入包含数值预测分布的关键信息
  • 探索了LLM内部编码数值任务不确定性的方式

方法论

训练回归探针,直接从LLM的内部表征预测数值输出分布的统计函数(例如,均值、中位数、分位数)。

原文摘要

Large Language Models (LLMs) have recently been successfully applied to regression tasks -- such as time series forecasting and tabular prediction -- by leveraging their in-context learning abilities. However, their autoregressive decoding process may be ill-suited to continuous-valued outputs, where obtaining predictive distributions over numerical targets requires repeated sampling, leading to high computational cost and inference time. In this work, we investigate whether distributional properties of LLM predictions can be recovered without explicit autoregressive generation. To this end, we study a set of regression probes trained to predict statistical functionals (e.g., mean, median, quantiles) of the LLM's numerical output distribution directly from its internal representations. Our results suggest that LLM embeddings carry informative signals about summary statistics of their predictive distributions, including the numerical uncertainty. This investigation opens up new questions about how LLMs internally encode uncertainty in numerical tasks, and about the feasibility of lightweight alternatives to sampling-based approaches for uncertainty-aware numerical predictions.

标签

LLM 预测分布 回归 内部表征 不确定性

arXiv 分类

cs.LG cs.AI