AGCD: Agent-Guided Cross-Modal Decoding for Weather Forecasting
AI 摘要
AGCD提出一种利用多智能体和跨模态解码进行天气预报的框架,提升预测精度和物理一致性。
主要贡献
- 提出Agent-Guided Cross-modal Decoding (AGCD) 框架
- 利用MLLMs生成状态条件物理先验知识
- 设计跨模态区域交互解码,有效注入物理先验
方法论
使用多智能体生成物理先验,通过跨模态解码将先验知识注入到天气预报模型中,提高预测精度。
原文摘要
Accurate weather forecasting is more than grid-wise regression: it must preserve coherent synoptic structures and physical consistency of meteorological fields, especially under autoregressive rollouts where small one-step errors can amplify into structural bias. Existing physics-priors approaches typically impose global, once-for-all constraints via architectures, regularization, or NWP coupling, offering limited state-adaptive and sample-specific controllability at deployment. To bridge this gap, we propose Agent-Guided Cross-modal Decoding (AGCD), a plug-and-play decoding-time prior-injection paradigm that derives state-conditioned physics-priors from the current multivariate atmosphere and injects them into forecasters in a controllable and reusable way. Specifically, We design a multi-agent meteorological narration pipeline to generate state-conditioned physics-priors, utilizing MLLMs to extract various meteorological elements effectively. To effectively apply the priors, AGCD further introduce cross-modal region interaction decoding that performs region-aware multi-scale tokenization and efficient physics-priors injection to refine visual features without changing the backbone interface. Experiments on WeatherBench demonstrate consistent gains for 6-hour forecasting across two resolutions (5.625 degree and 1.40625 degree) and diverse backbones (generic and weather-specialized), including strictly causal 48-hour autoregressive rollouts that reduce early-stage error accumulation and improve long-horizon stability.