AI Agents 相关度: 6/10

Compact Circulant Layers with Spectral Priors

Joseph Margaryan, Thomas Hamelryck
arXiv: 2602.21965v1 发布: 2026-02-25 更新: 2026-02-25

AI 摘要

研究紧凑的谱循环层及其变体,利用频域参数化实现高效神经网络和鲁棒性诊断。

主要贡献

  • 提出紧凑的谱循环层和BCCB层
  • 利用频域参数化实现结构化变分推断和精确谱范数计算
  • 提出实坐标下的 Hermitian-aware 低秩加对角变分后验

方法论

通过频域参数化滤波器,施加谱结构,并利用高斯过程先验构建变分后验,实现高效的贝叶斯神经网络。

原文摘要

Critical applications in areas such as medicine, robotics and autonomous systems require compact (i.e., memory efficient), uncertainty-aware neural networks suitable for edge and other resource-constrained deployments. We study compact spectral circulant and block-circulant-with-circulant-blocks (BCCB) layers: FFT-diagonalizable circular convolutions whose weights live directly in the real FFT (RFFT) half (1D) or half-plane (2D). Parameterizing filters in the frequency domain lets us impose simple spectral structure, perform structured variational inference in a low-dimensional weight space, and calculate exact layer spectral norms, enabling inexpensive global Lipschitz bounds and margin-based robustness diagnostics. By placing independent complex Gaussians on the Hermitian support we obtain a discrete instance of the spectral representation of stationary kernels, inducing an exact stationary Gaussian-process prior over filters on the discrete circle/torus. We exploit this to define a practical spectral prior and a Hermitian-aware low-rank-plus-diagonal variational posterior in real coordinates. Empirically, spectral circulant/BCCB layers are effective compact building blocks in both (variational) Bayesian and point estimate regimes: compact Bayesian neural networks on MNIST->Fashion-MNIST, variational heads on frozen CIFAR-10 features, and deterministic ViT projections on CIFAR-10/Tiny ImageNet; spectral layers match strong baselines while using substantially fewer parameters and with tighter Lipschitz certificates.

标签

循环神经网络 频域参数化 变分推断 模型压缩 鲁棒性

arXiv 分类

cs.LG