QuPAINT: Physics-Aware Instruction Tuning Approach to Quantum Material Discovery
AI 摘要
提出QuPAINT框架,利用物理先验知识提升量子材料光学图像识别能力。
主要贡献
- 提出Synthia物理驱动的合成数据生成器
- 构建QMat-Instruct大规模量子材料指令数据集
- 提出Physics-Aware Instruction Tuning (QuPAINT)框架
- 构建QF-Bench综合评测基准
方法论
利用物理先验生成合成数据,构建指令数据集微调多模态大模型,并设计物理信息注意力模块融合视觉特征。
原文摘要
Characterizing two-dimensional quantum materials from optical microscopy images is challenging due to the subtle layer-dependent contrast, limited labeled data, and significant variation across laboratories and imaging setups. Existing vision models struggle in this domain since they lack physical priors and cannot generalize to new materials or hardware conditions. This work presents a new physics-aware multimodal framework that addresses these limitations from both the data and model perspectives. We first present Synthia, a physics-based synthetic data generator that simulates realistic optical responses of quantum material flakes under thin-film interference. Synthia produces diverse and high-quality samples, helping reduce the dependence on expert manual annotation. We introduce QMat-Instruct, the first large-scale instruction dataset for quantum materials, comprising multimodal, physics-informed question-answer pairs designed to teach Multimodal Large Language Models (MLLMs) to understand the appearance and thickness of flakes. Then, we propose Physics-Aware Instruction Tuning (QuPAINT), a multimodal architecture that incorporates a Physics-Informed Attention module to fuse visual embeddings with optical priors, enabling more robust and discriminative flake representations. Finally, we establish QF-Bench, a comprehensive benchmark spanning multiple materials, substrates, and imaging settings, offering standardized protocols for fair and reproducible evaluation.