A Unified Multimodal Framework for Dataset Construction and Model-Based Diagnosis of Ameloblastoma
AI 摘要
构建多模态数据集,开发AI模型辅助成釉细胞瘤诊断与治疗决策。
主要贡献
- 构建了成釉细胞瘤多模态数据集
- 开发了基于多模态数据的深度学习模型
- 提高了成釉细胞瘤的分类和检测精度
方法论
采用NLP提取临床特征,进行图像预处理和增强。构建深度学习模型进行多模态数据融合和预测。
原文摘要
Artificial intelligence (AI)-enabled diagnostics in maxillofacial pathology require structured, high-quality multimodal datasets. However, existing resources provide limited ameloblastoma coverage and lack the format consistency needed for direct model training. We present a newly curated multimodal dataset specifically focused on ameloblastoma, integrating annotated radiological, histopathological, and intraoral clinical images with structured data derived from case reports. Natural language processing techniques were employed to extract clinically relevant features from textual reports, while image data underwent domain specific preprocessing and augmentation. Using this dataset, a multimodal deep learning model was developed to classify ameloblastoma variants, assess behavioral patterns such as recurrence risk, and support surgical planning. The model is designed to accept clinical inputs such as presenting complaint, age, and gender during deployment to enhance personalized inference. Quantitative evaluation demonstrated substantial improvements; variant classification accuracy increased from 46.2 percent to 65.9 percent, and abnormal tissue detection F1-score improved from 43.0 percent to 90.3 percent. Benchmarked against resources like MultiCaRe, this work advances patient-specific decision support by providing both a robust dataset and an adaptable multimodal AI framework.