Few-shot Writer Adaptation via Multimodal In-Context Learning
AI 摘要
提出了一种基于多模态上下文学习的少样本手写体风格迁移方法,无需参数更新即可实现。
主要贡献
- 提出了一种上下文驱动的HTR框架
- 设计了一个紧凑的CNN-Transformer模型
- 验证了结合上下文驱动和标准OCR训练策略的有效性
方法论
利用多模态上下文学习,通过少量目标书写者的样本,在推理时实现风格迁移,无需梯度计算和参数更新。
原文摘要
While state-of-the-art Handwritten Text Recognition (HTR) models perform well on standard benchmarks, they frequently struggle with writers exhibiting highly specific styles that are underrepresented in the training data. To handle unseen and atypical writers, writer adaptation techniques personalize HTR models to individual handwriting styles. Leading writer adaptation methods require either offline fine-tuning or parameter updates at inference time, both involving gradient computation and backpropagation, which increase computational costs and demand careful hyperparameter tuning. In this work, we propose a novel context-driven HTR framework3 inspired by multimodal in-context learning, enabling inference-time writer adaptation using only a few examples from the target writer without any parameter updates. We further demonstrate the impact of context length, design a compact 8M-parameter CNN-Transformer that enables few-shot in-context adaptation, and show that combining context-driven and standard OCR training strategies leads to complementary improvements. Experiments on IAM and RIMES validate our approach with Character Error Rates of 3.92% and 2.34%, respectively, surpassing all writer-independent HTR models without requiring any parameter updates at inference time.