LinguistAgent: A Reflective Multi-Model Platform for Automated Linguistic Annotation
AI 摘要
LinguistAgent是一个自动化语言标注平台,通过多模型架构和双Agent机制,提升复杂语义任务的标注效率。
主要贡献
- 提出了一个基于反射式多模型架构的自动化语言标注平台LinguistAgent
- 实现了双Agent(Annotator和Reviewer)工作流,模拟同行评审过程
- 支持Prompt Engineering、RAG和Fine-tuning三种标注范式的比较实验
方法论
采用了双Agent架构,模拟专家评审流程,结合Prompt工程、RAG和微调等技术,在隐喻识别任务上进行评估。
原文摘要
Data annotation remains a significant bottleneck in the Humanities and Social Sciences, particularly for complex semantic tasks such as metaphor identification. While Large Language Models (LLMs) show promise, a significant gap remains between the theoretical capability of LLMs and their practical utility for researchers. This paper introduces LinguistAgent, an integrated, user-friendly platform that leverages a reflective multi-model architecture to automate linguistic annotation. The system implements a dual-agent workflow, comprising an Annotator and a Reviewer, to simulate a professional peer-review process. LinguistAgent supports comparative experiments across three paradigms: Prompt Engineering (Zero/Few-shot), Retrieval-Augmented Generation, and Fine-tuning. We demonstrate LinguistAgent's efficacy using the task of metaphor identification as an example, providing real-time token-level evaluation (Precision, Recall, and $F_1$ score) against human gold standards. The application and codes are released on https://github.com/Bingru-Li/LinguistAgent.