LLM Reasoning 相关度: 8/10

LinGO: A Linguistic Graph Optimization Framework with LLMs for Interpreting Intents of Online Uncivil Discourse

Yuan Zhang, Thales Bertaglia
arXiv: 2602.04693v1 发布: 2026-02-04 更新: 2026-02-04

AI 摘要

LinGO利用语言图优化LLM,提升在线不文明言论意图识别准确性。

主要贡献

  • 提出了LinGO框架,用于多类意图不文明言论分类。
  • 分解语言为多步语言成分,针对性优化错误步骤。
  • 验证了RAG结合Gemini模型在不文明言论识别上的有效性。

方法论

LinGO将语言分解为多步语言成分,利用优化技术改进LLM的prompt和示例,提升意图识别的准确率。

原文摘要

Detecting uncivil language is crucial for maintaining safe, inclusive, and democratic online spaces. Yet existing classifiers often misinterpret posts containing uncivil cues but expressing civil intents, leading to inflated estimates of harmful incivility online. We introduce LinGO, a linguistic graph optimization framework for large language models (LLMs) that leverages linguistic structures and optimization techniques to classify multi-class intents of incivility that use various direct and indirect expressions. LinGO decomposes language into multi-step linguistic components, identifies targeted steps that cause the most errors, and iteratively optimizes prompt and/or example components for targeted steps. We evaluate it using a dataset collected during the 2022 Brazilian presidential election, encompassing four forms of political incivility: Impoliteness (IMP), Hate Speech and Stereotyping (HSST), Physical Harm and Violent Political Rhetoric (PHAVPR), and Threats to Democratic Institutions and Values (THREAT). Each instance is annotated with six types of civil/uncivil intent. We benchmark LinGO using three cost-efficient LLMs: GPT-5-mini, Gemini 2.5 Flash-Lite, and Claude 3 Haiku, and four optimization techniques: TextGrad, AdalFlow, DSPy, and Retrieval-Augmented Generation (RAG). The results show that, across all models, LinGO consistently improves accuracy and weighted F1 compared with zero-shot, chain-of-thought, direct optimization, and fine-tuning baselines. RAG is the strongest optimization technique and, when paired with Gemini model, achieves the best overall performance. These findings demonstrate that incorporating multi-step linguistic components into LLM instructions and optimize targeted components can help the models explain complex semantic meanings, which can be extended to other complex semantic explanation tasks in the future.

标签

不文明言论检测 语言图优化 LLM 意图识别 RAG

arXiv 分类

cs.CL cs.CY