LLM Reasoning 相关度: 8/10

Separate Before You Compress: The WWHO Tokenization Architecture

Kusal Darshana
arXiv: 2603.25309v1 发布: 2026-03-26 更新: 2026-03-26

AI 摘要

提出WWHO架构和SGPE算法,优化Abugida文字的LLM分词效率,降低Token Tax。

主要贡献

  • 提出WWHO三层架构和SGPE算法
  • 针对复杂Abugida文字的LLM分词问题
  • 显著降低Sinhala和Devanagari文字的Token数量

方法论

设计SGPE算法,结合语言规则和统计压缩,在30M数据集上训练WWHO模型,并在测试集上评估TWR指标。

原文摘要

Current Large Language Models (LLMs) mostly use BPE (Byte Pair Encoding) based tokenizers, which are very effective for simple structured Latin scripts such as English. However, standard BPE tokenizers struggle to process complex Abugida scripts due to their structural complexity. The problem is that these tokenizers break complex conjuncts, which are multi-codepoint grapheme clusters, into meaningless sub-character units. This degrades the LLM's reasoning efficiency by forcing it to learn basic orthographic structures at inference time and raises inference costs, resulting in a significant "Token Tax" for the Global South. We propose a new three-layer architecture, the WWHO (Where-What-How Often), and an algorithm named SGPE (Syllable-aware Grapheme Pair Encoding) that separates the linguistic rules of the script from the statistical compression process while enabling seamless multilingual tokenization. Using Sinhala and Devanagari (Hindi/Sanskrit) as highly complex Abugida scripts, we trained WWHO on a cleaned 30-million-sentence dataset and evaluated on a 1,499,950-sentence test set. For Sinhala, SGPE achieves a Token to Word Ratio (TWR) of 1.274 with 4.83 characters per token, representing a 61.7 percent reduction in tokens compared to OpenAI's o200k base. For Hindi, it achieves a TWR of 1.181 (27.0 percent reduction vs o200k). On the mixed-script (Sinhala, Devanagari, and English) dataset, SGPE achieves an overall TWR of 1.240, representing token reductions of 36.7 percent, 39.6 percent, and 60.2 percent relative to o200k base, Llama 4 Scout, and DeepSeek V3, respectively. This effectively extends the usable context window by up to 4.38 times for these Abugida languages while ensuring a Linguistic Zero-Breakage Guarantee, which ensures that no valid syllable is ever split across multiple tokens.

标签

LLM Tokenizer Abugida 分词 语言模型

arXiv 分类

cs.CL