LLM Memory & RAG 相关度: 7/10

Quecto-V1: Empirical Analysis of 8-bit Quantized Small Language Models for On-Device Legal Retrieval

Subrit Dikshit
arXiv: 2602.16640v1 发布: 2026-02-18 更新: 2026-02-18

AI 摘要

Quecto-V1是一个针对印度法律领域,使用8比特量化的专用小型语言模型,实现了高效的本地部署。

主要贡献

  • 设计并训练了针对印度法律领域的专用小型语言模型 Quecto-V1
  • 采用 8-bit 量化,将模型大小压缩到 150MB 以下,便于本地部署
  • 证明了在特定领域,量化后的专用模型可以优于通用模型

方法论

基于GPT-2架构,使用印度法律语料库从头开始训练,并采用后训练8比特量化(GGUF格式)进行模型压缩。

原文摘要

The rapid proliferation of Large Language Models (LLMs) has revolutionized Natural Language Processing (NLP) but has simultaneously created a "resource divide." State-of-the-art legal intelligence systems typically rely on massive parameter counts (7B+) and cloud-based inference, rendering them inaccessible to practitioners in resource-constrained environments and posing significant data sovereignty risks. This paper introduces Quecto-V1, a domain-specific Small Language Model (SLM) engineered to democratize access to Indian legal intelligence. Built upon a custom configuration of the GPT-2 architecture (124 million parameters), Quecto-V1 was trained from scratch exclusively on a corpus of Indian statutes, including the Indian Penal Code (IPC), the Code of Criminal Procedure (CrPC), and the Constitution of India. Unlike generalist models, which prioritize broad world knowledge, our approach maximizes "lexical density" within the legal domain. Furthermore, we address the deployment bottleneck by applying post-training 8-bit quantization (GGUF format), compressing the model to a memory footprint of under 150 MB. Our empirical analysis demonstrates that Quecto-V1 achieves high fidelity in retrieving statutory definitions and penal provisions, outperforming general-purpose SLMs in domain-specific exact match tasks while running entirely offline on consumer-grade CPUs. We further present an ablation study showing that 8-bit quantization yields a 74% reduction in model size with less than 3.5% degradation in retrieval accuracy compared to full-precision baselines. These findings suggest that for specialized, high-stakes domains like law, domain-specific training coupled with aggressive quantization offers a viable, privacy-preserving alternative to monolithic cloud models.

标签

小型语言模型 领域专用模型 量化 法律检索 本地部署

arXiv 分类

cs.CL