LightOnOCR-1B-1025-AIO-GGUF

The LightOnOCR-1B-1025 from lightonai is a compact 1B-parameter end-to-end vision-language model optimized for Optical Character Recognition (OCR) and document understanding, combining a Pixtral-based Vision Transformer encoder with a lightweight Qwen3-based text decoder distilled from high-quality VLMs, achieving state-of-the-art accuracy in its class on Olmo-Bench (76.1% overall, with 81.4% on ArXiv papers, 71.6% on old scans, and strong math/table handling) while processing 5.71 pages/second on a single H100 GPUโ€”5ร— faster than dots.ocr, 2ร— faster than PaddleOCR-VL-0.9B, and 1.73ร— faster than DeepSeekOCR at under $0.01 per 1,000 pages. Fully differentiable without external OCR pipelines, it excels at layout-aware extraction from high-resolution PDFs (rendered at ~1540px longest dimension), handling tables, receipts, forms, multi-column layouts, math notation, and multilingual text (Latin-dominant) under Apache 2.0 license, with compact 32k/16k vocab variants for European languages and vLLM/Transformers support for efficient inference and LoRA fine-tuning.

LightOnOCR-1B-1025 [GGUF]

File Name Quant Type File Size File Link
LightOnOCR-1B-1025.BF16.gguf BF16 1.51 GB Download
LightOnOCR-1B-1025.F16.gguf F16 1.51 GB Download
LightOnOCR-1B-1025.F32.gguf F32 3.01 GB Download
LightOnOCR-1B-1025.Q8_0.gguf Q8_0 805 MB Download
LightOnOCR-1B-1025.mmproj-bf16.gguf mmproj-bf16 821 MB Download
LightOnOCR-1B-1025.mmproj-f16.gguf mmproj-f16 819 MB Download
LightOnOCR-1B-1025.mmproj-f32.gguf mmproj-f32 1.64 GB Download
LightOnOCR-1B-1025.mmproj-q8_0.gguf mmproj-q8_0 437 MB Download

Quants Usage

(sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants)

Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better):

image.png

Downloads last month
91
GGUF
Model size
0.8B params
Architecture
qwen3
Hardware compatibility
Log In to view the estimation

8-bit

16-bit

32-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for prithivMLmods/LightOnOCR-1B-1025-AIO-GGUF

Quantized
(8)
this model