Skip to main content
Doclo

Industry · · 9 min read

Small OCR models are catching up. Here is what that means.

A new generation of open-source OCR now rivals cloud services on accuracy. For organizations that need to keep document data on-premise, the options just got significantly better.

Small OCR models are catching up. Here is what that means.

For the past several years, the choice for organizations processing documents at scale was straightforward: use a cloud OCR service from AWS, Google, or Microsoft. Open-source alternatives existed, but they struggled with tables, complex layouts, and anything beyond clean printed text. The accuracy gap was wide enough that the comparison barely warranted making.

That gap has closed. In some cases, it has reversed.

Three open-source OCR models released in early 2026 now match or exceed cloud services on the primary document parsing benchmarks. They run on standard hardware, support commercial use, and can be deployed entirely on-premise. For industries where document data cannot leave the building, this changes the conversation.

What happened

In March 2026, Zhipu AI released GLM-OCR, a compact model that scored highest on the OmniDocBench v1.5 document parsing benchmark, ahead of Gemini 3 Pro and GPT-5.2. A month earlier, Xiaohongshu's AI lab released dots.ocr-1.5, which halved the error rate of its predecessor across 126 languages. And Baidu shipped PaddleOCR v5, the latest version of the most widely deployed open-source OCR framework, with a mobile variant small enough to run on a phone.

These are not isolated results. October 2025 alone saw six major open-source OCR model releases. The underlying technology shifted from traditional step-by-step processing (detect text regions, then recognize characters, then assemble output) to models that read documents more like a person does, understanding layout, formatting, and structure simultaneously. That architectural shift is what closed the accuracy gap.

Three models, three profiles

Each model serves a different need. Understanding the tradeoffs matters more than the benchmark scores.

GLM-OCR is the accuracy leader. It tops the document parsing leaderboard while being small enough to run on a laptop. Compressed for deployment, it fits in roughly 500 megabytes. It excels at tables, structured data extraction (returning typed fields like amounts and dates, not just raw text), and seal recognition. It supports commercial use under an MIT license. The limitations: it only handles about eight languages well, and it struggles with degraded historical scans. It is also very new, released in March 2026, with limited production track record.

PaddleOCR v5 is the most proven. With over 72,000 stars on GitHub and years of deployment in banking, insurance, and logistics, it has the broadest real-world track record. Its mobile variant is just 21 megabytes and processes text in under 60 milliseconds on a standard processor. For phones and embedded devices, it remains the most practical option by a wide margin. GLM-OCR can technically run on recent flagship phones at around 1 GB compressed, but PaddleOCR's mobile pipeline is 50 times smaller and purpose-built for constrained hardware. Its newer vision-language model supports 111 languages and scores just behind GLM-OCR on benchmarks. The tradeoff is ecosystem friction: it runs on Baidu's own framework rather than the more common tools most AI teams already use, and documentation is primarily in Chinese.

dots.ocr-1.5 is the multilingual specialist. With support for 126 languages and the ability to process very high-resolution images natively, it is the strongest option for organizations handling documents across many countries and scripts. It can also convert charts and diagrams into vector graphics, a unique capability. The tradeoffs: it is the largest and slowest of the three, requires more powerful hardware, and has the least mature deployment tooling. It is also newer to the market and less battle-tested.

What this means for regulated industries

The significance of these models is not primarily about accuracy. It is about where they can run.

Cloud OCR services require sending documents to a third-party server. For many organizations, that is fine. For others, it creates compliance complexity that is expensive to manage and risky to get wrong. GDPR enforcement has totaled EUR 5.88 billion in cumulative fines since 2018. HIPAA violations carry penalties up to $1.5 million per category. These are not theoretical risks.

When OCR runs entirely on your own infrastructure, entire categories of compliance questions go away. No third-party data processing agreements. No ambiguity about where data traveled. No data transfer to audit.

In healthcare, patient records processed on an air-gapped hospital network never touch an external service. Claims documents can be parsed into structured fields (diagnosis codes, procedure codes, billing amounts) without any data leaving the facility.

In financial services, KYC documents, mortgage applications, and trade documentation all contain personally identifiable information that compliance frameworks scrutinize when sent to third parties. Processing locally eliminates that surface entirely. The speed advantage compounds too: under 100 milliseconds locally versus 1 to 3 seconds for a cloud round-trip, multiplied across thousands of documents per day.

In field operations, insurance adjusters, customs agents, and logistics workers can process documents on-site without reliable internet. OCR that runs in 60 milliseconds on a tablet is a fundamentally different capability than OCR that requires a network connection.

These use cases are not new. What is new is that the accuracy of on-premise options now matches the cloud services they are measured against. The deployment choice no longer requires an accuracy tradeoff.

The cost picture

Cloud OCR services charge roughly $1.50 per 1,000 pages for basic text extraction at lower volumes, with discounts at scale. Structured extraction (pulling out specific fields like names, amounts, and dates) costs $10 to $50 per 1,000 pages. Self-hosted OCR on GPU hardware costs roughly $0.14 per 1,000 pages with current open-source models.

At modest volumes, the difference is negligible and the operational convenience of a cloud service is worth the premium. At high volumes, the math shifts. An organization processing 10 million pages per month pays roughly $7,000 to $10,000 in cloud OCR fees for basic text extraction (after volume discounts from AWS or Google), and significantly more for structured extraction. The same basic workload self-hosted costs around $1,500 in compute. A 5 to 7x cost difference at scale, compounding monthly. The infrastructure investment is real (hardware, engineering time, monitoring), but the crossover point for high-utilization workloads typically arrives within 12 to 18 months.

All three models are free to use commercially. GLM-OCR and PaddleOCR use standard open-source licenses with no restrictions. dots.ocr-1.5 has a supplemental license that explicitly permits commercial and SaaS use.

The customization advantage

This is where the strategic case for open-source OCR becomes most compelling.

Cloud OCR services offer some customization. You can train custom models on Azure or build custom processors on Google. But you cannot change the underlying technology. You adapt to the vendor's system, not the other way around.

With open-source models, you can train the model on your exact document types. Your specific invoice layouts, your medical forms, your trade documentation formats. Research has demonstrated that this kind of domain-specific training can push field-level accuracy from 81% to 92% on invoices, cutting manual correction time by over 70%.

The practical approach is straightforward: collect 1,000 to 5,000 labeled examples of your document types, train the model, and implement a confidence-based workflow where high-confidence results go straight through and uncertain results route to human review.

Many organizations already have the training data they need without realizing it. Years of human-reviewed, corrected, and annotated documents sitting in existing systems (claims that were manually keyed, invoices that were verified by accounts payable, forms that were reviewed by underwriters) represent exactly the kind of labeled data that fine-tuning requires. That operational history is an asset. Organizations that have been processing documents manually for years are, in a real sense, better positioned to train a custom model than a startup with no document archive.

For organizations processing high volumes of repetitive document types (claims, loan applications, KYC packets), this is the path from "performs well on general benchmarks" to "performs well on our documents." It is also a path that cloud-only architectures cannot match with the same degree of control.

Where cloud services still lead

These models are not a drop-in replacement for cloud OCR in every scenario. Honesty about the gaps matters more than enthusiasm about the benchmarks.

Handwriting. Cloud providers have trained on enormous proprietary datasets of handwritten text. Open-source models are improving but remain inconsistent, particularly on cursive and stylized scripts. If your workflow involves significant handwritten content, cloud services are still the safer choice.

Rare languages. dots.ocr-1.5 covers 126 languages, but accuracy on less common scripts varies significantly. GLM-OCR only handles about eight languages well. If your documents span many languages, test carefully before committing.

Degraded documents. Old scans, faxes, photographs taken at angles. Cloud models with broader training data still handle these edge cases better. GLM-OCR scored only 37.6% on a test of historical scans, where top models reach roughly 79%.

Operational maturity. Cloud services come with SLAs, managed scaling, enterprise support, and monitoring built in. Self-hosting means your team owns the infrastructure, the updates, and the failure modes. For organizations without dedicated AI operations capacity, that overhead is significant.

Benchmarks are not production. Benchmarks measure accuracy on curated datasets. Production systems encounter coffee-stained documents, inconsistent scan quality, formats the model has never seen, and edge cases that no benchmark captures. Cloud services have years of exposure to this long tail. These newer models do not. The standard document parsing benchmark is already showing signs of saturation, meaning high scores tell you less than they appear to about real-world performance.

What to evaluate

The pace of improvement in open-source OCR is accelerating. Six major releases in October 2025. A new benchmark leader in March 2026. The interval between meaningful capability jumps is measured in weeks, not years.

For organizations considering their options, the evaluation approach is practical:

  1. Collect a representative sample of your actual production documents, including the difficult ones.
  2. Run them through your current OCR provider and through one or two of these open-source models.
  3. Compare field-level accuracy, not just overall text accuracy. A 99% character accuracy rate can still mean 5 to 10% of critical fields contain errors.
  4. Factor in your deployment requirements. If compliance mandates on-premise processing, that constraint narrows the field regardless of accuracy scores.
  5. Decide based on your numbers, not the benchmarks.

The accuracy is there. The deployment options are there. The licenses permit commercial use. The question is whether these models perform well enough on your specific documents, with your specific quality and format variation, to meet your operational threshold. That question can only be answered by testing.

For organizations where data sovereignty, processing costs at scale, or the ability to customize OCR for specific document types are priorities, these models have crossed the threshold from "interesting research" to "worth a serious pilot." The gap between open-source and cloud OCR is no longer about capability. It is about maturity, operational readiness, and fit for your specific use case.

Ready to solve your document challenges?

Talk to our team about how Doclo can fit into your workflow. No commitment, just a conversation.