ocr / CLAUDE.md
davanstrien's picture
davanstrien HF Staff
Fix dots-ocr-1.5.py for v1.5 model + document bbox coords
37431e4

OCR Scripts - Development Notes

Active Scripts

DeepSeek-OCR v1 (deepseek-ocr-vllm.py)

Production Ready (Fixed 2026-02-12)

Known issue (vLLM nightly, 2026-02-12): Some images trigger a crop dimension validation error:

ValueError: images_crop dim[2] expected 1024, got 640. Expected shape: ('bnp', 3, 1024, 1024), but got torch.Size([0, 3, 640, 640])

This is a vLLM bug: the preprocessor defaults to gundam mode (image_size=640), but the tensor validator expects 1024x1024 even when the crop batch is empty (dim 0). Hit 2/10 on davanstrien/ufo-ColPali, 0/10 on NLS Medical History. Likely depends on image aspect ratios. No upstream issue filed yet. Related feature request: vllm#28160 (no way to control resolution mode via mm-processor-kwargs).

LightOnOCR-2-1B (lighton-ocr2.py)

Production Ready (Fixed 2026-01-29)

Status: Working with vLLM nightly

What was fixed:

  • Root cause was NOT vLLM - it was the deprecated HF_HUB_ENABLE_HF_TRANSFER=1 env var
  • The script was setting this env var but hf_transfer package no longer exists
  • This caused download failures that manifested as "Can't load image processor" errors
  • Fix: Removed the HF_HUB_ENABLE_HF_TRANSFER=1 setting from the script

Test results (2026-01-29):

  • 10/10 samples processed successfully
  • Clean markdown output with proper headers and paragraphs
  • Output dataset: davanstrien/lighton-ocr2-test-v4

Example usage:

hf jobs uv run --flavor a100-large \
    -s HF_TOKEN \
    https://huggingface.co/datasets/uv-scripts/ocr/raw/main/lighton-ocr2.py \
    davanstrien/ufo-ColPali output-dataset \
    --max-samples 10 --shuffle --seed 42

Model Info:

  • Model: lightonai/LightOnOCR-2-1B
  • Architecture: Pixtral ViT encoder + Qwen3 LLM
  • Training: RLVR (Reinforcement Learning with Verifiable Rewards)
  • Performance: 83.2% on OlmOCR-Bench, 42.8 pages/sec on H100

PaddleOCR-VL-1.5 (paddleocr-vl-1.5.py)

Production Ready (Added 2026-01-30)

Status: Working with transformers

Note: Uses transformers backend (not vLLM) because PaddleOCR-VL only supports vLLM in server mode, which doesn't fit the single-command UV script pattern. Images are processed one at a time for stability.

Test results (2026-01-30):

  • 10/10 samples processed successfully
  • Processing time: ~50s per image on L4 GPU
  • Output dataset: davanstrien/paddleocr-vl15-final-test

Example usage:

hf jobs uv run --flavor l4x1 \
    -s HF_TOKEN \
    https://huggingface.co/datasets/uv-scripts/ocr/raw/main/paddleocr-vl-1.5.py \
    davanstrien/ufo-ColPali output-dataset \
    --max-samples 10 --shuffle --seed 42

Task modes:

  • ocr (default): General text extraction to markdown
  • table: Table extraction to HTML format
  • formula: Mathematical formula recognition to LaTeX
  • chart: Chart and diagram analysis
  • spotting: Text spotting with localization (uses higher resolution)
  • seal: Seal and stamp recognition

Model Info:

  • Model: PaddlePaddle/PaddleOCR-VL-1.5
  • Size: 0.9B parameters (ultra-compact)
  • Performance: 94.5% SOTA on OmniDocBench v1.5
  • Backend: Transformers (single image processing)
  • Requires: transformers>=5.0.0

DoTS.ocr-1.5 (dots-ocr-1.5.py)

Production Ready (Fixed 2026-03-14)

Status: Working with vLLM 0.17.1 stable

Model availability: The v1.5 model is NOT on HF from the original authors. We mirrored it from ModelScope to davanstrien/dots.ocr-1.5. Original: https://modelscope.cn/models/rednote-hilab/dots.ocr-1.5. License: MIT-based (with supplementary terms for responsible use).

Key fix (2026-03-14): Must pass chat_template_content_format="string" to llm.chat(). The model's tokenizer_config.json chat template expects string content (not openai-format lists). Without this, the model generates empty output (~1 token then EOS). The separate chat_template.json file handles multimodal content but vLLM uses the tokenizer_config template by default.

Bbox coordinate system (layout modes): Bounding boxes from layout-all and layout-only modes are in the resized image coordinate space, not original image coordinates. The model uses Qwen2VLImageProcessor which resizes images via smart_resize():

  • max_pixels=11,289,600, factor=28 (patch_size=14 × merge_size=2)
  • Images are scaled down so w×h ≤ max_pixels, dims rounded to multiples of 28
  • To map bboxes back to original image coordinates:
import math

def smart_resize(height, width, factor=28, min_pixels=3136, max_pixels=11289600):
    h_bar = max(factor, round(height / factor) * factor)
    w_bar = max(factor, round(width / factor) * factor)
    if h_bar * w_bar > max_pixels:
        beta = math.sqrt((height * width) / max_pixels)
        h_bar = math.floor(height / beta / factor) * factor
        w_bar = math.floor(width / beta / factor) * factor
    elif h_bar * w_bar < min_pixels:
        beta = math.sqrt(min_pixels / (height * width))
        h_bar = math.ceil(height * beta / factor) * factor
        w_bar = math.ceil(width * beta / factor) * factor
    return h_bar, w_bar

resized_h, resized_w = smart_resize(orig_h, orig_w)
scale_x = orig_w / resized_w
scale_y = orig_h / resized_h
# Then: orig_x = bbox_x * scale_x, orig_y = bbox_y * scale_y

Test results (2026-03-14):

  • 3/3 samples on L4: OCR mode working, ~147 toks/s output
  • 3/3 samples on L4: layout-all mode working, structured JSON with bboxes
  • 10/10 samples on A100: layout-only mode on NLS Highland News, ~670 toks/s output
  • Output datasets: davanstrien/dots-ocr-1.5-smoke-test-v3, davanstrien/dots-ocr-1.5-layout-test, davanstrien/dots-ocr-1.5-nls-layout-test

Prompt modes:

  • ocr — text extraction (default)
  • layout-all — layout + bboxes + categories + text (JSON)
  • layout-only — layout + bboxes + categories only (JSON)
  • web-parsing — webpage layout analysis (JSON) [new in v1.5]
  • scene-spotting — scene text detection [new in v1.5]
  • grounding-ocr — text from bounding box region [new in v1.5]
  • general — free-form (use with --custom-prompt) [new in v1.5]

Example usage:

hf jobs uv run --flavor l4x1 \
    -s HF_TOKEN \
    /path/to/dots-ocr-1.5.py \
    davanstrien/ufo-ColPali output-dataset \
    --model davanstrien/dots.ocr-1.5 \
    --max-samples 10 --shuffle --seed 42

Model Info:

  • Original: rednote-hilab/dots.ocr-1.5 (ModelScope only)
  • Mirror: davanstrien/dots.ocr-1.5 (HF)
  • Parameters: 3B (1.2B vision encoder + 1.7B language model)
  • Architecture: DotsOCRForCausalLM (custom code, trust_remote_code required)
  • Precision: BF16
  • GitHub: https://github.com/rednote-hilab/dots.ocr

Pending Development

DeepSeek-OCR-2 (deepseek-ocr2-vllm.py)

Production Ready (2026-02-12)

Status: Working with vLLM nightly (requires nightly for DeepseekOCR2ForCausalLM support, not yet in stable 0.15.1)

What was done:

  • Rewrote the broken draft script (which used base64/llm.chat/resolution modes)
  • Uses the same proven pattern as v1: llm.generate() with PIL images + NGramPerReqLogitsProcessor
  • Key v2 addition: limit_mm_per_prompt={"image": 1} in LLM init
  • Added addict and matplotlib as dependencies (required by model's HF custom code)

Test results (2026-02-12):

  • 10/10 samples processed successfully on L4 GPU
  • Processing time: 6.4 min (includes model download + warmup)
  • Model: 6.33 GiB, ~475 toks/s input, ~246 toks/s output
  • Output dataset: davanstrien/deepseek-ocr2-nls-test

Example usage:

hf jobs uv run --flavor l4x1 \
    -s HF_TOKEN \
    https://huggingface.co/datasets/uv-scripts/ocr/raw/main/deepseek-ocr2-vllm.py \
    NationalLibraryOfScotland/medical-history-of-british-india output-dataset \
    --max-samples 10 --shuffle --seed 42

Important notes:

  • Requires vLLM nightly (stable 0.15.1 does NOT include DeepSeek-OCR-2 support)
  • The nightly index (https://wheels.vllm.ai/nightly) occasionally has transient build issues (e.g., only ARM wheels). If this happens, wait and retry.
  • Uses same API pattern as v1: NGramPerReqLogitsProcessor, SamplingParams(temperature=0, skip_special_tokens=False), extra_args for ngram settings

Model Information:

Other OCR Scripts

Nanonets OCR (nanonets-ocr.py, nanonets-ocr2.py)

✅ Both versions working

PaddleOCR-VL (paddleocr-vl.py)

✅ Working


Future: OCR Smoke Test Dataset

Status: Idea (noted 2026-02-12)

Build a small curated dataset (uv-scripts/ocr-smoke-test?) with ~2-5 samples from diverse sources. Purpose: fast CI-style verification that scripts still work after dep updates, without downloading full datasets.

Design goals:

  • Tiny (~20-30 images total) so download is seconds not minutes
  • Covers the axes that break things: document type, image quality, language, layout complexity
  • Has ground truth text where possible for quality regression checks
  • All permissively licensed (CC0/CC-BY preferred)

Candidate sources:

Source What it covers Why
NationalLibraryOfScotland/medical-history-of-british-india Historical English, degraded scans Has hand-corrected text column for comparison. CC0. Already tested with GLM-OCR.
davanstrien/ufo-ColPali Mixed modern documents Already used as our go-to test set. Varied layouts.
Something with tables Structured data extraction Tests --task table modes. Maybe a financial report or census page.
Something with formulas/LaTeX Math notation Tests --task formula. arXiv pages or textbook scans.
Something multilingual (CJK, Arabic, etc.) Non-Latin scripts GLM-OCR claims zh/ja/ko support. Good to verify.
Something handwritten Handwriting recognition Edge case that reveals model limits.

How it would work:

# Quick smoke test for any script
uv run glm-ocr.py uv-scripts/ocr-smoke-test smoke-out --max-samples 5
# Or a dedicated test runner that checks all scripts against it

Open questions:

  • Build as a proper HF dataset, or just a folder of images in the repo?
  • Should we include expected output for regression testing (fragile if models change)?
  • Could we add a --smoke-test flag to each script that auto-uses this dataset?
  • Worth adding to HF Jobs scheduled runs for ongoing monitoring?

OCR Benchmark Coordinator (ocr-bench-run.py)

Status: Working end-to-end (2026-02-14)

Launches N OCR models on the same dataset via run_uv_job(), each pushing to a shared repo as a separate config via --config/--create-pr. Eval done separately with ocr-elo-bench.py.

Model Registry (4 models)

Slug Model ID Size Default GPU Notes
glm-ocr zai-org/GLM-OCR 0.9B l4x1
deepseek-ocr deepseek-ai/DeepSeek-OCR 4B l4x1 Auto-passes --prompt-mode free (no grounding tags)
lighton-ocr-2 lightonai/LightOnOCR-2-1B 1B a100-large
dots-ocr rednote-hilab/dots.ocr 1.7B l4x1 Stable vLLM (>=0.9.1)

Each model entry has a default_args list for model-specific flags (e.g., DeepSeek uses ["--prompt-mode", "free"]).

Workflow

# Launch all 4 models on same data
uv run ocr-bench-run.py source-dataset --output my-bench --max-samples 50

# Evaluate directly from PRs (no merge needed)
uv run ocr-elo-bench.py my-bench --from-prs --mode both

# Or merge + evaluate
uv run ocr-elo-bench.py my-bench --from-prs --merge-prs --mode both

# Other useful flags
uv run ocr-bench-run.py --list-models          # Show registry table
uv run ocr-bench-run.py ... --dry-run           # Preview without launching
uv run ocr-bench-run.py ... --wait              # Poll until complete
uv run ocr-bench-run.py ... --models glm-ocr dots-ocr  # Subset of models

Eval script features (ocr-elo-bench.py)

  • --from-prs: Auto-discovers open PRs on the dataset repo, extracts config names from PR title [config-name] suffix, loads data from refs/pr/N without merging
  • --merge-prs: Auto-merges discovered PRs via api.merge_pull_request() before loading
  • --configs: Manually specify which configs to load (for merged repos)
  • --mode both: Runs pairwise ELO + pointwise scoring
  • Flat mode (original behavior) still works when --configs/--from-prs not used

Scripts pushed to Hub

All 4 scripts have been pushed to uv-scripts/ocr on the Hub with --config/--create-pr support:

  • glm-ocr.py
  • deepseek-ocr-vllm.py
  • lighton-ocr2.py
  • dots-ocr.py

Benchmark Results

Run 1: NLS Medical History (2026-02-14) — Pilot

Dataset: NationalLibraryOfScotland/medical-history-of-british-india (10 samples, shuffled, seed 42) Output repo: davanstrien/ocr-bench-test (4 open PRs) Judge: Qwen/Qwen2.5-VL-72B-Instruct via HF Inference Providers Content: Historical English, degraded scans of medical texts

ELO (pairwise, 5 samples evaluated):

  1. DoTS.ocr — 1540 (67% win rate)
  2. DeepSeek-OCR — 1539 (57%)
  3. LightOnOCR-2 — 1486 (50%)
  4. GLM-OCR — 1436 (29%)

Pointwise (5 samples):

  1. DeepSeek-OCR — 5.0/5.0
  2. GLM-OCR — 4.6
  3. LightOnOCR-2 — 4.4
  4. DoTS.ocr — 4.2

Key finding: DeepSeek-OCR's --prompt-mode document produces grounding tags (<|ref|>, <|det|>) that the judge penalizes heavily. Switching to --prompt-mode free (now the default in the registry) made it jump from last place to top 2.

Caveat: 5 samples is far too few for stable rankings. The judge VLM is called once per comparison (pairwise) or once per model-sample (pointwise) via HF Inference Providers API.

Run 2: Rubenstein Manuscript Catalog (2026-02-15) — First Full Benchmark

Dataset: biglam/rubenstein-manuscript-catalog (50 samples, shuffled, seed 42) Output repo: davanstrien/ocr-bench-rubenstein (4 PRs) Judge: Jury of 2 via ocr-vllm-judge.pyQwen/Qwen2.5-VL-7B-Instruct + Qwen/Qwen3-VL-8B-Instruct on A100 Content: ~48K typewritten + handwritten manuscript catalog cards from Duke University (CC0)

ELO (pairwise, 50 samples, 300 comparisons, 0 parse failures):

Rank Model ELO W L T Win%
1 LightOnOCR-2-1B 1595 100 50 0 67%
2 DeepSeek-OCR 1497 73 77 0 49%
3 GLM-OCR 1471 57 93 0 38%
4 dots.ocr 1437 70 80 0 47%

OCR job times (all 50 samples each):

  • dots-ocr: 5.3 min (L4)
  • deepseek-ocr: 5.6 min (L4)
  • glm-ocr: 5.7 min (L4)
  • lighton-ocr-2: 6.4 min (A100)

Key findings:

  • LightOnOCR-2-1B dominates on manuscript catalog cards (67% win rate, 100-point ELO gap over 2nd place) — a very different result from the NLS pilot where it placed 3rd
  • Rankings are dataset-dependent: NLS historical medical texts favored DoTS.ocr and DeepSeek-OCR; Rubenstein typewritten/handwritten cards favor LightOnOCR-2
  • Jury of small models works well: 0 parse failures on 300 comparisons thanks to vLLM structured output (xgrammar). Majority voting between 2 judges provides robustness
  • 50 samples gives meaningful separation: Clear ELO gaps (1595 → 1497 → 1471 → 1437) unlike the noisy 5-sample pilot
  • This validates the multi-dataset benchmark approach — no single dataset tells the whole story

Run 3: UFO-ColPali (2026-02-15) — Cross-Dataset Validation

Dataset: davanstrien/ufo-ColPali (50 samples, shuffled, seed 42) Output repo: davanstrien/ocr-bench-ufo (4 PRs) Judge: Qwen/Qwen3-VL-30B-A3B-Instruct via ocr-vllm-judge.py on A100 (updated prompt) Content: Mixed modern documents (invoices, reports, forms, etc.)

ELO (pairwise, 50 samples, 294 comparisons):

Rank Model ELO W L T Win%
1 DeepSeek-OCR 1827 130 17 0 88%
2 dots.ocr 1510 64 83 0 44%
3 LightOnOCR-2-1B 1368 77 70 0 52%
4 GLM-OCR 1294 23 124 0 16%

Human validation (30 comparisons): DeepSeek-OCR #1 (same as judge), LightOnOCR-2 #3 (same). Middle pack (GLM-OCR #2 human / #4 judge, dots.ocr #4 human / #2 judge) shuffled.

Cross-Dataset Comparison (Human-Validated)

Model Rubenstein Human Rubenstein Kimi UFO Human UFO 30B
DeepSeek-OCR #1 #1 #1 #1
GLM-OCR #2 #3 #2 #4
LightOnOCR-2 #4 #2 #3 #3
dots.ocr #3 #4 #4 #2

Conclusion: DeepSeek-OCR is consistently #1 across datasets and evaluation methods. Middle-pack rankings are dataset-dependent. Updated prompt fixed the LightOnOCR-2 overrating seen with old prompt/small judges.

Note: NLS pilot results (5 samples, 72B API judge) omitted — not comparable with newer methodology.

Known Issues / Next Steps

  1. More samples needed — Done. Rubenstein run (2026-02-15) used 50 samples and produced clear ELO separation across all 4 models.
  2. Smaller judge model — Tested with Qwen VL 7B + Qwen3 VL 8B via ocr-vllm-judge.py. Works well with structured output (0 parse failures). Jury of small models compensates for individual model weakness. See "Offline vLLM Judge" section below.
  3. Auto-merge in coordinator--wait could auto-merge PRs after successful jobs. Not yet implemented.
  4. Adding more modelsrolm-ocr.py exists but needs --config/--create-pr added. deepseek-ocr2-vllm.py, paddleocr-vl-1.5.py, etc. could also be added to the registry.
  5. Leaderboard Space — See future section below.
  6. Result persistenceocr-vllm-judge.py now has --save-results REPO_ID flag. First dataset: davanstrien/ocr-bench-rubenstein-judge.
  7. More diverse datasets — Rankings are dataset-dependent (LightOnOCR-2 wins on Rubenstein, DoTS.ocr won pilot on NLS). Need benchmarks on tables, formulas, multilingual, and modern documents for a complete picture.
  8. Human validationocr-human-eval.py completed on Rubenstein (30/30). Tested 3 judge configs. Kimi K2.5 (170B) via Novita + updated prompt = best human agreement (only judge to match human's #1). Now default in ocr-jury-bench.py. See OCR-BENCHMARK.md for full comparison.

Offline vLLM Judge (ocr-vllm-judge.py)

Status: Working end-to-end (2026-02-15)

Runs pairwise OCR quality comparisons using a local VLM judge via vLLM's offline LLM() pattern. Supports jury mode (multiple models vote sequentially on the same GPU) with majority voting.

Why use this over the API judge (ocr-jury-bench.py)?

API judge (ocr-jury-bench.py) Offline judge (ocr-vllm-judge.py)
Parse failures Needs retries for malformed JSON 0 failures — vLLM structured output guarantees valid JSON
Network Rate limits, timeouts, transient errors Zero network calls
Cost Per-token API pricing Just GPU time
Judge models Limited to Inference Providers catalog Any vLLM-supported VLM
Jury mode Sequential API calls per judge Sequential model loading, batch inference per judge
Best for Quick spot-checks, access to 72B models Batch evaluation (50+ samples), reproducibility

Pushed to Hub: uv-scripts/ocr as ocr-vllm-judge.py (2026-02-15)

Test Results (2026-02-15)

Test 1 — Single judge, 1 sample, L4:

  • Qwen2.5-VL-7B-Instruct, 6/6 comparisons, 0 parse failures
  • Total time: ~3 min (including model download + warmup)

Test 2 — Jury of 2, 3 samples, A100:

  • Qwen2.5-VL-7B + Qwen3-VL-8B, 15/15 comparisons, 0 parse failures
  • GPU cleanup between models: successful (nanobind warnings are cosmetic)
  • Majority vote aggregation working ([2/2] unanimous, [1/2] split)
  • Total time: ~4 min (including both model downloads)

Test 3 — Full benchmark, 50 samples, A100 (Rubenstein Manuscript Catalog):

  • Qwen2.5-VL-7B + Qwen3-VL-8B jury, 300/300 comparisons, 0 parse failures
  • Input: davanstrien/ocr-bench-rubenstein (4 PRs from ocr-bench-run.py)
  • Produced clear ELO rankings with meaningful separation
  • See "Benchmark Results → Run 2" in the OCR Benchmark Coordinator section above

Usage

# Single judge on L4
hf jobs uv run --flavor l4x1 -s HF_TOKEN \
    ocr-vllm-judge.py davanstrien/ocr-bench-nls-50 --from-prs \
    --judge-model Qwen/Qwen2.5-VL-7B-Instruct --max-samples 10

# Jury of 2 on A100 (recommended for jury mode)
hf jobs uv run --flavor a100-large -s HF_TOKEN \
    ocr-vllm-judge.py davanstrien/ocr-bench-nls-50 --from-prs \
    --judge-model Qwen/Qwen2.5-VL-7B-Instruct \
    --judge-model Qwen/Qwen3-VL-8B-Instruct \
    --max-samples 50

Implementation Notes

  • Comparisons built upfront on CPU as NamedTuples, then batched to vLLM in single llm.chat() call
  • Structured output via compatibility shim: StructuredOutputsParams (vLLM >= 0.12) → GuidedDecodingParams (older) → prompt-based fallback
  • GPU cleanup between jury models: destroy_model_parallel() + gc.collect() + torch.cuda.empty_cache()
  • Position bias mitigation: A/B order randomized per comparison
  • A100 recommended for jury mode; L4 works for single 7B judge

Next Steps

  1. Scale test — Completed on Rubenstein Manuscript Catalog (50 samples, 300 comparisons, 0 parse failures). Rankings differ from API-based pilot (different dataset + judge), validating multi-dataset approach.
  2. Result persistence — Added --save-results REPO_ID flag. Pushes 3 configs to HF Hub: comparisons (one row per pairwise comparison), leaderboard (ELO + win/loss/tie per model), metadata (source dataset, judge models, seed, timestamp). First dataset: davanstrien/ocr-bench-rubenstein-judge.
  3. Integrate into ocr-bench-run.py — Add --eval flag that auto-runs vLLM judge after OCR jobs complete

Blind Human Eval (ocr-human-eval.py)

Status: Working (2026-02-15)

Gradio app for blind A/B comparison of OCR outputs. Shows document image + two anonymized OCR outputs, human picks winner or tie. Computes ELO rankings from human annotations and optionally compares against automated judge results.

Usage

# Basic — blind human eval only
uv run ocr-human-eval.py davanstrien/ocr-bench-rubenstein --from-prs --max-samples 5

# With judge comparison — loads automated judge results for agreement analysis
uv run ocr-human-eval.py davanstrien/ocr-bench-rubenstein --from-prs \
    --judge-results davanstrien/ocr-bench-rubenstein-judge --max-samples 5

Features

  • Blind evaluation: Two-tab design — Evaluate tab never shows model names, Results tab reveals rankings
  • Position bias mitigation: A/B order randomly swapped per comparison
  • Resume support: JSON annotations saved atomically after each vote; restart app to resume where you left off
  • Live agreement tracking: Per-vote feedback shows running agreement with automated judge (when --judge-results provided)
  • Split-jury prioritization: Comparisons where automated judges disagreed ("1/2" agreement) shown first — highest annotation value per vote
  • Image variety: Round-robin interleaving by sample so you don't see the same document image repeatedly
  • Soft/hard disagreement analysis: Distinguishes between harmless ties-vs-winner disagreements and genuine opposite-winner errors

First Validation Results (Rubenstein, 30 annotations)

Tested 3 judge configs against 30 human annotations. Kimi K2.5 (170B) via Novita is the only judge to match human's #1 pick (DeepSeek-OCR). Small models (7B/8B/30B) all overrate LightOnOCR-2 due to bias toward its commentary style. Updated prompt (prioritized faithfulness > completeness > accuracy) helps but model size is the bigger factor.

Full results and analysis in OCR-BENCHMARK.md → "Human Validation" section.

Next Steps

  1. Second dataset — Run on NLS Medical History for cross-dataset human validation
  2. Multiple annotators — Currently single-user; could support annotator ID for inter-annotator agreement
  3. Remaining LightOnOCR-2 gap — Still #2 (Kimi) vs #4 (human). May need to investigate on more samples or strip commentary in preprocessing

Future: Leaderboard HF Space

Status: Idea (noted 2026-02-14)

Build a Hugging Face Space with a persistent leaderboard that gets updated after each benchmark run. This would give a public-facing view of OCR model quality.

Design ideas:

  • Gradio or static Space displaying ELO ratings + pointwise scores
  • ocr-elo-bench.py could push results to a dataset that the Space reads
  • Or the Space itself could run evaluation on demand
  • Show per-document comparisons (image + side-by-side OCR outputs)
  • Historical tracking — how scores change across model versions
  • Filter by document type (historical, modern, tables, formulas, multilingual)

Open questions:

  • Should the eval script push structured results to a dataset (e.g., uv-scripts/ocr-leaderboard-data)?
  • Static leaderboard (updated by CI/scheduled job) vs interactive (evaluate on demand)?
  • Include sample outputs for qualitative comparison?
  • How to handle different eval datasets (NLS medical history vs UFO vs others)?

Incremental Uploads / Checkpoint Strategy — ON HOLD

Status: Waiting on HF Hub Buckets (noted 2026-02-20)

Current state:

  • glm-ocr.py (v1): Simple batch-then-push. Works fine for most jobs.
  • glm-ocr-v2.py: Adds CommitScheduler-based incremental uploads + checkpoint/resume. ~400 extra lines. Works but has tradeoffs (commit noise, --create-pr incompatible, complex resume metadata).

Decision: Do NOT port v2 pattern to other scripts. Wait for HF Hub Buckets instead.

Why: Two open PRs will likely make the v2 CommitScheduler approach obsolete:

  • huggingface_hub#3673 — Buckets API: S3-like mutable object storage on HF, no git versioning overhead
  • huggingface_hub#3807 — HfFileSystem support for buckets: fsspec-compatible, so pyarrow/pandas/datasets can read/write hf://buckets/ paths directly

What Buckets would replace: Once landed, incremental saves become one line per batch:

batch_ds.to_parquet(f"hf://buckets/{user}/ocr-scratch/shard-{batch_num:05d}.parquet")

No CommitScheduler, no CleanupScheduler, no resume metadata, no completed batch scanning. Just write to the bucket path via fsspec. Final step: read back from bucket, push_to_hub to a clean dataset repo (compatible with --create-pr).

Action items when Buckets ships:

  1. Test hf://buckets/ fsspec writes on one script (glm-ocr is the guinea pig)
  2. Verify: write performance, atomicity (partial writes visible?), auth propagation in HF Jobs
  3. If it works, adopt as the standard pattern for all scripts — simple enough to inline (~20 lines)
  4. Retire glm-ocr-v2.py CommitScheduler approach

Until then: v1 scripts stay as-is. glm-ocr-v2.py exists if someone needs resume on a very large job today.


Last Updated: 2026-02-20 Watch PRs:

  • HF Hub Buckets API (#3673): Core buckets support. Will enable simpler incremental upload pattern for all scripts.
  • HfFileSystem Buckets (#3807): fsspec support for hf://buckets/ paths. Key for zero-boilerplate writes from scripts.
  • DeepSeek-OCR-2 stable vLLM release: Currently only in nightly. Watch for vLLM 0.16.0 stable release on PyPI to remove nightly dependency.
  • nanobind leak warnings in vLLM structured output (xgrammar): Cosmetic only, does not affect results. May be fixed in future xgrammar release.