Intelligent Document Processing Leaderboard

Comprehensive document AI leaderboard comparing the best models across OCR, table extraction, key information extraction, and visual QA. Compare performance, accuracy, and cost.

This work is sponsored by Nanonets.

0 Benchmarks0 ModelsOpen evaluation
#
Model
Overall
OlmOCR
OmniDoc
IDP
Size
1Nanonets OCR2+Nanonets81.882.289.573.8
2Gemini-3-ProGoogle81.473.588.881.8
3Claude Sonnet 4.6Anthropic80.874.486.981.2
4Claude Opus 4.6Anthropic80.373.985.981.1
5Gemini-3-FlashGoogle79.969.290.180.5
6GPT-5.2OpenAI79.272.288.077.4
7GPT-5-MiniOpenAI70.856.782.573.3
8GPT-4.1OpenAI70.055.579.974.7
9Claude Haiku 4.5Anthropic69.656.279.672.9
10Ministral-8BMistral AI68.057.878.367.98B
11GLM-OCRZhipu AI63.666.769.254.9
12GPT-5-NanoOpenAI50.722.863.465.8
13Llama-3.2-Vision-11BMeta50.147.244.658.611B
14Pixtral-12BMistral AI46.036.842.359.012B

About the Leaderboard

The Intelligent Document Processing (IDP) Leaderboard provides a comprehensive evaluation framework for assessing the capabilities of various AI models in document understanding and processing tasks. All models are evaluated using identical prompts, images, and scoring pipelines. The overall score is the mean of all benchmark scores.