Intelligent Document Processing Leaderboard

Comprehensive document AI leaderboard comparing the best models across OCR, table extraction, key information extraction, and visual QA. Compare performance, accuracy, and cost.

0 Benchmarks0 ModelsOpen evaluation
#
Model
Overall
OlmOCR
OmniDoc
IDP
1Nanonets OCR-3Nanonets85.987.490.080.2
2Nanonets OCR2+Nanonets81.882.089.573.8
3GPT-5.4OpenAI81.073.485.384.4
4Qwen3-VL-PlusAlibaba80.177.982.579.8
5Qwen3-VL-235BAlibaba79.676.881.980.0
6Gemini-3-ProGoogle79.467.788.881.8
7Claude Sonnet 4.6Anthropic79.169.386.981.2
8Claude Opus 4.6Anthropic78.869.385.981.1
9Gemini-3-FlashGoogle78.665.390.180.5
10Gemini 3.1 ProGoogle78.560.785.389.6
11GPT-5.2OpenAI78.068.788.077.4
12Qwen3.5-9BAlibaba76.777.276.776.2
13Qwen3.5-4BAlibaba72.575.467.674.5
14GPT-5-MiniOpenAI71.759.382.573.3
15Mistral Small 4Mistral AI71.569.676.468.5
16Claude Haiku 4.5Anthropic70.258.279.672.9
17Ministral-8BMistral AI69.558.778.371.7
18GPT-4.1OpenAI69.554.079.974.7
19GLM-OCRZhipu AI64.268.469.254.9
20Qwen3.5-2BAlibaba62.671.948.767.1
21Qwen3.5-0.8BAlibaba57.864.847.361.2
22GPT-5-NanoOpenAI52.026.863.465.8
23Llama-3.2-Vision-11BMeta50.849.144.658.6
24Pixtral-12BMistral AI46.538.342.359.0

About the Leaderboard

The Intelligent Document Processing (IDP) Leaderboard provides a comprehensive evaluation framework for assessing the capabilities of various AI models in document understanding and processing tasks. The overall score is the mean of all benchmark scores.