PDF-Extract-Kit: High-Quality PDF Extraction Toolkit🔥🔥🔥
Easier to use: Just grab MinerU Desktop. No coding, no login, just a simple interface and smooth interactions. Enjoy it without any fuss!🚀🚀🚀
- 2025/06/13 2.0.0 Released
- MinerU 2.0 represents a comprehensive reconstruction and upgrade from architecture to functionality, delivering a more streamlined design, enhanced performance, and more flexible user experience.
- New Architecture: MinerU 2.0 has been deeply restructured in code organization and interaction methods, significantly improving system usability, maintainability, and extensibility.
- Removal of Third-party Dependency Limitations: Completely eliminated the dependency on
pymupdf
, moving the project toward a more open and compliant open-source direction. - Ready-to-use, Easy Configuration: No need to manually edit JSON configuration files; most parameters can now be set directly via command line or API.
- Automatic Model Management: Added automatic model download and update mechanisms, allowing users to complete model deployment without manual intervention.
- Offline Deployment Friendly: Provides built-in model download commands, supporting deployment requirements in completely offline environments.
- Streamlined Code Structure: Removed thousands of lines of redundant code, simplified class inheritance logic, significantly improving code readability and development efficiency.
- Unified Intermediate Format Output: Adopted standardized
middle_json
format, compatible with most secondary development scenarios based on this format, ensuring seamless ecosystem business migration.
- Removal of Third-party Dependency Limitations: Completely eliminated the dependency on
- New Model: MinerU 2.0 integrates our latest small-parameter, high-performance multimodal document parsing model, achieving end-to-end high-speed, high-precision document understanding.
- Small Model, Big Capabilities: With parameters under 1B, yet surpassing traditional 72B-level vision-language models (VLMs) in parsing accuracy.
- Multiple Functions in One: A single model covers multilingual recognition, handwriting recognition, layout analysis, table parsing, formula recognition, reading order sorting, and other core tasks.
- Ultimate Inference Speed: Achieves peak throughput exceeding 10,000 tokens/s through
sglang
acceleration on a single NVIDIA 4090 card, easily handling large-scale document processing requirements. - Online Experience: You can experience this model online on our Hugging Face demo:
- Incompatible Changes Notice: To improve overall architectural rationality and long-term maintainability, this version contains some incompatible changes:
- Python package name changed from
magic-pdf
tomineru
, and the command-line tool changed frommagic-pdf
tomineru
. Please update your scripts and command calls accordingly. - For modular system design and ecosystem consistency considerations, MinerU 2.0 no longer includes the LibreOffice document conversion module. If you need to process Office documents, we recommend converting them to PDF format through an independently deployed LibreOffice service before proceeding with subsequent parsing operations.
- Python package name changed from
- New Architecture: MinerU 2.0 has been deeply restructured in code organization and interaction methods, significantly improving system usability, maintainability, and extensibility.
- MinerU 2.0 represents a comprehensive reconstruction and upgrade from architecture to functionality, delivering a more streamlined design, enhanced performance, and more flexible user experience.
History Log
2025/05/24 Release 1.3.12
- Added support for PPOCRv5 models, updated
ch_server
model toPP-OCRv5_rec_server
, andch_lite
model toPP-OCRv5_rec_mobile
(model update required)- In testing, we found that PPOCRv5(server) has some improvement for handwritten documents, but has slightly lower accuracy than v4_server_doc for other document types, so the default ch model remains unchanged as
PP-OCRv4_server_rec_doc
. - Since PPOCRv5 has enhanced recognition capabilities for handwriting and special characters, you can manually choose the PPOCRv5 model for Japanese-Traditional Chinese mixed scenarios and handwritten documents
- You can select the appropriate model through the lang parameter
lang='ch_server'
(Python API) or--lang ch_server
(command line):ch
:PP-OCRv4_server_rec_doc
(default) (Chinese/English/Japanese/Traditional Chinese mixed/15K dictionary)ch_server
:PP-OCRv5_rec_server
(Chinese/English/Japanese/Traditional Chinese mixed + handwriting/18K dictionary)ch_lite
:PP-OCRv5_rec_mobile
(Chinese/English/Japanese/Traditional Chinese mixed + handwriting/18K dictionary)ch_server_v4
:PP-OCRv4_rec_server
(Chinese/English mixed/6K dictionary)ch_lite_v4
:PP-OCRv4_rec_mobile
(Chinese/English mixed/6K dictionary)
- In testing, we found that PPOCRv5(server) has some improvement for handwritten documents, but has slightly lower accuracy than v4_server_doc for other document types, so the default ch model remains unchanged as
- Added support for handwritten documents through optimized layout recognition of handwritten text areas
- This feature is supported by default, no additional configuration required
- You can refer to the instructions above to manually select the PPOCRv5 model for better handwritten document parsing results
- The
huggingface
andmodelscope
demos have been updated to versions that support handwriting recognition and PPOCRv5 models, which you can experience online
2025/04/29 Release 1.3.10
- Added support for custom formula delimiters, which can be configured by modifying the
latex-delimiter-config
section in themagic-pdf.json
file in your user directory.
2025/04/27 Release 1.3.9
- Optimized formula parsing functionality, improved formula rendering success rate
2025/04/23 Release 1.3.8
- The default
ocr
model (ch
) has been updated toPP-OCRv4_server_rec_doc
(model update required)PP-OCRv4_server_rec_doc
is trained on a mixture of more Chinese document data and PP-OCR training data based onPP-OCRv4_server_rec
, adding recognition capabilities for some traditional Chinese characters, Japanese, and special characters. It can recognize over 15,000 characters and improves both document-specific and general text recognition abilities.- Performance comparison of PP-OCRv4_server_rec_doc/PP-OCRv4_server_rec/PP-OCRv4_mobile_rec
- After verification, the
PP-OCRv4_server_rec_doc
model shows significant accuracy improvements in Chinese/English/Japanese/Traditional Chinese in both single language and mixed language scenarios, with comparable speed toPP-OCRv4_server_rec
, making it suitable for most use cases. - In some pure English scenarios,
PP-OCRv4_server_rec_doc
may have word adhesion issues, whilePP-OCRv4_server_rec
performs better in these cases. Therefore, we've kept thePP-OCRv4_server_rec
model, which users can access by adding the parameterlang='ch_server'
(Python API) or--lang ch_server
(command line).
2025/04/22 Release 1.3.7
- Fixed the issue where the lang parameter was ineffective during table parsing model initialization
- Fixed the significant speed reduction of OCR and table parsing in
cpu
mode
2025/04/16 Release 1.3.4
- Slightly improved OCR-det speed by removing some unnecessary blocks
- Fixed page-internal sorting errors caused by footnotes in certain cases
2025/04/12 Release 1.3.2
- Fixed dependency version incompatibility issues when installing on Windows with Python 3.13
- Optimized memory usage during batch inference
- Improved parsing of tables rotated 90 degrees
- Enhanced parsing of oversized tables in financial report samples
- Fixed the occasional word adhesion issue in English text areas when OCR language is not specified (model update required)
2025/04/08 Release 1.3.1
- Fixed several compatibility issues
- Added support for Python 3.13
- Made final adaptations for outdated Linux systems (such as CentOS 7) with no guarantee of continued support in future versions, installation instructions
2025/04/03 Release 1.3.0
- Installation and compatibility optimizations
- Resolved compatibility issues caused by
detectron2
by removinglayoutlmv3
usage in layout - Extended torch version compatibility to 2.2~2.6 (excluding 2.5)
- Added CUDA compatibility for versions 11.8/12.4/12.6/12.8 (CUDA version determined by torch), solving compatibility issues for users with 50-series and H-series GPUs
- Extended Python compatibility to versions 3.10~3.12, fixing the issue of automatic downgrade to version 0.6.1 when installing in non-3.10 environments
- Optimized offline deployment process, eliminating the need to download any model files after successful deployment
- Resolved compatibility issues caused by
- Performance optimizations
- Enhanced parsing speed for batches of small files by supporting batch processing of multiple PDF files (script example), with formula parsing speed improved by up to 1400% and overall parsing speed improved by up to 500% compared to version 1.0.1
- Reduced memory usage and improved parsing speed by optimizing MFR model loading and usage (requires re-running the model download process to get incremental updates to model files)
- Optimized GPU memory usage, requiring only 6GB minimum to run this project
- Improved running speed on MPS devices
- Parsing effect optimizations
- Updated MFR model to
unimernet(2503)
, fixing line break loss issues in multi-line formulas
- Updated MFR model to
- Usability optimizations
- Completely replaced the
paddle
framework andpaddleocr
in the project by usingpaddleocr2torch
, resolving conflicts betweenpaddle
andtorch
, as well as thread safety issues caused by thepaddle
framework - Added real-time progress bar display during parsing, allowing precise tracking of parsing progress and making the waiting process more bearable
- Completely replaced the
2025/03/03 1.2.1 released
- Fixed the impact on punctuation marks during full-width to half-width conversion of letters and numbers
- Fixed caption matching inaccuracies in certain scenarios
- Fixed formula span loss issues in certain scenarios
2025/02/24 1.2.0 released
This version includes several fixes and improvements to enhance parsing efficiency and accuracy:
- Performance Optimization
- Increased classification speed for PDF documents in auto mode.
- Parsing Optimization
- Improved parsing logic for documents containing watermarks, significantly enhancing the parsing results for such documents.
- Enhanced the matching logic for multiple images/tables and captions within a single page, improving the accuracy of image-text matching in complex layouts.
- Bug Fixes
- Fixed an issue where image/table spans were incorrectly filled into text blocks under certain conditions.
- Resolved an issue where title blocks were empty in some cases.
2025/01/22 1.1.0 released
In this version we have focused on improving parsing accuracy and efficiency:
- Model capability upgrade (requires re-executing the model download process to obtain incremental updates of model files)
- The layout recognition model has been upgraded to the latest
doclayout_yolo(2501)
model, improving layout recognition accuracy. - The formula parsing model has been upgraded to the latest
unimernet(2501)
model, improving formula recognition accuracy.
- The layout recognition model has been upgraded to the latest
- Performance optimization
- On devices that meet certain configuration requirements (16GB+ VRAM), by optimizing resource usage and restructuring the processing pipeline, overall parsing speed has been increased by more than 50%.
- Parsing effect optimization
- Added a new heading classification feature (testing version, enabled by default) to the online demo (mineru.net/huggingface/modelscope), which supports hierarchical classification of headings, thereby enhancing document structuring.
2025/01/10 1.0.1 released
This is our first official release, where we have introduced a completely new API interface and enhanced compatibility through extensive refactoring, as well as a brand new automatic language identification feature:
- New API Interface
- For the data-side API, we have introduced the Dataset class, designed to provide a robust and flexible data processing framework. This framework currently supports a variety of document formats, including images (.jpg and .png), PDFs, Word documents (.doc and .docx), and PowerPoint presentations (.ppt and .pptx). It ensures effective support for data processing tasks ranging from simple to complex.
- For the user-side API, we have meticulously designed the MinerU processing workflow as a series of composable Stages. Each Stage represents a specific processing step, allowing users to define new Stages according to their needs and creatively combine these stages to customize their data processing workflows.
- Enhanced Compatibility
- By optimizing the dependency environment and configuration items, we ensure stable and efficient operation on ARM architecture Linux systems.
- We have deeply integrated with Huawei Ascend NPU acceleration, providing autonomous and controllable high-performance computing capabilities. This supports the localization and development of AI application platforms in China. Ascend NPU Acceleration
- Automatic Language Identification
- By introducing a new language recognition model, setting the
lang
configuration toauto
during document parsing will automatically select the appropriate OCR language model, improving the accuracy of scanned document parsing.
- By introducing a new language recognition model, setting the
2024/11/22 0.10.0 released
Introducing hybrid OCR text extraction capabilities:
- Significantly improved parsing performance in complex text distribution scenarios such as dense formulas, irregular span regions, and text represented by images.
- Combines the dual advantages of accurate content extraction and faster speed in text mode, and more precise span/line region recognition in OCR mode.
2024/11/15 0.9.3 released
Integrated RapidTable for table recognition, improving single-table parsing speed by more than 10 times, with higher accuracy and lower GPU memory usage.
2024/11/06 0.9.2 released
Integrated the StructTable-InternVL2-1B model for table recognition functionality.
2024/10/31 0.9.0 released
This is a major new version with extensive code refactoring, addressing numerous issues, improving performance, reducing hardware requirements, and enhancing usability:
- Refactored the sorting module code to use layoutreader for reading order sorting, ensuring high accuracy in various layouts.
- Refactored the paragraph concatenation module to achieve good results in cross-column, cross-page, cross-figure, and cross-table scenarios.
- Refactored the list and table of contents recognition functions, significantly improving the accuracy of list blocks and table of contents blocks, as well as the parsing of corresponding text paragraphs.
- Refactored the matching logic for figures, tables, and descriptive text, greatly enhancing the accuracy of matching captions and footnotes to figures and tables, and reducing the loss rate of descriptive text to near zero.
- Added multi-language support for OCR, supporting detection and recognition of 84 languages. For the list of supported languages, see OCR Language Support List.
- Added memory recycling logic and other memory optimization measures, significantly reducing memory usage. The memory requirement for enabling all acceleration features except table acceleration (layout/formula/OCR) has been reduced from 16GB to 8GB, and the memory requirement for enabling all acceleration features has been reduced from 24GB to 10GB.
- Optimized configuration file feature switches, adding an independent formula detection switch to significantly improve speed and parsing results when formula detection is not needed.
- Integrated PDF-Extract-Kit 1.0:
- Added the self-developed
doclayout_yolo
model, which speeds up processing by more than 10 times compared to the original solution while maintaining similar parsing effects, and can be freely switched withlayoutlmv3
via the configuration file. - Upgraded formula parsing to
unimernet 0.2.1
, improving formula parsing accuracy while significantly reducing memory usage. - Due to the repository change for
PDF-Extract-Kit 1.0
, you need to re-download the model. Please refer to How to Download Models for detailed steps.
- Added the self-developed
2024/09/27 Version 0.8.1 released
Fixed some bugs, and providing a localized deployment version of the online demo and the front-end interface.
2024/09/09 Version 0.8.0 released
Supporting fast deployment with Dockerfile, and launching demos on Huggingface and Modelscope.
2024/08/30 Version 0.7.1 released
Add paddle tablemaster table recognition option
2024/08/09 Version 0.7.0b1 released
Simplified installation process, added table recognition functionality
2024/08/01 Version 0.6.2b1 released
Optimized dependency conflict issues and installation documentation
2024/07/05 Initial open-source release
MinerU is a tool that converts PDFs into machine-readable formats (e.g., markdown, JSON), allowing for easy extraction into any format. MinerU was born during the pre-training process of InternLM. We focus on solving symbol conversion issues in scientific literature and hope to contribute to technological development in the era of large models. Compared to well-known commercial products, MinerU is still young. If you encounter any issues or if the results are not as expected, please submit an issue on issue and attach the relevant PDF.
pdf_zh_cn.mp4
- Remove headers, footers, footnotes, page numbers, etc., to ensure semantic coherence.
- Output text in human-readable order, suitable for single-column, multi-column, and complex layouts.
- Preserve the structure of the original document, including headings, paragraphs, lists, etc.
- Extract images, image descriptions, tables, table titles, and footnotes.
- Automatically recognize and convert formulas in the document to LaTeX format.
- Automatically recognize and convert tables in the document to HTML format.
- Automatically detect scanned PDFs and garbled PDFs and enable OCR functionality.
- OCR supports detection and recognition of 84 languages.
- Supports multiple output formats, such as multimodal and NLP Markdown, JSON sorted by reading order, and rich intermediate formats.
- Supports various visualization results, including layout visualization and span visualization, for efficient confirmation of output quality.
- Supports running in a pure CPU environment, and also supports GPU(CUDA)/NPU(CANN)/MPS acceleration
- Compatible with Windows, Linux, and Mac platforms.
If you encounter any installation issues, please first consult the FAQ.
If the parsing results are not as expected, refer to the Known Issues.
There are three different ways to experience MinerU:
Warning
Pre-installation Notice—Hardware and Software Environment Support
To ensure the stability and reliability of the project, we only optimize and test for specific hardware and software environments during development. This ensures that users deploying and running the project on recommended system configurations will get the best performance with the fewest compatibility issues.
By focusing resources on the mainline environment, our team can more efficiently resolve potential bugs and develop new features.
In non-mainline environments, due to the diversity of hardware and software configurations, as well as third-party dependency compatibility issues, we cannot guarantee 100% project availability. Therefore, for users who wish to use this project in non-recommended environments, we suggest carefully reading the documentation and FAQ first. Most issues already have corresponding solutions in the FAQ. We also encourage community feedback to help us gradually expand support.
Parsing Backend | pipeline | vlm-transformers | vlm-sgslang |
Operating System | windows/linux/mac | windows/linux | windows(wsl2)/linux |
Memory Requirements | Minimum 16GB+, 32GB+ recommended | ||
Disk Space Requirements | 20GB+, SSD recommended | ||
Python Version | 3.10-3.13 | ||
CPU Inference Support | ✅ | ❌ | ❌ |
GPU Requirements | Turing architecture or later, 6GB+ VRAM or Apple Silicon | Ampere architecture or later, 8GB+ VRAM | Ampere architecture or later, 24GB+ VRAM |
pip install --upgrade pip
pip install uv
uv pip install "mineru[core]>=2.0.0"
git clone https://github.com/opendatalab/MinerU.git
cd MinerU
uv pip install -e .[core]
To use sglang acceleration for VLM model inference, install the full version:
uv pip install "mineru[all]>=2.0.0"
Or install from source:
uv pip install -e .[all]
The simplest command line invocation is:
mineru -p <input_path> -o <output_path>
<input_path>
: Local PDF file or directory (supports pdf/png/jpg/jpeg)<output_path>
: Output directory
Get all available parameter descriptions:
mineru --help
Usage: mineru [OPTIONS]
Options:
-v, --version Show version and exit
-p, --path PATH Input file path or directory (required)
-o, --output PATH Output directory (required)
-m, --method [auto|txt|ocr] Parsing method: auto (default), txt, ocr (pipeline backend only)
-b, --backend [pipeline|vlm-transformers|vlm-sglang-engine|vlm-sglang-client]
Parsing backend (default: pipeline)
-l, --lang [ch|ch_server|... ] Specify document language (improves OCR accuracy, pipeline backend only)
-u, --url TEXT Service address when using sglang-client
-s, --start INTEGER Starting page number (0-based)
-e, --end INTEGER Ending page number (0-based)
-f, --formula BOOLEAN Enable formula parsing (default: on, pipeline backend only)
-t, --table BOOLEAN Enable table parsing (default: on, pipeline backend only)
-d, --device TEXT Inference device (e.g., cpu/cuda/cuda:0/npu/mps, pipeline backend only)
--vram INTEGER Maximum GPU VRAM usage per process (pipeline backend only)
--source [huggingface|modelscope|local]
Model source, default: huggingface
--help Show help information
MinerU automatically downloads required models from HuggingFace on first run. If HuggingFace is inaccessible, you can switch model sources:
mineru -p <input_path> -o <output_path> --source modelscope
Or set environment variable:
export MINERU_MODEL_SOURCE=modelscope
mineru -p <input_path> -o <output_path>
mineru-models-download --help
Or use interactive command-line tool to select models:
mineru-models-download
After download, model paths will be displayed in current terminal and automatically written to mineru.json
in user directory.
mineru -p <input_path> -o <output_path> --source local
Or enable via environment variable:
export MINERU_MODEL_SOURCE=local
mineru -p <input_path> -o <output_path>
mineru -p <input_path> -o <output_path> -b vlm-sglang-engine
- Start Server:
mineru-sglang-server --port 30000
- Use Client in another terminal:
mineru -p <input_path> -o <output_path> -b vlm-sglang-client -u http://127.0.0.1:30000
💡 For more information about output files, please refer to Output File Documentation
You can also call MinerU through Python code, see example code at: 👉 Python Usage Example
Community developers have created various extensions based on MinerU, including:
- Graphical interface based on Gradio
- Web API based on FastAPI
- Client/server architecture with multi-GPU load balancing
- MCP Server based on the official API
These projects typically offer better user experience and additional features.
For detailed deployment instructions, please refer to: 👉 Derivative Projects Documentation
- Reading order based on the model
- Recognition of
index
andlist
in the main text - Table recognition
- Heading Classification
- Code block recognition in the main text
- Chemical formula recognition
- Geometric shape recognition
- Reading order is determined by the model based on the spatial distribution of readable content, and may be out of order in some areas under extremely complex layouts.
- Vertical text is not supported.
- Tables of contents and lists are recognized through rules, and some uncommon list formats may not be recognized.
- Code blocks are not yet supported in the layout model.
- Comic books, art albums, primary school textbooks, and exercises cannot be parsed well.
- Table recognition may result in row/column recognition errors in complex tables.
- OCR recognition may produce inaccurate characters in PDFs of lesser-known languages (e.g., diacritical marks in Latin script, easily confused characters in Arabic script).
- Some formulas may not render correctly in Markdown.
Currently, some models in this project are trained based on YOLO. However, since YOLO follows the AGPL license, it may impose restrictions on certain use cases. In future iterations, we plan to explore and replace these with models under more permissive licenses to enhance user-friendliness and flexibility.
- PDF-Extract-Kit
- DocLayout-YOLO
- UniMERNet
- RapidTable
- PaddleOCR
- PaddleOCR2Pytorch
- layoutreader
- xy-cut
- fast-langdetect
- pypdfium2
- pdfminer.six
- pypdf
@misc{wang2024mineruopensourcesolutionprecise,
title={MinerU: An Open-Source Solution for Precise Document Content Extraction},
author={Bin Wang and Chao Xu and Xiaomeng Zhao and Linke Ouyang and Fan Wu and Zhiyuan Zhao and Rui Xu and Kaiwen Liu and Yuan Qu and Fukai Shang and Bo Zhang and Liqun Wei and Zhihao Sui and Wei Li and Botian Shi and Yu Qiao and Dahua Lin and Conghui He},
year={2024},
eprint={2409.18839},
archivePrefix={arXiv},
primaryClass={cs.CV},
url={https://arxiv.org/abs/2409.18839},
}
@article{he2024opendatalab,
title={Opendatalab: Empowering general artificial intelligence with open datasets},
author={He, Conghui and Li, Wei and Jin, Zhenjiang and Xu, Chao and Wang, Bin and Lin, Dahua},
journal={arXiv preprint arXiv:2407.13773},
year={2024}
}
Magic-Doc Fast speed ppt/pptx/doc/docx/pdf extraction tool
Magic-HTML Mixed web page extraction tool