DeepSeek-OCR GGUF (for llama.cpp PR #17400)
This repository provides GGUF model files required for running the DeepSeek-OCR / MTMD support introduced in the following llama.cpp pull request:
👉 llama.cpp PR #17400 https://github.com/ggml-org/llama.cpp/pull/17400
These models are only compatible with the PR branch and will not run on upstream llama.cpp main.
📥 Download
You can download the model files directly using:
huggingface-cli download <this-repo> --include "deepseek-ocr-f16.gguf" --local-dir gguf_models/deepseek-ai
huggingface-cli download <this-repo> --include "mmproj-deepseek-ocr-f16.gguf" --local-dir gguf_models/deepseek-ai
🛠 Build llama.cpp PR Branch
Clone llama.cpp and check out the PR branch:
git clone https://github.com/ggml-org/llama.cpp
cd llama.cpp
# Checkout PR branch (recommended: GitHub CLI)
gh pr checkout 17400
# or manually:
# git fetch origin pull/17400/head:pr17400
# git checkout pr17400
Build:
cmake -B build -DCMAKE_BUILD_TYPE=Release
cmake --build build -j
🚀 Run Example
Use the llama-mtmd-cli executable from the PR:
build/bin/llama-mtmd-cli \
-m gguf_models/deepseek-ai/deepseek-ocr-f16.gguf \
--mmproj gguf_models/deepseek-ai/mmproj-deepseek-ocr-f16.gguf \
--image tmp/mtmd_test_data/Deepseek-OCR-2510.18234v1_page1.png \
-p "<|grounding|>Convert the document to markdown." \
--chat-template deepseek
- Downloads last month
- 358
Hardware compatibility
Log In
to view the estimation
8-bit
16-bit
32-bit
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
🙋
Ask for provider support
Model tree for sabafallah/DeepSeek-OCR-GGUF
Base model
deepseek-ai/DeepSeek-OCR