SINQ: Sinkhorn-Normalized Quantization for Calibration-Free Low-Precision LLM Weights
Paper
•
2509.22944
•
Published
•
79
This is a 4-bit quantized version of DeepSeek-Prover-V1.5-Base using SINQ (Sinkhorn-Normalized Quantization).
BaseQuantizeConfig(
nbits=4,
group_size=128,
method="sinq",
tiling_mode="2D",
axis=1
)
pip install torch transformers
pip install git+https://github.com/huawei-csl/SINQ.git
import torch
from sinq.patch_model import AutoSINQHFModel
from transformers import AutoTokenizer
# Load quantized model
tokenizer = AutoTokenizer.from_pretrained("Minhdn/deepseek-prover-sinq-4bit")
model = AutoSINQHFModel.from_quantized_safetensors(
"Minhdn/deepseek-prover-sinq-4bit",
device="cuda:0",
compute_dtype=torch.bfloat16
)
# Generate
prompt = "theorem add_comm (a b : Nat) : a + b = b + a := by"
inputs = tokenizer(prompt, return_tensors="pt").to(model.device)
outputs = model.generate(**inputs, max_new_tokens=100)
result = tokenizer.decode(outputs[0], skip_special_tokens=True)
print(result)
For better quality at the cost of larger model size, consider:
If you use this model, please cite both the original DeepSeek-Prover paper and SINQ:
@article{deepseek2024prover,
title={DeepSeek-Prover: Advancing Theorem Proving in LLMs through Large-Scale Synthetic Data},
author={DeepSeek-AI},
journal={arXiv preprint arXiv:2405.14333},
year={2024}
}
@article{sinq2024,
title={SINQ: Sinkhorn-Normalized Quantization for Calibration-Free Low-Precision LLMs},
author={SINQ Authors},
journal={arXiv preprint arXiv:2509.22944},
year={2024}
}
This model inherits the MIT license from the original DeepSeek-Prover-V1.5-Base model.
Base model
deepseek-ai/DeepSeek-Prover-V1.5-Base