llm.create_chat_completion(
messages = "No input example has been defined for this model task."
)
ik_llama.cpp quantizations of DeepSeek-R1-0528
Quantized using ik_llama.cpp build = 3788 (4622fadc)
NOTE: These quants MUST be run using the llama.cpp fork, ik_llama.cpp
Credits to @ubergarm for his DeepSeek quant recipes for which these quants were based on.
| name | file size | quant type | bpw |
|---|---|---|---|
| DeepSeek-R1-0528-IQ4_KT | 322.355 GiB | IQ4_KT (97.5%) / Q8_0 (2.5%) |
4.127 |
- Downloads last month
- 3
Hardware compatibility
Log In to add your hardware
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support
Model tree for Kebob/DeepSeek-R1-0528-IK_GGUF
Base model
deepseek-ai/DeepSeek-R1-0528
# !pip install llama-cpp-python from llama_cpp import Llama llm = Llama.from_pretrained( repo_id="Kebob/DeepSeek-R1-0528-IK_GGUF", filename="", )