gghfez commited on
Commit
3528d6e
·
verified ·
1 Parent(s): 556b7a0

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +23 -0
README.md ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ quantized_by: gghfez
3
+ pipeline_tag: text-generation
4
+ base_model: deepseek-ai/DeepSeek-V3-0324
5
+ license: mit
6
+ base_model_relation: quantized
7
+ tags:
8
+ - mla
9
+ - imatrix
10
+ - deepseek_v3
11
+ - conversational
12
+ - ik_llama.cpp
13
+ ---
14
+
15
+ ## `ik_llama.cpp` imatrix MLA Quantizations of DeepSeek-V3-0324
16
+
17
+ This is an IQ3_KS quant of DeepSeek-V3-0324 using [ubergarm](https://huggingface.co/ubergarm)'s IQ3_KS recipe from [ubergarm/DeepSeek-TNG-R1T2-Chimera-GGUF](https://huggingface.co/ubergarm/DeepSeek-TNG-R1T2-Chimera-GGUF) and Imatrix file from [ubergarm/DeepSeek-V3-0324-GGUF](https://huggingface.co/ubergarm/DeepSeek-V3-0324-GGUF).
18
+
19
+ This quant collection **REQUIRES** [ik_llama.cpp](https://github.com/ikawrakow/ik_llama.cpp/) fork to support advanced non-linear SotA quants and Multi-Head Latent Attention (MLA). Do **not** download these big files and expect them to run on mainline vanilla llama.cpp, ollama, LM Studio, KoboldCpp, etc!
20
+
21
+ See [ubergarm/DeepSeek-V3-0324-GGUF](https://huggingface.co/ubergarm/DeepSeek-V3-0324-GGUF) for his other quants and more details about them.
22
+
23
+ I've uploaded the converted BF16 weights [gghfez/DeepSeek-V3-0324-256x21B-BF16](https://huggingface.co/gghfez/DeepSeek-V3-0324-256x21B-BF16) if I, or anyone else wants to create similar quants in the future.