bartowski commited on
Commit
b39773e
·
verified ·
1 Parent(s): 08edbff

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -0
README.md CHANGED
@@ -51,6 +51,7 @@ IQ2_XXS may not be final, the size increase is quite substantial so I may want t
51
  | [DeepSeek-V3-0324-IQ3_XXS.gguf](https://huggingface.co/bartowski/deepseek-ai_DeepSeek-V3-0324-GGUF/tree/main/deepseek-ai_DeepSeek-V3-0324-IQ3_XXS) | IQ3_XXS | 257.93GB | true | Lower quality, new method with decent performance, comparable to Q3 quants. |
52
  | [DeepSeek-V3-0324-Q2_K_L-V2.gguf](https://huggingface.co/bartowski/deepseek-ai_DeepSeek-V3-0324-GGUF/tree/main/deepseek-ai_DeepSeek-V3-0324-Q2_K_L-V2) | Q2_K_L | 247.35GB | true | Attempted to modify tensor quant levels for better performance. Also uses Q8_0 for embed and output weights. |
53
  | [DeepSeek-V3-0324-Q2_K_L.gguf](https://huggingface.co/bartowski/deepseek-ai_DeepSeek-V3-0324-GGUF/tree/main/deepseek-ai_DeepSeek-V3-0324-Q2_K_L) | Q2_K_L | 244.93GB | true | Uses Q8_0 for embed and output weights. Very low quality but surprisingly usable. |
 
54
  | [DeepSeek-V3-0324-IQ2_M.gguf](https://huggingface.co/bartowski/deepseek-ai_DeepSeek-V3-0324-GGUF/tree/main/deepseek-ai_DeepSeek-V3-0324-IQ2_M) | IQ2_M | 217.43GB | true | Relatively low quality, uses SOTA techniques to be surprisingly usable. |
55
  | [DeepSeek-V3-0324-IQ2_S.gguf](https://huggingface.co/bartowski/deepseek-ai_DeepSeek-V3-0324-GGUF/tree/main/deepseek-ai_DeepSeek-V3-0324-IQ2_S) | IQ2_S | 197.00GB | true | Low quality, uses SOTA techniques to be usable. |
56
  | [DeepSeek-V3-0324-IQ2_XXS-V2.gguf](https://huggingface.co/bartowski/deepseek-ai_DeepSeek-V3-0324-GGUF/tree/main/deepseek-ai_DeepSeek-V3-0324-IQ2_XXS-V2) | IQ2_XXS | 188.95GB | true | *Being replaced soon by 179GB version* Attempted to modify tensor quant levels for better performance. |
 
51
  | [DeepSeek-V3-0324-IQ3_XXS.gguf](https://huggingface.co/bartowski/deepseek-ai_DeepSeek-V3-0324-GGUF/tree/main/deepseek-ai_DeepSeek-V3-0324-IQ3_XXS) | IQ3_XXS | 257.93GB | true | Lower quality, new method with decent performance, comparable to Q3 quants. |
52
  | [DeepSeek-V3-0324-Q2_K_L-V2.gguf](https://huggingface.co/bartowski/deepseek-ai_DeepSeek-V3-0324-GGUF/tree/main/deepseek-ai_DeepSeek-V3-0324-Q2_K_L-V2) | Q2_K_L | 247.35GB | true | Attempted to modify tensor quant levels for better performance. Also uses Q8_0 for embed and output weights. |
53
  | [DeepSeek-V3-0324-Q2_K_L.gguf](https://huggingface.co/bartowski/deepseek-ai_DeepSeek-V3-0324-GGUF/tree/main/deepseek-ai_DeepSeek-V3-0324-Q2_K_L) | Q2_K_L | 244.93GB | true | Uses Q8_0 for embed and output weights. Very low quality but surprisingly usable. |
54
+ | [DeepSeek-V3-0324-IQ2_M-V2.gguf](https://huggingface.co/bartowski/deepseek-ai_DeepSeek-V3-0324-GGUF/tree/main/deepseek-ai_DeepSeek-V3-0324-IQ2_M-V2) | IQ2_M | 224.49GB | true | Attempted to modify tensor quant levels for better performance. |
55
  | [DeepSeek-V3-0324-IQ2_M.gguf](https://huggingface.co/bartowski/deepseek-ai_DeepSeek-V3-0324-GGUF/tree/main/deepseek-ai_DeepSeek-V3-0324-IQ2_M) | IQ2_M | 217.43GB | true | Relatively low quality, uses SOTA techniques to be surprisingly usable. |
56
  | [DeepSeek-V3-0324-IQ2_S.gguf](https://huggingface.co/bartowski/deepseek-ai_DeepSeek-V3-0324-GGUF/tree/main/deepseek-ai_DeepSeek-V3-0324-IQ2_S) | IQ2_S | 197.00GB | true | Low quality, uses SOTA techniques to be usable. |
57
  | [DeepSeek-V3-0324-IQ2_XXS-V2.gguf](https://huggingface.co/bartowski/deepseek-ai_DeepSeek-V3-0324-GGUF/tree/main/deepseek-ai_DeepSeek-V3-0324-IQ2_XXS-V2) | IQ2_XXS | 188.95GB | true | *Being replaced soon by 179GB version* Attempted to modify tensor quant levels for better performance. |