Update README.md
Browse files
README.md
CHANGED
@@ -51,6 +51,7 @@ IQ2_XXS may not be final, the size increase is quite substantial so I may want t
|
|
51 |
| [DeepSeek-V3-0324-IQ3_XXS.gguf](https://huggingface.co/bartowski/deepseek-ai_DeepSeek-V3-0324-GGUF/tree/main/deepseek-ai_DeepSeek-V3-0324-IQ3_XXS) | IQ3_XXS | 257.93GB | true | Lower quality, new method with decent performance, comparable to Q3 quants. |
|
52 |
| [DeepSeek-V3-0324-Q2_K_L-V2.gguf](https://huggingface.co/bartowski/deepseek-ai_DeepSeek-V3-0324-GGUF/tree/main/deepseek-ai_DeepSeek-V3-0324-Q2_K_L-V2) | Q2_K_L | 247.35GB | true | Attempted to modify tensor quant levels for better performance. Also uses Q8_0 for embed and output weights. |
|
53 |
| [DeepSeek-V3-0324-Q2_K_L.gguf](https://huggingface.co/bartowski/deepseek-ai_DeepSeek-V3-0324-GGUF/tree/main/deepseek-ai_DeepSeek-V3-0324-Q2_K_L) | Q2_K_L | 244.93GB | true | Uses Q8_0 for embed and output weights. Very low quality but surprisingly usable. |
|
|
|
54 |
| [DeepSeek-V3-0324-IQ2_M.gguf](https://huggingface.co/bartowski/deepseek-ai_DeepSeek-V3-0324-GGUF/tree/main/deepseek-ai_DeepSeek-V3-0324-IQ2_M) | IQ2_M | 217.43GB | true | Relatively low quality, uses SOTA techniques to be surprisingly usable. |
|
55 |
| [DeepSeek-V3-0324-IQ2_S.gguf](https://huggingface.co/bartowski/deepseek-ai_DeepSeek-V3-0324-GGUF/tree/main/deepseek-ai_DeepSeek-V3-0324-IQ2_S) | IQ2_S | 197.00GB | true | Low quality, uses SOTA techniques to be usable. |
|
56 |
| [DeepSeek-V3-0324-IQ2_XXS-V2.gguf](https://huggingface.co/bartowski/deepseek-ai_DeepSeek-V3-0324-GGUF/tree/main/deepseek-ai_DeepSeek-V3-0324-IQ2_XXS-V2) | IQ2_XXS | 188.95GB | true | *Being replaced soon by 179GB version* Attempted to modify tensor quant levels for better performance. |
|
|
|
51 |
| [DeepSeek-V3-0324-IQ3_XXS.gguf](https://huggingface.co/bartowski/deepseek-ai_DeepSeek-V3-0324-GGUF/tree/main/deepseek-ai_DeepSeek-V3-0324-IQ3_XXS) | IQ3_XXS | 257.93GB | true | Lower quality, new method with decent performance, comparable to Q3 quants. |
|
52 |
| [DeepSeek-V3-0324-Q2_K_L-V2.gguf](https://huggingface.co/bartowski/deepseek-ai_DeepSeek-V3-0324-GGUF/tree/main/deepseek-ai_DeepSeek-V3-0324-Q2_K_L-V2) | Q2_K_L | 247.35GB | true | Attempted to modify tensor quant levels for better performance. Also uses Q8_0 for embed and output weights. |
|
53 |
| [DeepSeek-V3-0324-Q2_K_L.gguf](https://huggingface.co/bartowski/deepseek-ai_DeepSeek-V3-0324-GGUF/tree/main/deepseek-ai_DeepSeek-V3-0324-Q2_K_L) | Q2_K_L | 244.93GB | true | Uses Q8_0 for embed and output weights. Very low quality but surprisingly usable. |
|
54 |
+
| [DeepSeek-V3-0324-IQ2_M-V2.gguf](https://huggingface.co/bartowski/deepseek-ai_DeepSeek-V3-0324-GGUF/tree/main/deepseek-ai_DeepSeek-V3-0324-IQ2_M-V2) | IQ2_M | 224.49GB | true | Attempted to modify tensor quant levels for better performance. |
|
55 |
| [DeepSeek-V3-0324-IQ2_M.gguf](https://huggingface.co/bartowski/deepseek-ai_DeepSeek-V3-0324-GGUF/tree/main/deepseek-ai_DeepSeek-V3-0324-IQ2_M) | IQ2_M | 217.43GB | true | Relatively low quality, uses SOTA techniques to be surprisingly usable. |
|
56 |
| [DeepSeek-V3-0324-IQ2_S.gguf](https://huggingface.co/bartowski/deepseek-ai_DeepSeek-V3-0324-GGUF/tree/main/deepseek-ai_DeepSeek-V3-0324-IQ2_S) | IQ2_S | 197.00GB | true | Low quality, uses SOTA techniques to be usable. |
|
57 |
| [DeepSeek-V3-0324-IQ2_XXS-V2.gguf](https://huggingface.co/bartowski/deepseek-ai_DeepSeek-V3-0324-GGUF/tree/main/deepseek-ai_DeepSeek-V3-0324-IQ2_XXS-V2) | IQ2_XXS | 188.95GB | true | *Being replaced soon by 179GB version* Attempted to modify tensor quant levels for better performance. |
|