Quantized GGUF models for UIGEN-T3-4B-Preview

This repository contains GGUF quantized versions of Tesslate/UIGEN-T3-4B-Preview. It uses 128 samples from the training dataset to calibrate iMatrix quants.

Available quantizations:

Full Precision

  • FP16 (full precision)
  • BF16 (bfloat16 precision)

imatrix Quantizations

These quantizations use importance matrix (imatrix) calibration for better quality:

  • IQ3_M-imat
  • IQ3_XXS-imat
  • Q4_K_M-imat
  • Q4_K_S-imat
  • IQ4_XS-imat
  • Q5_K_M-imat
  • Q5_K_S-imat
  • Q6_K-imat
  • Q8_0-imat

Standard Quantizations

  • Q2_K
  • Q3_K_L
  • Q3_K_M
  • Q3_K_S

About imatrix quantization

The imatrix quantizations in this repository use calibration data to preserve the most important weights during quantization, resulting in better model quality compared to standard quantization methods.

Original model

This is a quantized version of Tesslate/UIGEN-T3-4B-Preview.

Generated on

Wed Jun 11 18:08:41 UTC 2025

Downloads last month
141
GGUF
Model size
4.02B params
Architecture
qwen3
Hardware compatibility
Log In to view the estimation

2-bit

3-bit

4-bit

5-bit

6-bit

8-bit

16-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for qingy2024/UIGEN-T3-4B-Preview-128-GGUF

Base model

Qwen/Qwen3-4B-Base
Finetuned
Qwen/Qwen3-4B
Quantized
(7)
this model