Qwen3-4B-Instruct-2507-MPOA-GGUF

This model was converted to GGUF format from YanLabs/Qwen3-4B-Instruct-2507-MPOA using GGUF Forge.

Quants

The following quants are available: Q2_K, Q3_K_S, Q3_K_M, Q3_K_L, Q4_0, Q4_K_S, Q4_K_M, Q5_0, Q5_K_S, Q5_K_M, Q6_K, Q8_0

Links

  • Host your own GGUF Forge (Beta): GGUF Forge
  • llama.cpp Repository used for quantization: llama.cpp

Automatic Conversion by GGUF Forge

Downloads last month
92
GGUF
Model size
4B params
Architecture
qwen3
Hardware compatibility
Log In to add your hardware

2-bit

3-bit

4-bit

5-bit

6-bit

8-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for Akicou/Qwen3-4B-Instruct-2507-MPOA-GGUF

Quantized
(4)
this model

Collection including Akicou/Qwen3-4B-Instruct-2507-MPOA-GGUF