Llama3.2-3B-Explained (GGUF)
A fine-tuned version of meta-llama/Llama-3.2-3B-Instruct trained on Explained 0.41k alpaca data using Auto-SFT — an automated hyperparameter search and supervised fine-tuning pipeline.
The base model was adapted to follow the style and content of the Explained 0.41k alpaca dataset. Expect improved performance on tasks similar to those represented in the training data.
Model Details
| Property | Value |
|---|---|
| Base model | meta-llama/Llama-3.2-3B-Instruct |
| Training data | data/Explained-0.41k-alpaca.json |
| Fine-tuning epochs | 2 |
| Fine-tuning date | 2026-03-25 |
| Fine-tuning method | LoRA (merged to full 16-bit) |
Training Hyperparameters
LoRA
| Parameter | Value |
|---|---|
r |
4 |
alpha |
8 |
dropout |
0.0 |
target_modules |
['q_proj', 'v_proj', 'k_proj', 'o_proj'] |
Training
| Parameter | Value |
|---|---|
learning_rate |
1e-05 |
batch_size |
1 |
gradient_accumulation_steps |
2 |
warmup_ratio |
0.0 |
max_seq_length |
512 |
GGUF Files
These quantized GGUF files can be used directly with llama.cpp, Ollama, LM Studio, and other compatible runtimes.
| File | Description |
|---|---|
Llama3.2-3B-Explained-BF16.gguf |
BF16 |
Llama3.2-3B-Explained-Q8_0.gguf |
8-bit — near-lossless, larger file |
Llama3.2-3B-Explained-Q6_K.gguf |
6-bit — high quality |
Llama3.2-3B-Explained-Q5_K_M.gguf |
5-bit medium — good quality/size balance |
Llama3.2-3B-Explained-Q5_K_S.gguf |
Q5_K_S |
Llama3.2-3B-Explained-Q4_K_M.gguf |
4-bit medium — recommended for most use cases |
Llama3.2-3B-Explained-Q4_K_S.gguf |
Q4_K_S |
Llama3.2-3B-Explained-Q3_K_L.gguf |
Q3_K_L |
Llama3.2-3B-Explained-Q3_K_M.gguf |
Q3_K_M |
Llama3.2-3B-Explained-Q3_K_S.gguf |
Q3_K_S |
Llama3.2-3B-Explained-Q2_K.gguf |
2-bit — smallest size, lowest quality |
Llama3.2-3B-Explained-IQ4_NL.gguf |
IQ4_NL |
Generated by Auto-SFT
- Downloads last month
- 798
Hardware compatibility
Log In to add your hardware
2-bit
3-bit
4-bit
5-bit
6-bit
8-bit
16-bit
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support
Model tree for theprint/Llama3.2-3B-Explained-GGUF
Base model
meta-llama/Llama-3.2-3B-Instruct Adapter
theprint/Llama3.2-3B-Explained