Meta-Llama-3.1-Math-QA-finetuning-Group-3

This model is a fine-tuned version of Meta-Llama-3.1-8B on the MetaMathQA dataset for mathematical reasoning tasks. Training Details

Method: QLoRA (4-bit quantization with LoRA adapters) Framework: Unsloth for memory and time efficient fine-tuning Dataset: 50,000 randomly selected samples from MetaMathQA (seed=42) Hardware: Google Colab T4 GPU

Hyperparameters

# QLoRA Configuration
load_in_4bit = True
lora_r = 16
lora_alpha = 16
lora_dropout = 0

# Training Configuration
num_train_epochs = 5
max_steps = 50
learning_rate = 5e-5
per_device_train_batch_size = 2
gradient_accumulation_steps = 4
warmup_steps = 5
weight_decay = 0.001
lr_scheduler_type = "linear"
optim = "adamw_8bit"
seed = 3407

Training Results

  • 50th Epoch training loss: 0.551400
Downloads last month
20
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for youth-ai-initiative/Meta-Llama-3.1-Math-QA-finetuning-Group-3

Finetuned
(1656)
this model

Dataset used to train youth-ai-initiative/Meta-Llama-3.1-Math-QA-finetuning-Group-3