--- library_name: transformers tags: - trl - sft datasets: - PedroCJardim/QASports metrics: - perplexity base_model: - meta-llama/Meta-Llama-3-8B --- # Llama3-8B-fine tuned by Arham This model is a fine-tuned version of Llama 3, optimized for open-ended text generation. It has been trained on SportsQA dataset to improve its ability to generate more relevant, coherent, and informative responses for football related queries. ## Model Details **Base Model:** LLama 3 - 8B Version **Fine Tuned On:** Sports QA Dataset. Visit here: https://huggingface.co/datasets/PedroCJardim/QASports **Use Case:** Open Ended questions regarding football. ### About Base Model Llama 3 LLaMA 3 8B is a transformer-based LLM with 8 billion parameters, trained on 15 trillion tokens for improved language understanding. It supports a 128,000-token context window, making it capable of handling longer text sequences. Compared to LLaMA 2, it processes text more efficiently and generates more accurate responses. The model balances performance and resource efficiency, making it deployable on commercial hardware. It's ideal for NLP tasks like resume screening, offering strong text analysis without heavy computational demands. ### Performance Compared to the base Llama 3 model, this fine-tuned version shows improvements in: - More accurate and contextually relevant responses. - Better coherence and consistency in generated text. - Improved understanding and interpretation of prompts. ### Usage You can load and use this model with `transformers`: ```python from transformers import AutoModelForCausalLM, AutoTokenizer model_name = "arham-15/llama3_8B_finetuned_by_arham" model = AutoModelForCausalLM.from_pretrained(model_name) tokenizer = AutoTokenizer.from_pretrained(model_name) device = "cuda" if torch.cuda.is_available() else "cpu" input_text = "The greatest footballer of all time is" inputs = tokenizer(input_text, return_tensors="pt").to(device) output = model.generate(**inputs, max_length=50) print(tokenizer.decode(output[0], skip_special_tokens=True)) ``` ### Limitations & Future Improvements - The model may still generate repetitive responses in some cases. - Further fine-tuning can improve domain-specific knowledge. ### Contribute & Feedback If you find this model useful, feel free to share feedback or contribute improvements!