2shlee/llama3-8b-ko-chat-v1

ν•œκ΅­μ–΄ λŒ€ν™”λ₯Ό μœ„ν•΄ νŒŒμΈνŠœλ‹λœ LLaMA 3 8B λͺ¨λΈμž…λ‹ˆλ‹€.

Model Description

이 λͺ¨λΈμ€ Meta의 LLaMA 3 8B Instruct λͺ¨λΈμ„ 기반으둜 ν•œκ΅­μ–΄ 챗봇 μš©λ„λ‘œ LoRA νŒŒμΈνŠœλ‹λ˜μ—ˆμŠ΅λ‹ˆλ‹€.

Training Details

  • Base Model: meta-llama/Meta-Llama-3-8B-Instruct
  • Fine-tuning Method: LoRA (PEFT)
  • Target Modules: q_proj, k_proj, v_proj, o_proj

How to Use

With PEFT (Recommended)

from transformers import AutoTokenizer, AutoModelForCausalLM
from peft import PeftModel
import torch

# 베이슀 λͺ¨λΈ λ‘œλ“œ
base_model = "meta-llama/Meta-Llama-3-8B-Instruct"
tokenizer = AutoTokenizer.from_pretrained(base_model)
model = AutoModelForCausalLM.from_pretrained(
    base_model,
    torch_dtype=torch.float16,
    device_map="auto"
)

# LoRA μ–΄λŒ‘ν„° 적용
model = PeftModel.from_pretrained(model, "2shlee/llama3-8b-ko-chat-v1")

# μΆ”λ‘ 
messages = [{"role": "user", "content": "μ•ˆλ…•ν•˜μ„Έμš”!"}]
input_text = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
inputs = tokenizer(input_text, return_tensors="pt").to(model.device)

with torch.no_grad():
    outputs = model.generate(**inputs, max_new_tokens=256, temperature=0.7)
    
response = tokenizer.decode(outputs[0], skip_special_tokens=True)
print(response)

With vLLM (Production)

python -m vllm.entrypoints.openai.api_server \
    --model meta-llama/Meta-Llama-3-8B-Instruct \
    --enable-lora \
    --lora-modules ko-chat=2shlee/llama3-8b-ko-chat-v1

Intended Uses

  • ν•œκ΅­μ–΄ λŒ€ν™”ν˜• AI μ„œλΉ„μŠ€
  • 챗봇 μ–΄μ‹œμŠ€ν„΄νŠΈ
  • Q&A μ‹œμŠ€ν…œ
  • ν…μŠ€νŠΈ 생성

Limitations

  • 베이슀 λͺ¨λΈ(LLaMA 3)의 일반적인 ν•œκ³„μ  적용
  • ν•™μŠ΅ 데이터에 μ—†λŠ” λ„λ©”μΈμ—μ„œλŠ” μ„±λŠ₯이 μ €ν•˜λ  수 있음
  • μ‹€μ‹œκ°„ μ •λ³΄λ‚˜ μ΅œμ‹  지식이 ν•„μš”ν•œ μ§ˆλ¬Έμ—λŠ” 뢀적합

License

이 λͺ¨λΈμ€ Llama 3 Community Licenseλ₯Ό λ”°λ¦…λ‹ˆλ‹€.

Acknowledgements

Built with Meta Llama 3

Citation

@misc{2shlee_llama3_8b_ko_chat_v1},
  author = {shlee},
  title = {2shlee/llama3-8b-ko-chat-v1},
  year = {2026},
  publisher = {Hugging Face},
  howpublished = {\url{https://huggingface.co/2shlee/llama3-8b-ko-chat-v1}}
}
Downloads last month
15
Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support

Model tree for 2shlee/llama3-8b-ko-chat-v1

Adapter
(2200)
this model