โ›ฑ ํ•ด๋‹น ๋ชจ๋ธ์€์€ llama3.1 instruct๋ฅผ Foundation ๋ชจ๋ธ๋กœ ํ•˜๋Š” ํ•œ๊ตญ์–ด ๋ฐ

ํ•œ๊ตญ์˜ ๋‹ค์–‘ํ•œ ๋ฌธํ™”์— ์ ์šฉํ•  ์ˆ˜ ์žˆ๋„๋ก ํ•˜๊ธฐ ์œ„ํ•ด ๊ฐœ๋ฐœ ๋˜์—ˆ์œผ๋ฉฐ

์ž์ฒด ์ œ์ž‘ํ•œ 53์˜์—ญ์˜ ํ•œ๊ตญ์–ด ๋ฐ์ดํ„ฐ๋ฅผ ํ™œ์šฉํ•˜์—ฌ ํ•œ๊ตญ ์‚ฌํšŒ ๊ฐ€์น˜์™€ ๋ฌธํ™”๋ฅผ ์ดํ•ดํ•˜๋Š”

๋ชจ๋ธ ์ž…๋‹ˆ๋‹ค. Thanks for ktds โœŒ

โถ ํ•™์Šต ๋ฐ์ดํ„ฐ

  • ํ•ด๋‹น ๋ชจ๋ธ์€์€ ์ž์ฒด ๊ฐœ๋ฐœํ•œ ์ด 3.6GB ํฌ๊ธฐ์˜ ๋ฐ์ดํ„ฐ๋ฅผ ๋ฐ”ํƒ•์œผ๋กœ ํ•™์Šต๋˜์—ˆ์Šต๋‹ˆ๋‹ค. ๋ชจ๋‘ 233๋งŒ ๊ฑด์˜ QnA, ์š”์•ฝ, ๋ถ„๋ฅ˜ ๋“ฑ ๋ฐ์ดํ„ฐ๋ฅผ ํฌํ•จํ•˜๋ฉฐ, ๊ทธ ์ค‘ 133๋งŒ ๊ฑด์€ 53๊ฐœ ์˜์—ญ์˜ ๊ฐ๊ด€์‹ ๋ฌธ์ œ๋กœ ๊ตฌ์„ฑ๋˜์—ˆ์Šต๋‹ˆ๋‹ค. ์ด ์˜์—ญ์—๋Š” ํ•œ๊ตญ์‚ฌ, ์‚ฌํšŒ, ์žฌ๋ฌด, ๋ฒ•๋ฅ , ์„ธ๋ฌด, ์ˆ˜ํ•™, ์ƒ๋ฌผ, ๋ฌผ๋ฆฌ, ํ™”ํ•™ ๋“ฑ์ด ํฌํ•จ๋˜๋ฉฐ, Chain of Thought ๋ฐฉ์‹์œผ๋กœ ํ•™์Šต๋˜์—ˆ์Šต๋‹ˆ๋‹ค. ๋˜ํ•œ 130๋งŒ ๊ฑด์˜ ์ฃผ๊ด€์‹ ๋ฌธ์ œ๋Š” ํ•œ๊ตญ์‚ฌ, ์žฌ๋ฌด, ๋ฒ•๋ฅ , ์„ธ๋ฌด, ์ˆ˜ํ•™ ๋“ฑ 38๊ฐœ ์˜์—ญ์— ๊ฑธ์ณ ํ•™์Šต๋˜์—ˆ์Šต๋‹ˆ๋‹ค. ํ•™์Šต ๋ฐ์ดํ„ฐ ์ค‘ ํ•œ๊ตญ์˜ ์‚ฌํšŒ ๊ฐ€์น˜์™€ ์ธ๊ฐ„์˜ ๊ฐ์ •์„ ์ดํ•ดํ•˜๊ณ  ์ง€์‹œํ•œ ์‚ฌํ•ญ์— ๋”ฐ๋ผ ์ถœ๋ ฅํ•  ์ˆ˜ ์žˆ๋Š” ๋ฐ์ดํ„ฐ๋ฅผ ํ•™์Šตํ•˜์˜€์Šต๋‹ˆ๋‹ค.
  • ํ•™์Šต Instruction Datasets Format:
    {"prompt": "prompt text", "completion": "ideal generated text"}

โท ์‚ฌ์šฉ ์‚ฌ๋ก€

ํ•ด๋‹น ๋ชจ๋ธ์€ ๋‹ค์–‘ํ•œ ์‘์šฉ ๋ถ„์•ผ์—์„œ ์‚ฌ์šฉ๋  ์ˆ˜ ์žˆ์Šต๋‹ˆ๋‹ค. ์˜ˆ๋ฅผ ๋“ค์–ด:

  • ๊ต์œก ๋ถ„์•ผ: ์—ญ์‚ฌ, ์ˆ˜ํ•™, ๊ณผํ•™ ๋“ฑ ๋‹ค์–‘ํ•œ ํ•™์Šต ์ž๋ฃŒ์— ๋Œ€ํ•œ ์งˆ์˜์‘๋‹ต ๋ฐ ์„ค๋ช… ์ƒ์„ฑ.
  • ๋น„์ฆˆ๋‹ˆ์Šค: ๋ฒ•๋ฅ , ์žฌ๋ฌด, ์„ธ๋ฌด ๊ด€๋ จ ์งˆ์˜์— ๋Œ€ํ•œ ๋‹ต๋ณ€ ์ œ๊ณต ๋ฐ ๋ฌธ์„œ ์š”์•ฝ.
  • ์—ฐ๊ตฌ ๋ฐ ๋ฌธํ™”: ํ•œ๊ตญ ์‚ฌํšŒ์™€ ๋ฌธํ™”์— ๋งž์ถ˜ ์ž์—ฐ์–ด ์ฒ˜๋ฆฌ ์ž‘์—…, ๊ฐ์ • ๋ถ„์„, ๋ฌธ์„œ ์ƒ์„ฑ ๋ฐ ๋ฒˆ์—ญ.
  • ๊ณ ๊ฐ ์„œ๋น„์Šค: ์‚ฌ์šฉ์ž์™€์˜ ๋Œ€ํ™” ์ƒ์„ฑ ๋ฐ ๋งž์ถคํ˜• ์‘๋‹ต ์ œ๊ณต.
  • ์ด ๋ชจ๋ธ์€ ๋‹ค์–‘ํ•œ ์ž์—ฐ์–ด ์ฒ˜๋ฆฌ ์ž‘์—…์—์„œ ๋†’์€ ํ™œ์šฉ๋„๋ฅผ ๊ฐ€์ง‘๋‹ˆ๋‹ค.

โธ ํ•œ๊ณ„ โ›ˆโ›ˆ

  • ํ•ด๋‹น ๋ชจ๋ธ์€ ํ•œ๊ตญ์–ด์™€ ํ•œ๊ตญ ๋ฌธํ™”์— ํŠนํ™”๋˜์–ด ์žˆ์œผ๋‚˜, ํŠน์ • ์˜์—ญ(์˜ˆ: ์ตœ์‹  ๊ตญ์ œ ์ž๋ฃŒ, ์ „๋ฌธ ๋ถ„์•ผ)์˜ ๋ฐ์ดํ„ฐ ๋ถ€์กฑ์œผ๋กœ ์ธํ•ด ๋‹ค๋ฅธ ์–ธ์–ด ๋˜๋Š” ๋ฌธํ™”์— ๋Œ€ํ•œ ์‘๋‹ต์˜ ์ •ํ™•์„ฑ์ด ๋–จ์–ด์งˆ ์ˆ˜ ์žˆ์Šต๋‹ˆ๋‹ค. ๋˜ํ•œ, ๋ณต์žกํ•œ ๋…ผ๋ฆฌ์  ์‚ฌ๊ณ ๋ฅผ ์š”๊ตฌํ•˜๋Š” ๋ฌธ์ œ์— ๋Œ€ํ•ด ์ œํ•œ๋œ ์ถ”๋ก  ๋Šฅ๋ ฅ์„ ๋ณด์ผ ์ˆ˜ ์žˆ์œผ๋ฉฐ, ํŽธํ–ฅ๋œ ๋ฐ์ดํ„ฐ๊ฐ€ ํฌํ•จ๋  ๊ฒฝ์šฐ ํŽธํ–ฅ๋œ ์‘๋‹ต์ด ์ƒ์„ฑ๋  ๊ฐ€๋Šฅ์„ฑ๋„ ์กด์žฌํ•ฉ๋‹ˆ๋‹ค.

โบ ์‚ฌ์šฉ ๋ฐฉ๋ฒ•


  from transformers import AutoModel, AutoTokenizer
  
  tokenizer = AutoTokenizer.from_pretrained("SEOKDONG/llama3.1_korean_v0.1_sft_by_aidx")
  model = AutoModel.from_pretrained("SEOKDONG/llama3.1_korean_v0.1_sft_by_aidx")

    input_text =  """ ใ€Œ๊ตญ๋ฏผ๊ฑด๊ฐ•๋ณดํ—˜๋ฒ•ใ€์ œ44์กฐ, ใ€Œ๊ตญ๋ฏผ๊ฑด๊ฐ•๋ณดํ—˜๋ฒ• ์‹œํ–‰๋ นใ€์ œ19์กฐ,ใ€Œ์•ฝ๊ด€์˜ ๊ทœ์ œ์— ๊ด€ํ•œ ๋ฒ•๋ฅ ใ€์ œ5์กฐ, ใ€Œ์ƒ๋ฒ•ใ€์ œ54์กฐ ์ฐธ์กฐ ํŒ๋‹จ ํ•ด์ค˜"""
    inputs = tokenizer(input_text, return_tensors="pt")
  with torch.no_grad():
        outputs = model.generate(**inputs, max_length=1024,  temperature=0.5, do_sample=True, repetition_penalty=1.15)
  result = tokenizer.decode(outputs[0], skip_special_tokens=True)
  print(result)

Hereโ€™s the English version of the provided text:

Downloads last month
9
Safetensors
Model size
8B params
Tensor type
F32
ยท
Inference Providers NEW
Input a message to start chatting with SEOKDONG/llama3.1_korean_v0.1_sft_by_aidx.

Model tree for SEOKDONG/llama3.1_korean_v0.1_sft_by_aidx

Finetuned
(2433)
this model
Quantizations
3 models

Dataset used to train SEOKDONG/llama3.1_korean_v0.1_sft_by_aidx

Space using SEOKDONG/llama3.1_korean_v0.1_sft_by_aidx 1