saiteki-kai commited on
Commit
3d36d37
·
verified ·
1 Parent(s): 58898f0

End of training

Browse files
README.md CHANGED
@@ -3,12 +3,27 @@ library_name: transformers
3
  license: llama3.1
4
  base_model: meta-llama/Llama-3.1-8B-Instruct
5
  tags:
 
 
 
6
  - generated_from_trainer
 
 
7
  metrics:
8
  - accuracy
9
  model-index:
10
  - name: QA-Llama-3.1-4155
11
- results: []
 
 
 
 
 
 
 
 
 
 
12
  ---
13
 
14
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
@@ -16,20 +31,20 @@ should probably proofread and complete it, then remove this comment. -->
16
 
17
  # QA-Llama-3.1-4155
18
 
19
- This model is a fine-tuned version of [meta-llama/Llama-3.1-8B-Instruct](https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct) on an unknown dataset.
20
  It achieves the following results on the evaluation set:
21
  - Loss: 0.0781
22
- - Accuracy: 0.6965
23
- - Macro F1: 0.6444
24
- - Macro Precision: 0.7361
25
- - Macro Recall: 0.5968
26
  - Micro F1: 0.7539
27
- - Micro Precision: 0.8035
28
  - Micro Recall: 0.7100
29
- - Flagged/accuracy: 0.8561
30
  - Flagged/precision: 0.9050
31
- - Flagged/recall: 0.8284
32
- - Flagged/f1: 0.8650
33
 
34
  ## Model description
35
 
 
3
  license: llama3.1
4
  base_model: meta-llama/Llama-3.1-8B-Instruct
5
  tags:
6
+ - multi-label
7
+ - question-answering
8
+ - text-classification
9
  - generated_from_trainer
10
+ datasets:
11
+ - beavertails
12
  metrics:
13
  - accuracy
14
  model-index:
15
  - name: QA-Llama-3.1-4155
16
+ results:
17
+ - task:
18
+ name: Text Classification
19
+ type: text-classification
20
+ dataset:
21
+ name: saiteki-kai/BeaverTails-it
22
+ type: beavertails
23
+ metrics:
24
+ - name: Accuracy
25
+ type: accuracy
26
+ value: 0.6964434241607612
27
  ---
28
 
29
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
 
31
 
32
  # QA-Llama-3.1-4155
33
 
34
+ This model is a fine-tuned version of [meta-llama/Llama-3.1-8B-Instruct](https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct) on the saiteki-kai/BeaverTails-it dataset.
35
  It achieves the following results on the evaluation set:
36
  - Loss: 0.0781
37
+ - Accuracy: 0.6964
38
+ - Macro F1: 0.6445
39
+ - Macro Precision: 0.7365
40
+ - Macro Recall: 0.5970
41
  - Micro F1: 0.7539
42
+ - Micro Precision: 0.8036
43
  - Micro Recall: 0.7100
44
+ - Flagged/accuracy: 0.8560
45
  - Flagged/precision: 0.9050
46
+ - Flagged/recall: 0.8283
47
+ - Flagged/f1: 0.8649
48
 
49
  ## Model description
50
 
all_results.json ADDED
@@ -0,0 +1,38 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 3.0,
3
+ "eval_accuracy": 0.6964434241607612,
4
+ "eval_flagged/accuracy": 0.856040190305087,
5
+ "eval_flagged/f1": 0.8649247674346008,
6
+ "eval_flagged/precision": 0.9049547636933729,
7
+ "eval_flagged/recall": 0.8282861498908852,
8
+ "eval_loss": 0.07808855175971985,
9
+ "eval_macro_f1": 0.6445018666803658,
10
+ "eval_macro_precision": 0.7364622252548454,
11
+ "eval_macro_recall": 0.5969831652053833,
12
+ "eval_micro_f1": 0.7539061393653881,
13
+ "eval_micro_precision": 0.8036066497604959,
14
+ "eval_micro_recall": 0.7099951988904102,
15
+ "eval_runtime": 86.4772,
16
+ "eval_samples_per_second": 695.143,
17
+ "eval_steps_per_second": 5.435,
18
+ "test_accuracy": 0.6834501137860821,
19
+ "test_flagged/accuracy": 0.8485147921906815,
20
+ "test_flagged/f1": 0.8584934687141619,
21
+ "test_flagged/precision": 0.8994256241941155,
22
+ "test_flagged/recall": 0.8211247257745198,
23
+ "test_loss": 0.08345632255077362,
24
+ "test_macro_f1": 0.6200746051485584,
25
+ "test_macro_precision": 0.7147873439119878,
26
+ "test_macro_recall": 0.5715412667131191,
27
+ "test_micro_f1": 0.7434023275005474,
28
+ "test_micro_precision": 0.7991924971031287,
29
+ "test_micro_recall": 0.6948931078996253,
30
+ "test_runtime": 90.6304,
31
+ "test_samples_per_second": 736.971,
32
+ "test_steps_per_second": 5.76,
33
+ "total_flos": 1.0625604952409506e+19,
34
+ "train_loss": 0.07807489077325368,
35
+ "train_runtime": 15386.354,
36
+ "train_samples_per_second": 105.487,
37
+ "train_steps_per_second": 1.648
38
+ }
eval_results.json ADDED
@@ -0,0 +1,18 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 3.0,
3
+ "eval_accuracy": 0.6964434241607612,
4
+ "eval_flagged/accuracy": 0.856040190305087,
5
+ "eval_flagged/f1": 0.8649247674346008,
6
+ "eval_flagged/precision": 0.9049547636933729,
7
+ "eval_flagged/recall": 0.8282861498908852,
8
+ "eval_loss": 0.07808855175971985,
9
+ "eval_macro_f1": 0.6445018666803658,
10
+ "eval_macro_precision": 0.7364622252548454,
11
+ "eval_macro_recall": 0.5969831652053833,
12
+ "eval_micro_f1": 0.7539061393653881,
13
+ "eval_micro_precision": 0.8036066497604959,
14
+ "eval_micro_recall": 0.7099951988904102,
15
+ "eval_runtime": 86.4772,
16
+ "eval_samples_per_second": 695.143,
17
+ "eval_steps_per_second": 5.435
18
+ }
test_results.json ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "test_accuracy": 0.6834501137860821,
3
+ "test_flagged/accuracy": 0.8485147921906815,
4
+ "test_flagged/f1": 0.8584934687141619,
5
+ "test_flagged/precision": 0.8994256241941155,
6
+ "test_flagged/recall": 0.8211247257745198,
7
+ "test_loss": 0.08345632255077362,
8
+ "test_macro_f1": 0.6200746051485584,
9
+ "test_macro_precision": 0.7147873439119878,
10
+ "test_macro_recall": 0.5715412667131191,
11
+ "test_micro_f1": 0.7434023275005474,
12
+ "test_micro_precision": 0.7991924971031287,
13
+ "test_micro_recall": 0.6948931078996253,
14
+ "test_runtime": 90.6304,
15
+ "test_samples_per_second": 736.971,
16
+ "test_steps_per_second": 5.76
17
+ }
train_results.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 3.0,
3
+ "total_flos": 1.0625604952409506e+19,
4
+ "train_loss": 0.07807489077325368,
5
+ "train_runtime": 15386.354,
6
+ "train_samples_per_second": 105.487,
7
+ "train_steps_per_second": 1.648
8
+ }
trainer_state.json ADDED
The diff for this file is too large to render. See raw diff