Mixtral of Experts
Paper
•
2401.04088
•
Published
•
160
This is a merge of pre-trained language models created using mergekit.
This model is a merge of all of my SOVL models, in the hopes to create the most unhinged and wild model possible. But in Mixtral fashion!
It may be insane, it may be incoherent. I can't load it :3
This model was merged using the Mixture Of Experts method.
The following models were included in the merge:
The following YAML configuration was used to produce this model:
base_model: saishf/Ortho-SOVL-8B-L3
gate_mode: random
dtype: bfloat16
experts:
- source_model: saishf/Ortho-SOVL-8B-L3
- source_model: saishf/SOVLish-Maid-L3-8B
- source_model: saishf/Merge-Mayhem-L3-V2.1
- source_model: saishf/Merge-Mayhem-L3-V2
Detailed results can be found here
| Metric | Value |
|---|---|
| Avg. | 66.76 |
| AI2 Reasoning Challenge (25-Shot) | 61.95 |
| HellaSwag (10-Shot) | 79.38 |
| MMLU (5-Shot) | 65.49 |
| TruthfulQA (0-shot) | 51.48 |
| Winogrande (5-shot) | 75.69 |
| GSM8k (5-shot) | 66.57 |