ekurtic commited on
Commit
8d36c47
·
verified ·
1 Parent(s): d9f8d09

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +46 -0
README.md ADDED
@@ -0,0 +1,46 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language:
3
+ - en
4
+ - de
5
+ - fr
6
+ - it
7
+ - pt
8
+ - hi
9
+ - es
10
+ - th
11
+ license: llama3.3
12
+ pipeline_tag: text-generation
13
+ tags:
14
+ - facebook
15
+ - meta
16
+ - pytorch
17
+ - llama
18
+ - llama-3
19
+ - neuralmagic
20
+ - redhat
21
+ - speculators
22
+ - eagle3
23
+ ---
24
+
25
+ # Llama-3.3-70B-Instruct-speculator.eagle3
26
+
27
+ ## Model Overview
28
+ - **Verifier:** meta-llama/Llama-3.3-70B-Instruct
29
+ - **Speculative Decoding Algorithm:** EAGLE-3
30
+ - **Model Architecture:** Eagle3Speculator
31
+ - **Release Date:** 09/15/2025
32
+ - **Version:** 1.0
33
+ - **Model Developers:** RedHat
34
+
35
+ This is a speculator model designed for use with [meta-llama/Llama-3.3-70B-Instruct](https://huggingface.co/meta-llama/Llama-3.3-70B-Instruct), based on the [EAGLE-3](https://arxiv.org/abs/2503.01840) speculative decoding algorithm.
36
+ It was trained using the [speculators](https://github.com/vllm-project/speculators) library on a combination of the [Aeala/ShareGPT_Vicuna_unfiltered](https://huggingface.co/datasets/Aeala/ShareGPT_Vicuna_unfiltered) and the `train_sft` split of [HuggingFaceH4/ultrachat_200k](https://huggingface.co/datasets/HuggingFaceH4/ultrachat_200k) datasets.
37
+
38
+ ## Evaluations
39
+
40
+ Subset of GSM8k (math reasoning):
41
+ * acceptance_rate = [0.801, 0.637, 0.464]
42
+ * conditional_acceptance_rate = [0.801, 0.795, 0.729]
43
+
44
+ Subset of MTBench:
45
+ * acceptance_rate = [0.733, 0.537, 0.384]
46
+ * conditional_acceptance_rate = [0.733, 0.733, 0.715]