saurabh5 commited on
Commit
b3502c4
·
verified ·
1 Parent(s): 69d440c

Update config.json on branch main

Browse files
Files changed (1) hide show
  1. config.json +6 -5
config.json CHANGED
@@ -1,9 +1,10 @@
1
  {
2
  "architectures": [
3
- "Olmo2RetrofitForCausalLM"
4
  ],
5
  "attention_bias": false,
6
  "attention_dropout": 0.0,
 
7
  "eos_token_id": 100257,
8
  "hidden_act": "silu",
9
  "hidden_size": 4096,
@@ -44,7 +45,7 @@
44
  "full_attention"
45
  ],
46
  "max_position_embeddings": 65536,
47
- "model_type": "olmo2-retrofit",
48
  "num_attention_heads": 32,
49
  "num_hidden_layers": 32,
50
  "num_key_value_heads": 32,
@@ -61,8 +62,8 @@
61
  "rope_theta": 500000,
62
  "sliding_window": 4096,
63
  "tie_word_embeddings": false,
64
- "torch_dtype": "bfloat16",
65
- "transformers_version": "4.54.0.dev0",
66
- "use_cache": false,
67
  "vocab_size": 100278
68
  }
 
 
1
  {
2
  "architectures": [
3
+ "Olmo3ForCausalLM"
4
  ],
5
  "attention_bias": false,
6
  "attention_dropout": 0.0,
7
+ "dtype": "bfloat16",
8
  "eos_token_id": 100257,
9
  "hidden_act": "silu",
10
  "hidden_size": 4096,
 
45
  "full_attention"
46
  ],
47
  "max_position_embeddings": 65536,
48
+ "model_type": "olmo3",
49
  "num_attention_heads": 32,
50
  "num_hidden_layers": 32,
51
  "num_key_value_heads": 32,
 
62
  "rope_theta": 500000,
63
  "sliding_window": 4096,
64
  "tie_word_embeddings": false,
65
+ "transformers_version": "4.57.0",
66
+ "use_cache": true,
 
67
  "vocab_size": 100278
68
  }
69
+