| { | |
| "arch_specifier": "no-align+fused-gelu-mlp", | |
| "architectures": [ | |
| "PrismaticForConditionalGeneration" | |
| ], | |
| "hf_llm_id": "Qwen/Qwen2.5-1.5B", | |
| "image_resize_strategy": "resize-naive", | |
| "image_sizes": [ | |
| 224, | |
| 224 | |
| ], | |
| "llm_backbone_id": "qwen25-1_5b-extra", | |
| "llm_max_length": 32768, | |
| "model_type": "prismatic", | |
| "output_projector_states": false, | |
| "pad_to_multiple_of": 64, | |
| "pad_token_id": 151643, | |
| "text_config": { | |
| "bos_token_id": null, | |
| "eos_token_id": null, | |
| "hidden_size": 1536, | |
| "intermediate_size": 8960, | |
| "max_position_embeddings": 32768, | |
| "max_window_layers": 28, | |
| "model_type": "qwen2", | |
| "num_attention_heads": 12, | |
| "num_hidden_layers": 28, | |
| "num_key_value_heads": 2, | |
| "pad_token_id": 151643, | |
| "sliding_window": 4096, | |
| "torch_dtype": "bfloat16", | |
| "use_sliding_window": false, | |
| "vocab_size": 151936 | |
| }, | |
| "timm_model_ids": [ | |
| "vit_large_patch14_reg4_dinov2.lvd142m", | |
| "vit_so400m_patch14_siglip_224" | |
| ], | |
| "timm_override_act_layers": [ | |
| null, | |
| null | |
| ], | |
| "torch_dtype": "bfloat16", | |
| "transformers_version": "4.40.1", | |
| "use_fused_vision_backbone": true, | |
| "vision_backbone_id": "dinosiglip-vit-so-224px" | |
| } | |