--- tags: - generative - text-generation - openai - gpt-oss - reap - bf16 metrics: - perplexity - accuracy model-index: - name: gpt-oss-20b-reap-0.4-bf16 results: [] --- # gpt-oss-20b-reap-0.4-bf16 This repository contains a bfloat16 version of the `sandeshrajx/gpt-oss-20b-reap-0.4-mxfp4` model. ## Model Description This model is a bfloat16 version of the MXFP4 quantized `openai/gpt-oss-20b` model. - **Original Model:** `openai/gpt-oss-20b` - **Pruning Method:** `reap` with a compression ratio of `0.4` - **Original Quantization Method:** MXFP4 weight-only quantization - **Current Format:** bfloat16 - **Dataset used for pruning/quantization (if applicable):** `theblackcat102/evol-codealpaca-v1` ## Usage You can load this model using the `transformers` library: ```python from transformers import AutoModelForCausalLM, AutoTokenizer model_name = "sandeshrajx/gpt-oss-20b-reap-0.4-bf16" tokenizer = AutoTokenizer.from_pretrained(model_name) model = AutoModelForCausalLM.from_pretrained(model_name, torch_dtype=torch.bfloat16) # Use the model for inference # ... ``` ## License (Please specify the license of the original model and any modifications)