--- language: - en license: apache-2.0 library_name: transformers tags: - code - python - docstring - documentation - code-generation - local-llm - privacy - ollama - qwen3 - knowledge-distillation - developer-tools base_model: Qwen/Qwen3-0.6B pipeline_tag: text-generation model-index: - name: Distil-Localdoc-Qwen3-0.6B results: - task: type: text-generation name: Docstring Generation metrics: - type: accuracy value: 0.76 name: LLM-as-Judge Accuracy verified: false ---
distil labs
---
Homepage GitHub Hugging Face
LinkedIn Slack Twitter
# Distil-Localdoc-Qwen3-0.6B A small language model (SLM) fine-tuned by Distil Labs for generating high-quality Python docstrings in Google style. Optimized to run locally via Ollama, ensuring your proprietary code never leaves your infrastructure. *********** [GITHUB DEMO AND CODE](https://github.com/distil-labs/Distil-localdoc/) *********** ## Model Details - **Developed by**: Distil Labs GmbH - **License**: Apache 2.0 - **Finetuned from**: Qwen/Qwen3-0.6B - **Model Size**: 0.6B parameters - **Deployment**: Local inference via Ollama ## Use-case Given Python functions or methods without docstrings, the model generates complete, properly formatted documentation following Google style guide. **Before:** ```python def calculate_total(items, tax_rate=0.08, discount=None): subtotal = sum(item['price'] * item['quantity'] for item in items) if discount: subtotal *= (1 - discount) return subtotal * (1 + tax_rate) ``` **After:** ```python def calculate_total(items, tax_rate=0.08, discount=None): """ Calculate the total cost of items, applying a tax rate and optionally a discount. Args: items: List of item objects with price and quantity tax_rate: Tax rate expressed as a decimal (default 0.08) discount: Discount rate expressed as a decimal; if provided, the subtotal is multiplied by (1 - discount) Returns: Total amount after applying the tax Example: >>> items = [{'price': 10, 'quantity': 2}, {'price': 5, 'quantity': 1}] >>> calculate_total(items, tax_rate=0.1, discount=0.05) 22.5 """ subtotal = sum(item['price'] * item['quantity'] for item in items) if discount: subtotal *= (1 - discount) return subtotal * (1 + tax_rate) ``` The model handles: - **Functions**: Parameter descriptions, return values, exceptions, and usage examples - **Methods**: Instance and class method documentation with proper formatting - **Note**: The tool skips double underscore (dunder: __xxx__) methods ## Why Local? **Privacy & Security**: Proprietary codebases contain intellectual property and trade secrets. Cloud APIs create: - IP exposure risks - Compliance violations (GDPR, SOC 2, HIPAA) - Security audit failures - Dependency on external services **Speed & Cost**: Document entire codebases in minutes without API rate limits or per-token charges. ## Training The tuned model was trained using knowledge distillation, leveraging the teacher model GPT-OSS-120B. We used 28 diverse Python functions and classes as seed data and supplemented them with 10,000 synthetic examples covering various domains: - Data science and machine learning - Web development (Flask, FastAPI, Django) - DevOps and system utilities - Algorithm implementations - API clients and wrappers Training data includes examples with: - Various function complexities (simple to async patterns) - Error handling patterns - Async/await patterns - Different parameter types and return values ## Evaluation We evaluated the model on 250 held-out test examples using LLM-as-a-judge methodology to assess the overall quality of generated docstrings. | Model | Size | Accuracy | |--------------------|------|---------------| | GPT-OSS (thinking) | 120B | 0.81 ± 0.02 | | Qwen3 0.6B (tuned) | 0.6B | 0.76 ± 0.01 | | Qwen3 0.6B (base) | 0.6B | 0.55 ± 0.04 | The fine-tuned model achieves **94%** of the teacher model's performance while running entirely on local hardware with **zero API costs** and **complete privacy**. ## How to Use ### Installation Follow the instructions in the [Github repository](https://github.com/distil-labs/Distil-localdoc/) Quick start: ```bash # Install Ollama curl -fsSL https://ollama.com/install.sh | sh # Download and build the model pip install huggingface_hub hf download distil-labs/Distil-Localdoc-Qwen3-0.6B --local-dir distil-model cd distil-model ollama create localdoc_qwen3 -f Modelfile # Run on your code python localdoc_cli.py --file your_script.py ``` ### CLI Usage ```bash # Basic usage (generates Google-style docstrings) python localdoc_cli.py --file my_module.py # Use specific model python localdoc_cli.py --file my_module.py --model localdoc_qwen3 ``` The tool will: 1. Parse your Python file using AST 2. Identify all functions and methods without docstrings (skips dunder methods) 3. Generate appropriate docstrings based on code structure 4. Preserve all original code and existing docstrings 5. Output a new file with `_documented` suffix ## Model Sources - **Homepage**: [https://distillabs.ai](https://distillabs.ai) - **Repository**: [https://github.com/distil-labs/Distil-localdoc](https://github.com/distil-labs/Distil-localdoc) - **Contact**: contact@distillabs.ai ## Citation ```bibtex @software{distil_localdoc_2024, title = {Distil-Localdoc: Local Python Documentation Generation with SLMs}, author = {Distil Labs}, year = {2024}, url = {https://huggingface.co/distil-labs/Distil-Localdoc-Qwen3-0.6B} } ``` ## Community - Follow us on [LinkedIn](https://www.linkedin.com/company/distil-labs/) - Join our [Slack community](https://join.slack.com/t/distil-labs-community/shared_invite/zt-36zqj87le-i3quWUn2bjErRq22xoE58g) - Star us on [GitHub](https://github.com/distil-labs/Distil-localdoc)