Realistic Vision V6.0 B1 (Axera Optimized)

This repository contains an optimized derivative of Realistic Vision V6.0 B1, fused with LCM-LoRA for high-performance inference on Axera hardware (AX650N / LLM8850).

Note: This repository is a compilation of existing open-source works. The maintainer of this repository is not the original creator of the base models or the conversion tools, but has performed the compilation, graph surgery, and quantization for the Axera NPU platform.

Model Description

Performance Metrics

Benchmarks performed on a Raspberry Pi 5 (8GB) with an Axera AX650N M.2 Accelerator:

Component Latency
Text Encoder ~14 ms
U-Net (4 steps) ~1716 ms
VAE Decoder ~936 ms
Total Inference ~2.7 seconds
Model Loading ~15.7 seconds

Note: Inference time is for a 512x512 image with 4 steps.

Sample Gallery

Sample Image Prompt: A serene portrait of an elderly man with silver hair and a warm smile, wearing traditional embroidered clothing, sitting in a lush greenhouse surrounded by tropical plants, soft morning sunlight, 8k, highly detailed, realistic

Conversion Workflow Summary

  1. Export: Components exported from PyTorch/Diffusers to ONNX with LoRA fusion.
  2. Graph Surgery: Modified U-Net to expose the 1280-dim time embedding input (/down_blocks.0/resnets.0/act_1/Mul_output_0).
  3. Quantization: Compiled using Pulsar2 with u16/int8 mixed precision and representative calibration data.

Usage Instructions

To run this model on Axera hardware, you will need the axengine Python library.

1. Install Dependencies

pip install axengine transformers torch pillow numpy

2. Run Inference

Use the provided sample_inference.py script:

python sample_inference.py --prompt "A beautiful landscape, oil painting"

Credits and Citations

This work is made possible by the following upstream projects:

Optimization & Compilation: The artifacts in this repository were compiled and optimized for the Axera NPU using the Pulsar2 toolchain.

Additional Resources

Licensing and Restrictions

This model is subject to the CreativeML Open RAIL-M license. Please review the license for usage restrictions and safety guidelines.

Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support