NEW: Added NVFP4 quants of ERNIE-Image and ERNIE-Image-Turbo
Original Ernie Source: https://huggingface.co/baidu/ERNIE-Image
FireRed-Image-Edit-1.0 Quantizations (FP8 & NVFP4)
This repository provides quantized versions of the FireRed-Image-Edit-1.0 model, optimized for reduced VRAM usage while maintaining high performance.
We provide weights in FP8 and the cutting-edge NVFP4 (Nvidia FP4) formats, making it easier to run this powerful image editing model on a wider range of hardware. Original source (about the model): https://huggingface.co/FireRedTeam/FireRed-Image-Edit-1.0
Available Versions
| File Name | Format | Size | Description |
|---|---|---|---|
FireRed-Image-Edit-1.0_FP8.safetensors |
FP8 | ~20.4 GB | Standard 8-bit quantization for balanced performance. |
FireRed-Image-Edit-1_NVFP4.safetensors |
NVFP4 | ~11.6 GB | Highly compressed 4-bit format optimized for NVIDIA Blackwell/Ada architecture. |
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support