NEW: Added NVFP4 quants of ERNIE-Image and ERNIE-Image-Turbo

Original Ernie Source: https://huggingface.co/baidu/ERNIE-Image

FireRed-Image-Edit-1.0 Quantizations (FP8 & NVFP4)

This repository provides quantized versions of the FireRed-Image-Edit-1.0 model, optimized for reduced VRAM usage while maintaining high performance.

We provide weights in FP8 and the cutting-edge NVFP4 (Nvidia FP4) formats, making it easier to run this powerful image editing model on a wider range of hardware. Original source (about the model): https://huggingface.co/FireRedTeam/FireRed-Image-Edit-1.0

Available Versions

File Name Format Size Description
FireRed-Image-Edit-1.0_FP8.safetensors FP8 ~20.4 GB Standard 8-bit quantization for balanced performance.
FireRed-Image-Edit-1_NVFP4.safetensors NVFP4 ~11.6 GB Highly compressed 4-bit format optimized for NVIDIA Blackwell/Ada architecture.
Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support