Visual Generation Models
Collection
5 items • Updated
Production-ready Diffusers export of PixNerd-XL/16 class-conditional ImageNet checkpoints.
PixNerd-XL-16-256epoch%3D319-step%3D1600000_emainit.ckpt256x256PixNerd-XL-16-512res512_ft200k_epoch%3D325-step%3D1800000_emainit.ckpt512x512Both checkpoints are packaged with:
pipeline.pymodeling_pixnerd_transformer_2d.pyscheduling_pixnerd_flow_match.pytransformer/ weights + configscheduler/ configpip install torch diffusers
import torch
from diffusers import DiffusionPipeline
model_dir = "PixNerd-XL-16-256" # or PixNerd-XL-16-512
pipe = DiffusionPipeline.from_pretrained(
model_dir,
custom_pipeline=f"{model_dir}/pipeline.py",
torch_dtype=torch.float32,
).to("cpu") # use "cuda" if available
# Class-conditional generation: class label 207 (golden retriever)
images = pipe(
prompt=[207],
num_images_per_prompt=1,
height=256,
width=256,
num_inference_steps=25,
guidance_scale=4.0,
timeshift=3.0,
order=2,
).images
images[0].save("sample.png")
prompt for conditioning input.prompt=[207].height and width should match checkpoint intent (256 or 512), but custom sizes work if divisible by patch size.conversion_metadata.json.Source paper (ICLR 2026):
Source code:
@article{2507.23268,
Author = {Shuai Wang and Ziteng Gao and Chenhui Zhu and Weilin Huang and Limin Wang},
Title = {PixNerd: Pixel Neural Field Diffusion},
Year = {2025},
Eprint = {arXiv:2507.23268},
}