Spaces:
Running
on
Zero
Running
on
Zero
| import torch | |
| from PIL import Image | |
| from transformers import AutoProcessor | |
| from longcat_image.models import LongCatImageTransformer2DModel | |
| from longcat_image.pipelines import LongCatImageEditPipeline | |
| if __name__ == '__main__': | |
| device = torch.device('cuda') | |
| checkpoint_dir = './weights/LongCat-Image-Edit' | |
| text_processor = AutoProcessor.from_pretrained( checkpoint_dir, subfolder = 'tokenizer' ) | |
| transformer = LongCatImageTransformer2DModel.from_pretrained( checkpoint_dir , subfolder = 'transformer', | |
| torch_dtype=torch.bfloat16, use_safetensors=True).to(device) | |
| pipe = LongCatImageEditPipeline.from_pretrained( | |
| checkpoint_dir, | |
| transformer=transformer, | |
| text_processor=text_processor, | |
| ) | |
| pipe.to(device, torch.bfloat16) | |
| generator = torch.Generator("cpu").manual_seed(43) | |
| img = Image.open('assets/test.png').convert('RGB') | |
| prompt = 'ε°η«εζη' | |
| image = pipe( | |
| img, | |
| prompt, | |
| negative_prompt='', | |
| guidance_scale=4.5, | |
| num_inference_steps=50, | |
| num_images_per_prompt=1, | |
| generator=generator | |
| ).images[0] | |
| image.save('./edit_example.png') | |