| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
|
|
| import unittest |
|
|
| import numpy as np |
| import torch |
|
|
| from diffusers import VersatileDiffusionImageVariationPipeline |
| from diffusers.utils.testing_utils import load_image, require_torch_gpu, slow, torch_device |
|
|
|
|
| torch.backends.cuda.matmul.allow_tf32 = False |
|
|
|
|
| class VersatileDiffusionImageVariationPipelineFastTests(unittest.TestCase): |
| pass |
|
|
|
|
| @slow |
| @require_torch_gpu |
| class VersatileDiffusionImageVariationPipelineIntegrationTests(unittest.TestCase): |
| def test_inference_image_variations(self): |
| pipe = VersatileDiffusionImageVariationPipeline.from_pretrained("shi-labs/versatile-diffusion") |
| pipe.to(torch_device) |
| pipe.set_progress_bar_config(disable=None) |
|
|
| image_prompt = load_image( |
| "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/versatile_diffusion/benz.jpg" |
| ) |
| generator = torch.manual_seed(0) |
| image = pipe( |
| image=image_prompt, |
| generator=generator, |
| guidance_scale=7.5, |
| num_inference_steps=50, |
| output_type="numpy", |
| ).images |
|
|
| image_slice = image[0, 253:256, 253:256, -1] |
|
|
| assert image.shape == (1, 512, 512, 3) |
| expected_slice = np.array([0.0441, 0.0469, 0.0507, 0.0575, 0.0632, 0.0650, 0.0865, 0.0909, 0.0945]) |
|
|
| assert np.abs(image_slice.flatten() - expected_slice).max() < 1e-2 |
|
|