Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Lumina Next img2img #91

Closed
thelemuet opened this issue Jun 23, 2024 · 7 comments
Closed

Lumina Next img2img #91

thelemuet opened this issue Jun 23, 2024 · 7 comments

Comments

@thelemuet
Copy link

Hello! I have been playing with the SFT model and I'm really impressed to far!

Are there any plans to implement img2img for your image models? For regular img2img usage like SD but also I get the feeling this model would work really well for upscaling/refining since it's already able to generate very consistent/coherent images at higher resolutions (unlike ie: SDXL at higher than 1024x1024).

@gaopengpjlab
Copy link
Contributor

Can you suggest any popular image2image model? I will take a look at it. By the way, in lumina-t2x technique report, we illustrate that training-free image2image editing is possible.

@kijai
Copy link

kijai commented Jun 23, 2024

I'm also very interested in img2img and curious how the showcased image editing was achieved? I have tried the basics of using encoded image instead of random noise/augmented with noise, but that doesn't seem to work. Or did you use reverse sampling?

@thelemuet
Copy link
Author

Can you suggest any popular image2image model? I will take a look at it. By the way, in lumina-t2x technique report, we illustrate that training-free image2image editing is possible.

I believe all Stable Diffusion model can do img2img (have not tried SD3 but I assume it does).

Here is a quick example usage using SDXL. Starting from a simple image of a 3d render where I know want to keep the composition, lighting and colors. I would set it to generate 30 steps but skip a few steps depending on how much I want to retain from the original image and how much new details I want the model to generate. Here is what it looks like from original image and different results depending on starting step:

iti1

And an upscale example:

upscale1

@thelemuet thelemuet reopened this Jun 23, 2024
@gaopengpjlab
Copy link
Contributor

We will release a demo similar as stablediffusionimg2image pipeline.

https://huggingface.co/docs/diffusers/en/api/pipelines/stable_diffusion/img2img

@zhuole1025
Copy link
Contributor

Hi! We just update the code for img2img using Lumina next: https://github.com/Alpha-VLLM/Lumina-T2X/blob/main/lumina_next_t2i_mini/scripts/sample_img2img.sh
Here is the the demo:
image

@zhuole1025
Copy link
Contributor

I'm also very interested in img2img and curious how the showcased image editing was achieved? I have tried the basics of using encoded image instead of random noise/augmented with noise, but that doesn't seem to work. Or did you use reverse sampling?

Our implementation is identical to this in the diffusers, but you have to be cautious with the start timestep which is affected by the time-shifting scale. Welcome to add this into ComfyUI and have some tests!

@kijai
Copy link

kijai commented Jun 26, 2024

I'm also very interested in img2img and curious how the showcased image editing was achieved? I have tried the basics of using encoded image instead of random noise/augmented with noise, but that doesn't seem to work. Or did you use reverse sampling?

Our implementation is identical to this in the diffusers, but you have to be cautious with the start timestep which is affected by the time-shifting scale. Welcome to add this into ComfyUI and have some tests!

Got it. Thank you!
image

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

4 participants