"Native refiner swap inside one single k-sampler. The advantage is that now the refiner model can reuse the base model's momentum (or ODE's history parameters) collected from k-sampling to achieve more coherent sampling. In Automatic1111's high-res fix and ComfyUI's node system, the base model and refiner use two independent k-samplers, which means the momentum is largely wasted, and the sampling continuity is broken. Fooocus uses its own advanced k-diffusion sampling that ensures seamless, native, and continuous swap in a refiner setup."
This is so interesting and seems obvious in retrospect, but super impressive! The code is simple too, going to hack around with this over the weekend :)
As a frontend developer, this reads to me as technobabble you'd find in entertainment media. In general, I learn about things not directly related to my sphere of interests by osmosis, but this is on another level. Reminds me of the time when I started my computing journey. I wonder if I'll be able to understand this eventually just by reading a relevant comment or blog here and there.
Ah well. Hopefully it is soon. Also, on behalf of all Apple Silicon Mac users, would be nice if the author looked into implementing Metal FlashAttention [1].
For those who don't know, ControlNet is often used in conjunction with Stable Diffusion. It lets you add extra conditions to guide what is being generated. There are extensions for Automatic1111's stable diffusion webui that can make use of ControlNet. Some examples I've seen are copying the pose of a person/animal in an image and outputting a different person with the same pose (and extending to videos). Also taking line art drawings and filling it in with style.
> Learned from Midjourney, the manual tweaking is not needed, and users only need to focus on the prompts and images
Except prompt-based tweaking doesn’t work very well in MJ; certainly not as well as manually-directed in-painting and out-painting. It’s virtually impossible in MJ to hold one part of the image constant while adding to/modifying the remainder.
Interesting, and I look forward to using it, but I wish the distribution had kept the folder-name conventions of AUTOMATIC1111, so that we could more easily have used symbolic links for folders of LoRAs and checkpoints etc. that we'd rather not duplicate.
Apparently it uses the folder structure of ComfyUI - I just symlinked the models folder from that and it worked with no issues. (I also reused my ComfyUI venv, just had to do a pip install pygit2 to make it work)
Can't you symlink individual files? More effort but only a quick bit of scripting away.
(I've occasionally used a duplicate file eliminator that finds dups over a certain size and replaces them with symlinks. You can run it on an entire subtree or drive)
Tbh I’m (loosely) following commit message best practices in all of my projects out of irrational fear of being viewed as unprofessional. But never needed that effing prose in my workflow. Maybe a keyword from time to time. I’m using code, not messages to navigate history, and in a rare occasion. If all my messages turned into “i” I’d lose nothing, because all rationales and essentials are in code comments. I’d better seen dates (and related grouping) in a log by default and looked for a commit by some grepping patch contents rather than messages.
definitely the smoothest install process and relatively snappy on my local windows machine that I've come across.
I do hope to see some ControlNet integrations as that's become a key part of my workflow for exploring new images.
Just like I expected, I get this error when trying to run it on my AMD GPU...
"RuntimeError: Found no NVIDIA driver on your system. Please check that you have an NVIDIA GPU and installed a driver from http://www.nvidia.com/Download/index.aspx"
Maybe it can get modified to use DirectML? Although it looks like it's using PyTorch 2.0, and I think torch-directml only supports 1.13. Why is ML and GPGPU such a dependency mess?
[+] [-] erwannmillon|2 years ago|reply
This is so interesting and seems obvious in retrospect, but super impressive! The code is simple too, going to hack around with this over the weekend :)
[+] [-] Klaster_1|2 years ago|reply
[+] [-] liuliu|2 years ago|reply
[+] [-] airgapstopgap|2 years ago|reply
> Coming soon ...
Ah well. Hopefully it is soon. Also, on behalf of all Apple Silicon Mac users, would be nice if the author looked into implementing Metal FlashAttention [1].
1. https://github.com/philipturner/metal-flash-attention
[+] [-] sunpazed|2 years ago|reply
[+] [-] politelemon|2 years ago|reply
https://stable-diffusion-art.com/controlnet/
[+] [-] isoprophlex|2 years ago|reply
https://stable-diffusion-art.com/qr-code/
[+] [-] kashunstva|2 years ago|reply
Except prompt-based tweaking doesn’t work very well in MJ; certainly not as well as manually-directed in-painting and out-painting. It’s virtually impossible in MJ to hold one part of the image constant while adding to/modifying the remainder.
[+] [-] AbraKdabra|2 years ago|reply
[+] [-] k3liutZu|2 years ago|reply
I am not sure what I would have expected upon reading this comment, but I was not prepared.
[+] [-] dvrp|2 years ago|reply
[+] [-] Hard_Space|2 years ago|reply
[+] [-] erikprotagonist|2 years ago|reply
[+] [-] andybak|2 years ago|reply
(I've occasionally used a duplicate file eliminator that finds dups over a certain size and replaces them with symlinks. You can run it on an entire subtree or drive)
[+] [-] dvrp|2 years ago|reply
[+] [-] GaggiX|2 years ago|reply
[+] [-] netghost|2 years ago|reply
[+] [-] peddling-brink|2 years ago|reply
[+] [-] wruza|2 years ago|reply
[+] [-] yellow_postit|2 years ago|reply
[+] [-] ilkke|2 years ago|reply
My biggest discovery so far is using shuffle to guide the output style (and curating a folder of great style guide images).
[+] [-] captn3m0|2 years ago|reply
Anything better than X forwarding?
[+] [-] freeone3000|2 years ago|reply
[+] [-] unknown|2 years ago|reply
[deleted]
[+] [-] sorenjan|2 years ago|reply
"RuntimeError: Found no NVIDIA driver on your system. Please check that you have an NVIDIA GPU and installed a driver from http://www.nvidia.com/Download/index.aspx"
Maybe it can get modified to use DirectML? Although it looks like it's using PyTorch 2.0, and I think torch-directml only supports 1.13. Why is ML and GPGPU such a dependency mess?
[+] [-] bufferoverflow|2 years ago|reply
[+] [-] AuryGlenz|2 years ago|reply
For those that don’t know, the Adetailer extension for Auto1111 does a second pass on faces at a higher resolution and then inpaints them back in.
[+] [-] natch|2 years ago|reply
[+] [-] dvrp|2 years ago|reply
[+] [-] r-k-jo|2 years ago|reply
[+] [-] burkaygur|2 years ago|reply
[+] [-] brucethemoose2|2 years ago|reply
Lots of the changes just... make sense.