Comment on What's everybody using for AI art and video generation nowadays?
tal@lemmy.today 1 week ago
I don’t do video generation.
I’m mostly moved away from Automatic1111 to ComfyUI. If you’ve ever used an image processing program that uses a flowchart-style of operations to modify images, it looks kinda like that. Comfy’s more work to learn — you need to learn and understand some things that Automatic1111 is doing internally ---- but:
-
It’s much more capable at building up complex images and series of dependent processes that are re-generated when you make a change in a workflow.
-
It can run Flux. Last I looked, Automatic1111 could not. I understand that Forge can, and is a little more like Automatic1111, but I haven’t spent time with it. I’d call Flux and derived models are quite impressive from a natural language standpoint. My experience on SD and Pony-based models meant that most of the prompts I wrote were basically sequences of keywords. With Flux, it’s far more natural-language looking, and it can do some particularly neat stuff just from the prompt (“The image is a blended composite with a progression from left to right showing winter to spring to summer to autumn.”).
-
It has queuing. It may be that Automatic1111 has since picked it up, but I found it to be a serious lack back when I was using it.
-
ComfyUI scales up better if you’re using a lot of plugins. In Automatic1111, a plugin adds buttons and UI elements into each page. In Comfy, a plugin almost always just adds more nodes to the node library, doesn’t go fiddling with the existing UI.
That being said, I’m out-of-date on Automatic1111. But last I looked, the major selling point for me was the SD Ultimate Upscale plugin, and that’s been subsequently ported to ComfyUI.
For me, one major early selling point was that a workflow that I frequently wanted was to (a) generate an image and then (b) trivially perform an SD Ultimate Upscale. In Automatic1111, that required setting up txt2img and SD Ultimate Upscale in img2img, then running a txt2img operation to generate an image, waiting until it finished, manually clicking the button to send the image to img2img, and then running the upscale operation. With ComfyUI, I just save a workflow that does all that, and Comfy will rerun everything necessary based on any changes that I make (and won’t rerun things that aren’t). I can just disable the upscale portion of the workflow if I don’t need that bit. ComfyUI was a higher barrier to entry, but it made more-complex tasks much less time-consuming and require less nursemaiding.