Tell us what is missing in ControlNet Integrated #932
Replies: 38 comments 9 replies
-
There are two new controlnet-based architectures to try: One input with multiple preprocessor inputs Modified variants of some blocks that seem to perform well on SVD and SDXL. |
Beta Was this translation helpful? Give feedback.
-
instant id+inpaint PLZ |
Beta Was this translation helpful? Give feedback.
-
First of all I want to say - thanks for the Great tool. That the previous version was very good, that the new version looks very promising (judging by commits). Regarding controlnet.
|
Beta Was this translation helpful? Give feedback.
-
A feature request: segmentation tagging control net (not yet existing) Inputs: character in default frontal pose (photo), a segment map where segment color defines clothing region in frontal pose, a segment map with same colors but in target pose. Desired output: ControlNet should map clothing features from default pose into target pose using color-coded segments as strict guidance. That means - without adding features not present in segment in default pose, without symmetry breaks and so on. Probably specific color channel should hint about SYMMETRY for robustness. Target pose should not be restricted to "usual poses", theoretically upside-down poses, partially occluded poses, self occlusions can be resolved just fine by such 1-1 segmentation mapping |
Beta Was this translation helpful? Give feedback.
-
The outpaint of xinsir union-controlnet-promax ... if i could outpaint with SDXL it would be wonderful. Kepp up the good works ... |
Beta Was this translation helpful? Give feedback.
-
I hopefully update controlnet will work with 2 major extension : btw ur works is awsome |
Beta Was this translation helpful? Give feedback.
-
The biggest need here might simply be a document outlining how Forge's built-in CN differs in options and behavior. It's not much, but the will-not-adds and the workarounds to some of those would be useful for newcomers and converts from A1111. |
Beta Was this translation helpful? Give feedback.
-
First of all, Thank you for your work lllyasviel. |
Beta Was this translation helpful? Give feedback.
-
Well, I can't get it to work in the Photoshop Plugin/Extention... that's missing for me, but probably not something that can be helped since it's pretty much not getting updates on their end. I'd probably rather have it work faster in the UI than have a poor implementation in photoshop anyway at this point. Looking forward to seeing "The Chosen One" for character consistency soon as a CN model or extension of some kind. |
Beta Was this translation helpful? Give feedback.
-
Fooocus sdxl inpaint. Inpaint_v26.fooocus.patch never worked as well on Forge for me as it did on Foocus. |
Beta Was this translation helpful? Give feedback.
-
For hands you also have: And this is looking interesting as well: I've seen people add manual images to OpenPose, but ClickDiff does it automatically, and the other do too. Would be great if we can finally fix hands and feet with ControlNet.
|
Beta Was this translation helpful? Give feedback.
-
1.) SDXL / Pony inpainting would be amazing. 2.) A custom ControlNet which utilizes either photos or prompts to transfer outfits/clothing onto a generation. This could maybe be enhanced by utilizing layerdiffuse to overlay it correctly upon the subject through multiple passes. (Idea, not sure if could be implemented) 3.) I don't know if this is possible either, but instead of something like regional prompter or Forge Couple, creating a controlnet that instead recognizes separate individual characters interacting. I know there are things similar to this that exist, but they aren't quite what I'm looking for to generate multiple characters interacting easily while maintaining a level of versatility and not being constrained to one specific pose via OpenPose. If this idea is at all possible, a controlnet model could utilize a reference image to apply an interaction/action between multiple characters. Or instead of a reference image, it could utilize a syntax similar to Forge Couple's "NEWLINE" , but instead of generating in a new space of the 2d plane, applies character adherence to subject(s) in the scene. Then, an action could be specified, detailing how these characters are interacting. Basically, a prompt adherence helper for models that struggle with prompt adherence, especially as it relates to characteristics of specific individuals in a scene. Example: Again, not sure if this is even possible without a multimodal model, but I think it would be more useful to have something that helps ensure characters look a certain way than it is to make sure they are in a specific place/pose (most of the time.) |
Beta Was this translation helpful? Give feedback.
-
For me, what's missing is
2a. Dsine on independent controlnet for A1111 has a bug that if you use a different resolution/ratio than the image, the preprocessor will generate zoomed into the image rather than crop or resize. Its been a bug since the release of Dsine as a preprocessor. This happens regardless of settings. You essentially have to format images to be the same exact resolution and ratio as the preprocessor settings, otherwise it zooms in, which can be a pain if you're working with images of mixed resolutions. 2b. Propose a feature based on the last problem: A new feature that allows people to use a draggable and resizable selector box over the controlnet image to determine the focus area of the preprocessor. I'd imagine this would also be useful for other controlnet preprocessors too, as you can then take any part of any image and make it the focus of the preprocessor on the fly. So for example, you drag a 1:1 ratio box overlay on the image, have it be resizable, and then it'll use the inputted resolution, even if the image itself is smaller; So you could zoom into a face on a 512x512 image and make just the face area preprocess at 1024x1024. Simple and versatile way of handling controlnet inputs of varying resolutions and aspect ratios and fills the gap that "just resize" and "crop and resize" doesn't cover.
|
Beta Was this translation helpful? Give feedback.
-
Illyas, I'm very new to this and probably the most ignorant of all those involved here, but that doesn't stop me from telling you that I admire not only your great work in Forge, which by the way has been very useful for me as a graphic designer, but, above all, your selflessness. I'm sure that if there were more people like you in this world, things would be different, as we say in Chile. Thank you very much!! A big hug. |
Beta Was this translation helpful? Give feedback.
-
Ah! y lo nuevo de Forge funcionando muy bien en mi 3060ti, algo lento aún pero con muy buena calidad, Flux y SDXL |
Beta Was this translation helpful? Give feedback.
-
Hello, I'm new switch to ForgeUI but it seems the integrated controlnet batch function doen not work properly, for the batch option it created multiple masks but only generate 1 image like the picture shows. I really like to use batch controlnet, truly appreciated if this issue can be addressed. |
Beta Was this translation helpful? Give feedback.
-
For me, SUPIR as controlnet or extension option |
Beta Was this translation helpful? Give feedback.
-
Hi illya, just to thank you and tell you that Spaces is great, but what you've done with Diffusion in Low Bits is amazing, my 8G VRAM card no longer suffers hahaha |
Beta Was this translation helpful? Give feedback.
-
When i try to use controlnet for flux, it doesn't work at all. when there will be fix for it or it is a controlnet model problem? |
Beta Was this translation helpful? Give feedback.
-
I’d love to see your take on an upscaler utilizing multiple ControlNets. I know you can kind of do this in Img2Img, but something that needs less tinkering. Basically, something that utilizes multiple ControlNets like IP-Adapter, Tile etc to make an upscaled image. The ControlNets are already preselected and processed on the backend based on the base model (SD1.5,SDXL,Flux). |
Beta Was this translation helpful? Give feedback.
-
Appreciate this repo a lot! It would be great if the mask had a copy of the input image so that you can see clearly what you are masking. |
Beta Was this translation helpful? Give feedback.
-
I've been using ForgeUi recently. It's good to make presets settings. |
Beta Was this translation helpful? Give feedback.
-
I think you've been doing amazing so far I would really love it if you
could get Regional prompter extension to work again, But I figure that you
are getting stuff squared away at your own pace.
…On Fri, Aug 23, 2024 at 3:15 PM SunGreen777 ***@***.***> wrote:
I've been using ForgeUi recently. It's good to make *presets settings.*
—
Reply to this email directly, view it on GitHub
<#932 (comment)>,
or unsubscribe
<https://github.com/notifications/unsubscribe-auth/AHN7U7YEHNNT4OO5FTNP5ODZS6YBPAVCNFSM6AAAAABMAFM6XCVHI2DSMVQWIX3LMV43URDJONRXK43TNFXW4Q3PNVWWK3TUHMYTANBTGQ4TGNQ>
.
You are receiving this because you are subscribed to this thread.Message
ID:
<lllyasviel/stable-diffusion-webui-forge/repo-discussions/932/comments/10434936
@github.com>
|
Beta Was this translation helpful? Give feedback.
-
deep anything v2 😬😁 |
Beta Was this translation helpful? Give feedback.
-
i mention some suggestions here: #1564 , but mainly for controlnets i hope you add ability to use photomaker2 with other controlnets like openpose....etc, and i hope we have a refine for sdxl outpainting, something kinda close to photoshop outpainting will be amazing and outpaint even without describing the outpointed area details, i know we can use inpainting to outpaint but the way we can do this with, and the results are so bad, and thanks a lot for your really amazing work and brilliant ideas like the new spaces idea which i hope we have more new spaces on it. |
Beta Was this translation helpful? Give feedback.
-
I would really like to see it Pix2Pix |
Beta Was this translation helpful? Give feedback.
-
It's gonna be a looong month.... -_- |
Beta Was this translation helpful? Give feedback.
-
Hello, and thanks for your awesome work ! |
Beta Was this translation helpful? Give feedback.
-
PuLID + EVA-CLIP support would be cool. PuLID EVA-CLIP |
Beta Was this translation helpful? Give feedback.
-
The integrated ControlNet is not updated for a while, and we are going to make it a bit more up-to-date.
However that will happen after some other newer experiments. Old features are relatively lower priority - so that we can experiment with more new ideas first.
I personally know some models like uni-controlnet (or called promax by some non-research people) and some preprocessors like DSINE/DepthAnything/etc. Those are very easy to add – after I get other things done, those only need me 1 or 2 hours to add them all.
However I will not add the below things:
By the way, several experiments will be implemented in a “Forge Space” (that I will add later, similar to some local HuggingFace Spaces) rather than that ControlNet extension. So many functionalities will happen elsewhere.
This post is only for controlnets. Experiments with other newer diffusion models will happen before controlnets.
Please tell us what is missing in ControlNet Integrated. Please do not talk about things that are already mentioned in this post – I will take a look several days later when I finish some other things.
Update Sep 1:
The rewrite of ControlNet Intergrated will start at about Sep 29. The estimated finish date is about Oct 7. When this note is announced, the main targets include some diffusers formatted Flux ControlNets and some community implementation of Union ControlNets. However, this may be extended if stronger models come out after this note.
About IPAdapter:
Note that the CLIP Vision names of
sd-webui-controlnet
are non-standard. Forge uses correct names:Beta Was this translation helpful? Give feedback.
All reactions