John Delves Richardson
Tribune
I enjoyed reading the colourful descriptions of how it went wrong.
Seriously though, I'm glad you tried it and shared your process experience. It's surprising how you managed to run it with a 6Gb VRAM video card. I'm not sure how much it'd affect the usability, but there's always a cloud option (which the plugin natively supports) in case you decide to make serious works with the setup.
Anyway, from what I gathered from your post, it looks like the setup went correctly, but you may need a bit more time to get yourself familiar with the tool.
I thought it might be helpful if I could create a render using the workflow you tried. So, here's a brief explanation of how I used the workflow you mentioned below:
The line drawing doesn't have to be very detailed, so I just scribbled this on a blank paint layer:
View attachment 1399791
There are a few ways to refine this control map image using an iterative method, but I'll just use this rudimentary drawing as it is this time. The important thing is to add a control layer and choose the right kind of model for the image. As shown above, I added a "Line Art" entry and made it point to the paint layer containing my drawing.
Then I added the prompt text and hit generate to get this image:
View attachment 1399792
You can keep generating until you stumble upon an image that gives you the right vibe. The overall atmosphere is more important than the details in this stage since you can't expect to get all the details right in a single pass (especially if you have to use 512x512 resolution).
As expected, you can see the image contains several problematic areas. The AI apparently didn't recognise the rifle in my crude drawing, so I selected the area and refined the prompt:
View attachment 1399790
You may have noted that I also included the girl's face area instead of just selecting the rifle part. The reason is that it usually works better when the AI has sufficient context to understand what is happening in the image. A great thing about the AI plugin is that it supports the usual layer-based workflow of Krita itself, meaning I can always delete unnecessary parts in a generated layer or merge several of them.
After generating an image, you can choose "Apply" to save it as a layer, and you can erase everything except for the good parts of it. Also, note that I changed my prompt to describe only what is depicted in the selected area so that the AI wouldn't get distracted by other irrelevant terms.
The "strength" I highlighted in the above image is "denoising strength" in other AI frontends, meaning how much of the underlying image you want to change. Beware that setting it 100% has a special meaning with the plugin, so try to lower it if you don't get what you want.
Now all I need to do is repeat the process to refine all the areas I want to change. Using a control net to keep the composition while changing a region is common. I could use the Line Art layer again, but it's based on a crude drawing which must have a lot of errors in proportions and may not even match the current image exactly.
In this case, I can generate a new control map image using the current image. And, I chose the depth map model because it's good for preserving the composition without keeping the details which I want to change rather than preserve them. You can use the highlighted button shown below to generate a depth map layer:
View attachment 1399788
Also, I wanted to add body hair to suit my personal preferences, which can be difficult to achieve without using Loras. So I temporarily changed the settings to add relevant Loras in the preference dialogue as shown below:
View attachment 1399789
When all was done, I switched to the upscale mode to increase the resolution to 2048x2048, which resulted in this final output:
(I already reached the maximum number of images limitation, so I just linked the image from the other thread.)
Hope this could help. Please feel free to let me know in case you need further assistance.
I have to admit that with this post you have captured my interest and attention. I will definitely do some experiment myself. Thanks @malins , too.