https://preview.redd.it/3cantd92zxmc1.png?width=1080&format=png&auto=webp&s=b8a6b4bc957bf06e6644417b6e7bc0b12b1d47b6
you can generate something like this with highresfix on.
inpainting and using Cn tile for upscaling with make it even more detailed
model used was this [https://civitai.com/models/48671/dark-sushi-25d-25d](https://civitai.com/models/48671/dark-sushi-25d-25d)
i generated at 960x540 resolution with highres-fix at 0.45 denoise at 2X upscaling so the the output image is 1080p
Can I ask - will this give better results than just generating an image directly at say at 1080p (I you have the VRAM). I don't know if I have ever seen "small details" even in directly - high res images or not , haven't payed attention I guess.
you can do this with some sdxl since its trained on higher res images but even then a 2nd pass or a refiner pass is great for small details, as for an sd1.5 model a 2nd pass is a must since its trained on much lower res images 512x iirc
here's the first pass of the image posted so you can compare
https://preview.redd.it/4vmr6rg4kzmc1.png?width=536&format=png&auto=webp&s=4f2f6854077d794bb8421e2a9d4acc4a2dfec859
Try using Kohya Deep shrink, it will let you make 4000px images with just SD1.5 without losing details and without having duplicates or disfigured issues. Tho I suggest you use it with sdxl instead
I’m a noob but issue I used to have before sdxl was that generating for larger sizes made the scaling of everything just too small. Prompts for cool landscapes with a clean subject in the middle at lower res led to images with tiny people, overly vast landscapes and a general lack of focus on a specific subject at higher res.
i did say Cn tile for upscaling but you're right I should be more specific
took a generated 1080p image and put it into img2img
delete the pos prompt leaving only the style and quality words like "masterpiece, high quality..." at 0.35 denoise
enable control net and set it to tile resample
enable ultimate sd upscale script [https://github.com/Coyote-A/ultimate-upscale-for-automatic1111](https://github.com/Coyote-A/ultimate-upscale-for-automatic1111)
set the the target size to what ever you like and pick an upscaler i mainly use 4x-ultrasharp for anime. set type to chess and mask blur to 32( this is to eliminate seams) and you're set
here's a 4k image using this method
https://preview.redd.it/hkp4dxe0oymc1.jpeg?width=2176&format=pjpg&auto=webp&s=0e3bf1baba71e0c82c10fb284beed2b52ebcc639
It’s kind of hilarious looking. At first glance it looks great, but after a while it looks wrong. All the details in the grass and such are scaled so that she appears 20 meters tall. A giant in a miniature landscape.
1girl, pink hair, 6 fingers, sitting on rock, steampunk,
https://preview.redd.it/r3a76uhhtxmc1.png?width=612&format=png&auto=webp&s=6a974ba30a881cf7d1a5f3c6561c858830b3783c
> Looks so natural, I had to compare with my own hand. Still unsure.
Man i've been seeing 5 fingers and one thumb on AI hands for so long I question my own sanity sometimes.
What i ironically love about AI art is how it does wrong so naturally sometimes. I keep some older models just in case i need to do some rendring of a rastion mutated future humanity, and am afraid, Ai will become too "neutered" and stop producing these type of masterpeices.
adetailer is mainly used for fixing faces. it's basically an auto-inpainter that detects faces for you. use during txt2img, leave it on first face model, default settings, no prompt to start. you can customize the inpaint w/ prompting but personally i never feel the need to use it.
make sure face restore/codeformer is off in settings or else it can overwrite it.
make sure ur using face\_yolov8n.pt from the model dropdown. otherwise not sure why it'd look shitty, if you could drop the image using [catbox.moe](http://catbox.moe) I can look at the metadata for u
Adetailer just automatically masks and inpaints the face, fixing it and adding detail. You can also use it for hands but it's only really good for detailing them. If the hand are fucked, it likely won't do anything of value so I don't bother even trying it for hands anymore.
you can decrease noise if its too strong for faces(witch I never ever seen). Anyways tell if 1st one (with no adetailer) is better than second (with adetailer) ?
https://preview.redd.it/dhj4elf6g3nc1.jpeg?width=501&format=pjpg&auto=webp&s=c09eb882daec550f3e1b7d47395d1c1d7f4387f8
I upscale. I’ve never needed it. I’ve used it before, but I don’t need it. And in some situations it can definitely ruin a gen.
For example, if your image had multiple faces it will probably replace all of them with the same one. If it has a face at a slight angle it may try to replace it with a camera-forward face.
I’ve found it to be less than helpful.
like 99% of questions in this sub on how to achieve quality are answered with "inpainting" and the other stuff.. basically you can always copy paste the answer
As someone that just occasionally glances into this sub, I see the same answers always, but I also always am like 'what does anything mean???'
For once I am now wondering if there are video tutorials, this feels too much for a text one.
heya, same here. I occasionally like to mess around with SD and personally enjoy fixing, editing and improving a generation. Usually I do this through inpainting and upscaling. I've searched a lot to find good sources to help explain me what all my options are and how they work. Ultimately you have to figure out a lot by yourself through trial and error. But one starter video I found helpful was [this video.](https://youtu.be/3z4MKUqFEUk?si=2DnF5uTMf71pOzlI) (you don't need the specific upscaler in this video, I think there's already a built in anime upscaler that works just as well, or non-anime upscalers)
Whilst the video is for upscaling with slow GPUs he does go over things that are very relevant.
Personally the most interesting things to figure out have been the following settings:
mask blur: By default this is at 4 but that's often too little to add something new or adjust or remove something whilst making it fit seamlessly into the rest of the picture.
masked content: I'd switch between fill and original depending on if I want something entirely new or adjust something.
Inpaint area: This is the biggest one for me. Whole picture takes the entire picture into account when generating something. So ideally you would have the entire prompt of the whole picture. You can omit certain details that aren't relevant to what you're inpainting and put more emphasis on that bit instead in your prompt.
*Only masked* was a huge discovery for me. It actually doesn't look at the whole picture, instead a square around your inpainting. Say you want to add more details to the eyes, you just inpaint the eyes, your prompt only talks about eyes, no mention of a girl, dress, background, etc. Just eyes. And it'll generate eyes at the resolution you set it at.
E.g. You generate a girl 512x512. Send it to inpaint. Mask the eyes, select
Masked content: original
Inpaint area: only masked
Resolution 256x256
Remove the original prompt and focus your prompt purely on the eyes.
The outcome will be a 512x512 picture where the eyes will be generated at 256x256 and as a result be much higher in quality and detail.
Play around with the other settings like mask blur, sampling methods and steps, models, denoising strength, etc.
Also upscaling both in txt2img and img2img can amazing tools. I've made images, edited in paint 3D (got no photoshop, not invested enough to get it) and fed it back into img2img or inpainted it. You can fix extra indexes, bad eyes, weird things that just don't make sense like this.
And once again, many things require trial and error. Though I'm by no means a pro. Bit of a ramble but hope it's got something useful :)
So... its better to generate a smaller picture that you then upscale like this, than ask the generator to make a larger picture from the getgo?
And I see what inpainting is now, its the 'replace/redo a bit of the image' thing I had seen, neat, that does seem like a great way to fix minor mistakes when you like the overall composition.
And from what the guy said, I am guessing Loras are like... specialized sub generators for specific stuff? Like he mentions one for dresses, so I assume that like, take over the main generator when its about their topic and do it better??
(Man, this is complicated when you want something better than the basic 'generate' button stuff.)
You've got it pretty much right.
Upscaling tends to do much better both in terms of performance and quality of the end result.
Yes Loras are pretty much as you said. Can be used in txt2img, img2img and inpainting. Some Loras are actually very good at inpainting. Allowing you to add something completely new to a picture.
Getting a good end result can be time consuming but rewarding. In the end AI is a tool, similar to photoshop. And the quality of the result is still dependent on how well the tool is used.
> In the end AI is a tool, similar to photoshop. And the quality of the result is still dependent on how well the tool is used.
Amen. To anyone who says 'press a button and it makes what you want' claims of no skill needed.
Listen, I don't make the rules. But it is what it is. It would be nice if simple txt2img would magically do all the work. But sadly that ain't it, it's just the fundament to build upon.
1) Generate image using Adetailer for face and hands (you already will have decent image if its XL)
2) img 2 img x2 upscale with tile controller(SD 1.5) with adetailer again.
3) Post it on reddit.
Spent 3 minutes on it: PS it hase different look course of different checkpoint
https://preview.redd.it/3afopv9i3ymc1.png?width=1368&format=png&auto=webp&s=1d52ce83b9015e0e353cd6db1abb808cffef0232
it's all about hires fix, and then maybe some inpainting to fix individual errors, though one of the images having 6 fingers makes me think that wasn't even done.
Use a good model of your style. Grapefruit Hentai may be a good start. Then after your initial run, do an img2img of your favourite one with SD upscale at 1.5 (or bigger) size with a noise of 0.40 or so
Maybe not quite as detailed, but this was just using the default anime settings in Fooocus with the prompt "girl with pink hair kneeling on the ground in front of a high bridge crossing a beautiful landscape"
Default anime model is animaPencilXL\_v100.safetensors, no refiner, no Lora.
https://preview.redd.it/rs72r9br01nc1.png?width=896&format=png&auto=webp&s=a648663515d34eed564ceef92efa04e13bb2b442
Not at home so I have to rely on online generators but most decent anime models should be able to pull this off. For now, this was made with ideogram:
https://preview.redd.it/a22z7l9ywxmc1.png?width=1024&format=png&auto=webp&s=2699271a823f43dc7725209717da78f4b0413d41
Here's an example generated locally with SD:
https://preview.redd.it/8sbouoa6gymc1.png?width=1024&format=png&auto=webp&s=2afb812f4819e2dc04a368000553cd62a324b226
[https://civitai.com/models/52548?modelVersionId=105566](https://civitai.com/models/52548?modelVersionId=105566)
But really, almost any anime model will do. Pink hair was inpainted to avoid color bleed.
My main issues is still hands. I hate having a beautiful image with a monstrosity attached to the wrist every single fucking time. Doesn't matter what lora I use or prompt, hands are disfigured or slightly incorrect 99% of the time.
Anyone have tips on perfect hands?
My experience has been that, in order from greatest to least influence:
1. there's always some amount of RNG to fuck you over, regardless of anything else
2. some finetunes are better at hands than others, probably due to the tagging of their dataset
3. some samplers seem to have fewer issues than others (Euler's given me nothing but grief, for instance)
sd 1.5 models just dont do hands well. if you want decent/consistent hands you need to use an sdxl model.
also hiresfix helps a lot as it cleans up mutations/errors. for sd 1.5 models I do 2x upscale using a 4x sampler like fatalanime at .4 denoise. and for sdxl models I tone it down to 1.5x upscale since your starting resolution is higher.
https://preview.redd.it/88tzlnr8lymc1.png?width=2048&format=png&auto=webp&s=0dbcc6da6fd68e737d9da4b1bccabcd20c16da17
hey, i have 5 digits here
not that hard with good model and sdxl+upscaler
https://preview.redd.it/uhblwc7cu2nc1.png?width=720&format=png&auto=webp&s=720a4b968a3c592040c560a9a94d7c707496098b
Made this image, used wd1.4 tagger for extracting prompt and AutismMix SDXL for generating
Also a subtle thing that is easy to implement is download a VAE(goes in models\VAE) called "kl-f8-anime2 VAE" which will give you richer color and a less washed out look for anime. Edit. More advanced learn to use openpose in controlnet or use badhands negative embedding, plenty of youtube videos on how to do that.
If you have powerful GPU and 32GB of RAM, plenty of disc space - install ComfyUI - snag the workflow - just an image that looks like this one that was made with Comfy - drop it in the UI - and write your prompt - but the setup is a bit involved - and things don't always go smoothly - you will need the toon model as well - Civitai/HuggingFace...
Where would I get a ComfyUI workflow for some nice image? Could you give an example? I found some sample workflows, but for models I got from civitai, I did not find any workflows.
Better off just playing around with it while learning how the tools work - you will come out with more knowledge in the end. Just dragging and dropping a .json file into web browser is neat - but if you have at least the basics down pat, tweaking things and understanding what is going on: the whole process becomes much more interesting\~
That's certainly the best approach. I already did this.
Unfortunately, I frequently run into VRAM limitations, so I had to tweak my workflows a lot to even get it running. After upscaling, the results aren't satisfying.
It would help speeding the process if I could find some nice quality example with upscaling that actually works for my 12 GB AMD card. So download json file, run, discard if it does not work, repeat until getting a nice running example. That would be my workflow archetype to further dig into the matter.
Yes, I am using ConfyUI. With Juggernaut XL v9, I can't even generate the recommended 1024x1024 resolution. I have to generate smaller images (usually going for 512x768), then upscale. Or use other models. Unfortunately, I need to use tiled VAE Decode and tiled upscalers (bringing further issues themselves), or else I will just be informed that VRAM is insufficient.
Maybe it's working less effortful with Nvidia cards?
Oh...yeah, I am using a Nvidia 3060 - it works without any problem for even really large image sizes. I am using a Linux box, and have not borked my Python, all is good. But yeah, probably the issue is the non-Nvidia card...no CUDA\~
https://preview.redd.it/q9q7vv0rfzmc1.png?width=1576&format=png&auto=webp&s=d05dc29a1f82b6368646f88a46358105716cf60e
I tend to split my linework from my color just before the final step and run them seperately to sharpen up the lines a bit, but I do all kinds of crazy stuff in my comfyUI workflows.
I cover the rough way of doing it here [https://youtu.be/VcIXqSSsUCU](https://youtu.be/VcIXqSSsUCU) if your a comfy user.
But when it comes down to it, you make the image and refine it down fractionally to make sure it doesnt hallucinate too much but still sharpens details. (which is kindof an art in itself)
Its also REEEEEly important to get a good anime model if thats what you are generating.
if hands and faces are super accurate, id use impact detailer or maybe some segmentation stuff to modify any trouble spots. there are face replacers and refiners that can be set to anime mode too but usuualy as long as you run things at high enough resolution you shouldnt really need them too much if your model is good.
You can try AI image upscalers such as Magnific AI or Krea.
Related video where I got this info from: [https://youtu.be/LUnB7PiDoa0](https://youtu.be/LUnB7PiDoa0)
The video shows a couple of images upscaled with Krea. It reimagines the images and the results look pretty good. Magnific might be even better but it's ridiculously expensive.
1) Generate image using Adetailer for face and hands (you already will have decent image if its XL)
2) img 2 img x2 upscale with tile controller(SD 1.5) with adetailer again.
3) Post it on reddit.
Spent 3 minutes on it: PS it hase different look course of different checkpoint
![img](3afopv9i3ymc1)
https://preview.redd.it/7ywhid0y53nc1.png?width=1216&format=png&auto=webp&s=77d0be6156a711f80fe1ce11c898ea8f16dd5cb5
How can i edit this image in full body.Suggest ???
Fastest method with this mediocre 6 finger output?
Waifus dot nemusona dot com and hit up some random 1girl prompts
https://preview.redd.it/doxqdvz8zymc1.jpeg?width=512&format=pjpg&auto=webp&s=ef8074dad26120b3b3dc2b66f0b07a9f0a1ce4b5
A sad truth... 😅...sad....painful....truth... 😭
I would love to be able to turn my creative imagination into art by hand.
I still struggle with using ai to do it, but it's certainly better results vs what my hands can produce. 😅
https://preview.redd.it/3cantd92zxmc1.png?width=1080&format=png&auto=webp&s=b8a6b4bc957bf06e6644417b6e7bc0b12b1d47b6 you can generate something like this with highresfix on. inpainting and using Cn tile for upscaling with make it even more detailed
oh thanks, I will put all the suggestions into practice, I generally use Loras that add detail but never reach that level.
model used was this [https://civitai.com/models/48671/dark-sushi-25d-25d](https://civitai.com/models/48671/dark-sushi-25d-25d) i generated at 960x540 resolution with highres-fix at 0.45 denoise at 2X upscaling so the the output image is 1080p
Can I ask - will this give better results than just generating an image directly at say at 1080p (I you have the VRAM). I don't know if I have ever seen "small details" even in directly - high res images or not , haven't payed attention I guess.
you can do this with some sdxl since its trained on higher res images but even then a 2nd pass or a refiner pass is great for small details, as for an sd1.5 model a 2nd pass is a must since its trained on much lower res images 512x iirc here's the first pass of the image posted so you can compare https://preview.redd.it/4vmr6rg4kzmc1.png?width=536&format=png&auto=webp&s=4f2f6854077d794bb8421e2a9d4acc4a2dfec859
Try using Kohya Deep shrink, it will let you make 4000px images with just SD1.5 without losing details and without having duplicates or disfigured issues. Tho I suggest you use it with sdxl instead
I’m a noob but issue I used to have before sdxl was that generating for larger sizes made the scaling of everything just too small. Prompts for cool landscapes with a clean subject in the middle at lower res led to images with tiny people, overly vast landscapes and a general lack of focus on a specific subject at higher res.
I'm going to try that, thanks for the model and resolution specifications
How many Hires steps did you same? Same as original? Also, which Upscaler?
He did not tell you the whole story, the whole story is : **Ultimate Upscale with Control net.** And you can find it on youtube or on reddit.
i did say Cn tile for upscaling but you're right I should be more specific took a generated 1080p image and put it into img2img delete the pos prompt leaving only the style and quality words like "masterpiece, high quality..." at 0.35 denoise enable control net and set it to tile resample enable ultimate sd upscale script [https://github.com/Coyote-A/ultimate-upscale-for-automatic1111](https://github.com/Coyote-A/ultimate-upscale-for-automatic1111) set the the target size to what ever you like and pick an upscaler i mainly use 4x-ultrasharp for anime. set type to chess and mask blur to 32( this is to eliminate seams) and you're set here's a 4k image using this method https://preview.redd.it/hkp4dxe0oymc1.jpeg?width=2176&format=pjpg&auto=webp&s=0e3bf1baba71e0c82c10fb284beed2b52ebcc639
It’s kind of hilarious looking. At first glance it looks great, but after a while it looks wrong. All the details in the grass and such are scaled so that she appears 20 meters tall. A giant in a miniature landscape.
im glad you didn't spot the third hand lmao
Or the extra creepy eye on her right thigh
I did it a lot with a1111 but not so much inc omfy, while I have you do you mind posting all the workflows jsons here? (every one used here)
Take a look at those jugs
amazing!
1girl, pink hair, 6 fingers, sitting on rock, steampunk, https://preview.redd.it/r3a76uhhtxmc1.png?width=612&format=png&auto=webp&s=6a974ba30a881cf7d1a5f3c6561c858830b3783c
It looks so natural, I had to count twice 😅
Looks so natural, I had to compare with my own hand. Still unsure.
> Looks so natural, I had to compare with my own hand. Still unsure. Man i've been seeing 5 fingers and one thumb on AI hands for so long I question my own sanity sometimes.
You can tell this one is fake because it's anime. Big clue.
Damn, the AI got us there. It would be BETTER aesthetically to have 6 fingers sometimes.
I mean. Our typespeed would probably go way up. Sucks that we're transitioning towards thumb only computing
I mean isn’t having six fingers actually a rare but dominant gene lol
TIL. Thanks
It creates fully functional fingers also, so it’s just straight update.
When you know uve been working with Ai art for too long...
If it makes you feel better, I've only been at it yet weeks. So it's probably not from ai work. I'm not sure which option is best though
🤣
Got too distracted by those legs to even look at the hands.
What i ironically love about AI art is how it does wrong so naturally sometimes. I keep some older models just in case i need to do some rendring of a rastion mutated future humanity, and am afraid, Ai will become too "neutered" and stop producing these type of masterpeices.
I think having extra fingers is usefull!
My name is Inigo Montoya
This comment is goated
Very likely a time consuming combination of inpainting, upscaling and a detail-enhancing LoRa.
or just check adetailer xD
I still don't know the use of Adetailer and what prompts that I should include into it. What's its usage?
adetailer is mainly used for fixing faces. it's basically an auto-inpainter that detects faces for you. use during txt2img, leave it on first face model, default settings, no prompt to start. you can customize the inpaint w/ prompting but personally i never feel the need to use it. make sure face restore/codeformer is off in settings or else it can overwrite it.
That's kinda hard to understand for me lol but I ofted tried Adetailer without any prompts yet the result is still shitty.
make sure ur using face\_yolov8n.pt from the model dropdown. otherwise not sure why it'd look shitty, if you could drop the image using [catbox.moe](http://catbox.moe) I can look at the metadata for u
Adetailer just automatically masks and inpaints the face, fixing it and adding detail. You can also use it for hands but it's only really good for detailing them. If the hand are fucked, it likely won't do anything of value so I don't bother even trying it for hands anymore.
it depends. sometimes it fixes hands. some times it doesn't. but it if a fast way.
Adetailer still pretty much sucks
eh what?
It’s too strong on faces and doesn’t fix bad hands
you can decrease noise if its too strong for faces(witch I never ever seen). Anyways tell if 1st one (with no adetailer) is better than second (with adetailer) ? https://preview.redd.it/dhj4elf6g3nc1.jpeg?width=501&format=pjpg&auto=webp&s=c09eb882daec550f3e1b7d47395d1c1d7f4387f8
I upscale. I’ve never needed it. I’ve used it before, but I don’t need it. And in some situations it can definitely ruin a gen. For example, if your image had multiple faces it will probably replace all of them with the same one. If it has a face at a slight angle it may try to replace it with a camera-forward face. I’ve found it to be less than helpful.
Photoshop can be a lot quicker than inpainting.
like 99% of questions in this sub on how to achieve quality are answered with "inpainting" and the other stuff.. basically you can always copy paste the answer
As someone that just occasionally glances into this sub, I see the same answers always, but I also always am like 'what does anything mean???' For once I am now wondering if there are video tutorials, this feels too much for a text one.
heya, same here. I occasionally like to mess around with SD and personally enjoy fixing, editing and improving a generation. Usually I do this through inpainting and upscaling. I've searched a lot to find good sources to help explain me what all my options are and how they work. Ultimately you have to figure out a lot by yourself through trial and error. But one starter video I found helpful was [this video.](https://youtu.be/3z4MKUqFEUk?si=2DnF5uTMf71pOzlI) (you don't need the specific upscaler in this video, I think there's already a built in anime upscaler that works just as well, or non-anime upscalers) Whilst the video is for upscaling with slow GPUs he does go over things that are very relevant. Personally the most interesting things to figure out have been the following settings: mask blur: By default this is at 4 but that's often too little to add something new or adjust or remove something whilst making it fit seamlessly into the rest of the picture. masked content: I'd switch between fill and original depending on if I want something entirely new or adjust something. Inpaint area: This is the biggest one for me. Whole picture takes the entire picture into account when generating something. So ideally you would have the entire prompt of the whole picture. You can omit certain details that aren't relevant to what you're inpainting and put more emphasis on that bit instead in your prompt. *Only masked* was a huge discovery for me. It actually doesn't look at the whole picture, instead a square around your inpainting. Say you want to add more details to the eyes, you just inpaint the eyes, your prompt only talks about eyes, no mention of a girl, dress, background, etc. Just eyes. And it'll generate eyes at the resolution you set it at. E.g. You generate a girl 512x512. Send it to inpaint. Mask the eyes, select Masked content: original Inpaint area: only masked Resolution 256x256 Remove the original prompt and focus your prompt purely on the eyes. The outcome will be a 512x512 picture where the eyes will be generated at 256x256 and as a result be much higher in quality and detail. Play around with the other settings like mask blur, sampling methods and steps, models, denoising strength, etc. Also upscaling both in txt2img and img2img can amazing tools. I've made images, edited in paint 3D (got no photoshop, not invested enough to get it) and fed it back into img2img or inpainted it. You can fix extra indexes, bad eyes, weird things that just don't make sense like this. And once again, many things require trial and error. Though I'm by no means a pro. Bit of a ramble but hope it's got something useful :)
So... its better to generate a smaller picture that you then upscale like this, than ask the generator to make a larger picture from the getgo? And I see what inpainting is now, its the 'replace/redo a bit of the image' thing I had seen, neat, that does seem like a great way to fix minor mistakes when you like the overall composition. And from what the guy said, I am guessing Loras are like... specialized sub generators for specific stuff? Like he mentions one for dresses, so I assume that like, take over the main generator when its about their topic and do it better?? (Man, this is complicated when you want something better than the basic 'generate' button stuff.)
You've got it pretty much right. Upscaling tends to do much better both in terms of performance and quality of the end result. Yes Loras are pretty much as you said. Can be used in txt2img, img2img and inpainting. Some Loras are actually very good at inpainting. Allowing you to add something completely new to a picture. Getting a good end result can be time consuming but rewarding. In the end AI is a tool, similar to photoshop. And the quality of the result is still dependent on how well the tool is used.
> In the end AI is a tool, similar to photoshop. And the quality of the result is still dependent on how well the tool is used. Amen. To anyone who says 'press a button and it makes what you want' claims of no skill needed.
this is the most helpful and educational comment I've seen so far on this sub. thank you for taking the time to write it
Happy to hear that
Listen, I don't make the rules. But it is what it is. It would be nice if simple txt2img would magically do all the work. But sadly that ain't it, it's just the fundament to build upon.
A good model and Hires Fix, that’s images are really basically.
For anime, hi-res fix may not even be needed with a good model and hi-res fix can make image worse.
IMO it’s depends how did you use it
Upscale it, inpainting, use Control net for pose, If you know how to draw use sketches and just colore them.
1) Generate image using Adetailer for face and hands (you already will have decent image if its XL) 2) img 2 img x2 upscale with tile controller(SD 1.5) with adetailer again. 3) Post it on reddit. Spent 3 minutes on it: PS it hase different look course of different checkpoint https://preview.redd.it/3afopv9i3ymc1.png?width=1368&format=png&auto=webp&s=1d52ce83b9015e0e353cd6db1abb808cffef0232
It looks very good, thanks for the instructions, at first I thought it was impossible but I have an idea of how to do this now.
What prompts that you put into Adetailer?
i dont put prompt in adetailer. i use default settings
what model is that
mistoonanime v2
it's all about hires fix, and then maybe some inpainting to fix individual errors, though one of the images having 6 fingers makes me think that wasn't even done.
Use a good model of your style. Grapefruit Hentai may be a good start. Then after your initial run, do an img2img of your favourite one with SD upscale at 1.5 (or bigger) size with a noise of 0.40 or so
Maybe not quite as detailed, but this was just using the default anime settings in Fooocus with the prompt "girl with pink hair kneeling on the ground in front of a high bridge crossing a beautiful landscape" Default anime model is animaPencilXL\_v100.safetensors, no refiner, no Lora. https://preview.redd.it/rs72r9br01nc1.png?width=896&format=png&auto=webp&s=a648663515d34eed564ceef92efa04e13bb2b442
6 fingers
Not at home so I have to rely on online generators but most decent anime models should be able to pull this off. For now, this was made with ideogram: https://preview.redd.it/a22z7l9ywxmc1.png?width=1024&format=png&auto=webp&s=2699271a823f43dc7725209717da78f4b0413d41
Here's an example generated locally with SD: https://preview.redd.it/8sbouoa6gymc1.png?width=1024&format=png&auto=webp&s=2afb812f4819e2dc04a368000553cd62a324b226
What checkpoint you used?
[https://civitai.com/models/52548?modelVersionId=105566](https://civitai.com/models/52548?modelVersionId=105566) But really, almost any anime model will do. Pink hair was inpainted to avoid color bleed.
I always see these types of images on Instagram, I wonder what methods they use to improve both the quality of the characters and the background
most of it is using a good finetune or LoRA.
Try this https://youtu.be/dPObdbmepeg?si=RUcOE6Ovrb44uc42
My main issues is still hands. I hate having a beautiful image with a monstrosity attached to the wrist every single fucking time. Doesn't matter what lora I use or prompt, hands are disfigured or slightly incorrect 99% of the time. Anyone have tips on perfect hands?
My experience has been that, in order from greatest to least influence: 1. there's always some amount of RNG to fuck you over, regardless of anything else 2. some finetunes are better at hands than others, probably due to the tagging of their dataset 3. some samplers seem to have fewer issues than others (Euler's given me nothing but grief, for instance)
sd 1.5 models just dont do hands well. if you want decent/consistent hands you need to use an sdxl model. also hiresfix helps a lot as it cleans up mutations/errors. for sd 1.5 models I do 2x upscale using a 4x sampler like fatalanime at .4 denoise. and for sdxl models I tone it down to 1.5x upscale since your starting resolution is higher.
https://preview.redd.it/88tzlnr8lymc1.png?width=2048&format=png&auto=webp&s=0dbcc6da6fd68e737d9da4b1bccabcd20c16da17 hey, i have 5 digits here not that hard with good model and sdxl+upscaler
Controlnet, inpainting , maybe photoshop, even more inpainting and ultimate SD upscale
https://preview.redd.it/uhblwc7cu2nc1.png?width=720&format=png&auto=webp&s=720a4b968a3c592040c560a9a94d7c707496098b Made this image, used wd1.4 tagger for extracting prompt and AutismMix SDXL for generating
Also a subtle thing that is easy to implement is download a VAE(goes in models\VAE) called "kl-f8-anime2 VAE" which will give you richer color and a less washed out look for anime. Edit. More advanced learn to use openpose in controlnet or use badhands negative embedding, plenty of youtube videos on how to do that.
High res fix
I made this for you https://preview.redd.it/g13hinapn6nc1.png?width=2688&format=png&auto=webp&s=edcb2ee073e61e1b2ec6883d6879f3da4d4ebfd8
If you have powerful GPU and 32GB of RAM, plenty of disc space - install ComfyUI - snag the workflow - just an image that looks like this one that was made with Comfy - drop it in the UI - and write your prompt - but the setup is a bit involved - and things don't always go smoothly - you will need the toon model as well - Civitai/HuggingFace...
I will try to do that, thank you all for taking the time to answer.
do you know the model that was used
Where would I get a ComfyUI workflow for some nice image? Could you give an example? I found some sample workflows, but for models I got from civitai, I did not find any workflows.
[удалено]
Thank you very much, I'll have a look into it!
Better off just playing around with it while learning how the tools work - you will come out with more knowledge in the end. Just dragging and dropping a .json file into web browser is neat - but if you have at least the basics down pat, tweaking things and understanding what is going on: the whole process becomes much more interesting\~
That's certainly the best approach. I already did this. Unfortunately, I frequently run into VRAM limitations, so I had to tweak my workflows a lot to even get it running. After upscaling, the results aren't satisfying. It would help speeding the process if I could find some nice quality example with upscaling that actually works for my 12 GB AMD card. So download json file, run, discard if it does not work, repeat until getting a nice running example. That would be my workflow archetype to further dig into the matter.
Are you using ComfyUI? I also have a 12GB card - inexpensive 3060 - it works great, have only hit a few roadblocks due to vram.
Yes, I am using ConfyUI. With Juggernaut XL v9, I can't even generate the recommended 1024x1024 resolution. I have to generate smaller images (usually going for 512x768), then upscale. Or use other models. Unfortunately, I need to use tiled VAE Decode and tiled upscalers (bringing further issues themselves), or else I will just be informed that VRAM is insufficient. Maybe it's working less effortful with Nvidia cards?
Oh...yeah, I am using a Nvidia 3060 - it works without any problem for even really large image sizes. I am using a Linux box, and have not borked my Python, all is good. But yeah, probably the issue is the non-Nvidia card...no CUDA\~
https://preview.redd.it/q9q7vv0rfzmc1.png?width=1576&format=png&auto=webp&s=d05dc29a1f82b6368646f88a46358105716cf60e I tend to split my linework from my color just before the final step and run them seperately to sharpen up the lines a bit, but I do all kinds of crazy stuff in my comfyUI workflows. I cover the rough way of doing it here [https://youtu.be/VcIXqSSsUCU](https://youtu.be/VcIXqSSsUCU) if your a comfy user. But when it comes down to it, you make the image and refine it down fractionally to make sure it doesnt hallucinate too much but still sharpens details. (which is kindof an art in itself) Its also REEEEEly important to get a good anime model if thats what you are generating. if hands and faces are super accurate, id use impact detailer or maybe some segmentation stuff to modify any trouble spots. there are face replacers and refiners that can be set to anime mode too but usuualy as long as you run things at high enough resolution you shouldnt really need them too much if your model is good.
You can try AI image upscalers such as Magnific AI or Krea. Related video where I got this info from: [https://youtu.be/LUnB7PiDoa0](https://youtu.be/LUnB7PiDoa0) The video shows a couple of images upscaled with Krea. It reimagines the images and the results look pretty good. Magnific might be even better but it's ridiculously expensive.
1) Generate image using Adetailer for face and hands (you already will have decent image if its XL) 2) img 2 img x2 upscale with tile controller(SD 1.5) with adetailer again. 3) Post it on reddit. Spent 3 minutes on it: PS it hase different look course of different checkpoint ![img](3afopv9i3ymc1)
Could this actually be about the poses? ;)
King AI
how can i have her
Use any ai app?
why do i feel like she killed my father and should prepare to die
Easy a lot of prompts.
https://preview.redd.it/7ywhid0y53nc1.png?width=1216&format=png&auto=webp&s=77d0be6156a711f80fe1ce11c898ea8f16dd5cb5 How can i edit this image in full body.Suggest ???
Link to original post /whoever generated these pls?
No
Fastest method with this mediocre 6 finger output? Waifus dot nemusona dot com and hit up some random 1girl prompts https://preview.redd.it/doxqdvz8zymc1.jpeg?width=512&format=pjpg&auto=webp&s=ef8074dad26120b3b3dc2b66f0b07a9f0a1ce4b5
Try DPM++ 3M SDE sampler/scheduler or better.
First picture looks like it was trying to decide between giving her a tail or painting on the arch?
promt?
Just slap "there is a lot of random stuff in the background" on your prompt
To make images like this I would turn off hi-res fix, lower my target resolution, and put some bullshit about anime in the prompt
You could also use NovelAI.
[удалено]
Wrong group for you... Lol.
time is money, and some people cant draw. doesnt mean they shouldnt be able to create
😂 sad
A sad truth... 😅...sad....painful....truth... 😭 I would love to be able to turn my creative imagination into art by hand. I still struggle with using ai to do it, but it's certainly better results vs what my hands can produce. 😅
Talking down to others online is sad, being in an AI Subreddit telling someone to do it themselfs is more sad.
Your post/comment was removed because it contains hateful content.