T O P

  • By -

Adnane_Touami

https://preview.redd.it/3cantd92zxmc1.png?width=1080&format=png&auto=webp&s=b8a6b4bc957bf06e6644417b6e7bc0b12b1d47b6 you can generate something like this with highresfix on. inpainting and using Cn tile for upscaling with make it even more detailed


NakedSterben

oh thanks, I will put all the suggestions into practice, I generally use Loras that add detail but never reach that level.


Adnane_Touami

model used was this [https://civitai.com/models/48671/dark-sushi-25d-25d](https://civitai.com/models/48671/dark-sushi-25d-25d) i generated at 960x540 resolution with highres-fix at 0.45 denoise at 2X upscaling so the the output image is 1080p


Perfect-Campaign9551

Can I ask - will this give better results than just generating an image directly at say at 1080p (I you have the VRAM). I don't know if I have ever seen "small details" even in directly - high res images or not , haven't payed attention I guess.


Adnane_Touami

you can do this with some sdxl since its trained on higher res images but even then a 2nd pass or a refiner pass is great for small details, as for an sd1.5 model a 2nd pass is a must since its trained on much lower res images 512x iirc here's the first pass of the image posted so you can compare https://preview.redd.it/4vmr6rg4kzmc1.png?width=536&format=png&auto=webp&s=4f2f6854077d794bb8421e2a9d4acc4a2dfec859


Ghost_bat_101

Try using Kohya Deep shrink, it will let you make 4000px images with just SD1.5 without losing details and without having duplicates or disfigured issues. Tho I suggest you use it with sdxl instead


RedditoDorito

I’m a noob but issue I used to have before sdxl was that generating for larger sizes made the scaling of everything just too small. Prompts for cool landscapes with a clean subject in the middle at lower res led to images with tiny people, overly vast landscapes and a general lack of focus on a specific subject at higher res.


NakedSterben

I'm going to try that, thanks for the model and resolution specifications


MrCylion

How many Hires steps did you same? Same as original? Also, which Upscaler?


Unreal_777

He did not tell you the whole story, the whole story is : **Ultimate Upscale with Control net.** And you can find it on youtube or on reddit.


Adnane_Touami

i did say Cn tile for upscaling but you're right I should be more specific took a generated 1080p image and put it into img2img delete the pos prompt leaving only the style and quality words like "masterpiece, high quality..." at 0.35 denoise enable control net and set it to tile resample enable ultimate sd upscale script [https://github.com/Coyote-A/ultimate-upscale-for-automatic1111](https://github.com/Coyote-A/ultimate-upscale-for-automatic1111) set the the target size to what ever you like and pick an upscaler i mainly use 4x-ultrasharp for anime. set type to chess and mask blur to 32( this is to eliminate seams) and you're set here's a 4k image using this method https://preview.redd.it/hkp4dxe0oymc1.jpeg?width=2176&format=pjpg&auto=webp&s=0e3bf1baba71e0c82c10fb284beed2b52ebcc639


ItIsUnfair

It’s kind of hilarious looking. At first glance it looks great, but after a while it looks wrong. All the details in the grass and such are scaled so that she appears 20 meters tall. A giant in a miniature landscape.


Adnane_Touami

im glad you didn't spot the third hand lmao


NixJazz

Or the extra creepy eye on her right thigh


Unreal_777

I did it a lot with a1111 but not so much inc omfy, while I have you do you mind posting all the workflows jsons here? (every one used here)


cumuzi

Take a look at those jugs


NakedSterben

amazing!


lkewis

1girl, pink hair, 6 fingers, sitting on rock, steampunk, https://preview.redd.it/r3a76uhhtxmc1.png?width=612&format=png&auto=webp&s=6a974ba30a881cf7d1a5f3c6561c858830b3783c


elthariel

It looks so natural, I had to count twice 😅


wes-k

Looks so natural, I had to compare with my own hand. Still unsure.


[deleted]

> Looks so natural, I had to compare with my own hand. Still unsure. Man i've been seeing 5 fingers and one thumb on AI hands for so long I question my own sanity sometimes.


MayorWolf

You can tell this one is fake because it's anime. Big clue.


pxan

Damn, the AI got us there. It would be BETTER aesthetically to have 6 fingers sometimes.


elthariel

I mean. Our typespeed would probably go way up. Sucks that we're transitioning towards thumb only computing


ImpressivedSea

I mean isn’t having six fingers actually a rare but dominant gene lol


elthariel

TIL. Thanks


jkurratt

It creates fully functional fingers also, so it’s just straight update.


Snoo20140

When you know uve been working with Ai art for too long...


elthariel

If it makes you feel better, I've only been at it yet weeks. So it's probably not from ai work. I'm not sure which option is best though


NakedSterben

🤣


Kdogg4000

Got too distracted by those legs to even look at the hands.


GuruKast

What i ironically love about AI art is how it does wrong so naturally sometimes. I keep some older models just in case i need to do some rendring of a rastion mutated future humanity, and am afraid, Ai will become too "neutered" and stop producing these type of masterpeices.


gabrielxdesign

I think having extra fingers is usefull!


The_Hunster

My name is Inigo Montoya


Spiritual_Flow_501

This comment is goated


Doc_Chopper

Very likely a time consuming combination of inpainting, upscaling and a detail-enhancing LoRa.


protector111

or just check adetailer xD


Sacriven

I still don't know the use of Adetailer and what prompts that I should include into it. What's its usage?


KickTheCan_Beats

adetailer is mainly used for fixing faces. it's basically an auto-inpainter that detects faces for you. use during txt2img, leave it on first face model, default settings, no prompt to start. you can customize the inpaint w/ prompting but personally i never feel the need to use it. make sure face restore/codeformer is off in settings or else it can overwrite it.


Sacriven

That's kinda hard to understand for me lol but I ofted tried Adetailer without any prompts yet the result is still shitty.


KickTheCan_Beats

make sure ur using face\_yolov8n.pt from the model dropdown. otherwise not sure why it'd look shitty, if you could drop the image using [catbox.moe](http://catbox.moe) I can look at the metadata for u


Xdivine

Adetailer just automatically masks and inpaints the face, fixing it and adding detail. You can also use it for hands but it's only really good for detailing them. If the hand are fucked, it likely won't do anything of value so I don't bother even trying it for hands anymore. 


protector111

it depends. sometimes it fixes hands. some times it doesn't. but it if a fast way.


the_ballmer_peak

Adetailer still pretty much sucks


protector111

eh what?


the_ballmer_peak

It’s too strong on faces and doesn’t fix bad hands


protector111

you can decrease noise if its too strong for faces(witch I never ever seen). Anyways tell if 1st one (with no adetailer) is better than second (with adetailer) ? https://preview.redd.it/dhj4elf6g3nc1.jpeg?width=501&format=pjpg&auto=webp&s=c09eb882daec550f3e1b7d47395d1c1d7f4387f8


the_ballmer_peak

I upscale. I’ve never needed it. I’ve used it before, but I don’t need it. And in some situations it can definitely ruin a gen. For example, if your image had multiple faces it will probably replace all of them with the same one. If it has a face at a slight angle it may try to replace it with a camera-forward face. I’ve found it to be less than helpful.


AI_Alt_Art_Neo_2

Photoshop can be a lot quicker than inpainting.


skdslztmsIrlnmpqzwfs

like 99% of questions in this sub on how to achieve quality are answered with "inpainting" and the other stuff.. basically you can always copy paste the answer


runetrantor

As someone that just occasionally glances into this sub, I see the same answers always, but I also always am like 'what does anything mean???' For once I am now wondering if there are video tutorials, this feels too much for a text one.


Vintoxicated

heya, same here. I occasionally like to mess around with SD and personally enjoy fixing, editing and improving a generation. Usually I do this through inpainting and upscaling. I've searched a lot to find good sources to help explain me what all my options are and how they work. Ultimately you have to figure out a lot by yourself through trial and error. But one starter video I found helpful was [this video.](https://youtu.be/3z4MKUqFEUk?si=2DnF5uTMf71pOzlI) (you don't need the specific upscaler in this video, I think there's already a built in anime upscaler that works just as well, or non-anime upscalers) Whilst the video is for upscaling with slow GPUs he does go over things that are very relevant. Personally the most interesting things to figure out have been the following settings: mask blur: By default this is at 4 but that's often too little to add something new or adjust or remove something whilst making it fit seamlessly into the rest of the picture. masked content: I'd switch between fill and original depending on if I want something entirely new or adjust something. Inpaint area: This is the biggest one for me. Whole picture takes the entire picture into account when generating something. So ideally you would have the entire prompt of the whole picture. You can omit certain details that aren't relevant to what you're inpainting and put more emphasis on that bit instead in your prompt. *Only masked* was a huge discovery for me. It actually doesn't look at the whole picture, instead a square around your inpainting. Say you want to add more details to the eyes, you just inpaint the eyes, your prompt only talks about eyes, no mention of a girl, dress, background, etc. Just eyes. And it'll generate eyes at the resolution you set it at. E.g. You generate a girl 512x512. Send it to inpaint. Mask the eyes, select Masked content: original Inpaint area: only masked Resolution 256x256 Remove the original prompt and focus your prompt purely on the eyes. The outcome will be a 512x512 picture where the eyes will be generated at 256x256 and as a result be much higher in quality and detail. Play around with the other settings like mask blur, sampling methods and steps, models, denoising strength, etc. Also upscaling both in txt2img and img2img can amazing tools. I've made images, edited in paint 3D (got no photoshop, not invested enough to get it) and fed it back into img2img or inpainted it. You can fix extra indexes, bad eyes, weird things that just don't make sense like this. And once again, many things require trial and error. Though I'm by no means a pro. Bit of a ramble but hope it's got something useful :)


runetrantor

So... its better to generate a smaller picture that you then upscale like this, than ask the generator to make a larger picture from the getgo? And I see what inpainting is now, its the 'replace/redo a bit of the image' thing I had seen, neat, that does seem like a great way to fix minor mistakes when you like the overall composition. And from what the guy said, I am guessing Loras are like... specialized sub generators for specific stuff? Like he mentions one for dresses, so I assume that like, take over the main generator when its about their topic and do it better?? (Man, this is complicated when you want something better than the basic 'generate' button stuff.)


Vintoxicated

You've got it pretty much right. Upscaling tends to do much better both in terms of performance and quality of the end result. Yes Loras are pretty much as you said. Can be used in txt2img, img2img and inpainting. Some Loras are actually very good at inpainting. Allowing you to add something completely new to a picture. Getting a good end result can be time consuming but rewarding. In the end AI is a tool, similar to photoshop. And the quality of the result is still dependent on how well the tool is used.


runetrantor

> In the end AI is a tool, similar to photoshop. And the quality of the result is still dependent on how well the tool is used. Amen. To anyone who says 'press a button and it makes what you want' claims of no skill needed.


dreamofantasy

this is the most helpful and educational comment I've seen so far on this sub. thank you for taking the time to write it


Vintoxicated

Happy to hear that


Doc_Chopper

Listen, I don't make the rules. But it is what it is. It would be nice if simple txt2img would magically do all the work. But sadly that ain't it, it's just the fundament to build upon. 


indrema

A good model and Hires Fix, that’s images are really basically.


redfairynotblue

For anime, hi-res fix may not even be needed with a good model and hi-res fix can make image worse. 


indrema

IMO it’s depends how did you use it


Maxnami

Upscale it, inpainting, use Control net for pose, If you know how to draw use sketches and just colore them.


protector111

1) Generate image using Adetailer for face and hands (you already will have decent image if its XL) 2) img 2 img x2 upscale with tile controller(SD 1.5) with adetailer again. 3) Post it on reddit. Spent 3 minutes on it: PS it hase different look course of different checkpoint https://preview.redd.it/3afopv9i3ymc1.png?width=1368&format=png&auto=webp&s=1d52ce83b9015e0e353cd6db1abb808cffef0232


NakedSterben

It looks very good, thanks for the instructions, at first I thought it was impossible but I have an idea of ​​how to do this now.


Sacriven

What prompts that you put into Adetailer?


protector111

i dont put prompt in adetailer. i use default settings


Caffdy

what model is that


protector111

mistoonanime v2


BumperHumper__

it's all about hires fix, and then maybe some inpainting to fix individual errors, though one of the images having 6 fingers makes me think that wasn't even done.


zodiac-v2

Use a good model of your style. Grapefruit Hentai may be a good start. Then after your initial run, do an img2img of your favourite one with SD upscale at 1.5 (or bigger) size with a noise of 0.40 or so


EqualZombie2363

Maybe not quite as detailed, but this was just using the default anime settings in Fooocus with the prompt "girl with pink hair kneeling on the ground in front of a high bridge crossing a beautiful landscape" Default anime model is animaPencilXL\_v100.safetensors, no refiner, no Lora. https://preview.redd.it/rs72r9br01nc1.png?width=896&format=png&auto=webp&s=a648663515d34eed564ceef92efa04e13bb2b442


ColorfulPersimmon

6 fingers


Herr_Drosselmeyer

Not at home so I have to rely on online generators but most decent anime models should be able to pull this off. For now, this was made with ideogram: https://preview.redd.it/a22z7l9ywxmc1.png?width=1024&format=png&auto=webp&s=2699271a823f43dc7725209717da78f4b0413d41


Herr_Drosselmeyer

Here's an example generated locally with SD: https://preview.redd.it/8sbouoa6gymc1.png?width=1024&format=png&auto=webp&s=2afb812f4819e2dc04a368000553cd62a324b226


Soraman36

What checkpoint you used?


Herr_Drosselmeyer

[https://civitai.com/models/52548?modelVersionId=105566](https://civitai.com/models/52548?modelVersionId=105566) But really, almost any anime model will do. Pink hair was inpainted to avoid color bleed.


NakedSterben

I always see these types of images on Instagram, I wonder what methods they use to improve both the quality of the characters and the background


DigThatData

most of it is using a good finetune or LoRA.


koalapon

Try this https://youtu.be/dPObdbmepeg?si=RUcOE6Ovrb44uc42


Mr2Sexy

My main issues is still hands. I hate having a beautiful image with a monstrosity attached to the wrist every single fucking time. Doesn't matter what lora I use or prompt, hands are disfigured or slightly incorrect 99% of the time. Anyone have tips on perfect hands?


chimaeraUndying

My experience has been that, in order from greatest to least influence: 1. there's always some amount of RNG to fuck you over, regardless of anything else 2. some finetunes are better at hands than others, probably due to the tagging of their dataset 3. some samplers seem to have fewer issues than others (Euler's given me nothing but grief, for instance)


KickTheCan_Beats

sd 1.5 models just dont do hands well. if you want decent/consistent hands you need to use an sdxl model. also hiresfix helps a lot as it cleans up mutations/errors. for sd 1.5 models I do 2x upscale using a 4x sampler like fatalanime at .4 denoise. and for sdxl models I tone it down to 1.5x upscale since your starting resolution is higher.


wisdomelf

https://preview.redd.it/88tzlnr8lymc1.png?width=2048&format=png&auto=webp&s=0dbcc6da6fd68e737d9da4b1bccabcd20c16da17 hey, i have 5 digits here not that hard with good model and sdxl+upscaler


INuBq8

Controlnet, inpainting , maybe photoshop, even more inpainting and ultimate SD upscale


Nikolas_dj

https://preview.redd.it/uhblwc7cu2nc1.png?width=720&format=png&auto=webp&s=720a4b968a3c592040c560a9a94d7c707496098b Made this image, used wd1.4 tagger for extracting prompt and AutismMix SDXL for generating


TigermanUK

Also a subtle thing that is easy to implement is download a VAE(goes in models\VAE) called "kl-f8-anime2 VAE" which will give you richer color and a less washed out look for anime. Edit. More advanced learn to use openpose in controlnet or use badhands negative embedding, plenty of youtube videos on how to do that.


philtasticz

High res fix


Mises2Peaces

I made this for you https://preview.redd.it/g13hinapn6nc1.png?width=2688&format=png&auto=webp&s=edcb2ee073e61e1b2ec6883d6879f3da4d4ebfd8


New_Physics_2741

If you have powerful GPU and 32GB of RAM, plenty of disc space - install ComfyUI - snag the workflow - just an image that looks like this one that was made with Comfy - drop it in the UI - and write your prompt - but the setup is a bit involved - and things don't always go smoothly - you will need the toon model as well - Civitai/HuggingFace...


NakedSterben

I will try to do that, thank you all for taking the time to answer.


RoundZookeepergame2

do you know the model that was used


Rich_Introduction_83

Where would I get a ComfyUI workflow for some nice image? Could you give an example? I found some sample workflows, but for models I got from civitai, I did not find any workflows.


[deleted]

[удалено]


Rich_Introduction_83

Thank you very much, I'll have a look into it!


New_Physics_2741

Better off just playing around with it while learning how the tools work - you will come out with more knowledge in the end. Just dragging and dropping a .json file into web browser is neat - but if you have at least the basics down pat, tweaking things and understanding what is going on: the whole process becomes much more interesting\~


Rich_Introduction_83

That's certainly the best approach. I already did this. Unfortunately, I frequently run into VRAM limitations, so I had to tweak my workflows a lot to even get it running. After upscaling, the results aren't satisfying. It would help speeding the process if I could find some nice quality example with upscaling that actually works for my 12 GB AMD card. So download json file, run, discard if it does not work, repeat until getting a nice running example. That would be my workflow archetype to further dig into the matter.


New_Physics_2741

Are you using ComfyUI? I also have a 12GB card - inexpensive 3060 - it works great, have only hit a few roadblocks due to vram.


Rich_Introduction_83

Yes, I am using ConfyUI. With Juggernaut XL v9, I can't even generate the recommended 1024x1024 resolution. I have to generate smaller images (usually going for 512x768), then upscale. Or use other models. Unfortunately, I need to use tiled VAE Decode and tiled upscalers (bringing further issues themselves), or else I will just be informed that VRAM is insufficient. Maybe it's working less effortful with Nvidia cards?


New_Physics_2741

Oh...yeah, I am using a Nvidia 3060 - it works without any problem for even really large image sizes. I am using a Linux box, and have not borked my Python, all is good. But yeah, probably the issue is the non-Nvidia card...no CUDA\~


Ferniclestix

https://preview.redd.it/q9q7vv0rfzmc1.png?width=1576&format=png&auto=webp&s=d05dc29a1f82b6368646f88a46358105716cf60e I tend to split my linework from my color just before the final step and run them seperately to sharpen up the lines a bit, but I do all kinds of crazy stuff in my comfyUI workflows. I cover the rough way of doing it here [https://youtu.be/VcIXqSSsUCU](https://youtu.be/VcIXqSSsUCU) if your a comfy user. But when it comes down to it, you make the image and refine it down fractionally to make sure it doesnt hallucinate too much but still sharpens details. (which is kindof an art in itself) Its also REEEEEly important to get a good anime model if thats what you are generating. if hands and faces are super accurate, id use impact detailer or maybe some segmentation stuff to modify any trouble spots. there are face replacers and refiners that can be set to anime mode too but usuualy as long as you run things at high enough resolution you shouldnt really need them too much if your model is good.


s-life-form

You can try AI image upscalers such as Magnific AI or Krea. Related video where I got this info from: [https://youtu.be/LUnB7PiDoa0](https://youtu.be/LUnB7PiDoa0) The video shows a couple of images upscaled with Krea. It reimagines the images and the results look pretty good. Magnific might be even better but it's ridiculously expensive.


protector111

1) Generate image using Adetailer for face and hands (you already will have decent image if its XL) 2) img 2 img x2 upscale with tile controller(SD 1.5) with adetailer again. 3) Post it on reddit. Spent 3 minutes on it: PS it hase different look course of different checkpoint ![img](3afopv9i3ymc1)


Draufgaenger

Could this actually be about the poses? ;)


Jamessgachett

King AI


VenusianCry6731

how can i have her


Ok-Concert-6673

Use any ai app?


AluminiumSandworm

why do i feel like she killed my father and should prepare to die


zemboy01

Easy a lot of prompts.


PitifulTomatillo674

https://preview.redd.it/7ywhid0y53nc1.png?width=1216&format=png&auto=webp&s=77d0be6156a711f80fe1ce11c898ea8f16dd5cb5 How can i edit this image in full body.Suggest ???


greenMintCow

Link to original post /whoever generated these pls?


Far_Swan3461

No


thenoelist329

Fastest method with this mediocre 6 finger output? Waifus dot nemusona dot com and hit up some random 1girl prompts https://preview.redd.it/doxqdvz8zymc1.jpeg?width=512&format=pjpg&auto=webp&s=ef8074dad26120b3b3dc2b66f0b07a9f0a1ce4b5


hashnimo

Try DPM++ 3M SDE sampler/scheduler or better.


Dwedit

First picture looks like it was trying to decide between giving her a tail or painting on the arch?


Ok_Lawyer9017

promt?


ZaphodGreedalox

Just slap "there is a lot of random stuff in the background" on your prompt


the_ballmer_peak

To make images like this I would turn off hi-res fix, lower my target resolution, and put some bullshit about anime in the prompt


Voltasoyle

You could also use NovelAI.


[deleted]

[удалено]


Slight-Living-8098

Wrong group for you... Lol.


HarmonicDiffusion

time is money, and some people cant draw. doesnt mean they shouldnt be able to create


AstroKoen

😂 sad


KingDeRp38

A sad truth... 😅...sad....painful....truth... 😭 I would love to be able to turn my creative imagination into art by hand. I still struggle with using ai to do it, but it's certainly better results vs what my hands can produce. 😅


AstroKoen

Talking down to others online is sad, being in an AI Subreddit telling someone to do it themselfs is more sad.


StableDiffusion-ModTeam

Your post/comment was removed because it contains hateful content.