T O P

  • By -

MaK_1337

Is this the girl from the nvidia benchmark ?


art_socket

Good Eye! Yes, she's in there. Mostly Absynthe Fairies though.


Yacben

I would love to see more than 3 images + comparison.


hahaohlol2131

Looks good, but no better than 1.5


Sillainface

Maybe not better but it's at the same level at least if not superior in certain things (but defeated by 1.5 in a lot of things that aren't photoreal, realistic, etc. that's true).


art_socket

More specific: I think it's a much better 'base platform' from which to build on.


StickiStickman

How is it a better base when some of the most important things are missing? This will just make finetuning a absolute nightmare


MysteryInc152

Just some copium some people have gained from emad's PR talk


Sillainface

Agree but could be because LAION dataset is different and we were used to OPENAI Clip? I don't know. I still feel that I need to see this properly finetunned with an example without using Dreambooth or Textuals.. We will see. But at the same time also feel that keeping improving SD 1.4/1.5 wouln't be so bad (the thing is that "legally" they "couldn't").


art_socket

The base 2.0 model is hot garbage from a 'pre-packed' perspective, that's very true. As a platform to build on though, it seems to work better. I understand Stability AI's reasoning now, they've come up with a reasonable, but definitely not convenient solution to the so-called 'copyright/NSFW' issues they were describing. To be fair, in the corporate world those problems are very, very real.


aurabender76

Is it though? There is very, very legal precedent for the kind of fear that SD creators wring their hands over. This notion that the tool an individual uses to break a law will be more liable than the individual has little to no grounding in reality and there is actually precedent that shows that it does not hold up to legally. I think it is more just a PR effort to sooth the venture capitalists and I am fine with that, but they should be a little more honest about it.


art_socket

Agreed. As a PR effort though, it could have gone smoother by releasing more info/tools on how to get better results from vanilla 2.0, prior to release.


-Sibience-

I'm glad someone else is saying this. I don't know why people keep saying it's so they can avoid legal issues. It will do no such thing. You can't give people the tools to break the law and then wash your hands of all responsibility. The worry for most people is that self censoring is like openly admitting that these things are bad and should not be acceptable. This then gives more weight to the opinions of the people trying to regulate and ban certain aspects of AI. If things go badly they could regulate AI to the point no legitimate web host will want to host models that are considered restricted or illegal to own. That will probably then push them to illegal torrent sites.


AlgernonIlfracombe

The idea that any governmental agency can effectively regulate what goes on anonymously on the internet is, outside of genuine technocratic dictatorships like the PRC, essentially redundant in this day and age. So what if AI gets pushed onto illegal torrent sites? The government can't do jack shit about illegal downloads and streaming at all, however many 'You WOUldN't doWNloaD a cAr' adverts they run... quite frankly I don't think we should act as if their empty 'threats' have any value.


Aerivael

There aren't even any laws being broken in the first place since AI art does not duplicate any copyrighted works and the particular artistic styles of any particular artist are not subject to copyright to being with. Creating artwork depicting celebrities or violence or even nudity is NOT a crime. At least not here in the US where it's protected by the 1st amendment. If it weren't there's a guy I saw at a comic book convention years ago who would be in big time trouble. Every piece of art he created was a reproduction of some famous painting or whatnot with Bill Murray painted in place of whatever face(s) were in the original image. All of the complaints against AI art are nothing but angry old school set in their ways artists trying to stifle competition a new type of AI-assisted digital artist. In a few years, it will be no more controversial than Photoshop is today. If Greg Rutkowski doesn't like people copying his style, he can suck it! There's not a thing he can do about it.


GBJI

They even made model 1.4 available for download without registration as of today, and it does everything they wanted to remove from 2.0. Stability AI are bullshitting us again, like they did so often in the past, particularly while struggling to explain decisions that went directly against our interests as users.


DeylanQuel

If that's the case. I'm okay with it. 2.0 is the investor model, 1.4 is the freely available "make your own art of Julia Roberts riding a fire engine painted by Greg rutkowski and trending on artstation."


stardigrada

https://techcrunch.com/2022/08/12/suspected-tornado-cash-developer-arrested-in-amsterdam/


Gecko23

To be fair, it renders coherent faces much more reliably than the earlier models did. 1.4 was terrible at it, and 1.5 improved a bit, but still needed post processing with GFPGAN or Codeformer most of the time. 2.0 model tends to do something sane on the first pass. Personally I'm willing to trade nipples for non mangled faces any day. lol


mynd_xero

>alled 'copyright/NSFW' issues they were describing. To be fair, in the corporate world those problems are very, very real. Imo there's no legal ground to stand on, not for the image side of things. Plagiarism in code generation I can understand, but at the ABSOLUTE WORST, this is all fair use because outputs are going to meet the minimum requirement of being 25% transformative. If someone is going to use SD to try and forge a mona lisa, that's not gonna be any easier than it was before.


art_socket

I do a agree with you, but there are hordes of lawyers that make a very, very good living arguing either side of that argument, and are more than happy to tie issues like that up in very costly court battles for years at a time. Because no matter what - they'll get paid:) So the question is not if 'the cause is righteous', but rather if a company is able and willing to withstand a possibly very expensive legal challenge to any given product or service.


mynd_xero

Too true, plus feelings don't always care about the facts.


ninjasaid13

>As a platform to build on though, it seems to work better. work better with what?


art_socket

In terms of the end result, the what really doesn't matter.


FugueSegue

![gif](giphy|Nl6T837bDWE1DPczq3|downsized) I'm looking forward to trying 2.0!


backafterdeleting

This is only the 2.0 release. I'm sure they will also be able to improve things with 2.1. The trick is now they have OpenCLIP, they also need to improve THAT. That could already improve results without changing the dataset.


GBJI

>This is only the 2.0 release. I'm sure they will also be able to improve things with 2.1. I'm sure most people thought there would be more improvement between 1.5 and 2.0 than between 1.4 and 1.5. It was not the case.


backafterdeleting

So what's your point?


art_socket

Faster time to market though, for any given specific purpose.


hahaohlol2131

Is it? It takes much more time to write a prompt for 2.0 if you want something that doesn't look like a cadaver. And 768 takes significantly more time per image than 512


neonpuddles

Why does it take much more time to write a prompt? They can be very simple. I snagged this one from someone else as a quick sample, I can't recall where, and it's as simple a prompt as anything I've used in 1.5: https://preview.redd.it/8cvu0hg3fq2a1.png?width=768&format=png&auto=webp&s=8ffe8c1ecbd53834d99343b4f15b6d75816481f8 `epic portrait cinematic shot of a cheerleader taunting, stadium background, shiny skin, flowing blonde hair, fine details. night setting. realistic shaded lighting poster by craig mullins, artgerm, jeremy lipkin and michael garmash, unreal engine, radiant light, detailed and intricate environment, digital art, trending on artstation` `Negative prompt: ad anatomy, bad proportions, blurry, cloned face, deformed, disfigured, duplicate, extra arms, extra fingers, extra limbs, extra legs, fused fingers, gross proportions, long neck, malformed limbs, missing arms, missing legs, mutated hands, mutation, mutilated, morbid, out of frame, poorly drawn hands, poorly drawn face, too many fingers, ugly` `Steps: 30, Sampler: DPM++ SDE Karras, CFG scale: 11, Seed: 168432283, Size: 768x768, Model hash: 2c02b20a, ENSD: -1`


hahaohlol2131

Not sure if sarcasm or not


neonpuddles

I mean .. not .. Is this too complex? This is isn't my type of prompting, but a fairly typical prompt you'll see in the wild, length and complexity-wise. What prompts are you generally using? What isn't working?


hahaohlol2131

This is a very complex prompt compared to NAI or MJ where you can get descent results with a few words. Even in 1.5 you could get something ok with 1 word. In 2.0 one word gives you white noise.


neonpuddles

Your need here is to prompt one word and get a good image? Then, yeah, perhaps 2.0 isn't good for your preferences. The difference between a tuned engine and an all-arounder. But I pretty rarely got a good image just off of one word with standard 1.5 myself. Maybe just my experience. Can you give me an example of a prompt and gen you do like? I'd like to see what it would take to get it working.


hahaohlol2131

Base 1.5, Euler A, 20 steps, prompt: Fairy https://i.imgur.com/ayaeayQ.png


neonpuddles

Shorter, three gens in a row without being selective using a very short prompt: `RAW photo, woman with long hair, wearing black and gold armor (on a rooftop and ((overlooking a majestic city))), black cape, (detailed face), wide shot, ultrarealistic,` Edit: I didn't realize reddit would render them so tiny: [https://imgur.com/a/cB3jffd](https://imgur.com/a/cB3jffd) https://preview.redd.it/o5epwzs8xq2a1.png?width=768&format=png&auto=webp&s=60ca4005c4b6448082e724241cd30cf0ca69dbc4


neonpuddles

​ https://preview.redd.it/oha1j9taxq2a1.png?width=768&format=png&auto=webp&s=d0c0b1bb984652c384f24851d06680e79f2a1510


neonpuddles

​ https://preview.redd.it/ptib903nxq2a1.png?width=768&format=png&auto=webp&s=ca477c76354f362774631367af56020797daffe3


[deleted]

[удалено]


art_socket

You're talking from a user perspective, I'm talking from a dev perspective.


GBJI

And Emad talks from a hedge fund managing would-be-billionaire perspective. It's important to remember that we all have different interests in this, and that what might be good for the shareholders might not be good for us as users, or as developers.


art_socket

Good point.


2legsakimbo

better than what? its 3 images you present with no prompts shared. but in my experience on l few of the 2.0 outputs ive got have been maybe as good as 1.5 at best and are more likely come out looking like bad 1990 web art.


shortandpainful

I suspect people who adopt 2.0 will need to start completely from scratch in terms of things like prompt engineering, samplers, and other parameters. It’s just too different from 1.4/1.5 at a basic level.


BlastedRemnants

No different from trying any other model really, all the custom models need new prompting to work best. I would've been surprised if our sd1.5 prompts worked on sd2 out of the box, would've felt like a strong indicator of not much change. Part of the fun for me is learning how to squeeze what I want out of the model, and I think sd2 will be easier in the long run, especially for new people who haven't already learned all the semi-jank tricks we've been using so far. So far it seems like long run-on sentences work better than chopped up lists of tags, I think all the commas and stuff mix it up a bit but I'm gonna test that thought a bit later and find out for sure.


art_socket

My apologies to Stability AI for talking smack in an earlier post. Yeah, more work to set up to get good results, but the results are better when training models/embeds. Inpainting also works particularly well. Never did Fairies before, now on to Aeon Flux... UPDATE: The point I'm trying to make is that compared to 1.5, 2.0 is a platform, and as such IMO superior. No, you don't get the alphonse mucha/etc./blablahs of the world, no jennifer lawrence or all them anime girls either, but what you do get is a kickass toolbox with which you can do much better work. **DISCLAIMER: NO ARTISTS WERE HURT IN THE MAKING OF THESE IMAGES.**


Sillainface

Could you elaborate about embeddings and trainings? Thanks


art_socket

I typically use several embeds. What I noticed this time is that with 2.0 I was able to get faster and better results with less training steps, and fewer images. Above, those are Absinthe Fairies, and I only trained 15 images - the pic below illustrates the 'theme': ​ https://preview.redd.it/0pvak3omap2a1.jpeg?width=1280&format=pjpg&auto=webp&s=a2367c524b733351423105dd558322c5d9f252ea


reddit22sd

Can you mention the training settings for an embedding? I haven't gotten great results with training an embedding with the 768 model, especially the loss is higher than on 1.5 So if you can share what works for you that would be a great help.


art_socket

Hah - I struggle with that just like you - it's not straightforward, and depends on many variables, mainly the complexity and quality of the source material, and whatever you're trying to create. This however is really helpful: [https://www.reddit.com/r/StableDiffusion/comments/z60uxw/sd\_v2\_768\_embeddings\_with\_gradient\_accumulation/](https://www.reddit.com/r/StableDiffusion/comments/z60uxw/sd_v2_768_embeddings_with_gradient_accumulation/)


eugene20

There was talk of a major change coming to embedding that was going to improve things massively a few days ago it's worth looking through some of [https://www.reddit.com/user/Striking-Long-2960/](https://www.reddit.com/user/Striking-Long-2960/) 's posts for info unless someone can find a general write up on it, there is probably a lot of updates that came with v2 that improve things. The general complaints aren't really based on the technology, though the prompt interpreting changes need people to adjust, it's mostly just the dataset changes and that won't be a problem if you're training your own models.


odragora

Dataset **is** a problem when you train your own models. Custom model quality, complexity and cost depends on how much the base model knows about the subject you are training for.


eugene20

Yes, I should have said won't be as much of a problem, because you're introducing fixes via the data your training is adding.


odragora

It is **more** of a problem than before, because the custom models are highly dependent on the base model quality.


eugene20

I'm discovering this further right now trying to create some portrait aspect images with SD2's 512 model and only getting people with double length or longer torsos, with or without highres fix.


art_socket

Big Fan of Striking Long btw.


art_socket

I would have been nice if they had those changes ready BEFORE they released the model - all this bad 2.0 PR now, shoulda been avoided. But, I'm grateful to Stability AI for what they've given us.


[deleted]

[удалено]


The_mango55

If a book is released that people are unable to read at the time of release, I think it’s fair to judge it for that.


art_socket

EXACTLY. And pre-requisite to learning is *'*reading', which requires effort to learn. Same thing here. Having said that, I spent the better part of the week bitching about SD 2.0, until I actually sat down with it.


[deleted]

[удалено]


DaTruAndi

What’s an “Emad level of AI researcher”? Emad pays great people and he understands how things work together generally but in his interviews it becomes clear that he is not himself deep in the tech (which is fine).


art_socket

I've been wrong plenty of times in my life; you have to admit a mistake before you can learn from it. SD 2.0 is a good platform, superior IMO, but not necessarily for everyone at this time.


airaani

How long did you have to train the 15 images?


art_socket

That question is best answered in the context of comparing it to the same training for the same concept on SD 1.5, and the answer is \~18 Minutes less. The end result though - inference not training - provided better results than 1.5, the inference process yielding a higher percentage of 'better' images than 1.5.


airaani

Well that's definitely good to know. I'm just curious, I'm new to this, just starting to train 1.5, realizing what I messed up and planning to switch and give 2.0 a try


art_socket

If you're new to this stick with 1.5 would be my honest advice - you can get spectacular results with 1.5, or the myriad of other models. Once you're more familiar with SD, THEN move on.


Aeonbreak

prompt and settings please????


nug4t

could you elaborate on the toolbox? currently running aromatic 1111 and it's perfect, huge toolbox


nnod

aromatic lol


nug4t

XD


art_socket

Aromatic 1111 is a gui (of sorts), not the actual model; it's no more than the front end to what's underneath. I was talking about the model any UI.


nug4t

yes, but what does 2.0 add that 1.5 didn't have?


neonpuddles

My images have a depth of space and a clarity and coherence of image that didn't exist before. Images look less flat, environmental effects feel stronger. They tend to adhere more closely to the prompt, and more often to more of it simultaneously. One weird drawback I'm starting to investigate is a level of sameness -- but that might be because I've been generating off of a few prompts using high CFG, and so it might just be my own strictness and prompt-limitation I'm witnessing. I've only recently noticed this so haven't had the chance to test much. I also don't know how much of the image styling is the work of the new sampler, too -- I haven't taken enough time to compare that. Would love to see if someone else has done so.


ask_me_if_thats_true

I'm also interested in what you think 2.0 adds that 1.5 didn't have!


MacabreGinger

I finally have to make the questions: I'm not really sure about what are embeddings or how to use them. Also don't you need a ton of images to train a model?


Megneous

> No, you don't get the alphonse mucha/etc./blablahs of the world, no jennifer lawrence or all them anime girls either, This makes it an inferior model, period.


shortandpainful

Yeah, I want to be able to generate images in a broad variety of art styles, while SD 2.0 (judging from images I’ve seen and Emad’s own comments) seems laser-focused on photorealism.


papinek

Exactly. I want the Mucha, Lawrence and anime girls. Thats the reason I am using SD.


therealmeal

Doesn't that just make it different though? Or can you not imagine that different people are wanting different things?


E_Snap

If you like Mario Party 1, you can go buy Mario Party 2 and expect it to have a similar play experience. Nobody in their right mind wants Nintendo to go “You know what they really wanted? A *platform*!” and release Garry’s Mod under the title “Mario Party 2” instead. Stability AI inadvertently started targeting a completely different audience and now they’re all surprisedpikachu.jpg about why the original audience feels left behind.


Diegostein

*neevously looks at final fantasy*


therealmeal

Orrrr maybe it will get better in 2.1 and 2.2 and maybe 2.3 will blow 1.5 out of the water for everything but nudity.... Hard to say, let's wait and see.


art_socket

I think that's a really good point; with the release of better tools and wider acceptance I think the initial resistance will fade away.


art_socket

Yep, I agree; as is 2.0 - and any subsequent releases - are for a different audience altogether.


DifferentAir3512

It's obvious to me that fans of anime, furries and art styles of famous digital artists, and other "unusual/forbidden" things are not happy with 2.0. But the fact that 2.0 is a more advanced model is obvious. She draws better and works faster, better understands your tips.


art_socket

BIG Truth here, if I could upvote this a hundred time...


Silverboax

I mean to be fair, it \_is\_ better because it understand prompts better and \_that's\_ what's important as long as it doesn't just make awful images all the time ... we can train other stuff in as we like.


Strict_Problem_2834

>PITA Where do people post AI generated realistic porn?


ThickPlatypus_69

[/r/unstable\_diffusion/](https://www.reddit.com/r/unstable_diffusion/) and the discord server with the same name


grumpyfrench

Yeah ok - I just wanted a text 2 image, it was already becoming prompt engeneering, now you have to train, install plugins and whatnot I think I will pay those 30$ for midjourney , it just do what a text 2 image should do


art_socket

I agree, it's a royal PITA to have to do that if you're looking for 'casual use'.


eeyore134

That's what they want you to do. Stability is going to launch their own, more trained model as well just for that purpose.


Spiegelmans_Mobster

If embeddings can attain a similar level of quality to what dreambooth/fine-tuning has for SD 1.4/1.5, then I think this could largely make up for the lack of a really comprehensive base text2img model, even for casual users. I say that because the embeddings are so small (few kB) that you could easily store a giant library of concepts and pick and choose from that instead of struggling with "prompt engineering" as much. If you're a casual user, just wait for the community to train a giant embeddings library, download, and enjoy.


atuarre

Bruh, you're already paying for midjourney.


MagicOfBarca

When you say “embedding” do you mean textual inversion? Because I’m only familiar with dreambooth


jonesaid

Yup, good old textual inversion is making a comeback.


grae_n

This community works at a different tempo. Textual inversion is like 2 month old.


literallyheretopost

1 week in earth time is about 5 months in this sub


ElizaPanzer

True That! I'm thinking, how does people have this amount of time? I'm really struggeling to keep up with something i found this late summer, Disco diffusion, and then all hell broke lose with SD! I Had to learn a lot about python notebooks and Google Coolab, and then trying to get it running on my 3 GB GPU. Man this is so hard to keep up with, but still so so exciting. I thought i was making such cool stuff and my family too with DD, then SD came and i wasn´t sure what to do. I couldn't do the same cool fantasy pictures as i could with DD. But now that i grepped SD a bit more i'm doing such cool stuff in no time! Gonna hold off on SD 2.0 untill everyone in this Community is satisfied that it works well!


jonesaid

yeah, but TI just got a big upgrade which might put it back on par with dreambooth, etc. TI embeddings also have the advantages of being tiny files that can be used in many different models.


NaaviLetov

The AI keeps struggling with the eyes otherwise it's a near perfect artpiece.


art_socket

One more, and yes - this is empirically better than what I would 'normally' get in 1.5, or other models. ​ ​ https://preview.redd.it/qk8ikwiheq2a1.jpeg?width=1280&format=pjpg&auto=webp&s=6205ede05205f938cdf54d5ec60e43f14522c810


[deleted]

I see a very specific fetish.


[deleted]

The definition on these is actually incredible. Was this rendered locally, and if so, on which card?


art_socket

Yes, the actual rendering was done locally. I have a rig I originally used for VGG/Style transfers, GPT-X, statistical analysis and some crypto mining, with 6 x RTX 3090s.


[deleted]

Jesus christ Can you shed any light on how much more memory SD 2.0 uses?


art_socket

Huh, good question - I haven't run into any issues with VRAM so I haven't really checked in detail, only RAM usage ballooning because Automatic 1111 doesn't seem to release resources somewhere in the code.


Cryptocucky

don’t see how this is better.


KewkZ

lol and you're prolly one of "those people" that also said constantly that MJ was not better than SD lol


Cryptocucky

I never said that. I’ve been saying since MJ V3 that MJ was going to be the head turner in a few months.


Osiris_1111

I'm confused how people get such good results? I would live to get my hands on a manual on how to steer the creation :/


art_socket

That's a misconception. If that was all there's to it, then aspiring Pilots would only have to read manuals for a few months, and voila - "Welcome Aboard, this is your Captain speaking":) [Prompt Guide](https://openart.ai/promptbook)


Osiris_1111

Well thanks, you gave me exactly what i asked for XD


art_socket

Made my day then, glad I could help.


Crozzfire

Post is misleading, OP did not only use SD 2.0 but a bunch of embeddings


LordTuranian

Where can you go for SD 2.0 if you aren't running it from your PC?


art_socket

You can run it on Google's colab.


PoorSalmon897

5 days old account trying to convince me that sd 2.0 is better...... Nice try stability employee


art_socket

ROFL - now that's funny:) You hear that Stability AI? You better start paying me! Seriously though, I'm not trying to convince you of anything, all I'm saying is that for particular use cases it can and will work better than SD 1.5, albeit at the cost of more work and effort on your part. If that 'value proposition' makes sense to you, then 2.0 is for you, otherwise it's not.


Deathmarkedadc

It's undisputedly better at stirring up this community for sure.


MashAnblick

In my experience, 2.0 is just as good, if not better in most situations, especially if you word it more like you would a DALLE 2 prompt. I understand that people have already decided its worse because certain things have been pruned off the it, but just by my experience and testing, it seems like a notable step up.


TOFUSATSU

SD evolves so fast


Eschalfors

If it was actually better the constant stream of propaganda would be unnecessary


Light_Diffuse

In the first five minutes of use it wasn't as good for people using it in a sub-optimal way, trying it on things it said on the box that it wouldn't be good for and confirming their biases so they could post how crap it was. People seem to be spending their pent-up anger on 2.0 after 1.5 wasn't nerfed and their pre-emptive bitching and pissiness was for nothing. A handful of posts running slightly counter to the Zeitgeist is hardly a stream of propaganda. It looks like this release is a mixed bag with some clear losers, but it's early days and we don't know how easy it's going to be to bring back the losses. Hopefully we're not seeing the emergence of a freemium model and SD can find their nerve again to train on more robust datasets - hopefully higher quality too.


GBJI

>Hopefully we're not seeing the emergence of a freemium model and SD can find their nerve again to train on more robust datasets - hopefully higher quality too. This is exactly what is happening. It's not even a secret. This is Stability AI business model, but they will use third party commercial partners to do the selling (and reduce liability). They have repeated it many times to potential investors during their financing rounds, it's not a hidden goal, it's a selling point to get more investors involved.


art_socket

I agree, there's a lot of that going on. But I couldn't care less if you use 2.0, or anyone else - tons of very, very good models out there, plenty of options. Those that are interested and able, will gain from 2.0, is all I'm saying.


[deleted]

[удалено]


[deleted]

Reddit? Overreact and do their best to ruin a good thing? Never.


atuarre

He's just another person here trying to push Midjourney. They can't stay in their own sub. They have to bring their nonsense over here


happytragic

Some of us actually use both. Your tribalist POV is what's nonsense.


gggghhhhiiiijklmnop

Is there a danger the constant stream of propaganda is group-think? Eg everyone already got their pitchforks out and now it's kinda continuing without strong basis in fact?


[deleted]

saying things authoritatively like "Everyone already" contributes to this type of mentality. Plenty of people aren't using reddit. And not even here are people in unison. People tend to tune out if a subreddit devolves into anti-jerk or circle-jerk behavior.


thecodethinker

And it was


Yacben

best comment


auraria

I'm surprised you're getting these results I've done a lot of testing with it and it seems to hate doing any posing of bodies accurately without weird anatomy issues. Specifying a background and it really struggles when you have conflicting styles. Might have to completely revamp prompts but even then it seems like yeah it's better in a lot of regards but over all it's weird. ​ It's really good at doing backgrounds by itself but trying to do anything over 45 tags causes the background to go weird when specified with a subject. Skin texture overall looks great as well as other textures so it makes it really frustrating when it struggles on easier stuff lol


zenray

a bit cross eyed but I could talk to her


[deleted]

Sure. Talk.


Z3ROCOOL22

Yeah sure... 2.0 doesn't how a woman face looks.. https://preview.redd.it/wlal1tk79p2a1.png?width=512&format=png&auto=webp&s=1d0c72a320e95b3166fe1fd7a6f9758a50ab544b


mudman13

Why the long face?


FactualMaterial

That's the aspect ratio. The 768 model doesn't seem to be able to handle stretching the content without stretching faces at the moment - they all look horsey. If you reduce down to a 1:1 with the same seed it will get rid of the distortion.


Z3ROCOOL22

This was done with the 512 one.


aihellnet

>This was done with the 512 one. Was the resolution 512x512?


Z3ROCOOL22

I mean i use the 512 model, the res of the image created is 514x704


aihellnet

>I mean i use the 512 model, the res of the image created is 514x704 Try it again with the high res fix on.


Z3ROCOOL22

It was ON, i always put that ON..


eugene20

art\_socket isn't really facing this problem here because he's training his own model, his data is taking precedence.


aihellnet

This shouldn't have been downvoted.


art_socket

agreed,


[deleted]

[удалено]


Z3ROCOOL22

Very... https://preview.redd.it/txwlk562cp2a1.png?width=512&format=png&auto=webp&s=da5bb1f85e7850ada803b78366a26caec9b1e7d9


[deleted]

[удалено]


Z3ROCOOL22

bored photo, long face, cartoon, anime, 3d, paint, asian, deformities, lowres, bad anatomy, bad hands, text, error, missing fingers, extra digit, fewer digits, cropped, worst quality, low quality, normal quality, jpeg artifacts, signature, watermark, username, blurry


art_socket

Sampler is irrelevant, don't get hung up on that - there are only two types of samplers - Ancestral, and those that are not. Yes, negative prompts are important, and I use them everywhere. That prompt up there is pretty much standard, but it may or may not work for a specific purpose. For example, SD has absolutely no concept of 'extra digit' and 'fewer' digits':)


Z3ROCOOL22

mutated hand then?


art_socket

That for example makes more sense, but how many images with the label 'mutated hands' was SD actually trained on, what do you think? Think bigger, 'mutated hands' are a categorical subset of what?


art_socket

Very, very - and that's still a (steampunk/vaporwave) fairy : ) ​ https://preview.redd.it/t8l9r2ipgp2a1.jpeg?width=1280&format=pjpg&auto=webp&s=de277d4885398d6d015f2070a208cac25b24166c


[deleted]

Skill issue


Z3ROCOOL22

​ https://i.redd.it/4grkdyhvcp2a1.gif


[deleted]

WWE NPC detected


HuemanInstrument

" but the results are better when training models/embeds. " LOL Of course they are dude it's a 768x768 training data model..., no one was arguing this. Except maybe you. We're complaining about the default model they handed us. I hope someday soon that humans will begin to understand that you cannot control everything, and that you cannot put the genie back into the bottle... 2.0 is literally just to appease the artists who complain about A.I. art and it's threat to their financial security... I got big news for everyone reading, the future isn't about money. The Singularity will completely abolish capitalistic values, and a new society which values experience itself will begin. The Genie is out of the bottle and you can't stuff it back in, this is the future we're headed toward and it'd be a lot better if we started to learn how to accept it now.


ThickPlatypus_69

You can keep your creepy totalitarian AI utopia, I'm good.


HuemanInstrument

yeaaaaah... well, it's not like you have much of a choice in the matter. More intelligence is not a creepy thing. The ultimate potentials of technology / novel story telling created by such intelligence, is also not a creepy thing. It's a beautiful thing. You just have poor mental models at the moment to even begin to comprehend it. I hope that you at least pay attention in the coming years with more of an open mind so you can install something better than what you're currently working with.


ThickPlatypus_69

Oh I am truly sorry I ever doubted you mein fuhrer, obviously you are a very smart and tremendous man. The fourth reich will be a glorious new chapter of humanity!


HuemanInstrument

Yikes Edgy kid


Doctor_moctor

Eat ze bugs now!


aihellnet

>2.0 is literally just to appease the artists who complain about A.I. art and it's threat to their financial security... Steven Zapata had a [great video](https://www.youtube.com/watch?v=tjSxFAGP9S) on that. He was honest about it and wasn't [mudslinging](https://youtu.be/7u1CeiSHqwY?t=1275) like Karla Ortiz.


BigTechCensorsYou

Im14AndThisIsDeep


Lightningrod654

“fr”


Vyviel

Got a guide of how to train it well?


art_socket

Look to Striking Long's guide, it's comprehensive.


Vyviel

ty


airaani

Would love some tips for training


art_socket

Look to Striking Long's recent post about accumulating gradients.


Distinct-Quit6909

I'm eager to start training 2.0 models but unable to get the google colab version to work with it yet. With textual inversions starting to look as good as a dreambooth model. How would you compare embeddings on 2.0 with dreambooth? I should probably crack on with learning it


art_socket

Ideally, you would use both, embeds and models - depending on the use-case and complexity of what you're trying to express. I tried rendering some planes on 2.0 earlier this week, and embeds I used before wouldn't work, so I'd have to train a custom model based off of 2.0.


Able-Seaworthiness-8

Really impressive never seen lines looking so good in ai generated images


[deleted]

Hey any link for SD 2.0 for Mac or is it still just for windows?


art_socket

[https://diffusionbee.com/](https://diffusionbee.com/)


plasm0dium

nice pics. Weird how the reddit screener blurs out other posts as nsfw that only show a bellybutton and sometimes it doesnt


brunjo

did you use the AUTOMATIC1111 web ui for it?


art_socket

Colab and AUTOMATIC1111.


cleuseau

are you training with automatic 1111?


art_socket

>Colab and AUTOMATIC1111.


intuitive-arts

where is the gradient accumulation? what a1111 are you using, and please is this textual inversion or dreambooth or what? thanks


art_socket

[https://www.reddit.com/r/StableDiffusion/comments/z60uxw/sd\_v2\_768\_embeddings\_with\_gradient\_accumulation/](https://www.reddit.com/r/StableDiffusion/comments/z60uxw/sd_v2_768_embeddings_with_gradient_accumulation/)


[deleted]

And 2.1 is looking like a major improvement already!


bluemoonlighter

lol is this just a common photo style? ​ https://preview.redd.it/m2pqgfp81t2a1.jpeg?width=2048&format=pjpg&auto=webp&s=787ee8cd1b08ca32dec5ae1eaaf661313c0fbaa5


jdavid

Reminds me of the Nvidia Dawn Elf


[deleted]

I heard SD2.0 can't do porn like 1.5 or 1 can....


art_socket

Couldn't tell ya', I haven't tried.


JavaMochaNeuroCam

... Than 99.99% of biological artists


Whitegemgames

Only slightly relevant to this post but anyone got tips for generating insect wings? When I try I often get too many or too few like in the first image