Stable Diffusion, NovelAI, Machine Learning Art - AI art generation discussion and image dump

  • 🐕 I am attempting to get the site runnning as fast as possible. If you are experiencing slow page load times, please report it.
Thanks, but I'm having trouble getting faces to not be a horrible nightmare. Apologies if this sort of stuff has been explained earlier in the thread, but any tips for non-nightmare faces?
If you're using the 1.5 UI there is an option you can check that attempts to fix faces.

one_weird_trick.png
 
Thanks, but I'm having trouble getting faces to not be a horrible nightmare. Apologies if this sort of stuff has been explained earlier in the thread, but any tips for non-nightmare faces?
From what I've generally seen, if the character is not in the foreground and the main focus of the image you're generating isn't their face, odds are decent their face may get messed up. Here's a few examples from before my recent re-install, if it's working never ever update.
00167-3661778266.jpg
In this particular case, I'm using prompts like "close-up, extreme closeup" and "portrait, face focus". Basically, the face is the main focus and I think I've very rarely seen one of these with a heavily distorted face. At worst it's usually a miscolored eye which isn't hard to fix with 30 seconds in an image editor.
00026-169138885.jpg
So, as you can see, this particular face didn't turn out so well and I believe running face correction didn't help at all in this particular case. As a general guideline, I tend to think of it in terms of "the farther the character is from the point of view, the more likely it is their face will get distorted". Inpainting/img2img on the face and photoshopping may have been able to fix it, but decided not to bother in this case.
00032-699618329.jpg00034-699618331.jpg
Pretty sure these two images had the same prompt as the one prior; as you can see, much closer to the point of view, vastly superior faces.
00035-3592518446.jpg
Another general rule; in my opinion most models really don't perform well when the character is oriented/facing away from the point of view. Straight on to the back of the head seems fine, but any other angle facing away often seems to have issues.

Overall, this seems to be more of an issue when using realistic/painting styles than anime styles, but I may be either remembering incorrectly or more fortunate in my experimentation.

TL;DR -
1. Characters should face you and if possible be oriented towards you.
2. Characters (or their faces) should dominate the work and take up the majority of it for best results.
 
"Adolf Hitler with a catgirl"
Done using everything 4.5
 

Attachments

  • 00038-224852318.png
    00038-224852318.png
    430.6 KB · Views: 24
  • 00039-224852319.png
    00039-224852319.png
    412.6 KB · Views: 22
  • 00040-224852320.png
    00040-224852320.png
    404.4 KB · Views: 25
  • 00041-224852321.png
    00041-224852321.png
    435.6 KB · Views: 18
  • 00042-224852322.png
    00042-224852322.png
    358.1 KB · Views: 19
  • 00043-224852323.png
    00043-224852323.png
    431.2 KB · Views: 21
  • 00044-224852324.png
    00044-224852324.png
    449.7 KB · Views: 16
  • 00045-224852325.png
    00045-224852325.png
    437.1 KB · Views: 17
  • 00046-224852326.png
    00046-224852326.png
    393.5 KB · Views: 22
  • 00047-224852327.png
    00047-224852327.png
    384.4 KB · Views: 18
  • 00048-224852328.png
    00048-224852328.png
    352.3 KB · Views: 14
  • 00049-224852329.png
    00049-224852329.png
    435.9 KB · Views: 20
This is cool shit, but fuck if it hasn't made me depressed about learning drawing. Whats the point if a machine can do it better?
I mean, to me it's like playing chess. AI crushed humans in chess forever and I don't think a human will ever beat a chess machine ever again except if the computer is nerfed to shit. That doesn't make playing chess any less fun.

Though I guess it's different since it's not like chess-playing employs millions. But honestly, if you really love drawing, a machine doing it better shouldn't demoralize you any more than a human master doing it better.

On that note, a lot of the anti-AI bros on Twitter are already throwing in the towel and asking to just have AI art be segregated (snigger) from them now. I bet anything they'll be angry as soon as another big studio uses AI, but for Twitter, it's surprisingly not as vitriolic as I was expecting.
 
Even with all the improvements in this technology we have seen in the last months it's really difficult to impossible to make it do detailing in a certain way or really just draw a specific motive. With some work you can get roughly what you wanted, but never exactly besides sheer luck. As long as that's the case, I personally don't see the direct competition with human artists.
 
Even with all the improvements in this technology we have seen in the last months it's really difficult to impossible to make it do detailing in a certain way or really just draw a specific motive. With some work you can get roughly what you wanted, but never exactly besides sheer luck. As long as that's the case, I personally don't see the direct competition with human artists.
Exactly. AI art is good for churning out pictures quickly, and can produce some surprisingly good results at times, given the right combo of model, prompt, and seed. However, when it comes to specifics, that's where you'll either need to try and wrangle it further with inpainting and Photoshop and such, or just accept that you're not going to get it exactly how you want and leave it at "good enough."

That said, you  can use it as a springboard to get a rough concept of your idea, and then take the output to a human artist to refine and redraw with the details you want. Instead of ideas being trapped in your head and being unable to show them without spending months or years learning, or having to hope that an artist can understand your description and translate correctly, you can have the computer get you most of the way and the artist the rest.

So to @alpha889boba and any other artists-in-training, don't see AI as a threat. It's a tool like any other, one that you'll probably be able to get a lot of use out of too.
 
I just powered up my old laptop, and I found some images generated in April 2022 using Looking Glass. I'd completely forgotten about this.
lg8_-content-images--.pnglg0_-content-images--.png
lg5_-content-images--.pnglg9_-content-images--.png

Compare that to some of the recent images I've generated with Stable Diffusion
sample-16500.png00043-1454296993-the full moon in the night sky, qhmtg style.png

It's fucking ridiculous seeing how far the tech has come in less than a year.
 
Last edited:
I just powered up my old laptop, and I found some images generated in April 2022 using Looking Glass. I'd completely forgotten about this.
View attachment 4572665View attachment 4572668
View attachment 4572689View attachment 4572693

It's fucking ridiculous seeing how far the tech has come in less than a year.
Looking Glass was Google's and through their servers only, right? Now you can generate something far better not even for a premium on someone's enterprise hardware but on a consumer pc built for gaming from like 3 years back. I hope we see this kind of advancement in text ai.
 
Looking Glass was Google's and through their servers only, right? Now you can generate something far better not even for a premium on someone's enterprise hardware but on a consumer pc built for gaming from like 3 years back. I hope we see this kind of advancement in text ai.
Looking glass was a prototype of Dreambooth that used ruDALL-E as the base model. In this case, I fed Harold McNeill's Circle of Protection paintings and some Fullmetal Alchemist fan alchemy circles as training data.

I believe you were able to run the code locally since the project was hosted on colab, but it's been so long I can't remember.

And I fully agree. I think OpenAI not sharing ChatGPT is highly unethical, since it's stifling the tech's evolution and going against the company's original mission.
 
And I fully agree. I think OpenAI not sharing ChatGPT is highly unethical, since it's stifling the tech's evolution and going against the company's original mission.
It goes against the company's own name. Why did they keep that when they decided to sell out?
I've been critical of OpenAI for years and I don't remember noticing the convoluted for-profit transformation until recently:

They have been whining about the dangerousness of their own tools since at least GPT-2. Partly because of Musk killer robot rhetoric when he was a bigger influence there, the usual fake news and deepfake reasons, and now marketing and protection of the golden eggs they are laying. This shit is so good we can't even release it to the general public, yo.

There was a ton of press behind the OpenAI name even before the GPT-3/ChatGPT craze. OpenAI sounds nice and... open. I don't think they are going to change the name anytime soon, no matter how closed they get.
 
Thanks, but I'm having trouble getting faces to not be a horrible nightmare. Apologies if this sort of stuff has been explained earlier in the thread, but any tips for non-nightmare faces?
If you're using the 1.5 UI there is an option you can check that attempts to fix faces.

View attachment 4555146
That option basically just runs a different, face restoration "AI". It does work but it will make the faces look a lot different and quite same-y.
by far the best way i've found to "fix" faces is just using hires fix. if you aren't aware, almost every one of these models is trained on images at a 512x512 resolution. the more you stray from that, the more likely it is to start introducing weirdness to your images. basically once it gets to 512 it "restarts" the generation, and anything beyond that will be duplicated or tiled. this is what leads to those images you see where it looks like it gave up halfway through making a character and started generating a new body where their neck should be or things like background objects and characters repeating. ideally you would generate the initial image as close to 512 or less as you can, but because of the low resolution it can struggle with finer details in things like faces or hands. hires fix works by first generating a smaller image at the resolution you specify, and then essentially doing img2img passes over that image at a higher resolution. it works very well both for fixing imperfections and for adding more details to your image, depending on the upscaler you're using.

a couple examples:
1.png2.png
not terrible but kind of derpy
3.png4.png
much better!

it doesn't even have to be a huge difference in image size, 1.25x should be plenty just to get better faces. play around with different upscalers for different results. latent tends to add a lot of detail but can only really be used at higher denoise values and it sometimes changes the image a lot. ESRGAN works well at lower denoise, around 0.3 should be fine.

really, try upscaling your images in general. you might be surprised how much different the final results can be!
initial image:
5.png
1.25x hires fix:
6.png
2x img2img:
7.png
this was actually an upscaler meant for anime images but i liked how the contrast and colors turned out
same settings as the first one
8.png9.png10.png
 
Last edited:
As long as that's the case, I personally don't see the direct competition with human artists.
I think it will probably be soon at generating more visually compelling procedurally generated content for games. I bet we see another tantrum when that starts happening.
 
I think it will probably be soon at generating more visually compelling procedurally generated content for games. I bet we see another tantrum when that starts happening.
Friend just finished fulfilling her kickstarter after she allocated the art budget to a pair of A4000s and generated all the art with a custom dreambooth model. Lowered her total budget from $10k for commissioning artists to $2k for the GPUs and $4k on printing books and game pieces.

Nobody knows the art was AI generated, especially since a lot of the imperfections are lost in printing.
 
I think it will probably be soon at generating more visually compelling procedurally generated content for games. I bet we see another tantrum when that starts happening.
I think first we'll see about a metric fuckton of low effort shovelware using this to add graphics, probably with poor and disjointed results.

Not saying it can't be used in a good way, just the kind of people who want to go the low-effort road don't lend themselves to well, effort.

Still, from voice generation to graphics generation, this is going to be a boon for video game devs. I wish I had that shit 20 years ago when I wrote my early version of stellaris/elite crossover. Nobody wants to play that stuff text-only.
 
Back