Stable Diffusion, NovelAI, Machine Learning Art - AI art generation discussion and image dump

  • 🐕 I am attempting to get the site runnning as fast as possible. If you are experiencing slow page load times, please report it.
To expand on this, I recommend avoiding SDXL models if you have an 8 GB VRAM card or not more than 16 GB RAM; you can get away with it, but it's going to be a lot slower than using a non-XL model and upscaling 3-3.5x, plus your computer won't be good for much else during the process.
I find that the average operating system with a desktop environment strips off around 2GB of VRAM for web browsers and such. That leaves most people with 6GB. Now if you want to cheese it you could run --medvram. The other more interesting alternative is to just boot into shell or kill xserver on a Linux distro. Then just launch SD with --listen and access it using your phone or another machine on your network.
 
I find that the average operating system with a desktop environment strips off around 2GB of VRAM for web browsers and such. That leaves most people with 6GB. Now if you want to cheese it you could run --medvram. The other more interesting alternative is to just boot into shell or kill xserver on a Linux distro. Then just launch SD with --listen and access it using your phone or another machine on your network.
I hope they fixed the glitch where you had to have a X server running for Cuda to work. But a bare X server uses barely any VRAM so it wasn't a big issue.
"sudo X" possibly with a sparse config file if you also have no display attached.
But a lightweight desktop environment doesn't use much until you get into stuff like Chrome. Now, Gnome or KDE, I have no idea, I use XFCE. And you can start Chrome with --disable-gpu as I recall to be able to use it safely.
My 8GB 3070 with XFCE is only a hair slower than my 16GB 4060 Ti. 90 seconds per image vs 85 for 50 steps at native resolution and then 20 upscaled steps with output at 1664x2434.
This is all ComfyUI, no idea how Automatic is doing these days.
 
  • Informative
Reactions: IAmNotAlpharius
Coming to a theater near you
IMG_1541.jpeg
Also fantasy shit
IMG_1547.jpeg
An exciting movie
IMG_1549.jpeg
Team effort
IMG_1557.jpeg
 
Last edited:
VAE in simple terms converts the output of the model, a Latent image, into an actual human viewable image.
Would I be right in guessing that the reason not to bake a VAE into the model is that it's normally a standard VAE between a lot of models and baking it in just needlessly increases the size of the model when you could just have all your models be smaller and add the VAE yourself as the final stage?
 
So yeah RunPod needs to get its shit together with this SCP stuff. Spent a solid fucking hour just getting it to work after learning I had to input a custom docker command.
 
"Taylor Swift AI" is trending on X. Some people on 4chan decided to make AI-generated Taylor Swift memes and they went viral. Swifties are very upset over some low-quality, absurd AI images of Taylor Swift in compromising positions with muppets and Mr. Krabs.

Out of an abundance of caution, I'm not going to upload any images to the farms. Taylor Swift can probably hire better lawyers than Russell Greer. Here are a few archived links though (NSFW)


Archive of the /pol/ thread claiming to have initiated the trend: https://archive.fo/ikggc
 
"Taylor Swift AI" is trending on X. Some people on 4chan decided to make AI-generated Taylor Swift memes and they went viral. Swifties are very upset over some low-quality, absurd AI images of Taylor Swift in compromising positions with muppets and Mr. Krabs.

Out of an abundance of caution, I'm not going to upload any images to the farms. Taylor Swift can probably hire better lawyers than Russell Greer. Here are a few archived links though (NSFW)


Archive of the /pol/ thread claiming to have initiated the trend: https://archive.fo/ikggc
Holy fuck. Can you show me how to get, how to get to sesame street? White liberal Swifties are going full 1488 over this. This shit is hilarious!
 
I'm pretty sure Taylor Swift is probably one of the top 5 people on Civitai by number of models made of them.
So, did someone say Furry Taylor Swift?
SNEED_TS_2.jpgSNEED_TS_1.jpgSNEED_TS_0.jpg

One smart thing I did with my system is make a sqlite database to store all image metadata so I can always find where I left the furries images of a specific character.
 
"Taylor Swift AI" is trending on X. Some people on 4chan decided to make AI-generated Taylor Swift memes and they went viral. Swifties are very upset over some low-quality, absurd AI images of Taylor Swift in compromising positions with muppets and Mr. Krabs.

Out of an abundance of caution, I'm not going to upload any images to the farms. Taylor Swift can probably hire better lawyers than Russell Greer. Here are a few archived links though (NSFW)


Archive of the /pol/ thread claiming to have initiated the trend: https://archive.fo/ikggc
Holy fuck. Can you show me how to get, how to get to sesame street? White liberal Swifties are going full 1488 over this. This shit is hilarious!
If teenage ”Swifties” can somehow still blame all men for creating this kind of A.I. art while still listening to a woman admit on record that she broke up with multiple boyfriends on record to make hit songs, then this is proof that irony is (almost) dead.
 
It's one higher.

Fine.

XL means it's based on and only runs with the SDXL model(s) and software(and VRAM) that can handle it.
So, the model needs to match the LORA and whatever software you use has to support it.

SDXL is generally trained on 1024x1024 images compared to the base Stable Diffusion at 512x512. So that's the optimal size to start with.

You can run it with an 8GB video card, but it's a tight fit, and in my case I have to close most of my web browser windows as they use too much GPU memory, fucking Chrome.

Few more notes, the base SDXL comes in 2 parts. The main model and a refiner model. It's optimized for those two and most software supports the 2 stages. Almost every other model you'll find doesn't use the refiner as they've been trained without it.

You will sometimes see models that say "Baked VAE" or "Needs VAE" in this case you may need to tell your software to use the VAE in the model(Baked) or a separate one. Using the wrong option causes all kinds of visual artifacts. There is a standard SDXL VAE that's used for most models where it's not "Baked". VAE in simple terms converts the output of the model, a Latent image, into an actual human viewable image.
How long does SDXL takes to generate an image on a 3080 with 12GB VRAM?
 
Back