Stable Diffusion, NovelAI, Machine Learning Art - AI art generation discussion and image dump

  • 🐕 I am attempting to get the site runnning as fast as possible. If you are experiencing slow page load times, please report it.
3090s have nvlink and 2 together gets you 48GB of VRAM for your LLM. Not possible with 4090 or 5090 or anything consumer going forward.
You're short 8GB with a 24 VRAM to run Illama 60B with 4Q guff. That's really the best the enthusiast consumer should be shooting for. Upping it to 12GB or 16GB on the second card really should be enough to maybe get to 5Q or an expanded context. Fuck running two 3 slot cards running 350w inside a single case. You need a PSU upgrade for that as well.
 
That's what you're using it for?
...not exclusively.
I'm a permaDM, it's kind of nice to just be a player for once, even if I do have to keep track of all the rules and plot anyway because the DM is an amazing storyteller but just super scatterbrained.
 
  • Like
Reactions: Baraadmirer
You're short 8GB with a 24 VRAM to run Illama 60B with 4Q guff. That's really the best the enthusiast consumer should be shooting for. Upping it to 12GB or 16GB on the second card really should be enough to maybe get to 5Q or an expanded context. Fuck running two 3 slot cards running 350w inside a single case. You need a PSU upgrade for that as well.


Worse then that you probably want an enterprise board with large amounts of ram then cooling, etc. at that point might as well keep adding 3090s what you are after is a decommissioned coin miner/cracking/hashing open air frame rig. 1 hour per 3090 to repad them too since temps will be bad and used market 3090s were peak crypto shortage and beaten like rented mules. But, people are doing it for a reason. The good news is you know in advance Nvidia will never cut you a break on consumer AI.

 
  • Thunk-Provoking
Reactions: Vecr
Worse then that you probably want an enterprise board with large amounts of ram then cooling, etc. at that point might as well keep adding 3090s what you are after is a decommissioned coin miner/cracking/hashing open air frame rig. 1 hour per 3090 to repad them too since temps will be bad and used market 3090s were peak crypto shortage and beaten like rented mules. But, people are doing it for a reason. The good news is you know in advance Nvidia will never cut you a break on consumer AI.

You are really going to see diminishing returns past 70B so slamming a third 3090 in there is kind of pointless. The 4th is just retarded.
 
Also looks like Pony v7 is being baked, but it's gonna be a while.

Screenshot_20241112-191204.jpg
 
Worse then that you probably want an enterprise board with large amounts of ram then cooling, etc. at that point might as well keep adding 3090s what you are after is a decommissioned coin miner/cracking/hashing open air frame rig. 1 hour per 3090 to repad them too since temps will be bad and used market 3090s were peak crypto shortage and beaten like rented mules. But, people are doing it for a reason. The good news is you know in advance Nvidia will never cut you a break on consumer AI.

This is basically what I did, except with older server GPUs. Many (older) Supermicro motherboards, even some of the non-standard form factor ones, will still use standard ATX PSU sockets. From there, you can either adapt an existing chassis to fit one of these boards, or simply build your own.

As far as mining rigs go: Many of these only have 1x PCIe bandwidth, which will be a problem as AI is bandwidth intensive if you are using multiple GPUs. So if you get a mining rig, you will need to swap out both the motherboard and most likely all the risers too if you want to use the bandwidth that your cards offer.
 
This is basically what I did, except with older server GPUs. Many (older) Supermicro motherboards, even some of the non-standard form factor ones, will still use standard ATX PSU sockets. From there, you can either adapt an existing chassis to fit one of these boards, or simply build your own.

As far as mining rigs go: Many of these only have 1x PCIe bandwidth, which will be a problem as AI is bandwidth intensive if you are using multiple GPUs. So if you get a mining rig, you will need to swap out both the motherboard and most likely all the risers too if you want to use the bandwidth that your cards offer.
The trick is to use a standard atx motherboard with a shitload of pciex16 slots and mount it in a 6U case with a split bracket on top for the GPUs. Then just run ribbons up and/or split the 16x into 8s. Only issue is that for it to work you are gonna need two broadwell CPUs which only runs pciev3 or a threadripper.
 
The trick is to use a standard motherboard with a shitload of pciex16 slots and mount it in a 6U with a split bracket on top for the GPUs. Only issue is that for it to work you are gonna need two broadwell CPUs which only runs pciev4 or a threadripper.
Yep, that works. You can even use a separate PSU for all the GPUs if you get a PCIe slot power isolator like this one here: https://benchlab.io/products/elmorlabs-pmd-pci-e-slot-power-measurement-adapter
 
I asked Flux to generate me Crazy Frog flying a plane into the World Trade Center and it has failed me. I don't think this AI thing will be panning out.
 
I would highly recommend ZoinksNoob for image generation if you still have a taste for Stable Diffusion XL. It is compatible with Pony/SDXL Loras. It's one of the best models I've worked with, and the artist selection is amazing.
 
Saw the 5xxx leaks, nvidia its still being stingy with the VRAM, guessing its so their enterprise customers wont go and buy cheaper consumer cards.
The 4070 is 3 times faster than the 3090 for AI
Where you getting those benchmarks from? just asking.
Worse then that you probably want an enterprise board with large amounts of ram then cooling, etc. at that point might as well keep adding 3090s what you are after is a decommissioned coin miner/cracking/hashing open air frame rig. 1 hour per 3090 to repad them too since temps will be bad and used market 3090s were peak crypto shortage and beaten like rented mules. But, people are doing it for a reason. The good news is you know in advance Nvidia will never cut you a break on consumer AI.

Funny how almost 5k sounds like a lot now but back in the 90's it only got you 2 average PCs. The problem with ex-mining cards besides the wear is that most were very poorly maintained, at least that's my experience after looking at tons of cryptotards who went all in during the last craze, they knew jack shit about basic PC hardware and just bought the rigs ready-made and plugged them, didn't even bother to dust the things once in a while so the fans were caked. Another problem is that the vast majority used radeon cards so its not great for AI, most of the rigs that used geforce have long been taken apart and sold piecemeal to gamers.
 
The problem with ex-mining cards besides the wear is that most were very poorly maintained, at least that's my experience after looking at tons of cryptotards who went all in during the last craze, they knew jack shit about basic PC hardware and just bought the rigs ready-made and plugged them, didn't even bother to dust the things once in a while so the fans were caked.
I knew a tard who burned his house down by doing both this and shoving way too many rigs into a small space.
 
OpenAI is at war with its own Sora video testers following brief public leak (archive)
OpenAI has cut off testing access to its Sora video generation platform after a group of artists briefly shared their own early access in a publicly usable webpage Tuesday. The group, going by the moniker PR Puppets, claims the stunt was a protest against being asked to work as unpaid R&D and bug testers while participating in "art washing" of AI tools. But OpenAI says participation in the early alpha test is voluntary and has no requirements that testers provide feedback.

Saw the 5xxx leaks, nvidia its still being stingy with the VRAM, guessing its so their enterprise customers wont go and buy cheaper consumer cards.
There's a rumor that after releasing a 5080 16 GB, they will make a 24 GB version sometime after 3 GB GDDR7 modules become available. Initial cards come with 2 GB GDDR7 modules, 3 GB GDDR7 modules increase capacity without changing the bus width. We could end up seeing weird cards later in the generation like a 5070 refresh with 18 GB, or some 96-bit card with 9 GB instead of 6 GB.

Meanwhile, the 5090 should have 32 GB, making it the prosoomer AI king. Just stop being poor. The more you buy, the more you save.
 
There's a rumor that after releasing a 5080 16 GB, they will make a 24 GB version
I remember hearing the same about the 4070 Super having 16GB and it only got 12 like the regular 4070 and the Ti.

I'll believe it when I see it.
Meanwhile, the 5090 should have 32 GB
With the 5080 practically confirmed to be just 16GB I seriously doubt this is going to be 32, more likely 24.
Just stop being poor.
I'm going to retire in my early 40's precisely because I don't do dumb shit like burning money on rapidly depreciating assets like this piece of future e-waste.
 
I'll believe it when I see it.
What I'm talking about would be like a refresh which could come out much later. The 5080 will definitely start out with 16 GB. Nvidia has the option to boost any card's VRAM capacity by 50% using 3 GB GDDR7 modules if they so choose. Maybe they would call it a 5080 Ti/Super.

First generation of GDDR7 graphics cards sticking to 16Gbit (2GB) modules, 3GB on roadmaps
With the 5080 practically confirmed to be just 16GB I seriously doubt this is going to be 32, more likely 24.
It's practically confirmed that the 5090 will have 32 GB. It's such old news that people are talking about the die size (744mm^2) and 5090D now.

When you look at other leaked specs, the 5090 appears to be double the card of the 5080 in most ways, e.g. 21,760 vs. 10,752 CUDA cores. I thought China sanctions explained the huge chasm between the two cards, but the existence of a China-exclusive 5090D would make that unlikely. It's just Nvidia's strategy: make a desirable Titan-like flagship with a huge die near the reticle limit, and then have the 80-class card proportionately cheaper so price/performance doesn't change much.
rapidly depreciating assets like this piece of future e-waste.
It's a tool, and the only people who ought to buy it are those professionals/"prosumers" who will use it to make money (accounting for the price and the power bill). Definitely don't buy it if you don't have a better reason than playing Cyberpunk or generating anime tiddies. The apparent ~600W TDP alone will cause problems, and it will be comically large.
 
Back