Stable Diffusion, NovelAI, Machine Learning Art - AI art generation discussion and image dump

  • 🐕 I am attempting to get the site runnning as fast as possible. If you are experiencing slow page load times, please report it.
Lenovo touting AI image generation for non-binary girls who "first started using the Internet seven years ago [when I was 12]" in order to be their Authentic Selves. Including creepy uncanny-valley avatar with Annoying Orange style lips.


View attachment 6144292

Why Lenovo, a business-focused vendor, thought this was a good ad for them I don't know. 3.1M views but only 84 Likes. This is a text book example of why YouTube took away downvotes.

Well, scratch Lenovo off the list of any future purchases! I hope the kid isn't too scarred by this in the future.

Both of these are really sad. A conversation is an exchange of emotions and ideas. When you remove yourself from that conversation and use a proxy to communicate, you are robbing the other person of the ability to affect you, to change your mind. All while expecting them to bow to your ideology or emotions.

A concerned parent has more righteous conviction than a confused teen who thinks the discord mods are right. Yet an AI cares not for conviction.
 
Seems like almost every online AI tool or generation site has "sign in with Google" (never works), is nonfunctional, has limited free use, or some combination of such BS.
hotpot.ai has a low max res but allows virtually unlimited free use (50 per day but simply clearing your cache resets it).

If you're using it for waifus or otherwise censored content, use the trick of typing the censored words in another language like French.
 
I'm having problems running automatic1111 webui. I use to be able to run this thing but now I'm getting this error:

File "/home/name/Documents/automatic/stable-diffusion-webui/modules/launch_utils.py", line 386, in prepare_environment
raise RuntimeError(

RuntimeError: Torch is not able to use GPU; add --skip-torch-cuda-test to COMMANDLINE_ARGS variable to disable this check

I've installed the latest torch version, my gpu has the latest cuda version etc....

Do you use comfyUI or automatic 1111?

and whats this about a config file? What is it do you need when training stable diffusion or any AI like this for say silly tavern? I got images that should have captions in .txt files, stable diffusion 1.5 pruned safetensorfile. What else do you need?
Still using AUTOMATIC1111; I'm on version v1.9.3. Did you do a clean installation?
 
So what's the deal with those new AI accelerator chips? are those good for all AI or just local LLM? I saw one dude running LLaMA on an Rpi5 with a coral TPU which is just 4 TOPS and yet it ran decent enough.

BTW are there any benchmark programs to see how many TOPS your GPU can get?
What's hilarious is how SD ate shit over the Taylor Swift deepfakes thing, when most of the images were generated using Microsoft's tools by abusing their shitty filtering
I don't recall anyone mentioning SD instead of copilot. I do remember the MSFT CEO had to go out to apologize and promise they would fix it.
And also apparently you can use an AI Queen Latifah
That fatso hasn't been relevant in like 20 years, why hire her?
Looks like they did another with a fat japaneese woman:
I thought japanese waters were whale-free...
 
So what's the deal with those new AI accelerator chips? are those good for all AI or just local LLM? I saw one dude running LLaMA on an Rpi5 with a coral TPU which is just 4 TOPS and yet it ran decent enough.
LLMs are bound by memory speed. TPUs generally use system memory which is slower than GPU memory modules. They are not good for LLMs of high quality. In addition they can supposedly only do 8bit ops. That makes them good for some inference, but not for training which generally uses fp16 or fp32.
 
  • Informative
Reactions: cybertoaster
Both of these are really sad. A conversation is an exchange of emotions and ideas. When you remove yourself from that conversation and use a proxy to communicate, you are robbing the other person of the ability to affect you, to change your mind. All while expecting them to bow to your ideology or emotions.
In both cases it's someone who feels they have a "real" self inside that is truly them. Fantasy is an important part of growing up, maybe an idealised self-image can help that. But I think AI tools that assist on building that alternate life into some kind of facsimile of reality and perpetuating it, encouraging someone to try and live out that life and relate to their family members via it rather than their actual self is unhealthy. The crux of it is that "actual self". The girl sees her "Spider" identity as her actual self and the real biological reality as not her true identity. Very transhumanist, transcend the flesh and all that. But it can be a trap, same way some dudes live out their fantasy female self and eventually troon out.

Taking this this far indicates an unhappiness or disconnect with who you actually are. And the focus should be on becoming happy with yourself. I think asking your family members to relate to you via the fantasy version of yourself is a bit much. But... she's 19. You're still sorting stuff out at 19, well, some people are. I have more of a problem with Lenovo exploiting it for an ad opportunity and preserving it online for eternity. Though I suppose I did also share it here to illustrate that I didn't like what Lenovo were doing. So in absolute terms I guess I'm guilty of popularising it too. Maybe I should take down the post copy of it.

So what's the deal with those new AI accelerator chips? are those good for all AI or just local LLM? I saw one dude running LLaMA on an Rpi5 with a coral TPU which is just 4 TOPS and yet it ran decent enough.
Like the NPUs built into new processors? They're adequate for small models - they use system RAM and they're really for mobile devices that lack discrete GPU so that the OS can still do light AI tasks. Microsoft have set 40TOPs as a minimum and good enough for minor enhancements like handling visual filters on video calls, doing eye positioning adjustments, some natural language processing to work with Recall (when they re-add it). But not really up to the challenge of serious image generation. They also only accommodate low-bit maths, so 8-bit, I believe. Whereas many models use higher such as 16 or 32. So the models need to be heavily quantised down to 8bit. So the NPUs are pretty cool and look like they'll be pretty good for their intended purpose. But not so much for running LLMs or Image Generation locally. At least for now. Obviously all this refers to inference. Actually training models on one would be right out.

I'm quite surprised at hearing LLama working well on a Rasberry Pi. Really? Link at all? I'm presuming this is something heavily cut down.
BTW are there any benchmark programs to see how many TOPS your GPU can get?
They're coming. I think GeekBench have added it
That fatso hasn't been relevant in like 20 years, why hire her?
Hey now. Is there any reason to be mean to her? She was great as Mama Morton in Chicago.
 
I'm having problems running automatic1111 webui. I use to be able to run this thing but now I'm getting this error:

File "/home/name/Documents/automatic/stable-diffusion-webui/modules/launch_utils.py", line 386, in prepare_environment
raise RuntimeError(

RuntimeError: Torch is not able to use GPU; add --skip-torch-cuda-test to COMMANDLINE_ARGS variable to disable this check

I've installed the latest torch version, my gpu has the latest cuda version etc....

Do you use comfyUI or automatic 1111?

and whats this about a config file? What is it do you need when training stable diffusion or any AI like this for say silly tavern? I got images that should have captions in .txt files, stable diffusion 1.5 pruned safetensorfile. What else do you need?
Step 1 for debugging is to do a "git pull" or download the latest version of the UI, and then delete your venv and rebuild it.

Part 4 here explains the basic settings for training, like steps, epochs, types of LoRA, etc: https://civitai.com/articles/3522/valstrixs-crash-course-guide-to-lora-and-lycoris-training
 
  • Agree
Reactions: Baraadmirer
Smells more like it got viewbotted. It has 55 dislikes as of right now.
If it's an autoplay video on another website or used as an advertisement, that's the reason why it has a bunch of views and barely any interaction. Nobody would viewbot a video and then not pay for engagement along with it. Dislike viewer extensions aren't accurate and only reflect the dislikes of people who use the extension, there's no way of accessing the like/dislike ratio anymore unless you're the uploader.
 
Like the NPUs built into new processors?
Nah I mean these tiny boards you can put on an M.2 socket or a mPCIe in other versions:
1720025785744.png
Seems like a good way to add AI acceleration to "old" (as in right now as most CPUs don't come with NPUs yet) PCs, specially laptops, so you don't have to trash it since we haven't had socketed laptop CPUs since the 4th gen i7 IIRC.
ome natural language processing to work with Recall (when they re-add it).
Recall is a neat feature I just don't trust msft with that, and there are open source alternatives like Openrecall but I still need the horsepower to run those, hence the accelerators.
They also only accommodate low-bit maths, so 8-bit, I believe. Whereas many models use higher such as 16 or 32.
Well that sucks, and I checked and none of these boards have built-in memory, no idea why since there's space for a RAM chip there.
But not so much for running LLMs
Which is what I wanted, oh well...
I'm quite surprised at hearing LLama working well on a Rasberry Pi. Really? Link at all? I'm presuming this is something heavily cut down.

Rpi is selling a kit too, comes with a hailo accelerator tho which I heard the documentation is limited and I don't like companies founded by ex-spooks, specially foreign ones, but on paper its over 3x better than the coral TPU which for some reason google has not updated in years.

Still most of the demos are CV stuff which quite frankly is not that impressive anymore. Guess the only way to have a mostly usable local LLM on the go is a gaming laptop. Sucks that Framework went with radeon instead of geforce for their graphics module tho.
Hey now. Is there any reason to be mean to her?
Yeah I overdid it, my bad.
 
Nah I mean these tiny boards you can put on an M.2 socket or a mPCIe in other versions:
1720025785744.png
Oh well in that case take everything I said with a pinch of salt. I've no experience with them though I imagine much of what I said holds true. I'd be quite careful around driver support and making sure it will work with your choice of OS. But great if it does.

I just looked up some here:

The fact that the requirements say "Windows 10" and don't mention 11, or that it specifies Python 3.6 - 3.9 rather than current Python versions just further make me feel I would be double-checking support before purchase. The installation instructions tell you install their specific python3-pycoral library. I don't think this is something that you could just plug in and forget. It sounds like you might need to be specifically programming for it. I'm not at all saying this is a bad product. I am saying this is a "I'm a programmer experimenting with AI" product rather than a "I downloaded A1111 and am following the How To page" product.

Seems like a good way to add AI acceleration to "old" (as in right now as most CPUs don't come with NPUs yet) PCs,
Maybe. I'd want to know MS explicitly supported cards like this for that purpose though because unless they do, things like the Windows 11 AI features wont use it whether it can handle it or not.

Well that sucks, and I checked and none of these boards have built-in memory, no idea why since there's space for a RAM chip there.
I would imagine they're designed to use system memory, same as an onboard NPU is.

Sorry if all this is bad news, btw. But if it helps, I've used Runpod.io for some of my playing around and for a fraction of the price of actual hardware I can play around with some pretty high end stuff for hours at a time. Sure, I don't own it but if I can pay £10 and be playing around with some 48GB VRAM card for a week or so, it would take me months and months of that just for the price of a medium discrete GPU. Might be something to consider and I certainly would recommend if you're dipping your toe in the water to consider trying that. I mean, it's far better to spend £10-20 and see if you're actually interested in this stuff or if you get bored after a month before sinking hundreds into your own hardware.

Recall is a neat feature I just don't trust msft with that, and there are open source alternatives like Openrecall but I still need the horsepower to run those, hence the accelerators.

I think the hysteria is a bit too much on that, honestly. The first logical question is how much of a vulnerability is it. It certainly does create a "pot of gold" for anyone who compromises your system but it's not like there's no gold on the system without it. So you have to evaluate it on the delta between the two. Frankly, I'm not convinced for most people it's that much of a difference. Remember, someone has to be able to access the system under your existing user account permissions to compromise it and if they can do that they can likely get your banking information etc. by other means. And it already by design excludes capturing private windows and passwords, I believe. The second question after that would be if MS are doing it in a particularly insecure way. It sounds bad when you say they have a "SQLite database" running it, but that sort of searchability and single-user scaling it what SQLite is designed for and good at. As I said, it's all locked under someone's user credentials and device is encrypted by default. So again, it's a question of the delta - if they can compromise your user credentials is it that much worse than if they can do that without Recall? And when both of those are answered, the question is balancing that against the offered utility. And honestly, combine the ability to search an OCR'd database of your use with natural language queries handled by the AI side - "Show me a restaurant I told Sarah about sometime in the last couple of months", Hell yeah I can see that being useful to a lot of people.

Windows also has very good phone integration these days so if you choose to, it could handle that with any other messaging apps you use, all in one place. The security concerns are valid. But I don't think the hysterical reaction was.

Rpi is selling a kit too, comes with a hailo accelerator tho which I heard the documentation is limited and I don't like companies founded by ex-spooks, specially foreign ones, but on paper its over 3x better than the coral TPU which for some reason google has not updated in years.

Still most of the demos are CV stuff which quite frankly is not that impressive anymore. Guess the only way to have a mostly usable local LLM on the go is a gaming laptop...

Neat - watching it now. Maybe I'm wrong about the capabilities of this stuff. I'll watch and correct myself if I'm out of touch. Cheers!
 
  • Like
Reactions: cybertoaster
I've used Runpod.io for some of my playing around and for a fraction of the price of actual hardware I can play around with some pretty high end stuff for hours at a time. Sure, I don't own it but if I can pay £10 and be playing around with some 48GB VRAM card for a week or so, it would take me months and months of that just for the price of a medium discrete GPU. Might be something to consider and I certainly would recommend if you're dipping your toe in the water to consider trying that. I mean, it's far better to spend £10-20 and see if you're actually interested in this stuff or if you get bored after a month before sinking hundreds into your own hardware.
Do these services only charge you while your AI is running? like if I've an LLM in there its only charging me when I actually use it? I'm asking because these cloud companies can be a real POS with pricing, recall one that offered easy wordpress hosting a friend used once, they wanted to charge him like $15,000 because he forgot to cancel the service even tho he had never had anything running there at all.
 
Do these services only charge you while your AI is running? like if I've an LLM in there its only charging me when I actually use it? I'm asking because these cloud companies can be a real POS with pricing, recall one that offered easy wordpress hosting a friend used once, they wanted to charge him like $15,000 because he forgot to cancel the service even tho he had never had anything running there at all.
So I have used Runpod.io so will speak specifically about that. You front-load your balance with a set amount paying at the time. It ticks down as you incur charges and you can see your current balance. So it's not like AWS where you give them your credit card and they tell you later that you incurred way more charges than you thought because you left some fucking service running or loaded it up more than you thought and bang there goes your money. If you only want to spend £10 with runpod, then only put £10 on your account. This was the immediate and most appealing thing about them and it's been good.

For only being charged whilst you're running the AI - almost. That part is true but they will also charge you (at a much lower rate) for persistent storage. Which is optional, but you may not want to start every session by downloading the same 7GB model to your instance. So you might pay for 20GB of storage and that will continue to trickle down your balance over time for as long as you have it, using it or not. But the larger cost is actually running things and that applies only the time you are using it. Nor by processor cycles to be clear - if you fire up an instance with a 4090 and don't do anything with it, you're still paying for it for as long as you have it reserved. But why would you - just shut it down when done. It's quick to fire something up and attach your storage to it and vice versa.

I'm about half way through your Rp5 video, btw. I'm surprised how effective it is. I've noticed the uptime value in the top-right jumping not in time with the video so he's editing the response time for his viewers, but still, surprised it's this capable.
 
For only being charged whilst you're running the AI - almost. That part is true but they will also charge you (at a much lower rate) for persistent storage. Which is optional, but you may not want to start every session by downloading the same 7GB model to your instance. So you might pay for 20GB of storage and that will continue to trickle down your balance over time for as long as you have it, using it or not. But the larger cost is actually running things and that applies only the time you are using it. Nor by processor cycles to be clear - if you fire up an instance with a 4090 and don't do anything with it, you're still paying for it for as long as you have it reserved. But why would you - just shut it down when done. It's quick to fire something up and attach your storage to it and vice versa.
The right play is to spin up an ubuntu VM with the lowest GPU available, and then wget all of your models to persistent storage. Then stop it, and start up your intended GPU VM when you have enough downloaded to start. If you are running anything that needs splicing, splice it locally then SCP/SSH it to the server and stash it in persistent storage. They also have CPU VMs, but those are hard to procure in the right regions.
 
@cybertoaster Interesting video. A testament to how good Rasbery Pi's are. That said, the editing on the video hides how long some of those earlier responses take. I saw the uptime value in the top right jump by nearly eight minutes on one of them. And it looks like I was right about the AI accelerator. Though I couldn't tell from the video if he was actually using it - I'm not wholly convinced he actually set it up to use it.

It's impressive. Makes me think I should get a Rasberry Pi 5. Not for AI though, for other things.
The right play is to spin up a VM with the lowest GPU available, and then wget all of your models to persistent storage. If you are running anything that needs splicing, splice it locally then SCP/SSH it to the server and stash it in persistent storage. They also have CPU VMs, but those are hard to procure in the right regions.
That... is very solid advice. 👍
 
If you only want to spend £10 with runpod, then only put £10 on your account. This was the immediate and most appealing thing about them and it's been good.
Good.
So you might pay for 20GB of storage and that will continue to trickle down your balance over time for as long as you have it, using it or not
I take they don't let you use offsite storage.
I'm about half way through your Rp5 video, btw. I'm surprised how effective it is. I've noticed the uptime value in the top-right jumping not in time with the video so he's editing the response time for his viewers, but still, surprised it's this capable.
Interesting video. A testament to how good Rasbery Pi's are. That said, the editing on the video hides how long some of those earlier responses take. I saw the uptime value in the top right jump by nearly eight minutes on one of them. And it looks like I was right about the AI accelerator. Though I couldn't tell from the video if he was actually using it - I'm not wholly convinced he actually set it up to use it.
Yeah its speed up, he seems to be using the coral which is nearly obsolete and the USB version too so you have to add the delays from the interface. That company hailo is going to offer a new M.2 module with 40TOPS, the magic number msft wants. How well it will do on the Rpi I've no idea, would be interesting to see what these accelerator chips could make if combined in series with some fast RAM on a PCIe card. They seem to be way cheaper and more power efficient than GPUs for AI, tho it makes you wonder why there aren't more companies churning these out given how much money nvidia is making.
The right play is to spin up an ubuntu VM with the lowest GPU available, and then wget all of your models to persistent storage. Then stop it, and start up your intended GPU VM when you have enough downloaded to start. If you are running anything that needs splicing, splice it locally then SCP/SSH it to the server and stash it in persistent storage. They also have CPU VMs, but those are hard to procure in the right regions.
How expensive is to train a model on these services? because that's the kind of stuff you need to spend a lot on hardware to get it done in non-glacial time. I remember some guy on reddit said he trained his LLaMA model on runpod but he didn't specify which llama version, size, what the model was or more importantly in this case the cost, just that it took 3 days but given all the tiers its meaningless unless he used the highest one in which case it was (I guess) $352.

You can buy a good used 3080 with 12GB of VRAM for that, not the same than that MI300X but you own it.
Not for AI though, for other things.
With rpis you have to be more specific, I seen any number of things made with those, tho its mostly retrogaming tbh.
 
I take they don't let you use offsite storage.
They don't block it (that I know of, but you have an SSH tunnel to the instance so I don't think they could anyway) but would it be useful? Could you have some local drive and open up a connection and do some sort of network mount from the instance in Runpod of your local file system? Sure. But when the GPU needs to load the model you have stored locally into its VRAM, that's going to be a looooong process. You need storage local to the GPU. And if you're talking just loading it up to the included instance storage you get when you fire the instance up, that returns us to my initial point - do you want to be doing this every time you want to play around with this stuff? Storage is a cost but it's a low cost compared to actually running instances. Either way, that's something you have to accept if you aren't doing things locally. Everything has upsides and downsides.

Yeah its speed up, he seems to be using the coral
I'm not sure he is. He shows it at one point later on but I don't think he was using it. Also, my admittedly very cursory look suggested that you needed to explicitly use their Python package for it and I don't know that he tweaked anything to do so.

which is nearly obsolete and the USB version too so you have to add the delays from the interface. That company hailo is going to offer a new M.2 module with 40TOPS, the magic number msft wants. How well it will do on the Rpi I've no idea, would be interesting to see what these accelerator chips could make if combined in series with some fast RAM on a PCIe card. They seem to be way cheaper and more power efficient than GPUs for AI, tho it makes you wonder why there aren't more companies churning these out given how much money nvidia is making.
If they're cheaper than GPUs it'll be because they don't have a load of high-speed memory next to the chip that a discrete GPU card has. And other factors.

And again, I sound a note of caution about support for these and how automatic use of them is. I'd double-check anything you might want to use them with and see if they'll be used by the software you plan to run or if you need to start fiddling with the code to make it use them. Again, I have no experience with those cards it's just the sort of thing I wouldn't take for granted. Sorry if this is pouring a little cold water on an idea - might be fine. Just, be sure it's what you think before you buy.

How expensive is to train a model on these services? because that's the kind of stuff you need to spend a lot on hardware to get it done in non-glacial time.
I'm afraid training models is outside my experience. Though I can say it's something of a "how long is a piece of string" question. To answer a question like that, the question needs to contain specifics, really. I can say that if you can find some stats from someone they're going to translate 1:1 to something like runpod. Or at least should. If someone says "This model and this amount of data took me X hours on my MI300X" then X hours should be the same on runpod if you get an MI300X.

(And I'm in runpod right now so for fun, an MI300X with 192GB VRAM and 24vCPU is available for $4.89/hr presently. I don't know what the retail price is on one of those things (probably doesn't have a fixed one - all part of some enterprise deals), but lets say it's US$10,000. You'd have to use it continuously 24/7 for about three months before it became more economical to just buy your own. And that's before electricity costs. That's why I say if you're just learning this stuff, using something like runpod is a much cheaper way to dip your toe in the water and learn what you actually need or if you're going to stick with it).

You can buy a good used 3080 with 12GB of VRAM for that, not the same than that MI300X but you own it.
Yeah, but they're not remotely comparable in terms of power. And if you're doing something where they were comparable then you should just be using the 3080 in the first place. I can't see a 3080 available on there right now the closest they have is a 3090 for $0.44/hr. So three days of that would be $32.
 
How expensive is to train a model on these services? because that's the kind of stuff you need to spend a lot on hardware to get it done in non-glacial time. I remember some guy on reddit said he trained his LLaMA model on runpod but he didn't specify which llama version, size, what the model was or more importantly in this case the cost, just that it took 3 days but given all the tiers its meaningless unless he used the highest one in which case it was (I guess) $352.
The memory required scales with the size of the model. Supposedly you used to be able to train llama2 13B with 24 VRAM with some hacks. There's no way to do that now with llama3 as the context is twice as large even if it is a 8B now. That's not even getting into Illama2 70B which supposedly takes several hundred to a thousand GB of VRAM. It's unfathomable how much memory you would need for Illama 3 70B.
 
But when the GPU needs to load the model you have stored locally into its VRAM, that's going to be a looooong process. You need storage local to the GPU. And if you're talking just loading it up to the included instance storage you get when you fire the instance up, that returns us to my initial point - do you want to be doing this every time you want to play around with this stuff?
Good point.
I'm not sure he is. He shows it at one point later on but I don't think he was using it. Also, my admittedly very cursory look suggested that you needed to explicitly use their Python package for it and I don't know that he tweaked anything to do so.
On a side note here's another guy running LLMs on a rockchip which does have an NPU:

It runs......okay, I seen copilot running at similar speeds so its tolerable. The board has 16GB of RAM but its available with 32GB. Not as cheap as an rpi tho, easily twice the price.
If they're cheaper than GPUs it'll be because they don't have a load of high-speed memory next to the chip that a discrete GPU card has. And other factors.
The chips look way smaller too, no way those are as expensive to manufacture.
Yeah, but they're not remotely comparable in terms of power. And if you're doing something where they were comparable then you should just be using the 3080 in the first place. I can't see a 3080 available on there right now the closest they have is a 3090 for $0.44/hr. So three days of that would be $32.
I seen people running SDXL on a 3080 just fine, no idea about training SD models tho. The main limiting factor is that gaming cards rarely go over 12GB of VRAM with the new $1000 4080super having only 16GB. I'm surprised nobody has been modifying these cards with bigger RAM chips like they did with some modded radeons I seen for sale.
That's not even getting into Illama2 70B which supposedly takes several hundred to a thousand GB of VRAM. It's unfathomable how much memory you would need for Illama 3 70B.
Fuuuuuuuck, so the trillion dollar datacenter wasn't a joke. Are LLMs about to hit a ceiling? how can this keep going?
 
Back