Stable Diffusion, NovelAI, Machine Learning Art - AI art generation discussion and image dump

  • 🐕 I am attempting to get the site runnning as fast as possible. If you are experiencing slow page load times, please report it.
AFAIK sudowrite is using an old davinci finetune. If still the case, then the service is kind of a rip off. I'd look into llama 2 70b. Not only can you run it locally but it's definitively smarter in text related tasks. From my limited testing I'd even go as far to say it's better at such tasks than turbo, but mostly because OpenAI cheated to extend Turbo's context (which they haven't admitted but is easy to test) which makes it not "see" the entire context properly anymore. (look into RoPE for llama, they did something similar to Turbo, which ups perplexity)

Just like stable diffusion, with creative writing on language models, it's pretty much all in the prompt. Ask it to copy the style of an author you like. (e.g. use William Gibson and you'll see a lot of references to rain and neon lights) "Preseed" the context with text passages. If it's an instruct model, iterate yourself to the solution by describing characters, locations, and motivations until having the LM tie it all together in a final scene. There's a lot of scientific papers out there that pretty much prove that the texts output quality and reasoning gets better the more the model gets to "think" about the problem in it's context. It's all noise in the end, but the more guidance you deliver, the more accurate the "predictions" will be. LMs hallucinate by their very nature, but for creative endeavors that's a good thing.

Just like SD, the entire enthusiast field is full of coomers and other dregs of society, so it's hard to find serious discussion about this topic which doesn't revolve around sexualizing twelve year old anime girls - but it's out there. I saw some impressive model writing even from 70b I wouldn't have immediately clocked as machine generated. It does take talent steering though.

Facebook is apparently planning to kneecap the competition by releasing a GPT4 killer (llama 3 is already WIP) for free to level the playing field, and as a result we all will win. The Meta chief AI guy is an old, french white computer scientist who is outspoken anti censorship and anti AI doom, that's probably why the llama foundation models are all not censored.
 
Just like SD, the entire enthusiast field is full of coomers and other dregs of society
Do you really think that we would be seeing the amount of progress on this tech in the short amount of time it's been out, if that weren't the case?

War and Pornography are the nexuses around which technological progress revolves.
 
There was some discussion of generative AI on today's MATI, talking about RAND's report on the national security implications of generative AI, and Null rated my Stable Diffusion pizza Hilters.

For training LoRAs, does anyone have good resources or guides on how to train them more effectively? I have had decent success with character LoRAs, but the training process still feels like spinning random knobs and pushing buttons, hoping for the best.

In terms of tooling, I have mostly been training with this: https://github.com/derrian-distro/LoRA_Easy_Training_Scripts It works well on Linux, unlike Kohya.

Also, if anyone has a good idea for a KF-related LoRA and is willing to find images and tag them, I'd be willing to do the training on my GPU. No problem with SD 1.5 or SDXL.
 
For training LoRAs, does anyone have good resources or guides on how to train them more effectively? I have had decent success with character LoRAs, but the training process still feels like spinning random knobs and pushing buttons, hoping for the best.
It really is. As someone who has done a lot of Hypernetwork and Lora training. Here are my basic tips for LoRAs.
-More is better when generating a Lora for a style.
-Most LoRAs are done baking around 30 epochs.
-Multiple aspect ratios can be used for training, but they should be consistent and grouped and cropped cleanly ergo 3 pictures of 2:3 and 10 of 1:1. Instead of 13 pictures with random aspect ratios.
- Use the wd-tagger extension in automatic1111 to tag your images with anything with a certainty above .35. Then try just running it. If your results are bad prune the tags.
 
The Meta chief AI guy is an old, french white computer scientist who is outspoken anti censorship and anti AI doom, that's probably why the llama foundation models are all not censored.
The LLaMA models leaked, and I wouldn't count on Meta to look out for the little guy in the long run.

There was some discussion of generative AI on today's MATI, talking about RAND's report on the national security implications of generative AI, and Null rated my Stable Diffusion pizza Hilters.
When Congress gets bipartisan, get scared:
 
The LLaMA models leaked, and I wouldn't count on Meta to look out for the little guy in the long run.
True for Llama 1. The Llama 2 model series was officially released. https://ai.meta.com/llama/ you can download it here, even commercial usage is allowed. (I would just download quantitized versions from hugginface tho, point still stands)

I don't count on Meta to look out for the little guy, but to make a business decision. Point is right now that OpenAI has not only the AI market cornered but tries it's very best to pull up the ladder behind it. They do not have serious competition in the form of an actual every-day-usable product for a good price right now and a huge headstart. Even if Meta cooks up a GPT4 equivalent and offers it as a service, it would be hard to cut into that market, because as we all know, lots of things run on inertia and if Meta doesn't have massive incentives, businesses are just not gonna switch their provider like that, especially not away from something like OAI with it's low prices and brand recognition. While Meta has very deep pockets, there are also very, very deep pockets behind OAI.

So the next best thing you can do in this situation is to make GPT4 the default everyone with the right compute can just run by himself. With that you effectively level the playing field, dethrone OAI, and put everyone back at the start of the race. I do think this is Metas goal with Llama 3. Of course, some corps still will remain "sticky" and stick to OAI as provider of their choice, ("because we've always done it like this" - who hasn't worked in such a place?) but I think there's a lot of organizations that would be interested in running an inhouse model, (potentially with the right contractual support) where they can adjust every lever and knob themselves and aren't beholden to send potentially very sensitive data to a third party to process, which might in some cases not even be a legal thing to do. Then, when Meta can close the gap, it'll have a foot in the door. Of course it'd be a gamble, but well, the headstart of OAI is there.

Also releasing these models has already been proven to be free R&D. I'm sure some things the open source crowd worked out will find their way back to Meta, as was probably intended.

Might also all be bullshit and Meta might try to outcompete OAI directly in the market with a slightly worse GPT4 and fail miserably. Point still stands. Also I just really like Yann LeCun. He seems reasonable. Not a very common thing in this sphere.
 
Last edited:
So the next best thing you can do in this situation is to make GPT4 the default everyone with the right compute can just run by himself. With that you effectively level the playing field, dethrone OAI, and put everyone back at the start of the race. I do think this is Metas goal with Llama 3.
The way I see it is that any company trying to succeed in this space needs two things:

1. Models, architectures, algorithms, etc. (OpenAI has a decent lead here, but it's small enough that they are worried)
2. Data (Meta, Microsoft, Google, etc have a big lead here)

So from Meta's perspective, the more they can chip away at OpenAI's lead in models, the more they can leverage their insane amount of consumer data and try to monetize it. Once they start getting close to GPT-4 performance, they will likely start doing more closed-source work, but in the meantime they will use the open source community as a source of free labor and ideas until they have caught up.
 
StableBeluga2 70B is nice, and is really nearing ChatGPT in quality. I would recommend it if you have the machine to run it.

As for commercial usage, I think that LLMs are going to be fine tuned to leverage preexisting workflows already used by companies. Essentially you will ask it to help you buy a plane ticket to New York or ask it for a nice beach hotel to spend a weekend at, and it will hit a preexisting system for ticket prices and give it you. Then you can talk it through a purchase workflow and you are good. It will send you an email with your ticket.
 
Is it me or the free version of ChatGPT is incredible cucked? I asked it to give me some financial advice on stocks and refused, I asked it to give me list of serial killers and their manifestos and refused, I told him to explain the word nigger and refused, I asked it to make a poem about transgenders and refused.

Extremely cucked. I didn't remember it was this cucked before.
 

Here's a list of the currently existing open models with their evals, most of them are llama finetunes but there was also Falcon 180b released recently as open source model, made by some company in the emirates. Especially the top results need to be approached with a bit of caution as everyone wants to get the highscore and there might be tuning/overfitting on the eval questions.(meaning the model would do good in the eval but actually be shit) There was already a case where somebody just 1:1 copied another model and claimed it was theirs, so everything's possible. That list is also a nice insight of what all is going on in that sphere right now. Most of the top 70b llama 2 finetunes have in common that they're already beating Turbo/ChatGPT in the important evals. From subjective impression of running them I'd say that the shoe indeed fits and that they come across as more contextually aware, "smarter" if you will, although they do not have ChatGPTs knowledge.

The small models fit even on relatively modest hardware through the wonders of quantization, which is a technique that reduces the models size but costs "intelligence", the more it is reduced. The big development in this space was that you can run models partially or fully on RAM/CPU now, although it will not be fast. For 70b, I'd advise 48 GB of VRAM minimum, and yes, you can combine several cards to get there. (or get one A6000) With language models, it's all about memory bandwidth, not necessarily processing speed.
 
The small models fit even on relatively modest hardware through the wonders of quantization, which is a technique that reduces the models size but costs "intelligence", the more it is reduced. The big development in this space was that you can run models partially or fully on RAM/CPU now, although it will not be fast. For 70b, I'd advise 48 GB of VRAM minimum, and yes, you can combine several cards to get there. (or get one A6000)
Double 3090s seems to be the most popular budget setup for running 70b and similar models. The sweet spot for single card setups is ~35b and unfortunately they still have not released the 35b Llama 2 version they were talking about. Although they did release Code Llama 34b which is pretty good for code completion and summarization tasks.
 
35b Llama 2 version they were talking about
my guess is it died too often in the training, but yeah, strange.

Since I've already summarized so much I'll throw this info in too, besides the basic llama 2 models, Meta also released special "coding" models a bit later: https://ai.meta.com/blog/code-llama-large-language-model-coding/ I haven't tried them much but from all I heard results are pretty mixed even though benchmarks indicate otherwise (the 34b one apparently swings pretty close to gpt4 there, although people said in practice it is anything but)

Interestingly Code llamas are capable of massive contexts (16-100k) which I guess is useful for summarization and code.
 
Interestingly Code llamas are capable of massive contexts (16-100k) which I guess is useful for summarization and code.
Really? That's a lot.

I don't care about coding, but I do care about having a Virtual Memory. 100k tokens is a lot and if I could use that I could train it to be a lawyer, doctor, engineer or even a research scientist.

Imagine an AI who can do medical diagnostics with just symptoms, temperature and your picture. That's a multibillion dollar idea right there.
 
Really? That's a lot.

I don't care about coding, but I do care about having a Virtual Memory. 100k tokens is a lot and if I could use that I could train it to be a lawyer, doctor, engineer or even a research scientist.

Imagine an AI who can do medical diagnostics with just symptoms, temperature and your picture. That's a multibillion dollar idea right there.
Yeah, the question is - how well can it use that context? You need a lot of VRAM to host that kind of context, too. Processing it isn't free either. So far all the llamas have an almost autistic obsession with what they learned via context (you can easily correct 70b on a logical mistake by explaining something, and it'll actually correct it's answers with that knowledge. Impossible to do with Turbo as anyone who ever tried it knows) so that speaks for the usefulness of the architecture for such a context, but I'm not sure you'd want any language model to be a lawyer or a doctor at this stage.
Amidst the vast expanse of the border worlds, the air reeks of uncertainty, loneliness, and danger. Yet, for those brave enough to traverse these treacherous waters, opportunity awaits—opportunity that can only be seized by the rugged and bold. To conquer the darkness, one must have a trusty steed; a vessel as relentless as they are, ready to endure the perils of deep space exploration.

Introducing, from the makers of legendary spacecrafts, the Aethon class ships by Chronos Enterprises. Handcrafted and built to last, these sleek, sturdy vessels offer unparalleled agility and maneuverability. Equipped with state-of-the-art technology and an array of weaponry, the Aethon ships boast impenetrable defenses capable of withstanding even the harshest cosmic storms.

Built for both trade and combat, these vessels offer spacious cargo bays, perfect for hauling exotic goods across the galaxy. But don't let its size fool you; when the time comes, the Aethon ship can unleash a furious torrent of firepower on unsuspecting foes.

Join the ranks of elite privateers and experience the freedom that only Chronos Enterprises' Aethon class ships can provide. The stars await your command; embrace destiny and leave your mark upon the universe.

Remember, when you take control of an Aethon ship, you are not just securing a livelihood; you become part of a legacy—one built upon honor, courage, and victory. Welcome aboard, commander, and prepare to conquer the unknown.

It is important to note that while Chronos Enterprises takes pride in the engineering and performance of their Aethon Class Ships, customers should be aware that deep space travel presents inherent risks and dangers. Chronos Enterprises shall not be held liable for any loss, damage, injury, or death resulting from misuse or unforeseen circumstances beyond their control. Purchasing an Aethon Class Ship does not guarantee success in either trading or combat endeavors, nor does it guarantee safety during interstellar voyages. Proper training and caution are required when operating such vehicles, and pilots should adhere to all local laws and regulations when engaging in trade or conflict within respective territories. By acquiring an Aethon Class Ship, the owner acknowledges these risks and agrees to accept full responsibility for their actions and outcomes while utilizing the product. Remember, in the vastness of space, the consequences of your choices lie solely in your hands. Chronos Enterprises wishes you the best of luck in your cosmic ventures, but ultimately, your fate is yours to write.

(written by llama 2 upstage instruct-v2 70b)
 
And apparently, it will also be open-sourced
No, it will not be. Irregardless of Meta's deliberately misleading and homosexual definition of Open source, llama 2 is not open source, and I have no doubt Illama 3 will not be. Currently the only viable open source AI model is Falcon licensed under Apache 2, and that's only the 40B and lower versions of the model.
 
Last edited:
Until they find a real use for the models they're going to keep dipping into open source. The actual use for these systems is the issue. 99% of paperwork people do is bullshit so doing that faster makes the problem worse not better.

The only thing they can really do better than some asshole for 30$ an hour is astroturfing and volume. Platforms are pushing back with IDing and limiting 1 on 1 communication.

I really have a hard time seeing the use case for LLMs in the near future. Any company is going to be a hard sell on "oh we just feed all your garbage quality company data to this box" but that's something you need to be able to do. You couldn't hire someone and not tell them anything.

Currently AI is only useful for low grade marketing
 
  • Dumb
Reactions: Roland TB-303
New audio AI dropped, infinite lo-fi hip hop AI hell is now very plausible:


Check "Prompt: lofi hip hop beat melodic chillhop 85 BPM" (it's a .M4A)


Trance, Ibiza, Beach, Sun, 4 AM, Progressive, Synthesizer, 909, Dramatic Chords, Choir, Euphoric, Nostalgic, Dynamic, Flowing


Synthpop, Big Reverbed Synthesizer Pad Chords, Driving Gated Drum Machine, Atmospheric, Moody, Nostalgic, Cool, Club, Striped-back, Pop Instrumental, 100 BPM


 
Last edited:

"This was the point where AI-generated art passed the Turing Test for me."​

- Y-Combinator co-founder and frequent social media tech commentator Paul Graham

spiral-town-different-approach-to-qr-monster-v0-t37hzfc4henb1.jpgcheckertown1.jpgspiral_art.jpg

checker_girl_1.jpggirl_with_pearl_controlnet.jpgspiral_lady_2.jpg

QR codes and merchant memes rediscovered.
 
Last edited:
Back