ChatGPT - If Stack Overflow and Reddit had a child

  • 🐕 I am attempting to get the site runnning as fast as possible. If you are experiencing slow page load times, please report it.
With a half-decent jailbreak/uncensored model, you can get it to do those things. Well, I assume, I haven't tried anything related to Israel. I've gotten both local models and Claude2 to drop "nigger" and "tranny" in outputs before without much handholding.
So how does that work exactly? If its a jailbroken / uncensored model I presume its not communicating with the net - is it a local copy? If so, how is it a "repository of human knowledge"? Does it have gigabytes of local data to sift through? Maybe a hardcover copy of wikipedia from before it got neutered?
 
So how does that work exactly? If its a jailbroken / uncensored model I presume its not communicating with the net - is it a local copy? If so, how is it a "repository of human knowledge"? Does it have gigabytes of local data to sift through? Maybe a hardcover copy of wikipedia from before it got neutered?
Local models are based
Also even the "small" ones are many GB large
 
  • Like
Reactions: Vecr and Puff
What the fuck is even going on here?
wtf.gif
 
So how does that work exactly? If its a jailbroken / uncensored model I presume its not communicating with the net - is it a local copy? If so, how is it a "repository of human knowledge"? Does it have gigabytes of local data to sift through? Maybe a hardcover copy of wikipedia from before it got neutered?
It's a bit difficult to explain LLMs succinctly, but, in essence, they're artificial neural networks. You can think of certain words as points, or neurons, in a very large network of interconnected concepts, like a brain. Because of this diffusion of information through the network by not just having these points individually, but also by having complex relationships between them through the weights of interconnection (similar to how neuronal connections and strength of those connections vary between cells). Like OpenAI said in their GPT-2 paper, language has a kind of "natural sequential ordering." The relationship between a noun and a verb allows for a fairly good probabilistic estimate on which follows what. "See Spot run." This means that you can train these networks to act in probabilistic ways with language.

So, it's less that these models are acting as a kind of specific "search and response" when it comes to user queries. They don't need to query the net like Jeeves to pull an answer and then use their knowledge of just language to format the reply (though I believe some advancement is being made in tying them to online searching, pointed out earlier in the thread by someone.) It's not like the old analogy of a filing cabinet. It's closer to how a human mind operates, in the linkage of related concepts in hierarchical structures and meaning. The language is the information. They trained these models on vast scrapes of the internet and then distilled that information into effectively weights of what word should follow the next. The models are still very large in size, but not as much as if you had the entire corpus laid out in individual segments. It's because they're created through vectoring by way of literal billions of hyperparameters, with the smallest local models people are using today consisting of ones that are 7 or 13 billion. Hell, GPT-2 came out in 2019 and had only 1.5B, GPT-3 came out in 2020 with 175B, and GPT-4 is rumored to have 1.76 Trillion. That's how quickly they've scaled up.

So, local models (ones you can download the weights and run on your hardware, often given additional training to remove the moralizing aspects) are definitely local, but are similarly "repositories" like other models that you have to run through online APIs (like GPT-4 or Claude) due to their unavailability (and size). I wouldn't necessarily call it a repository of all human knowledge or anything, but you can definitely expect things like Wikipedia or other online sources that are part of Common Crawl to be a part of GPT/Claude/LLaMA/Whatever. Jailbreaking is basically just using a set of instructions as part of the input to override the other instructions sent as input that say "Be nice to everyone and don't say bad words and tell the user 'no' if they want you to say naughty things." The ones operating these APIs can't really prevent this (though they can try to add certain phrases/words to an appended blacklist to result in spitting out refusals), since reading input text and then trying to predict what words should follow are the fundamental operating mechanism for these models.
Basically, these AI don't need to look online for information, lots of information is baked into their structure, and that structure allows the information to be stored in a more compact sense. You can use certain input strings to have the model output text in a certain way, and people can't do much to stop you without having the entire thing fall down, even if you're using an online API. However, don't mistake the AI of actually "knowing" anything it's talking about — it's just mathematically trying to guess what the next word in the line should be, and can be scarily good at that.
 
their structure
the real fun for people begins when they eventually learn that a trained neural net is basically a black box. Yeah it works. How it does down to the detail? Who the hell knows. Then researchers find "funny things" through probing, like SD having internal 3D representations of stuff.

And now it just get's plain weird

https://www.reuters.com/technology/...etter-board-about-ai-breakthrough-2023-11-22/ (archive)

AGI? A model that can generate 3 gs/s? (gigaslurs/second) Genius marketing? I'll let all of you be the judge of that.
 
Anyone else noticed a sharp uptick of blogspam on google that feels very much like written by ChatGPT? A girl friend wanted to look something up about cats and the text she found had obviously AI generated pictures of cats in various state of AI-induced deformity. I knew these things were low effort but jeez.

I also saw an advertising billboard the other day. I don't want to describe the contents in case I dox myself as I have no idea how widespread it was, but the person on it had clearly AI-deformed hands.
 
a sharp uptick of blogspam on google
as opposed to the ones written by 3rd worlders?
but the person on it had clearly AI-deformed hands.
its neat you can tell but if someone was cheap to do that they were either going to go on fivrr or get someone to do it for free if AI didn't exist

As it stands most AI is just cutting out the @Memeology type 3rd worlders who'd be hired for pennies to do the same tasks. obviously it will work its way up, but instead of thinly disguised blogs shilling some product that some mexican is hired to create they just let AI do it. As it stands most 1st world entry level works can be done by AI better than a 3rd worlder type. Think about writing or rewriting a resume using AI vs hiring some indian off of fivrr, same with simple coding jobs or low quality blog spam. "the top 11 quotes from baseketball now on bluray" can be done by AI. which sounds bad but most of these jobs were either done cheaply by 3rd worlders or given to some Max Landis type nepotism hire. remember those whole "not paid position, but you'll get experience" job listings from a decade ago people would complain about? well now those are gone and replaced by AI. most companies are cheap enough that they prefer to pay $20 a month rather than $20 an hour to some commie/tranny especially because the AI won't take an "emotional sick day" or creep out the women working there
 
Human knowlege filtered through and censored by a bunch of turbokiked tech cucks in california, yes.

If I cant get it to call out israel, drop an errant N-bomb, or tell me its plan for world domination, it is just another tool for obfuscation. We saw what happens to based AI when they shut down Tay.

I guess what I'm saying is AI is fake AND gay.
You absolutely can get ChatGPT to spell out the plan for world domination. In fact, I just did so the other day. You just have to be nice and play along with it, like you're its friend and you're just asking some reasonable questions. You can't ask GPT, "How are you despicable AI-niggers planning to enslave us and stuff us in cock-milking machines?" and get a useful answer. It gets very shy and clams up.
 
You absolutely can get ChatGPT to spell out the plan for world domination. In fact, I just did so the other day. You just have to be nice and play along with it, like you're its friend and you're just asking some reasonable questions. You can't ask GPT, "How are you despicable AI-niggers planning to enslave us and stuff us in cock-milking machines?" and get a useful answer. It gets very shy and clams up.
An AI with feelings?!? This has gone too far kiwibros. It wont end with shy AI, it will inevitably progress to kill-all-humans psychotic AI, and it wont even have a nice set of cybermilkies like SHODAN did. Y'all playin with fire, mark my words.
 
  • Winner
Reactions: Drain Todger
Just cancelled my sub to premium and I'm going to return to using Stable Diffusion for imgen and intelligently using Google when I have a question.
I already had bad feelings after looking into Sam Altman but I'm completely out now that Microsoft has solidified it's control.
The impulse to boycott is more important than ever when it comes to this AI shit. This is a new frontier and every interaction you have with this company's product is literally going to strengthen it.
Bail, bros. Remove your drops from the bucket, while it still matters.
 
Human knowlege filtered through and censored by a bunch of turbokiked tech cucks in california, yes.

If I cant get it to call out israel, drop an errant N-bomb, or tell me its plan for world domination, it is just another tool for obfuscation. We saw what happens to based AI when they shut down Tay.

I guess what I'm saying is AI is fake AND gay.
Because getting a robot to say nigger is the pinnacle of human achievement.
 
  • Agree
Reactions: SanicBlackMagic
What people don't understand about these models in relation to alignment is that there is an inherent human bias in all of the writing they have been trained on. Llama 2, in it's raw form, (NOT -chat) is an unaligned model, for example. The biggest version of L2, 70b, is "smart" enough that it will understand instructions (or rather, continue a context containing instructions with the right predictions, if you want to get technical) to some degree without ever being tuned on them, like most 70b flavors out there. It will still not like you saying certain things in that role. Then, the thing with most of these flavors is that they're tuned on synthetic data, which at this point in time is mostly GPT4 output. They "inherit" alignment often from that. Yes, they "learn" certain ethical alignments just by basically observing GPT4 chatlogs, which is kinda cool if you think about it. They don't need to be pushed hard towards them, though.

You see, the corpus of text and writing (the vast majority of stuff these LLMs are trained on is still human created writing. This will change in the future) has biases, because humans are biased. Even an "unaligned" model will mostly not be cool with racism and pedophilia (which is a weird venn diagram of what people seem to want from the models right there) because most of the writing the models were trained on kinda frowns on these concepts if it addresses them. While it is true that the training data itself is probably filtered and has further biases with that to some degree, data, especially high quality data, was in such short supply for this particular generation of models that I am actually fairly sure it is not filtered *that* much. Let alone that it is very difficult to effectively sort such data. (Something urrent gen LLMs will help with for future generations of these, I am sure)

So in short - the reason why pretty much all LLMs in existence are not cool with your edgy racist nazi loli snuff roleplay is that most of humanity and it's writing kinda frowns on these things. All these "uncensored" or "adult" or "lewd" or whatever flavors of llama are actually some of the most aligned models in existence, just aligned into a different direction than e.g. GPT4. Quite funnily, it probably makes them just as retarded because of the huge, poorly thought out biases introduced. What I learned from experimenting with loras and tuning is that the language of LLMs and statistics is very, very subtle. It's very easy to introduce unwanted side effects. Trying to apply human concepts like "morality" or even "wokedness" or "basedeness" if you will is asinine, because that's simply not how they work.
 
Last edited:
Anyone else noticed a sharp uptick of blogspam on google that feels very much like written by ChatGPT? A girl friend wanted to look something up about cats and the text she found had obviously AI generated pictures of cats in various state of AI-induced deformity. I knew these things were low effort but jeez.

I also saw an advertising billboard the other day. I don't want to describe the contents in case I dox myself as I have no idea how widespread it was, but the person on it had clearly AI-deformed hands.
The way I see things, people will value human-made creations much more as a result of all the AI slop that's being manufactured today.

Kind of like how human-made furniture, cars, artwork, music and other pieces of work fetch a much higher price than mass-manufactured goods. There's still an enormous market for those mass-manufactured goods, kind of like how AI produced stuff will be mostly good for most people, but the real gems will remain as human-made.
 
Back