I'd like to see what would happen if we somehow set up distributed computing with something like a few million GPUs simulating a larger (and probably slower) neural network or hypernetwork, if that makes sense.
The latency and "bad actors" would utterly kill it. The new models are already slow with the blazing fast RAM speeds we have today and even incredibly high end, commercial GPUs designed for AI usage are just "good enough". The problem with this stuff is that it's basically all about the training and there's still many question marks how to do it well. It all boils down to us needing better hardware, really. And software/architecture optimization. (there's apparently tons of room)
---
Slightly related, an interesting pre-print paper I read somewhen, interesting because it agrees with my subjective impression of GPT 3.5, GPT 4, and anthropics "Claude" (all "aligned" models):
tl;dr:
alignment is pointless. Training it for alignment via reinforcement learning from human feedback (RLHF) probably actually makes things worse as like a psychopath in therapy, all the model does is learn to be "bad" better. Inherent alignment hopeless with current tech (maybe never?) if you want good models. The bigger a model's "memory" is, the easier it can slip free, the harder it is to reign itself back in. Memory apparently an inherent problem in general with this. Also but just implied and as I understood: the bigger and "smarter" (good evals) the model, the more easily it can just "decide" to bypass it's alignment.
Also, The future is wendy's automated drive-through calling customers niggers over and over & online banking chatbots with avtars of blonde, nubile western women sexting with the banks' customer ( I jest)
I got to play around with GPT4-35k (35k token memory vs. 4k token memory ChatGPT/Turbo has) and the difference in perceived intelligence is striking. If you don't use that huge context to prime it or fill it with things to categorize, you notice grasps at the "bigger picture" of the conversation as the context fills, smart questions asked in return unprompted, curiosity or at least acknowledgement regarding a chronological quality the chat might possess. (Which all other models I tried so far didn't have) It's still all statistics and autocomplete, but it is frankly quite eerie. I think for a hint of simulated personality, memory is key. Anthropic just released a version of Claude with 100k token context, but as I understood it, it's kinda fake and they basically just strapped a database/search engine to it. I haven't tried it yet and don't know how good or ungood it is.
Also a leaked google memo claims open source is probably the future for this tech and proprietary solutions won't be able to compete, especially if they don't drop the filtering:
(first link I grabbed off google)
After all the fears of replacing jobs & Skynet, I think everyone is overseeing the *vast* potential for entertainment this tech has. Somebody threw chatgpt into skyrim, making the NPCs answer with ChatGPT generated responses. Video's somewhere on youtube. I see a big potential for the "unlimited worlds with unlimited choices" videogames have promised us for so many decades but ultimately always failed to deliver. That stuff is truly gonna be grand, when people just all have their neverending fantasy worlds on their desktops. This staunch "no fun allowed" stance of the silicon valley bugmen is incomprehensible to me. Especially considering that the home computer revolution basically started with "well it might not be great for many things, but you can play vidya on it".