So, this is the why all the megacorps suddenly building AI data-centers and purchase nuclear power plants to feed them with electricity and book all the DRAM and VRAM production and in the whole world.

The reason is this. Remember Google reCAPTCHA - that window when you are forced to select all the cars or traffic lights to view the website you wanted (for an expected revard through a dopamine release).

What you actually doing is cleaning up Google’s wast data-set of images for free, verifying the previous labeling.

Here is how it works – there are some vague, error prone automatic labeling, based on patten-recognition (so called Computer Vision), which produced some labels for a dataset, which have some greater than 0,5 probability to be correct. Just this, no semantic knowledge whatsoever. Then they load it into reCAPTCHA, and you, by clicking on thumbals, systematically and consistently increase the probability of correct labeling (they could run the same image set through hundreds of iterations) for free.

At so-called “internet scale” this is not just individual monkeys with typewriters, trying to come up with the complete works of Shakespeare, but millions of monkeys, each one pick up a typewriter where the previous one left t, and there are millions of such typewriters, working in parallel, to increase the theoretical proability of hypothetical convergence to “To Be, Or Not To Be”.

Now exactly the same processes, but on a much larger scale, are being run by AI divisions of major tech corporations . There are millions of monkeys (Yous) are continuously, in massive parallel architecture, are refining slop-generator inner representations, so it, hopefully, will eventually converge on something meaningful (instad of producing millions of local-optima dead-ends).

They stated with an utter crap taken from Github and somehow turned into a “coding model” (which produce an utter trash, lost context between iterations and hallucinates and makes up shit all the time). Neverheless, to (You) it appears as a meaningful activity, and that you are making a progress or even learn something, improving, and above all, have the same neuromodulator spikes running, as with a real activity. but much intense, since it is all compressed in a tight feedback loop.

The outcome of all this is that each your textual feedback notches the knobs little bit in the “right direction” - the prompt, the prompt you gave, the resulting slop and your reaction and cognitive feedback to it is being feed back as the training data, so the utter crap undergoes a slow gradual process of continuous refining (of the was inner Artificial Neural Nets representations). The scale is enormous – of every single possible token as an “entry point”.

Theoretically, every single “session” with the slop generator improves its representation on the topic (subject) of conversation, without any semantic understanding – just slightly increasing the probabilities of being “right” (above 0,5), and then it back-propagates, without understanding, of course.

This is why they are all in. Their bets is that it will eventually converge onto something “more than the sum of its parts” (nope), and if not, they will steal half of the money by the time it bursts. Strictly MBA thinking.

One more time – there is no understanding or meaning involved whatsoever – the process is to find better numbers than they are right now, assuming that these numbers represents the actual probabilities of “more correct answers”. No one knows how, they just observe (and even “measure” by the rate of “successes” to “failures”) that the slop appears to be more consistent and coherent than before. That is it.

The real optimization begins when (You) are going to interact with the slop generator and give it your cognitive feedback, whatever it may be (no one notices or cares, it will eventually move to the right direction, just as an implication of the law of big numbers – it will converge not on an overage, but on more than 0,5 by design).

See, it is not even them, but the millions of monkeys who actually improve their AI representations by just “pushing it in the right direction all together”.

It worth repeating again – the slop it produces do not even compile, leave alone come up as correct or even consistent. This is not the goal. The goal is to make (You) to correct it somehow (not important) and to feed it back as the new training data in an automated process. The slop then automatically becomes more and more usable, with [probabilistically] a bit less errors than before, without any understanding of any kind whatsoever.

The whole shitshow runs on your very own dopamine and serotonin spikes and crashes, though (You) are not a battery, (You) are a monkey with a typewriter, because the appearance (Maya) is very persuasive, it appears to be “intelligent”, and even more intelligent than (You), until you “wake up” and try to actually examine and validate the slop, which turns out of being a fucking trash that doesn’t even compile, and breaks with each new “iteration” by losing the coherence from the previous contexts.

This is your AI, in principle.