For a quite long time already I sort of “ran in the background” this question – “how they bridge the semantic gap between verbiage and the code?” In the context of a well-written textbook (extremely rare, just 50 or so in existence) the code examples immediately follow or even intersperse the explanations. Most of the internet content is nothing like that.
The semantically closest entities in the code are specially formatted comments that are used to generate [an appearance of] “documentation”, with a modern fashion to mention some core concepts being used. This kind of well-documented (enforced by the strict rules) code one would find in placed like Google’s monorepo and other megacorp inner code bases. Almost noting this strict can be found on Github (the primary source of all training).
Now I see what they are doing. While stated from the Github crap, they now annotate the code in markdown in a systematic way, and then, of course, feed it (the verbiage and the code) back as new training examples (after people seemingly accepted it is “working” or “acceptable” solution).
They just use you to build a much better slop, just as they used captchas to make you manually label their dataset or validate the pre-existing labeling (which both are very costly processes). Now you are generate new training examples for them, not just for free, but some are paid for each token. No wonder both bigtech and Wall street are “all in” without looking back.
Another thing I noticed is that it already speaks in cliches an even “slogans”. When I tried to force it to carefully clarify and refine my handwritten GEMINI.md, it replaced all my slightly convoluted subtleties with crystal- clear short sentences and well-define common [mostly meme] terminology.
This is not a random coincidence and is actually expected – when you recycle the same slop over and over again, it will very quickly converge at some “local optima” – literally a concave-shaped, localized “echochamber”, just like any religious or scientific sect, constantly parroting their maxims and tenets.
But if you put yourself into the loop, clarifying and refining the slop, you could reach some fixed-point in terms of precision, brevity, clarity and conciseness, almost math-like – the language one used to read with the theorems and proofs .
This kind of writing is a lost craft since publishing became an industry.
And yet, the quality of the verbiage (slop) it spews out, when properly constrained, and over a few iterations putting it back into a new prompt is absolutely amazing. I spent, literally, a lifetime to learn to write the way I do (which is far from being satisfactory), and it just spit into your face what you hardly could write unless struggle for days, continuously refiling (which I, thank god, never do).
So, they trained on your slop as we speak (and call it major advancement in LLM training techniques), and when “personalized”, it quickly creates your own, personal Jesus, sorry, an echochamber, with all the known cognitive biases at work (when something else appear to be and being recognized and interpreted as something familiar).
One more time, the “magic” and appeal of LLMs is that they produce heavily cognitive biased appearances, which most people gladly and easily take “for real”. This is why the bubble wasn’t burst till now.
Just put back some your own subtleties it replaced with commonplace memes, and one day there will be nothing more to remove (clarify). The obsession of some famous MIT guys with fixed-points is not, of course, coincidental.
The trick, William Potter, is not minding that it hurts.
By the way, the proper way of prompting is to build a context very similar to what it has “already seen” and thus will easily extend, not some kind of “You are a helpful assistant” antrophomorphic bullshit, but once this very crap is constantly being feed back, it soon will become the only way to prompt.