Reading less social media bullshit and using ones own well-trained mind sometimes helps to clarify complex and noisy topics.

Here is why, in principle, any LLM coding (I cannot call this crap “programming” because there is no uderstanding involved) will always yield sub-par bullshit.

LLMs operate on tokens, which are abstract numbers, if you will. Parts of words of a human language are being associated with a distinct token, and then a probabilistic graph-like structure is created out of these tokens, using a fundamental back-propagation algorithm.

Then, at an inference phase, the system (a binary-encoded probabilistic mathematical model) appears (to an uneducated observer) to behave as if it is “intelligent” and “capable of doing reasoning”. This is, of course, an illusion, created by a vastly complex computation, which can be easily mistaken as an intelligence (yes, yes, the classic “a rope for a snake”) .

It the context of programming everything is much worse. Operating, in principle, on “symbols” associated with words that encode entities, attributes, properties and relations (nouns,verbs and adjectives are out there for a reason – they capture the “structure” of What Is) it cannot see the underlying patterns, generalize from them, and capture and name abstractions, in principle – they lacks the machinery to do so.

Here is how. Any undergraduate, I hope, knows that \[x \rightarrox x + 1\] and \[y \rightarrow y + 1\] are the same function. leave alone that \[x \rightarrow x + x\] and \[x \rigjtarrow x * 2\] are the same function, but different procedures (implementations).

Such mathematical notions are the most fundamental and has been taken, vial the Lambda Calculus, into a serious Computer Science – they have formally specified the notion of a bound (in a given context) and of a free variable, and precise rules for consistent renaming to avoid an unintended name capture or shadowing. The linguistic notions of shadowing and overloading has been formally defend too.

No conditional-propability based model, no matter what “architecture” (a Transformer or what not) could “see” or capture such notions, leave alone use them. Again, there is simply nothing to talk about – they just cannot, there is neither “awareness” nor “memory of experience” nor “instantiation of rules”.

So, the only context in which such models are marginally useful is in filling up the “blanks” in verbose, over-“engineered” OO APIs in crappy OO languages like C++ or Java. This is the only valid “use-case”. The long names, verbose syntax of types, fully qualified method access – all this can be generated (without any underlying understanding).

This (and only this) is how Microsoft, lest say, claims that it codes some percentage of some its software with AI – they feed their wast libraries into a more-or-less standard model and it performs a “glorified autocomplete” (indeed!).

This is the only use the mere mortals like you and me could have. If you read carefully these overflowing with excitement “vibe coding” shitposts, you will realize that the only task in which AI “succeeds” is, literally, coding some structured and formated crap (think of XML, JSON, CSS, HTML) and simple CRUD code which produces or consumes such (variously) encoded data, provided there were lots and lots of such examples in the training set.

Again, how a model “recognizes” a procedure? Only by its names and surrounding names (and there is a notion of order of arguments which it implicitly captures as a probability of a the next token. It is that “next”.

So, a properly trained model could generate something that looks like it has “seen” before, no more, no less. It does not perform, in principle, due to the lack of a “mathematical machinery” any kind of “understading”, “reasoning”, “inference”, “deduction”, “induction”, “generalizing”, “specializing”, and so on. It may appear as if it does (to a stupid internet imposter).

The next time you see some narcissistic asshole on youtube claims that he “vibe-coded” something, come back here and re-read these few paragraphs above.