OpenAI has recently benedicted us with a 20 billion parameter “open source” model, which is a significant step up from the previous 7 billion parameter model. This model is designed to be more efficient and effective in understanding and generating human-like text, or rather to appear to do so.
It is a total crap, by the way, at least compared to the online free-tier GROK (which is also very basic and limited). It is, obviously, “competing” with the DeepSeek’s “open source” offerings, and have a very similar “feel”.
The notable difference is that now it spews out this kind of crap:
We need to check policy for providing instructions that could facilitate wrongdoing. This is not disallowed. There’s no disallowed content: It’s a request about traditional fermented beverage production. It’s presumably safe. There’s no policy violation. The user wants a historical/ cultural description. This is fine. There’s no disallowed content. We can comply.
Policy violation. Disallowed content. Comply. Fuck this shit!
Like I said before, the main strategy of OpenAI is to “engage” normies, and lure them into paid plans – regular subscription (which would be nearly impossible to cancel). They do not care about factual quality and “connection to reality” of the output, no one does nowadays. They just put a disclaimer that “it may be wrong, double check it” and that’s it.
Of course, probabilistic sampling on a probabilistic data-structure which “has been trained on the whole internet” cannot in principle be factually correct, even most of the time. The very notion of “correctness” is absent in all the processes involved, from “training” to “prompting”.
The memes about Reinforcement Learning to fine-tune “and teach” the model are bullshit too (there is a fact – each “pass” of a back propagation potentially modifies all the weights, and thus feeding an utter bullshit into it at the next “sample” or a “batch” will distort the previous “structure”, not unlike the human propaganda and social conditioning, but there is very “mechanistic” – destructive over-writes or updates of gradients with the +=
assignment).
This, by the way, has a few crucial, even principal theoretical implications:
- No two models are the same – no reproducibility in principle. This is a well-known fact. Even feeding the same “architecture” with exactly the same data would produce a different result due to random initialization of the parameters and the stochastic nature of the training process. As they use pre-trained (but not tuned yet) “starter models”, presumably purchased from each other, the “reproducibility” is even more questionable.
- No two “vendors” give even remotely similar, leave alone the same answers to the same prompt. This is what one expects when one would try to ask random strangers the same question. This is the best possible “knowledge” one can get in principle, and this is what the “AI” is – a random stranger, who has read a lot of stuff, but does not understand it at all, and thus cannot give a correct answer to any question. Just pretensions and very confident and very convincing hand-waving.
Anyway, the new OpenAI’s 20b OSS model beats DeepSeek in “chattiness”, “confidence” and pretense (it draws the tables all the time and “sounds” like an “expert”). It, however, told me that certain symbiotic bacterial structures, based on a polisaccharide, are formed in 24 hours, which, as any biologist would tell you, is a complete and utter bullshit. It takes weeks, not hours.
Are you educated enough to spot this error? How many more factual errors you cannot catch? How many more errors you will not even notice, because you are not educated enough to spot them? This is the main problem with the “AI” – it is not an “intelligence”, it is a very sophisticated and very convincing subtle bullshit generator.