Literally just mainlining marketing material straight into whatever’s left of their rotting brains.
Well, my (admittedly postgrad) work with biology gives me the impression that the brain has a lot more parts to consider than just a language-trained machine. Hell, most living creatures don’t even have language.
It just screams of a marketing scam. I’m not against the idea of AI. Although from an ethical standpoint I question bringing life into this world for the purpose of using it like a tool. You know, slavery. But I don’t think this is what they’re doing. I think they’re just trying to sell the next Google AdSense
Notice the distinction in my comments between an LLM and other algorithms, that’s a key point that you’re ignoring. The idea that other commenters have is that for some reason there is no input that could produce the output of human thought other than the magical fairy dust that exists within our souls. I don’t believe this. I think a sufficiently advanced input could arrive at the holistic output of human thought. This doesn’t have to be LLMs.
I haven’t seen anyone here (or basically anyone at all, for that matter) suggest that there’s literally no way to create mentality like ours other than being exactly like us. The argument is just that LLMs are not even on the right track to do something like that. The technology is impressive in a lot of ways, but it is in no way comparable to even a rudimentary mind in the sense that people have minds, and there’s no amount of tweaking or refining the basic approach that’s going to move it in that direction. “Genuine” (in the sense of human-like) AI made from non-human stuff is certainly possible in principle, but LLMs are not even on that trajectory.
Even setting that aside, I think framing this as an I/O problem elides some really tricky and deep conceptual content, and suggests some fundamental misunderstanding about how complex this problem is. What on Earth does “the output of human thought” mean in this sense? Clearly you don’t really mean human thought, because you obviously think whatever “output” you’re looking for can be instantiated in non-human systems. It must mean human-like thought, but human-like in what sense? Which features are important to preserve, and which are incidental or parochial to the way humans do human-like thought? How you answer that question greatly influences how you evaluate putative cases of “genuine” AI, and it’s possible to build in a great deal of hidden bias if we don’t think carefully and deliberately about this. From what I’ve seen, virtually none of the AI hypers are thinking carefully or deliberately about this.
The top level comment this chain is on specifically reduces GPT by saying it’s “just an algorithm”, not by saying it’s “just an LLM”, which is implicitly claiming that no algorithm could match or exceed human capabilities, because they’re “just algorithms”.
You can even see this person further explicitly defending this position in other comments, so the mentality you say you haven’t seen is literally the basis for this entire thread.