mashbooq
russia is such a deeply unserious country
Knorozov was born in the village of Yuzhny near Kharkiv, at that time the capital of the newly formed Ukrainian SSR
Congress wouldn’t let him, iirc
My heart died a little at seeing my beautiful state slandered by being labelled as Arkansas
I’m not trained in formal computer science, so I’m unable to evaluate the quality of this paper’s argument, but there’s a preprint out that claims to prove that current computing architectures will never be able to advance to AGI, and that rather than accelerating, improvements are only going to slow down due to the exponential increase in resources necessary for any incremental advancements (because it’s an NP-hard problem). That doesn’t prove LLMs are end of the line, but it does suggest that additional improvements are likely to be marginal.
The one colleague using AI at my company produced (CUDA) code with lots of memory leaks that required two expert developers to fix. LLMs produce code based on vibes instead of following language syntax and proper coding practices. Maybe that would be ok in a more forgiving high level language, but I don’t trust them at all for low level languages.
School shootings? :shrug: Statues depicting the nude human form? “We must protect the children!” 🙄
Are you able to share what kinds of applications and what languages you write in? I’m still trying to grasp why LLM programming assistants seem popular despite the flaws I see in them, so I’m trying to understand the cases where they do work.
For example, my colleague was writing CUDA code to simulate optical physics, so it’s possible that the LLM’s failure was due in part to the niche application and a language that is unforgiving of deviations from the one correct way of writing things.