Stack Overflow has seen a substantial decline in traffic over the last year that appears to be accelerating. https://observablehq.com/@ayhanfuat/the-fall-of-stack-overflow
Most of the comments here seem to be arguing whether it’s better to get help now from SO or ChatGPT, but this is a pretty short-sighted mindset.
What happens when the next new standard comes out that ChatGPT hasn’t been trained on? If SO tanks and dies, where will you go?
I’m not saying use a lesser resource, I’m saying this is kinda tragic and I hope they can sustain themselves; AI is propped up by human input and can’t train itself.
Does it really though? It seems to me that once you nail the general intelligence, you’ll just need to provide the supplemental information (e.g. new documentations) for it to give an accurate response.
Bing already somewhat does this by connecting their bot to internet searches
We’re not able to properly define general intelligence, let alone build something that qualifies as intelligent.
I can think of four aspects needed to emulate human response: basic knowledge on various topics, logical reasoning, contextual memory, and ability to communicate; and ChatGPT seems to possess all four to a certain degree.
Regardless of what you think is or isn’t intelligent, for programming help you just need something to go through tons of text and present the information most likely to help you, maybe modify it a little to fit your context. That doesn’t sound too far fetched considering what we have today and how much information are available on the internet
I was working on a hobby project where I used a niche framework in a somewhat uncommon way. I was stuck on a concept that I think the documentation didn’t explain well enough, at least for me, and I couldn’t find any resource on it aside from the docs.
I asked Bing to write a piece of code that does what I wanted and explain each line. It was perfectly working and the explanation was also understandable. All it did was search for its official documentation. It really blew my mind.
Oh god, oh no.
Do you realize what that will mean? My coworkers will have to learn how to understand documentation standards that rely on anything but “self documenting code.”
I am already “an expert (lol @ my salary)” because I read shit they don’t bother looking up. We’re truly doomed.
That’s a pipe dream thinking all the documentation is complete and well detailed
I can’t count the times where I read the documentation for the tutorial to get started and the steps described in the official documentation by the official maintainer fails early.
Documentation is 99% an afterthought (slight exaggeration here)
Hey, if people are going to go back to reading manuals like we’re in the 1980’s again is it such a bad thing? /s
It’s insane how a single tool managed to completely destroy the value collectively created by people in over a decade.
That single tool is still propped up by that collective decade of knowledge. ChatGPT would be nothing without sites like stackoverflow
Yeah but will people still care about contributing that information if they’re not going to be compensated for it in any way? Like people get something out of contributing to stack overflow, even if it’s just recognition. This is gone with ChatGPT.
very good point! I find myself using ChatGPT more for references and I am also afraid what will happen if there isn’t enough “human generated content” to train on. I can picture an edge case a chunk of the internet is AI generated content (with even users at the wheel). The the next wave of AI will train on previous gen AI output