ChatGPT is hilariously incompetent… but on a serious note, I still firmly reject tools like copilot outside demos and the like because they drastically reduce code quality for short term acceleration. That’s a terrible trade-off in terms of cost.
I enjoy using copilot, but it is not made to think for you. It’s a better autocomplete, but don’t ever let it do more than a line at once.
The problem I have with it is that all the time it saves me I have to use on reading the code. I probably spend more time on that as once in a while the code it produces is broken in a subtle way.
I see some people swearing by it, which is the opposite of my experience. I suspect that if your coding was copying code from stack overflow then it indeed improved your experience as now this process is streamlined.
Biggest problem with it is that it lies with the exact same confidence it tells the truth. Or, put another way, it’s confidently incorrect as often as it is confidently correct - and there’s no way to tell the difference unless you already know the answer.
it’s kinda hilarious to me because one of the FIRST things ai researchers did was get models to identify things and output answers together with the confidence of each potential ID, and now we’ve somehow regressed back from that point
did we really regress back from that?
i mean giving a confidence for recognizing a certain object in a picture is relatively straightforward.
But LLMs put together words by their likeliness of belonging together under your input (terribly oversimplified).the confidence behind that has no direct relation to how likely the statements made are true. I remember an example where someone made chatgpt say that 2+2 equals 5 because his wife said so. So chatgpt was confident that something is right when the wife says it, simply because it thinks these words to belong together.
But LinkedIn bros and corporate people are gonna gobble it up anyways because it has the right buzzwords (including “AI”) and they can squeeze more (low quality) work from devs to brag about how many things they (the corporate owners) are doing.
It’s just a fad. There’s just a small bit that will stay after the hype is gone. You know, like blockchain, AR, metaverse, NFT and whatever it was before that. In a few years there will be another breakthrough with it and we’ll hear from it again for a short while, but for now it’s just a one trick pony.
I disagree, because unlike those things, ai actually has a use case. It needs a human supervisor and it isn’t always faster, but chat gpt has been the best educational resource I’ve ever had next to YouTube. It’s also decent at pumping out a lot of lazy work and writing so i don’t have to, or helping me break down a task into smaller parts. As long as you’re not expecting it to solve all your problems for you, it’s an amazing tool.
People said the same things about 3d printing and yeah, while it can’t create literally everything at industrial scale, and it’s not going to see much consumer use, it has found a place in certain applications like prototyping and small scale production.
Yeah, they think it can turn a beginner dev into an advanced dev, but really it’s more like having a team of beginner devs.
I’m still convinced that GitHub copilot is actively violating copyleft licenses. If not in word, then in the spirit.
they drastically reduce … quality for short term acceleration
Western society is built on this principle
Tell me about it…
I left my more mature company for a startup.
I feel like Tyler Durden sometimes.
Sort of. Nobody’s cutting corners on aviation structural components, for example. We’ve been pretty good at maximizing general value output, and usually that means lower quality, but not always.
It’s helped me a bit with resolving weird tomcat/Java issues when upgrading to RHEL8, though. It didn’t give me an answer, but it gave me ideas on where to look (in my case I didn’t realize fapolicyd replaced selinux)
That’s the point - you have the expertise to make proper sense of whatever it outputs. The people pushing for “AI” the most want to rely on it without any necessary expertise or just minimal efforts, like feeding it some of your financial reports and have generate a 5-year strategy only to fail miserably and have no one to blame this time (will still blame anyone else but themselves btw).
It’s not the most useless tool in the world by any means, but the mainstream talk is completely out of touch with reality on the matter, and so are mainstream actions (i.e. overrelying on it and putting way too much faith into it).
An unpopular opinion, I am sure, but if you’re a beginner with something - a new language, a new framework - and hate reading the docs, it’s a great way of just jumping into a new project. Like, I’ve been hacking away on a django web server for a personal project and it saved me a huge amount of time with understanding how apps are structured, how to interact with its settings, registering urls, creating views, the general development lifecycle of the project and the basic commands I need to do what I’m trying to do. God knows Google is a shitshow now and while Stackoverflow is fine and dandy (when it isn’t remarkably toxic and judgmental), the fact is that it cuts down on hours of fruitless research, assuming you’re not asking it to do anything genuinely novel or hyper-specific.
It helps a complete newbie like me get started and even learn while I do. Due to its restrictions and shortcoming, I’ve been having to learn how to structure and plan a project more carefully and thoughtfully, even creating design specs for programs and individual functions, all in order to provide useful prompts for ChatGPT to act on. I learn best by trial and error, with the ability to ask why things happened or are the way they are.
So, as a secondary teaching assistant, I think it’s very useful. But trying to use the API for ChatGPT 4 is…not worth it. I can easily blow through $20 in a few hours. So, I got a day and a half of use out of it before I gave up. :|
I predict that, within the year, AI will be doing 100% of the development work that isn’t total and utter bullshit pain-in-the-ass complexity, layered on obfuscations, composed of needlessly complex bullshit.
That’s right, within a year, AI will be doing .001% of programming tasks.
Legitimately could be a use case
“Attend this meeting for me. If anyone asks, claim that your camera and microphone aren’t working. After the meeting, condense the important information into one paragraph and email it to me.”
Here is a summary of the most important information from that meeting. Since there were two major topics, I’ve separated them into two paragraphs.
- It is a good morning today.
- Everyone is thanked for their time. Richard is looking forward to next week’s meeting.
The rest of the information was deemed irrelevant to you and your position.
Hell yes! I’ll join the front of the hype train if they can demo an AI fielding questions while a project manager reviews a card wall.
“look i registered my own domain name all by myself!”
the domain: “localhost”
I’m an elite hacker and I grabbed your IP address from this post. It’s 192.168.0.1 just so you know I’m not bluffing.
Heheh I’m ddossing them right now. Unfortunately the computer I’m doing it on is having a few connection issues
Engineering is about trust. In all other and generally more formalized engineering disciplines, the actual job of an engineer is to provide confidence that something works. Software engineering may employ fewer people because the tools are better and make people much more productive, but until everyone else trusts the computer more, the job will exist.
If the world trusts AI over engineers then the fact that you don’t have a job will be moot.
People don’t have anywhere near enough knowledge of how things work to make their choices based on trust. People aren’t getting on the subway because they trust the engineers did a good job; they’re doing it because it’s what they can afford and they need to get to work.
Similarly, people aren’t using Reddit or Adobe or choosing their cars firmware based on trust. People choose what is affordable and convenient.
In civil engineering public works are certified by an engineer; its literally them saying if this fails i am at fault. The public is trusting the engineer to say its safe.
What’s being discussed here is the hiring of engineers rather than consumer choices. Hiring an engineer is absolutely an expression of trust. The business trusts that the engineer will be able to concretely realize abstract business goals, and that they will be able to troubleshoot any deviations.
AI writing code is one thing, but intuitively trusting that an AI will figure out what you want for you and keep things running is a long way off.
In my hometown there’s two types of public transit: municipal and commercial. I was surprised to learn that a lot of folk, even the younger ones, only travel by former, even though the commercials are a lot faster, frequent and more comfortable. When asked why, the answer is the same: If anything happens on municipal transport - you can sue the transport company and even the city itself. If anything happens on a commercial line - there’s only a migrant driver and “Individual Enterpreneur John Doe” with a few leased buses to his name. Trust definitely plays a factor here, but you’re right that it’s definitely not based on technical knowledge.
Hmm. I’ve never thought about it that way. It took a long time for engineering to become that way IIRC - in the past anybody could build a bridge. The main obstacle to this, then, is that people might be a bit too risk-tolerant around AI at first. Hopefully this is where it ends up going, though.
As someone who works on the city side of development review, I can firmly say I’ll trust a puppy alone with my dinner than a Civil Engineer.
I just used copilot for the first time. It made me a ton of call to action text and website page text for various service pages inwas creating for a home builder. It was surprisingly useful, of course I modified the output a bit but overall saved me a ton of time.
Copilot has cut my workload by about 40% freeing me up for personal projects
Copilot is only dangerous in the hands of people who couldn’t program otherwise. I love it, it’s helped a ton on tedious tasks and really is like a pair programmer