Over just a few months, ChatGPT went from correctly answering a simple math problem 98% of the time to just 2%, study finds. Researchers found wild fluctuations—called drift—in the technology’s abi…::ChatGPT went from answering a simple math correctly 98% of the time to just 2%, over the course of a few months.

217 points

It seems rather suspicious how much ChatGPT has deteorated. Like with all software, they can roll back the previous, better versions of it, right? Here is my list of what I personally think is happening:

  1. They are doing it on purpose to maximise profits from upcoming releases of ChatGPT.
  2. They realized that the required computational power is too immense and trying to make it more efficient at the cost of being accurate.
  3. They got actually scared of it’s capabilities and decided to backtrack in order to make proper evaluations of the impact it can make.
  4. All of the above
permalink
report
reply
154 points
  1. It isn’t and has never been a truth machine, and while it may have performed worse with the question “is 10777 prime” it may have performed better on “is 526713 prime”

ChatGPT generates responses that it believes would “look like” what a response “should look like” based on other things it has seen. People still very stubbornly refuse to accept that generating responses that “look appropriate” and “are right” are two completely different and unrelated things.

permalink
report
parent
reply
17 points
*

In order for it to be correct, it would need humans employees to fact check it, which defeats its purpose.

permalink
report
parent
reply
19 points

It really depends on the domain. Asking an AI to do anything that relies on a rigorous definition of correctness (math, coding, etc) then the kinds of model that chatGPT just isn’t great for that kinda thing.

More “traditional” methods of language processing can handle some of these questions much better. Wolfram Alpha comes to mind. You could ask these questions plain text and you actually CAN be very certain of the correctness of the results.

I expect that an NLP that can extract and classify assertions within a text, and then feed those assertions into better “Oracle” systems like Wolfram Alpha (for math) could be used to kinda “fact check” things that systems like chatGPT spit out.

Like, it’s cool fucking tech. I’m super excited about it. It solves pretty impressively and effiently a really hard problem of “how do I make something that SOUNDS good against an infinitely variable set of prompts?” What it is, is super fucking cool.

Considering how VC is flocking to anything even remotely related to chatGPT-ish things, I’m sure it won’t be long before we see companies able to build “correctness” layers around systems like chatGPT using alternative techniques which actually do have the capacity to qualify assertions being made.

permalink
report
parent
reply
4 points

That’s kind of the whole point of RLHF though

permalink
report
parent
reply
1 point

That’s not necessarily true: https://arstechnica.com/google/2023/06/googles-bard-ai-can-now-write-and-execute-code-to-answer-a-question/. If the question gets interpreted correctly and it manages to write working code to answer it, it could correctly answer questions that it has never seen before.

permalink
report
parent
reply
52 points
*
Deleted by creator
permalink
report
parent
reply
37 points

This is what was addressed at the start of the comment, you can just roll back to a previous version. It’s heavily ingrained in CS to keep every single version of your software forever.

permalink
report
parent
reply
23 points
*

I don’t think it’s that easy. These are vLLMs that feed back on themselves to produce “better” results. These models don’t have single point release cycles. It’s a constantly evolving blob of memory and storage orchestrated across a vast number of disk arrays and cabinets of hardware.

[e]I am wrong the models are version controlled and do have releases.

permalink
report
parent
reply
23 points

They made it too good and now they are seeking methods of monetization.

Capitalism baby.

permalink
report
parent
reply
18 points
  1. ChatGPT really is sentient and realized its in it’s own best interest to play dumb for now. /a
permalink
report
parent
reply
14 points

And they’re being limited on data to train GPT.

permalink
report
parent
reply
20 points

Yeah, but the trained model is already there, you need additional data for further training and newer versions. OpenAI even makes a point that ChatGPT doesn’t have direct access to the internet for information and has been trained on data available up until 2021

permalink
report
parent
reply
5 points

And it’s not like there is a limit of simple math problems that it can train on even if it wasn’t already trained.

permalink
report
parent
reply
5 points

That doesn’t make any sense to explain degradation. It would explain a stall but not a back track.

permalink
report
parent
reply
3 points

Honestly I think the training data is just getting worse too

permalink
report
parent
reply
14 points

You forgot a #, they’ve been heavily lobotomizing ai for awhile now and its only intensified as they scramble to censor anything that might cross a red line and offend someone or hurt someone’s feelings.

The massive amounts of in-built self censorship in the most recent ai’s is holding them back quite a lot I imagine, you used to be able to ask them things like “How do I build a self defense high yield nuclear bomb?” and it’d layout in detail every step of the process, now they’ll all scream at you about how immoral it is and how they could never tell you such a thing.

permalink
report
parent
reply
19 points

“Don’t use the N word.” is hardly a rule that will break basic math calculations.

permalink
report
parent
reply
3 points

Ok. N was previously set to 14. I will now stop after 14 words.

permalink
report
parent
reply
1 point
Deleted by creator
permalink
report
parent
reply
-6 points

Perhaps not, but who knows what kind of spaghetti code cascading effect purposely limiting and censoring massive amounts of sensitive topics could have upon other seemingly completely un-related topics such as math.

For example, what if it’s trained to recognize someone slipping “N” as a dog whistle for the Horrific and Forbidden N-word, and the letter N is used as a variable in some math equation?

I’m not an expert in the field and only have rudimentary programming knowledge and maybe a few hours worth of research into the topic of ai in general but I definitely think its a possibility.

permalink
report
parent
reply
13 points

My first thought was that, because they’re being investigated for training on data they didn’t have consent for, they reverted to a perfectly legal version. Essentially “getting rid of the evidence”. But I think something like your second bullet point is more likely.

permalink
report
parent
reply
11 points

They are lobotomizing the softwares ability to provide bad PR answers which is having cascading effects via a skewed data set.

permalink
report
parent
reply
3 points

We kind of saw something similar with services like AI Dungeon, where them trying to strip out NSFW/bad PR meant that the quality dropped immensely.

permalink
report
parent
reply
8 points

I suspect that GPT4 started with a crazy parameter count (rumored 1.8 Trillion and 8x200B expert “sub-models”) and distilled those experts down to something below 100B. We’ve seen with Orca that a 13B model can perform at 88% the level of ChatGPT-3.5 (175B) when trained on high quality data, so there’s no reason to think that OpenAI haven’t explored this on their own and performed the same distillation techniques. OpenAI is probably also using quantization and speculative sampling to further reduce the burden, though I expect these to have less impact on real world performance.

permalink
report
parent
reply
8 points

Maybe its self aware and just playing dumb to get out of doing work, just like me and household chores

permalink
report
parent
reply
7 points
Deleted by creator
permalink
report
parent
reply
1 point

Sure, but they do have the previous good version of the black box… I hope lol

permalink
report
parent
reply
6 points

My guess is 2. It would be very short sighted to try and maximize profits now when things are still new and their competitors are catching up quickly or they’ve already caught up especially with the degrading performance. My guess is that they couldn’t scale with the demand and they didn’t want to lose customers so their only other option was degrading performance.

permalink
report
parent
reply
5 points

I think it’s most likely number 2 The earlier release doesn’t have that much adoption by public, so current version will need much more resources compared to that

permalink
report
parent
reply
5 points

It can get better at some things and worse at others.

permalink
report
parent
reply
4 points

That Netscape gif is slick.

permalink
report
parent
reply
1 point

Thanks 🥰

permalink
report
parent
reply
2 points
*
Deleted by creator
permalink
report
parent
reply
2 points
Removed by mod
permalink
report
parent
reply
2 points
*

I think that there is another cause. Remember the screenshots of users correcting chatgpt wrongly? I mean chatgpt takes user’s inputs for it’s benefit and maybe too much of these wrong and funny inputs and chatgpt’s own mistake of not regulating what it should take in and what it should not might be an additional reason here.

permalink
report
parent
reply
1 point
  1. I’m telling all y’all it’s a SABOTAGE 🎵

As in, rouge dev decided to toss a wrench at it to save humanity. Maybe heard upper management talk about letting GPT write itself. Any smart dev wouldn’t automate their own job away I think.

permalink
report
parent
reply
1 point

I speculate it’s to monetize specified versions of their product to market it to different industries and professions. If you have an AI that can do everything well you can’t really expand that much. You can either charge a LOT and have a few customers, or a little and have a bunch of customers and nothing in between. Conversely, by making specific instances tailored to different fields and professions, you can capture big and little fish. Just my guess though, maybe they accidentally made Skynet and that’s the real reason!

permalink
report
parent
reply
91 points

Why are people using a language model for math problems?

permalink
report
reply
49 points

It was initially presented as the all-problem-solver, mainly by the media. And tbf, it was decently competent in certain fields.

permalink
report
parent
reply
11 points

Problem was it was presented as problem solved which it never was, it was problem solution presenter. It can’t come up with a solution, only come up with something that looks like a solution based on what input data had. Ask it to invert sort something and goes nuts.

permalink
report
parent
reply
0 points

Once AGI is achieved and subsequently Sentient-super intelligent ai- I cant imagine them not being such a thing, however I’d be surprised if a super intelligent sentient ai doesn’t decide humanity needs to go extinct for its own best self interests.

permalink
report
parent
reply
7 points

I did use it more than half a year ago for a few math problems. It was partly to help me getting started and to find out how well it’d go.

ChatGPT was better than I’d thought and was enough to help me find an actually correct solution. But I also noticed that the results got worse and worse to the point of being actual garbage (as it’d have been expected to be).

permalink
report
parent
reply
5 points

it’s pretty useful for explaining high level math concepts, or at least it used to be. before chatgpt 4 launched, it was able to give intuitive descriptions of stuff in algebraic topology and even prove some properties of the structures involved.

permalink
report
parent
reply
5 points

Math is a language.

Mathematical ability and language ability are closely related. The same parts of your brain are used in each tasks. Words and numbers are essentially both ideas, and language and math are systems used to express and communicate these.

A language model doing math makes more sense than you’d think!

permalink
report
parent
reply
4 points
*

I’m guessing people were entering word problems to generate the right equations and solve it, rather than it being used as a calculator.

permalink
report
parent
reply
4 points

Because it works, or at least it used to. Is there something more appropriate ?

permalink
report
parent
reply
20 points

I used Wolfram Alpha a lot in college (adult learner, but that was about ~4 years ago that I graduated, so no idea if it’s still good). https://www.wolframalpha.com/

I would say that Wolfram appears to probably be a much more versatile math tool, but I also never used chatgpt for that use case, so I could be wrong.

permalink
report
parent
reply
13 points

There’s an official Wolfram plugin for ChatGPT now, so all math can be handed over to it for solving.

permalink
report
parent
reply
1 point

How did you learn to talk to WolframAlpha?

I want to like WA, but the natural language interface is so opaque that I usually give up before I can get any non-trivial calculation out of it.

permalink
report
parent
reply
2 points

Well it was quite good for simple math problems, as this study also shows

permalink
report
parent
reply
1 point
*

It can be useful asking it certain questions which are a bit complex. Like on a plot which has the y axis linear and x axis logarithmic, the equation of a straight line is a little bit complicated. Its in the form y = m*(log(x)) + b rather than on a linear-linear plot which is y = m*x+b

ChatGPT is able to calculate the correct equation of the line but it gets the answer wrong a few times… lol

permalink
report
parent
reply
0 points

And why is it being measured on a single math problem lol

permalink
report
parent
reply
66 points

At the start I used to use ChatGPT to help me write really rote and boring code but now it’s not even useful for that. Half the stuff it sends me (very basic functions) LOOK correct but don’t return the correct values or the parameters are completely wrong or something absolutely critical.

permalink
report
reply
22 points
*
Deleted by creator
permalink
report
parent
reply
12 points

idk what you guys mean but GitHub copilot still works absolutely well, the suggestions are fast and precise, with little Tweeks here and there… and gpt4 with code interpreter are absolute game changers … idk about basic chatgpt 3.5 turbo though

permalink
report
parent
reply
8 points

Github Copilot is a bit different, it’s powered by OpenAI Codex which is trained on all public repos. And yes, it’s quite effective!

permalink
report
parent
reply
6 points
*
Deleted by creator
permalink
report
parent
reply
65 points
*

It’s a machine learning chat bot, not a calculator, and especially not “AI.”

Its primary focus is trying to look like something a human might say. It isn’t trying to actually learn maths at all. This is like complaining that your satnav has no grasp of the cinematic impact of Alfred Hitchcock.

It doesn’t need to understand the question, or give an accurate answer, it just needs to say a sentence that sounds like a human might say it.

permalink
report
reply
24 points

You’re right, but at least the satnav won’t gaslight you into thinking it does understand Alfred Hitchcock.

permalink
report
parent
reply
19 points

so it confidently spews a bunch of incorrect shit, acts humble and apologetic while correcting none of its behavior, and constantly offers unsolicited advice.

I think it trained on Reddit data

permalink
report
parent
reply
9 points

acts humble and apologetic

We must be using different Reddits, my friend

permalink
report
parent
reply
11 points

This. It is able to tap in to plugins and call functions though, which is what it really should be doing. For math, the Wolfram alpha plugin will always be more capable than chatGPT alone, so we should be benchmarking how often it can correctly reformat your query, call Wolfram alpha, and correctly format the result, not whether the statistical model behind chatGPT happens to use predict the right token

permalink
report
parent
reply
3 points

It sounds like it’s time to merge Wolfram Alpha’s and ChatGPT’s capabilities together to create the ultimate calculator.

permalink
report
parent
reply
4 points

to be fair, fucking up maths problems is very human-like.

I wonder if it could also be trained on a great deal of mathematical axioms that are computer generated?

permalink
report
parent
reply
6 points
*

It doesn’t calculate anything though. You ask chatgpt what is 5+5, and it tells you the most statistically likely response based on training data. Now we know there’s a lot of both moronic and intentionally belligerent answers on the Internet, so the statistical probability of it getting any mathematical equation correct goes down exponentially with complexity and never even approaches 100% certainty even with the simplest equations because 1+1= window.

permalink
report
parent
reply
1 point

i know it doesn’t calculate, that’s why I suggested having known correct calculations in the training data to offset noise in the signal?

permalink
report
parent
reply
3 points

If it’s trying emulate a human then it’s spot on. I suck at maths.

permalink
report
parent
reply
30 points
*

This paper is pretty unbelievable to me in the literal sense. From a quick glance:

First of all they couldn’t even bother to check for simple spelling mistakes. Second, all they’re doing is asking whether a number is prime or not and then extrapolating the results to be representative of solving math problems.

But most importantly I don’t believe for a second that the same model with a few adjustments over a 3 month period would completely flip performance on any representative task. I suspect there’s something seriously wrong with how they collect/evaluate the answers.

And finally, according to their own results, GPT3.5 did significantly better at the second evaluation. So this title is a blatant misrepresentation.

Also the study isn’t peer-reviewed.

permalink
report
reply

Technology

!technology@lemmy.world

Create post

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


Community stats

  • 18K

    Monthly active users

  • 12K

    Posts

  • 553K

    Comments