3 points

This is the best summary I could come up with:


“And I try to help people understand there is an exponential here, and the unfortunate thing is you only get to sample it every couple of years because it just takes a while to build supercomputers and then train models on top of them.”

The laws suggest that simply scaling up model size and training data can lead to significant improvements in AI capabilities without necessarily requiring fundamental algorithmic breakthroughs.

The perception has been fueled by largely informal observations—and some benchmark results—about recent models like Google’s Gemini 1.5 Pro, Anthropic’s Claude Opus, and even OpenAI’s GPT-4o, which some argue haven’t shown the dramatic leaps in capability seen in earlier generations, and that LLM development may be approaching diminishing returns.

Scott’s stance suggests that tech giants like Microsoft still feel justified in investing heavily in larger AI models, betting on continued breakthroughs rather than hitting a capability plateau.

Some perceptions of slowing progress in LLM capabilities and benchmarking may be due to the rapid onset of AI in the public eye when, in fact, LLMs have been developing for years prior.

In the podcast interview, the Microsoft CTO pushed back against the idea that AI progress has stalled, but he acknowledged the challenge of infrequent data points in this field, as new models often take years to develop.


The original article contains 697 words, the summary contains 217 words. Saved 69%. I’m a bot and I’m open source!

permalink
report
reply
48 points

Microsoft CTO Kevin Scott is of course not a reliable source due to conflict of interest and his position in the US corporate world.

If anything, the fact that he is doing damage control PR around “LLM scaling laws” suggests something is amiss. Let’s see how things develop.

permalink
report
reply
32 points

Given Microsoft’s investment in OpenAI and strong marketing of its own Microsoft Copilot AI features, the company has a strong interest in maintaining the perception of continued progress, even if the tech stalls.

I believe this sums it up.

permalink
report
parent
reply
2 points
*

Yeah. There’s a very narrow corner that demands huge models, and that’s use cases where there’s no room for mistakes. That space is exciting, but also deeply bogged down in uncertainty, due both to laws and as-yet-undelivered, but 100% certainly coming-soon, law-creating-disasters.

Everywhere else, I suspect we’ve seen as good as we’re going to get, from current generation AI.

Tech firm CEOs know this too, but there’s not much interesting on the table to “bet the farm” on to court “swing for the fences” investors (gullible suckers) right now.

permalink
report
parent
reply
1 point
*

While I agree about the conflict of interest, I would largely say the same thing despite no such conflict of interest. However I see intelligence as a modular and many dimensional concept. If it scales as anticipated, it will still need to be organized into different forms of informational or computational flow for anything resembling an actively intelligent system.

On that note, the recent developments with active inference like RXinfer are astonishing given the current level of attention being paid. Seeing how llms are being treated, I’m almost glad it’s not being absorbed into the hype and hate cycle.

permalink
report
parent
reply
10 points

TLDR: he thinks the techniques are fine and you can just brute force them for the foreseeable future.

permalink
report
reply
1 point

Yeah… Why does it sound dumb in so many level?

permalink
report
parent
reply
1 point

lol I honestly needed to open the article to parse the title. That’s why I posted.

But I’m definitely of the belief that you need a hell of a lot more architecture than they have to go meaningfully further. Humans are a hell of a lot more complicated than a bit like of neurons.

permalink
report
parent
reply
4 points

Because he’s a salesman, and he’s selling you bullshit.

What the experts are now saying is that it looks like the LLM approach to AI will require exponentially larger amounts of training data (and data processing) to achieve linear growth. Next generation AI models will cost ten times as much to train, and the generation after that will cost ten times as much again.

The whole thing is a giant con. Kevin is just trying to keep investor confidence floating for a little longer.

permalink
report
parent
reply
1 point

And the harder the sell, the worse the product.

permalink
report
parent
reply

Technology

!technology@lemmy.world

Create post

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


Community stats

  • 18K

    Monthly active users

  • 12K

    Posts

  • 553K

    Comments