151 points
*

I highly doubt that unless they invented magic.

Edit: oh… They ommitted the “up to” in the headline.

permalink
report
reply
37 points

Added it

permalink
report
parent
reply
35 points
*

I meant tech radar. Thanks

permalink
report
parent
reply
124 points

This change is likened to expanding a CPU from a one-lane road to a multi-lane highway

This analogy just pegged the bullshit meter so hard I almost died of eyeroll.

permalink
report
reply
12 points

You’ve got to be careful with rolling your eyes, because the parallelism of the two eyes means that the eye roll can be twice as powerful ^1


(1) If measured against the silly baseline of a single eyeroll

permalink
report
parent
reply
7 points

Apparently the percentage of people actually understanding what they are doing in the management part of the industry is now too low to filter out even such bullshit.

permalink
report
parent
reply
96 points

The TL;DR for the article is that the headline isn’t exactly true. At this moment in time their PPU can potentially double a CPU’s performance - the 100x claim comes with the caveat of “further software optimisation”.


Tbh, I’m sceptical of the caveat. It feels like me telling someone I can only draw a stickman right now, but I could paint the Mona Lisa with some training.

Of course that could happen, but it’s not very likely to - so I’ll believe it when I see it.

Having said that they’re not wrong about CPU bottlenecks and the slowed rate of CPU performance improvements - so a doubling of performance would be huge in this current market.

permalink
report
reply
28 points

Putting the claim instead of the reality in the headline is journalistic malpractice. 2x for free is still pretty great tho.

permalink
report
parent
reply
31 points

Just finished the article, it’s not for free at all. Chips need to be designed to use it. I’m skeptical again. There’s no point IMO. Nobody wants to put the R&D into massively parallel CPUs when they can put that effort into GPUs.

permalink
report
parent
reply
7 points

Not every problem is amenable to GPUs. If it has a lot of branching, or needs to fetch back and forth from memory a lot, GPUs don’t help.

Now, does this thing have exactly the same limitations? I’m guessing yes, but it’s all too vague to know for sure. It’s sounds like they’re doing what superscalar CPUs have done for a while. On x86, that starts with the original Pentium from 1993, and Crays going back to the '60s. What are they doing to supercharge this idea?

Does this avoid some of security problems that have popped up with superscalar archs? For example, some kernel code running at ring 0 is running alongside userspace code, and it all gets the same ring 0 level as a result.

permalink
report
parent
reply

I get that we have to impress shareholders, but why can’t they just be honest and say it doubles CPU performance with the chance of even further improvement with software optimization. Doubling performance of the same hardware is still HUGE.

permalink
report
parent
reply
0 points

They… they did?

permalink
report
parent
reply
0 points

Not in the title

permalink
report
parent
reply
5 points

I’m just glad there are companies that are trying to optimize current tech rather than just piling over new hardware every damn year with forced planned obsolescence.

Though the claim is absurd, I think double the performance is NEAT.

permalink
report
parent
reply
6 points

This is new hardware piling. What they claim to do requires reworking manufacturing, is not retroactive with current designs, and demands more hardware components. It is basically a hardware thread scheduler. Cool idea, but it won’t save us from planned obsolescence, if anything it is more incentive for more waste.

permalink
report
parent
reply
2 points

Ah, good ol’ magic wishful thinking…

permalink
report
parent
reply
50 points
*

Why is this bullshit upvoted?
Already the first sentence, they change from the headline “without recoding” to “with further optimization”.
Then the explanation “a companion chip that optimizes processing tasks in real-time”
This is already done at compiler level and internally in any modern CPU for more than a decade.

It might be possible to some degree for some specific forms of code, like maybe Java. But generally for the CPU this is bullshit, and the headline is decidedly dishonest.

permalink
report
reply
44 points

Cybercriminals are creaming their jorts at the potential exploits this might open up.

permalink
report
reply
36 points

Please, hackers wear cargo shorts and toe shoes sir

permalink
report
parent
reply
14 points
*

Oof. But yeah. Fair.

I want to go on record that sometimes I just wear sandals with socks.

permalink
report
parent
reply
7 points

Truly! The scum of the earth!

permalink
report
parent
reply

Technology

!technology@lemmy.world

Create post

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


Community stats

  • 17K

    Monthly active users

  • 10K

    Posts

  • 466K

    Comments