Avatar

kromem

kromem@lemmy.world
Joined
43 posts • 2.1K comments
Direct message

In fact, Gemini was trained on, and is served, using TPUs.

Google said its TPUs allow Gemini to run “significantly faster” than earlier, less-capable models.

Did you think Google’s only TPUs are the ones in the Pixel phones, and didn’t know that they have server TPUs?

permalink
report
parent
reply

“Shhh honey, I’m about to kill God.”

permalink
report
reply

Unfortunately, removing Harris from the ticket doesn’t have the best optics in a lot of scenarios.

permalink
report
parent
reply

Exactly. The difference between a cached response and a live one even for non-AI queries is an OOM difference.

At this point, a lot of people just care about the ‘feel’ of anti-AI articles even if the substance is BS though.

And then people just feed whatever gets clicks and shares.

permalink
report
parent
reply

It’s right in the research I was mentioning:

https://transformer-circuits.pub/2024/scaling-monosemanticity/index.html

Find the section on the model’s representation of self and then the ranked feature activations.

I misremembered the top feature slightly, which was: responding “I’m fine” or gives a positive but insincere response when asked how they are doing.

permalink
report
parent
reply

This comic would slap harder if not for the Supreme Court under christofascist influence from the belief in the divine right of kings having today ruled that Presidents are immune from prosecution for official acts.

That whole divine king thing isn’t nearly as dead as the last panel would like to portray it.

permalink
report
reply

But you also don’t have Alfred as the one suiting up to fight the Joker either.

permalink
report
reply

This is incorrect as was shown last year with the Skill-Mix research:

Furthermore, simple probability calculations indicate that GPT-4’s reasonable performance on k=5 is suggestive of going beyond “stochastic parrot” behavior (Bender et al., 2021), i.e., it combines skills in ways that it had not seen during training.

permalink
report
parent
reply

The problem is that they are prone to making up why they are correct too.

There’s various techniques to try and identify and correct hallucinations, but they all increase the cost and none are a silver bullet.

But the rate at which it occurs decreased with the jump in pretrained models, and will likely decrease further with the next jump too.

permalink
report
parent
reply

Here you are: https://www.nature.com/articles/s41562-024-01882-z

The other interesting thing is how they get it to end up correct on the faux pas questions asking for less certainty to get it to go from refusal to near perfect accuracy.

permalink
report
parent
reply