You are viewing a single thread.
View all comments
84 points
*

Useful in the way that it increases emissions and hopefully leads to our demise because that’s what we deserve for this stupid technology.

permalink
report
reply
64 points

Surely this is better than the crypto/NFT tech fad. At least there is some output from the generative AI that could be beneficial to the whole of humankind rather than lining a few people’s pockets?

permalink
report
parent
reply
36 points

Unfortunately crypto is still somehow a thing. There is a couple year old bitcoin mining facility in my small town that brags about consuming 400MW of power to operate and they are solely owned by a Chinese company.

permalink
report
parent
reply
7 points

I recently noticed a number of bitcoin ATMs that have cropped up where I live - mostly at gas stations and the like. I am a little concerned by it.

permalink
report
parent
reply
-2 points

It takes living with a broken system to understand the fix for it. There are millions of people who have been saved by Bitcoin and the freedom that it brings, they are just mainly in the 2nd and 3rd worlds, so to many people they basically don’t exist.

permalink
report
parent
reply
-22 points
*

I’m crypto neutral.

But it’s really strange how anti-crypto ideologues don’t understand that the system of states printing money is literally destroying the planet. They can’t see the value of a free, fair, decentralized, automatable, accounting systems?

Somehow delusional chatbots wasting energy and resources are more worthwhile?

permalink
report
parent
reply
27 points

Printing currency isn’t destroying the planet…the current economic system is doing that, which is the same economic system that birthed crypto.

Governments issuing currency goes back to a time long before our current consumption at all cost economic system was a thing.

permalink
report
parent
reply
9 points

I’m fine doing away with physical dollars printed on paper and coins but crypto seems to solve none of the problems that we have with a fiat currency but instead continues to consume unnecessary amounts of energy while being driven by rich investors that would love nothing more than to spend and earn money in an untraceable way.

permalink
report
parent
reply
5 points

While the consumption for AI train can be large, there are arguments to be made for its net effect in the long run.

The article’s last section gives a few examples that are interesting to me from an environmental perspective. Using smaller problem-specific models can have a large effect in reducing AI emissions, since their relation to model size is not linear. AI assistance can indeed increase worker productivity, which does not necessarily decrease emissions but we have to keep in mind that our bodies are pretty inefficient meat bags. Last but not least, AI literacy can lead to better legislation and regulation.

permalink
report
parent
reply
13 points

The argument that our bodies are inefficient meat bags doesn’t make sense. AI isn’t replacing the inefficient meat bag unless I’m unaware of an AI killing people off and so far I’ve yet to see AI make any meaningful dent in overall emissions or research. A chatgpt query can use 10x more power than a regular Google search and there is no chance the result is 10x more useful. AI feels more like it’s adding to the enshittification of the internet and because of its energy use the enshittification of our planet. IMO if these companies can’t afford to build renewables to support their use then they can fuck off.

permalink
report
parent
reply
3 points
*

Using smaller problem-specific models can have a large effect in reducing AI emissions

Sure, if you consider anything at all to be “AI”. I’m pretty sure my spellchecker is relatively efficient.

AI literacy can lead to better legislation and regulation.

What do I need to read about my spellchecker? What legislation and regulation does it need?

permalink
report
parent
reply
4 points
*

Theoretically we could slow down training and coast on fine-tuning existing models. Once the AI’s trained they don’t take that much energy to run.

Everyone was racing towards “bigger is better” because it worked up to GPT4, but word on the street is that raw training is giving diminishing returns so the massive spending on compute is just a waste now.

permalink
report
parent
reply
2 points

Issue is, we’re reaching the limits of what GPT technologies can do, so we have to retrain them for the new ones, and currently available data have been already poisoned by AI generated garbage, which will make the adaptation of new technologies harder.

permalink
report
parent
reply
2 points

It’s a bit more complicated than that.

New models are sometimes targeting architecture improvements instead of pure size increases. Any truly new model still needs training time, it’s just that the training time isn’t going up as much as it used to. This means that open weights and open source models can start to catch up to large proprietary models like ChatGPT.

From my understanding GPT 4 is still a huge model and the best performing. The other models are starting to get close though, and can already exceed GPT 3.5 Turbo which was the previous standard to beat and is still what a lot of free chatbots are using. Some of these models are still absolutely huge though, even if not quite as big as GPT 4. For example Goliath is 120 billion parameters. Still pretty chonky and intensive to run even if it’s not quite GPT 4 sized. Not that anyone actually knows how big GPT 4 is. Word on the street is it’s a MoE model like Mixtral which run faster than a normal model for their size, but again no one outside Open AI actually can say with certainty.

You generally find that Open AI models are larger and slower. Wheras the other models focus more on giving the best performance at a given size as training and using huge models is much more demanding. So far the larger Open AI models have done better, but this could change as open source models see a faster improvement in the techniques they use. You could say open weights models rely on cunning architectures and fine tuning versus Open AI uses brute strength.

permalink
report
parent
reply

Technology

!technology@lemmy.world

Create post

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


Community stats

  • 17K

    Monthly active users

  • 12K

    Posts

  • 554K

    Comments