You are viewing a single thread.
View all comments View context
0 points

I think you live in a nonsense world. I literally use it everyday and yes, sometimes it’s shit and it’s bad at anything that even requires a modicum of creativity. But 90% of shit doesn’t require a modicum of creativity. And my point isn’t about where we’re at, it’s about how far the same tech progressed on another domain adjacent task in three years.

Lemmy has a “dismiss AI” fetish and does so at its own peril.

permalink
report
parent
reply
10 points

And my point isn’t about where we’re at, it’s about how far the same tech progressed on another domain adjacent task in three years.

First off, are you extrapolating the middle part of the sigmoid thinking it’s an exponential. Secondly, https://link.springer.com/content/pdf/10.1007/s11633-017-1093-8.pdf

permalink
report
parent
reply
3 points
*

I’ve written something vague in another place in this thread which seemed a good enough argument. But I didn’t expect that someone is going to link a literal scientific publication in the same very direction. Thank you, sometimes arguing in the Web is not a waste of time.

EDIT: Have finished reading it. Started thinking it was the same argument, in the middle got confused, in the end realized that yes, it’s the same argument, but explained well by a smarter person. A very cool article, and fully understandable for a random Lemming at that.

permalink
report
parent
reply
-5 points

Dismiss at your own peril is my mantra on this. I work primarily in machine vision and the things that people were writing on as impossible or “unique to humans” in the 90s and 2000s ended up falling rapidly, and that generation of opinion pieces are now safely stored in the round bin.

The same was true of agents for games like go and chess and dota. And now the same has been demonstrated to be coming true for languages.

And maybe that paper built in the right caveats about “human intelligence”. But that isn’t to say human intelligence can’t be surpassed by something distinctly inhuman.

The real issue is that previously there wasn’t a use case with enough viability to warrant the explosion of interest we’ve seen like with transformers.

But transformers are like, legit wild. It’s bigger than UNETs. It’s way bigger than ltsm.

So dismiss at your own peril.

permalink
report
parent
reply
5 points

But that isn’t to say human intelligence can’t be surpassed by something distinctly inhuman.

Tell me you haven’t read the paper without telling me you haven’t read the paper. The paper is about T2 vs. T3 systems, humans are just an example.

permalink
report
parent
reply
10 points

And I wouldn’t know where to start using it. My problems are often of the “integrate two badly documented company-internal APIs” variety. LLMs can’t do shit about that; they weren’t trained for it.

They’re nice for basic rote work but that’s often not what you deal with in a mature codebase.

permalink
report
parent
reply
-7 points

Again, dismiss at your own peril.

Because “Integrate two badly documented APIs” is precisely the kind of tasks that even the current batch of LLMs actually crush.

And I’m not worried about being replaced by the current crop. I’m worried about future frameworks on technology like greyskull running 30, or 300, or 3000 uniquely trained LLMs and other transformers at once.

permalink
report
parent
reply
-2 points
*

I’m with you. I’m a Senior software engineer and copilot/chatgpt have all but completely replaced me googling stuff, and replaced 90% of the time I’ve spent writing the code for simple tasks I want to automate. I’m regularly shocked at how often copilot will accurately auto complete whole methods for me. I’ve even had it generate a whole child class near perfectly, although this is likely primarily due to being very consistent with my naming.

At the very least it’s an extremely valuable tool that every programmer should get comfortable with. And the tech is just in it’s baby form. I’m glad I’m learning how to use it now instead of pooh-poohing it.

permalink
report
parent
reply
5 points
*

Are you a software developer? Or a hardware engineer? EDIT: Or anyone credible in evaluating my nonsense world against yours?

permalink
report
parent
reply
-6 points

Machine learning scientist.

permalink
report
parent
reply
6 points

That explains your optimism. Code generation is at a stage where it slaps together Stack Overflow answers and code ripped off from GitHub for you. While that is quite effective to get at least a crappy programmer to cobble together something that barely works, it is a far cry from having just anyone put out an idea in plain language and getting back code that just does it. A programmer is still needed in the loop.

I’m sure I don’t have to explain to you that AI development over the decades has often reached plateaus where the approach needed to be significantly changed in order for progress to be made, but it could certainly be the case where LLMs (at least as they are developed now) aren’t enough to accomplish what you describe.

permalink
report
parent
reply
-7 points

So close, but not there.

OK, you’ll know that I’m right when you somewhat expand your expertise to neighboring areas. Should happen naturally.

permalink
report
parent
reply

Technology

!technology@lemmy.world

Create post

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


Community stats

  • 18K

    Monthly active users

  • 12K

    Posts

  • 531K

    Comments