this post was submitted on 22 Dec 2024
410 points (96.2% liked)

Technology

60060 readers
2959 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] [email protected] 63 points 23 hours ago* (last edited 23 hours ago) (16 children)

There is this seeming need to discredit AI from some people that goes overboard. Some friends and family who have never really used LLMs outside of Google search feel compelled to tell me how bad it is.

But generative AIs are really good at tasks I wouldn't have imagined a computer doing just a few year ago. Even if they plateaued in place where they are right now it would lead to major shakeups in humanity's current workflow. It's not just hype.

The part that is over hyped is companies trying to jump the gun and wholesale replace workers with unproven AI substitutes. And of course the companies who try to shove AI where it doesn't really fit, like AI enabled fridges and toasters.

[–] [email protected] 5 points 4 hours ago* (last edited 4 hours ago)

Goldman Sachs, quote from the article:

“AI technology is exceptionally expensive, and to justify those costs, the technology must be able to solve complex problems, which it isn’t designed to do.”

Generative AI can indeed do impressive things from a technical standpoint, but not enough revenue has been generated so far to offset the enormous costs. Like for other technologies, It might just take time (remember how many billions Amazon burned before turning into a cash-generating machine? And Uber has also just started turning some profit) + a great deal of enshittification once more people and companies are dependent. Or it might just be a bubble.

As humans we're not great at predicting these things including of course me. My personal prediction? A few companies will make money, especially the ones that start selling AI as a service at increasingly high costs, many others will fail and both AI enthusiasts and detractors will claim they were right all along.

[–] [email protected] 2 points 5 hours ago

Like what outcome?

I have seen gains on cell detection, but it's "just" a bit better.

[–] [email protected] 16 points 16 hours ago

This is easy to say about the output of AIs.... if you don't check their work.

Alas, checking for accuracy these days seems to be considered old fogey stuff.

[–] [email protected] 63 points 23 hours ago (1 children)

The part that is over hyped is companies trying to jump the gun and wholesale replace workers with unproven AI substitutes. And of course the companies who try to shove AI where it doesn't really fit, like AI enabled fridges and toasters.

This is literally the hype. This is the hype that is dying and needs to die. Because generative AI is a tool with fairly specific uses. But it is being marketed by literally everyone who has it as General AI that can "DO ALL THE THINGS!" which it's not and never will be.

[–] [email protected] 7 points 18 hours ago (1 children)

The obsession with replacing workers with AI isn't going to die. It's too late. The large financial company that I work for has been obsessively tracking hours saved in developer time with GitHub Copilot. I'm an older developer and I was warned this week that my job will be eliminated soon.

[–] [email protected] 4 points 14 hours ago* (last edited 14 hours ago) (1 children)

The large financial company that I work for

So the company that is obsessed with money that you work for has discovered a way to (they think) make more money by getting rid of you and you're surprised by this?

At least you've been forewarned. Take the opportunity to abandon ship. Don't be the last one standing when the music stops.

[–] [email protected] 3 points 14 hours ago* (last edited 12 hours ago)

I never said that I was surprised. I just wanted to point out that many companies like my own are already making significant changes to how they hire and fire. They need to justify their large investment in AI even though we know the tech isn't there yet.

[–] [email protected] 38 points 23 hours ago (9 children)

Even if they plateaued in place where they are right now it would lead to major shakeups in humanity's current workflow

Like which one? Because it's now 2 years we have chatGPT and already quite a lot of (good?) models. Which shakeup do you think is happening or going to happen?

[–] [email protected] 1 points 38 minutes ago

I don’t know anything about the online news business but it certainly appears to have changed. Most of it is dreck, either way, and those organizations are not a positive contributor to society, but they are there, it is a business, and it has changed society

[–] [email protected] 10 points 22 hours ago (4 children)

Computer programming has radically changed. Huge help having llm auto complete and chat built in. IDEs like Cursor and Windsurf.

I’ve been a developer for 35 years. This is shaking it up as much as the internet did.

[–] [email protected] 27 points 21 hours ago* (last edited 21 hours ago) (6 children)

I quit my previous job in part because I couldn't deal with the influx of terrible, unreliable, dangerous, bloated, nonsensical, not even working code that was suddenly pushed into one of the projects I was working on. That project is now completely dead, they froze it on some arbitrary version.
When junior dev makes a mistake, you can explain it to them and they will not make it again. When they use llm to make a mistake, there is nothing to explain to anyone.
I compare this shake more to an earthquake than to anything positive you can associate with shaking.

[–] [email protected] 1 points 36 minutes ago

More business for me. As a DevOps guy, my job is to create automation to flag “ terrible, unreliable, dangerous, bloated, nonsensical, not even working code”

load more comments (5 replies)
[–] [email protected] 30 points 21 hours ago (8 children)

I hardly see it changed to be honest. I work in the field too and I can imagine LLMs being good at producing decent boilerplate straight out of documentation, but nothing more complex than that.

I often use LLMs to work on my personal projects and - for example - often Claude or ChatGPT 4o spit out programs that don't compile, use inexistent functions, are bloated etc. Possibly for languages with more training (like Python) they do better, but I can't see it as a "radical change" and more like a well configured snippet plugin and auto complete feature.

LLMs can't count, can't analyze novel problems (by definition) and provide innovative solutions...why would they radically change programming?

[–] [email protected] 0 points 3 hours ago (1 children)

I hardly see it changed to be honest. I work in the field too and I can imagine LLMs being good at producing decent boilerplate straight out of documentation, but nothing more complex than that.

I think one of the top lists on advent of code this year is a cheater that fully automated the solutions using LLMs. Not sure which LLM though, I use LLMs quite a bit and ChatGPT 4o frequently tells me nonsense like "perhaps subtracting by zero is affecting your results" (issues I thought were already gone in GPT 4, but I guess not, Sonnet 3.5 does a bit better in this regard).

[–] [email protected] 2 points 2 hours ago

Maybe some postmortem analysis will be interesting. The AoC is also a context in which the domain is self-contained and there is probably a ton of training material on similar problems and tasks. I can imagine LLM might do decently there.

Also there is no big consequence if they don't and it's probably possible to bruteforce (which is how many programming tasks have been solved).

load more comments (7 replies)
load more comments (2 replies)
load more comments (7 replies)
[–] [email protected] 19 points 20 hours ago (20 children)

Computers have always been good at pattern recognition. This isn't new. LLM are not a type of actual AI. They are programs capable of recognizing patterns and Loosely reproducing them in semi randomized ways. The reason these so-called generative AI Solutions have trouble generating the right number of fingers. Is not only because they have no idea how many fingers a person is supposed to have. They have no idea what a finger is.

The same goes for code completion. They will just generate something that fills the pattern they're told to look for. It doesn't matter if it's right or wrong. Because they have no concept of what is right or wrong Beyond fitting the pattern. Not to mention that we've had code completion software for over a decade at this point. Llms do it less efficiently and less reliably. The only upside of them is that sometimes they can recognize and suggest a pattern that those programming the other coding helpers might have missed. Outside of that. Such as generating act like whole blocks of code or even entire programs. You can't even get an llm to reliably spit out a hello world program.

load more comments (20 replies)
load more comments (10 replies)