204
this post was submitted on 25 Feb 2024
204 points (84.0% liked)
Technology
60008 readers
2639 users here now
This is a most excellent place for technology news and articles.
Our Rules
- Follow the lemmy.world rules.
- Only tech related content.
- Be excellent to each another!
- Mod approved content bots can post up to 10 articles per day.
- Threads asking for personal tech support may be deleted.
- Politics threads may be removed.
- No memes allowed as posts, OK to post as comments.
- Only approved bots from the list below, to ask if your bot can be added please contact us.
- Check for duplicates before posting, duplicates may be removed
Approved Bots
founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
Well. That's stupid.
Large language models are amazingly useful coding tools. They help developers write code more quickly.
They are nowhere near being able to actually replace developers. They can't know when their code doesn't make sense (which is frequently). They can't know where to integrate new code into an existing application. They can't debug themselves.
Try to replace developers with an MBA using a large language model AI, and once the MBA fails, you'll be hiring developers again - if your business still exists.
Every few years, something comes along that makes bean counters who are desperate to cut costs, and scammers who are desperate for a few bucks, declare that programming is over. Code will self-write! No-code editors will replace developers! LLMs can do it all!
No. No, they can't. They're just another tool in the developer toolbox.
I've been a developer for over 20 years and when I see Autogen generate code, decide to execute that code and then fix errors by making a decision to install dependencies, I can tell you I'm concerned. LLMs are a tool, but a tool that might evolve to replace us. I expect a lot of software roles in ten years to look more like an MBA that has the ability to orchestrate AI agents to complete a task. Coding skills will still matter, but not as much as soft skills will.
I really don't see it.
Think about a modern application. Think about the file structure, how the individual sources interrelate, how non-code assets are stored, how applications are deployed, and all the other bits and pieces that go into an application. An AI can't know any of that without being trained - by a human - on the specifics of that application's needs.
I use Copilot for my job. It's very nice, and makes my job easier. And if my boss fired me and the rest of the team and tried to do it himself, the application would be down in a day, then irrevocably destroyed in a week. Then he'd be fired, we'd be rehired, and we - unlike my now-former boss - would know things like how to revert the changes he made when he broke everything while trying to make Copilot create a whole new feature for the application.
AI code generation is pretty cool, but without the capacity to know what code actually should be generated, it's useless.
It's just going to create a summary story about the code base and reference that story as it implements features, not that different that a human. It's not necessarily something it can do now but it will come. Developers are not special, and I was never talking about Copilot.
I don't think most people grok just how hard implementing that kind of joined-up thinking and metacognition is.
You're right, developers aren't special, except in those ways all humans are, but we're a very long way indeed from being able to simulate them in AI - especially in large language models. Humans automatically engage in joined-up thinking, second-order logic, and so on, without having to consciously try. Those are all things a large language model literally can't do.
It doesn't know anything. It can't conceptualize a "summary story," or understand parts that it might get wrong in such a story. It's glorified autocomplete.
And that can be extraordinarily useful, but only if we're honest with ourselves about what it is and is not capable of.
Companies that decide to replace their developers with one guy using ChatGPT or Gemini or something will fail, and that's going to be true for the foreseeable future.
Try for a second to think beyond what they're able to do now and think about the future. Also, educate yourself on Autogen and CrewAI, you actually haven't addressed anything I said because you're too busy pontificating.
I am. In the future, they will need to be able to perform tasks using joined-up thinking, second-order logic, and metacognition if they're going to replace people like me with AI. And that is a very hard goal to achieve. Maybe not P = NP hard, but by no means trivial.
I have. My company looked at Autogen. We concluded it wasn't worth it. The solution to AI agents not being able to actually understand what they're doing isn't to amplify the problem by creating teams of them.
Every few years, something new comes along driven by incredible hype, and people declare programming to be dead. They insist a robot will be able to do my job. I have yet to see a technology that will plausibly do that in ten years, let alone now. And all the hype is built on a foundation of ignorance over how complicated a modern, enterprise-ready application is, and how necessary being able to think about its many moving parts is.
You know who doesn't suffer from that ignorance? Microsoft, the creators of Autogen. And they're currently hiring developers, not laying them off and replacing them with Autogen.
Well, I sometimes see a few tools at my job, which are supposed to be kinda usable by people like that. In reality they can't 90% of time.
That'd be because many people think that engineers deal in intermediate technical details, and the general idea is clear for this MBA. In fact it's not.
Ok.