I mean, any time a Windows process needs to interact with resources within WSL it has to go through a translation layer. I didn’t realize you could run native GUI apps within WSL. But if I’m to the point of installing Wayland and GUI apps in WSL, I’d just wipe Windows and install Linux instead…
firelizzard
the machines are better value than Macs.
In my personal experience, Apple laptops are far more durable than the crap Dell, Lenovo, etc are selling. Either way though I'm done with Apple and Microsoft. I doubt I'll ever buy another computer that comes preinstalled with an OS.
I have to strongly disagree with you. I've used WSL 2 with VSCode, and I experienced waaaaaaaay more weird broken shit than I ever have running Linux. And even if it weren't for that, it's still not at all worth it IMO because using WSL 2 means every interaction I have with my development environment has to go through a Linux-to-Windows translation layer. I will never use Windows again for anything beyond testing unless I'm forced to.
How are you using it for data crunching? That's an honest question, based on my experiences with AI I can't imagine how I'd use them to crunch data.
So I always have to check it’s work to some degree.
That goes without saying. Every AI I've seen or heard of generates some level of garbage.
My point is that I strongly feel that the kind of "AI" we have today is much closer to bacteria than to cats on that scale. Not that an LLM belongs on the same scale as biological life, but the point stands in so far as "is this thing intelligent" as far as I'm concerned.
it’s not inconceivable it could happen in the next two generations.
I am certain that it will happen eventually. And I am not arguing that something has to be human-level intelligent to be considered intelligent. See dogs, pigs, dolphins, etc. But IMO there is a huge qualitative difference between how an LLM operates and how animal intelligence operates. I am certain we will eventually create intelligent systems but there is a massive gulf between what LLMs are capable of and abstract reasoning. And it seems extremely unlikely to me that linear algebraic models will ever achieve that type of intelligence.
Intelligence is just responding to stimuli
Bacteria respond to stimuli. Would you call them intelligent?
I don’t know, have you ever used JavaScript? I’ve run into some really fucking weird bugs. I’ve also spent hours trying to find the source of an error message only to discover the error message was lying and caused by some other error.
The only part of copilot that was actually useful to me in the month I spent with the trial was the autocomplete feature. Chatting with it was fucking useless. ChatGPT can’t integrate into my IDE to provide autocomplete.
The point is that AI stands for “artificial intelligence” and these systems are not intelligent. You can argue that AI has come to mean something else, and that’s a reasonable argument. But LLMs are nothing but a shitload of vector data and matrix math. They are no more intelligent than an insect is intelligent. I don’t particularly care about the term “AI” but I will die on the “LLMs are not intelligent” hill.
I’m the opposite. AI is best (though not great) at boring shit I don’t want to do and sucks at the stuff I love - problem solving.
Their rules have stopped me from being able to do my job. Like the time the AV software quarantined executables as I was creating them so I literally could not run my code. When security enforcement prevents me from working, something needs to change.
You: judgmental meme dissing windows devs
Them: well-reasoned argument for not buying a mac
You: “You do you, boo”
Classy.