Reverting to RAM sticks is good, but not shutting down GPU line. GPU market needs more competiter, not less.
Technology
This is a most excellent place for technology news and articles.
Our Rules
- Follow the lemmy.world rules.
- Only tech related content.
- Be excellent to each another!
- Mod approved content bots can post up to 10 articles per day.
- Threads asking for personal tech support may be deleted.
- Politics threads may be removed.
- No memes allowed as posts, OK to post as comments.
- Only approved bots from the list below, to ask if your bot can be added please contact us.
- Check for duplicates before posting, duplicates may be removed
Approved Bots
Intel can't afford to keep making GPUs because it doesn't have the reliable CPU side to soak up the losses. The GPU market has established players and Intel, besides being a big name, didn't bring much to the table to build a place for itself in the market. Outside of good Linux support (I've heard, but not personally used) the Intel GPUs don't stand out for price or performance.
Intel is struggling with its very existence and doesn't have the money or time to explore new markets when their primary product is cratering their own revenue. Intel has a very deep problem with how it is run and will most likely be unable to survive as-is for much longer.
As a Linux user of an Intel Arc card. I can safely say that the support is outstanding. In terms of price to performance, I think it’s pretty good too. I mainly enjoy having 16GB of VRAM and not spending $450-$500+ to get that amount like Nvidia. I know AMD also has cards around the same price that have that amount of VRAM too though
It boggles the mind that AMD realized the importance of GPUs 20 years ago when they bought ATI and in all that time Intel still doesn’t have a competitive GPU.
Intel realized it back then too, but things didn't pan out the way they wanted.
nVidia and AMD were going to merge while ATi was circling the drain. Then Jensen and Hector Ruiz got into their shitfight about who was going to be CEO of the marged AMD/nVidia (it should have been Jensen, Hector Ruiz is an idiot) which eventually terminated the merger.
AMD, desperately needing a GPU side for their 'future is fusion' plans, bought the ailing ATi at a massive premium.
Intel was waiting for ATi to circle the drain a little more before swooping in and buying them cheap, AMD beat them to it.
That’s a slightly revisionist history. ATI was by no means “circling the drain”, they had a promising new GPU architecture soon to be released, and remember this because I bought ATI stock about 6 months before the merger.
Basically there is only money at the top of the gpu range. Everything else is a budget card with razor thin margins.
AI specific chips will take off over time but even then the ship is starting to sail . These are mostly data center projects.
coming up next: Intel fires 25% of their staff, CEO gets a quarterly bonus in the millions
Gelsinger said the market will have less demand for dedicated graphics cards in the future.
No wonder Intel is in such rough shape! Gelsinger is an idiot.
Does he think that the demand for AI-accelerating hardware is just going to go away? That the requirement of fast, dedicated memory attached to a parallel processing/matrix multiplying unit (aka a discreet GPU) is just going to disappear in the next five years‽
The board needs to fire his ass ASAP and replace him with someone who has a grip on reality. Or at least someone who has a some imagination of how the future could be.
Gelsinger said the market will have less demand for dedicated graphics cards in the future.
Reminds me of decades ago when intel didn't bother getting into graphics because they said pretty soon CPUs would be powerful enough for high-performance graphics rendering lmao
The short-sightedness of Intel absolutely staggers me.
CPUs would be powerful enough for high-performance graphics rendering lmao
And then they continued making 4 core desktop CPU's, even after phones were at deca-core. 🤣🤣🤣
To be fair, the arm SOCs on phones use BigLittle cores, where it will enable/disable cores on the fly and move software around so it's either running on the Big high performance cores or the Little low power cores based on power budget needs at that second. So effectively not all of those 6+ cores would be available and in use at the same time on phones
True, but I use the phone reference to show how ridiculous it is that Intel remained on 4 cores for almost 8 years.
Even Phenom was available with 6 good cores in 2010, yet Intel remained on 4 for almost 8 years until Coffee Lake came out late 2017, but only with 6 cores against the Ryzen 8.
Intel was pumping money from their near monopoly for 7 years, letting the PC die a slow death of irrelevancy. Just because AMD FX was so horrible their 8 Buldozer cores were worse than 4 Core2 from Intel. They were even worse than AMDs own previous gen Phenom.
It was pretty obvious when Ryzen came out that the market wanted more powerful processors for desktop computers.
It's been the same "vision" since the late 90s - the CPU is the computer and everything else is peripherals.
Does he think that the demand for AI-accelerating hardware is just going to go away? That the requirement of fast, dedicated memory attached to a parallel processing/matrix multiplying unit (aka a discreet GPU) is just going to disappear in the next five years‽
Maybe the idea is to put it on the CPU/NPU instead? Hence them going so hard on AI processors in the CPU, even though basically nothing uses it.
But if he wants npu then why not buff igpu too? I mean, igpu exclusive on CPU memory is good boost, look up intel i7 8709g they put AMD Radeon vega igpu and exclusive to igpu 4gb of hbm memory, it did wonders, now when AMD is winning in apu sector, they could utilise same ideas they did in the past
Seriously putting a couple gigs of on-package graphics memory would completely change the game, especially if it does some intelligent caching and uses RAM for additional memory as needed.
I want to see what happens if Intel or AMD seriously let a generation rip with on package graphics memory for the iGPU. The only real drawback I could see is if the power/thermal budget just isn't sufficient and it ends up with wonky performance (which I have seen on an overly thin and light laptop I have in my personal fleet. It's got a Ryzen 2600 by memory that's horribly thermally limited and because of that it leaves so much performance on the table)
Probably because APU's are getting better and more pc gamers are doing handhelds and apu laptops instead of dedicated desktops. PC gaming has gotten really expensive.
This is a non comparison for at least the next 5 years. A dedicated gpu is still a better choice hands down for gaming. Even going on a lower end build an older gpu will still beat the current best apu by a good amount, but in 10 years time it may not be so necessary to need a gpu over an apu. GPUs are getting too power hungry and expensive. Gamers gonna game, but they won't all want to spend an ever increasing amount of money to get better graphics, and arc would need at least another 5 years to be competition enough to claim a worthwhile market share from amd or nvidia and that's wishful thinking. Long time to bleed money on a maybe.
I don't think Lunar lake wasn't a "mistake" so much as it was a reaction. Intel couldn't make a competitive laptop chip to go up against Apple and Qualcomm. (There is a very weird love triangle between the three of them /s.) Intel had to go to TSMC to get a chip to market that satisfied this AI Copilot+ PC market boom(or bust). Intel doesn't have the ability to make a competitive chip in that space (yet) so they had to produce lunar lake as a one off.
Intel is very used to just giving people chips and forcing them to conform their software to the available hardware. We're finally in the era where the software defines what the cpu needs to be able to do. This is probably why Intel struggles. Their old market dominant strategy doesn't work in the CPU market anymore and they've found themselves on the back foot. Meanwhile new devices where the hardware and software are deeply integrated in design keep coming out while Intel is still swinging for the "here's our chip, figure it out for us" crowd.
In contrast to their desktop offerings, looking at Intel's server offerings shows that Intel gets it. They want to give you the right chips for the right job with the right accelerators.
He's not wrong that GPUs in the desktop space are going away because SoCs are inevitably going to be the future. This isn't because the market has demanded it or some sort of conspiracy, but literally we can't get faster without chips getting smaller and closer together.
Even though I'm burnt on Nvidia and the last two CPUs and GPUs I've bought have been all AMD, I'm excited to see what Nvidia and mediatek do next as this SOC future has some really interesting upsides to it. Projects like ashai Linux proton project and apple GPTK2 have shown me the SoC future is actually right around the corner.
Turns out, the end of the x86 era is a good thing?
contrast to their desktop offerings
That's because server offerings are real money, which is why Intel isn't fucking those up.
AMD is in the same boat: they make pennies on client and gaming (including gpu), but dumptrucks of cash from selling Epycs.
IMO, the Zen 5(%) and Arrow Lake bad-for-gaming results are because uarch development from Intel and AMD are entirely focused on the customers that pay them: datacenter and enterprise.
Both of those CPU families clearly show that efficiency and a focus on extremely threaded workloads were the priorities, and what do you know, that's enterprise workloads!
end of the x86 era
I think it's less the era of x86 is ended and more the era of the x86 duopoly putting consumer/gaming workloads first has ended because, well, there's just no money there relative to other things they could invest their time and design resources in.
I also expect this to happen with GPUs: AMD has already given up, and Intel is absolutely going to do that as soon as they possibly can without it being a catastrophic self-inflicted wound (since they want an iGPU to use). nVidia has also clearly stopped giving a shit about gaming - gamers get a GPU a year or two after enterprise has cards based on the same chip, and now they charge $2000* for them - and they're often crippled in firmware/software so that they won't compete with the enterprise cards as well as legally not being allowed to use the drivers in a situation like that.
ARM is probably the consumer future, but we'll see who and with what: I desperately hope that nVidia and MediaTek end up competitive so we don't end up in a Qualcomm oops-your-cpu-is-two-years-old-no-more-support-for-you hellscape, but well, nVidia has made ARM SOCs for like, decades, and at no point would I call any of the ones they've ever shipped high performance desktop replacements.
- Yes, I know there's a down-stack option that shows up later, but that's also kinda the point: the ones you can afford will show up for you... eventually. Very much designed to push purchasers into the top end.
He’s not wrong that GPUs in the desktop space are going away because SoCs are inevitably going to be the future. This isn’t because the market has demanded it or some sort of conspiracy, but literally we can’t get faster without chips getting smaller and closer together.
While I agree with you on a technical level, I read it as Pat Gelsinger intends to stop development of discrete graphics cards after Battlemage, which is disappointing but not surprising. Intel's GPUs while incredibly impressive simply have an uphill battle for desktop users and particularly gamers to ensure every game a user wishes to run can generally run without compatibility problems.
Ideally Intel would keep their GPU department going because they have a fighting chance at holding a significant market share now that they're past the hardest hurdles, but they're in a hard spot financially so I can't be surprised if they're forced to divest from discrete GPUs entirely
I would like to see further development but I always had a sneaking suspicion that its life was limited due to the fact that ARC does not come from Intel's fabs either. Like lunar lake, Arc is also made at TSMC.
Gelsinger said the market will have less demand for dedicated graphics cards in the future.
In other news, Intel is replaced by Nvidia in the Dow Jones, a company that exclusively produces dedicated graphics cards: https://lemmy.world/post/21576540
Nvidia does more than just GPUs.
Nvidia makes both SoCs like the Tegra series and server CPUs (Grace; ARM based to be used with their ML/AI cards with much higher bandwidths than regular CPUs).
Nvidia also just announced that they are working on a consumer desktop CPU.
Well I had the same thought a few years ago when APUs started getting better. But then I'm not the CEO of a huge tech company, so nobody lost their job because I was wrong.
And here I was thinking Arc and storage were the only semi-competitive wings of intel... They just needed a couple of years for adoption to increase
I've commented many times that Arc isn't competitive, at least not yet.
Although they were decent performers, they used twice the die size for similar performance compared to Nvidia and AMD, so Intel has probably sold them at very little profit.
Still I expected them to try harder this time, because the technologies to develop a good GPU, are strategically important in other areas too.
But maybe that's the reason Intel recently admitted they couldn't compete with Nvidia on high end AI?
Arcs are OK, and the competition is good. Their video encode performance is absolutely unworldly though, just incredible.
Mostly, they help bring the igpu graphics stack and performance up to full, and keep games targeting them well. They're needed for that alone if nothing else.
They were competitive for customers, but only because Intel sold them at no profit.
I have no confidence in Intel's long-term prospects.
Intel has been a mistake since 1978. But evil doesn't generally die.
Blaming loss on SoC? Lmfao. SoC is better. Just stop offering a lower tier and make all SoC 32gb+
… looking at you too, Apple.
I'm wondering, the integrated RAM like Intel did for Lunar Lake, could the same performance be achieved with the latest CAMM modules? The only real way to go integrated to get the most out of it is doing it with HBM, anything else seems like a bad trade-off.
So either you go HBM with real bandwidth and latency gains or CAMM with decent performance and upgradeable RAM sticks. But the on-chip ram like Intel did is neither providing the HBM performance nor the CAMM modularity.
I wonder why both isn't possible, build some into the chip but leave some DIMMs for upgradeability too at bit lower speed.
Especially with how normal memory tiering is nowadays, especially in the datacenter (Intel's bread and butter) now that you can stick a box of memory on a CXL network and put the memory from your last gen servers you just retired into said box for a third or fourth tier of memory before swapping. And the fun not tiered memory stuff the CXL enables. Really CXL just enables so much cool stuff that it's going to be incredible once that starts hitting small single row datacenters
The transfer speed isn't the big issue, it's the density and reliability. Packing more heat generating stuff onto the SoC package just makes it more difficult to dissipate. The transfer of data to where it needs to be is still the same, so the trade-off is pretty null in that sense except reduction of overall power consumption.
I see the idea of Intel dropping arc as good news for AMD. Intel was going to chip at AMD’s marketshare well before Nvidia’s. It would be better to have more competition though.
AMD would never close their GPU department because they sell their apu to Xbox, playstation, steam deck
Intel could have conceivably competed with them there too. If they were still a competent business and not in a crisis of mismanagement. Its amazing how much better AMD is managed compared to Intel.
I wasn’t saying AMD would shut down, but that Intel would take market share from them before truely affecting Nvidia’s market share. ie AMD and Intel would be fighting over the same 25ish% of the pc market.
With RAM access being the one big bottleneck of a modern PC, can anyone in the know tell me about those SoCs? How much RAM did they have, and was it faster than external DIMMs?
You shouldn't be comparing with DIMMs, those are a dead end at this point. CAMMs are replacing DIMMs and what future systems will use.
Intel likely designed Lunar lake before the LPCAMM2 standard was finalized and why it went on package. Now that LPCAMMs are a thing, it makes more sense to use those as they provide the same speed benefits while still allowing user replaceable RAM.