During the company's third-quarter 2024 earnings call, Intel confirmed that its future laptop chips will return to the traditional use of RAM sticks, reversing Lunar Lake's radical...
Panther Lake and Nova Lake laptops will return to traditional RAM sticks
I've commented many times that Arc isn't competitive, at least not yet.
Although they were decent performers, they used twice the die size for similar performance compared to Nvidia and AMD, so Intel has probably sold them at very little profit.
Still I expected them to try harder this time, because the technologies to develop a good GPU, are strategically important in other areas too.
But maybe that's the reason Intel recently admitted they couldn't compete with Nvidia on high end AI?
Arcs are OK, and the competition is good. Their video encode performance is absolutely unworldly though, just incredible.
Mostly, they help bring the igpu graphics stack and performance up to full, and keep games targeting them well. They're needed for that alone if nothing else.
True, yet Nvidia was a nobody that arrived out of nowhere with the Riva graphics cards, and beat everybody else thoroughly. ATi, S3, 3Dfx, Matrox etc.
But you are right, these things usually take time, and for instance Microsoft was prepared to spend 10 years without making money on Xbox, because they saw it had potential in the long run.
I'm surprised Intel consider themselves so hard pressed, they are already thinking of giving up.
Now that you mention it, I remember reading about that, but completely forgot.
I remembered it as the Riva coming out of nowhere. As the saying goes, first impressions last. And I only learned about NV1 much later.
But the third one stayed up!
👍 😋
But Intel also made the i815 GPU, So Arc isn't really the first.
Oof, yeah, they actually had another they didn't release, based off pentium cores with avx512, basically knights landing with software support for graphics.
They were canceling projects like it was going out of style, which is sad, that would have been amazing for Ai.
Yes, there was the Xeon Phi, Knights Landing, with up to 72 cores, and 4 threads per core!
The Knights Landing was put into production though, but it was more a compute unit than a GPU.
I'm not aware they tried to sell it as a GPU too? Although If I recall correctly they made some real time ray tracing demos.
So, trying not to dox myself, I worked with the architect twice.
Knights Ferry was derived directly from Larrabee (GPU), P54Cs with pre-AVX-512, .
KNC was a die shrink with more cores. Both of these were PCIe accelerators only.
KNL had full Airmont Atom cores with smt4, basically meaningful cores with proper AVX-512. Also you could boot them with linux, or as a PCIe accelerator.
KNM jadded ML instructions, basically 8/16bit float and faster SIMD.
They cancelled KNH.
I interviewed some of the actual Larrabee guys, they were wild, there was a lot of talk about dynamic translation, they were trying to do really complex things, but when people talk like that it makes me think they were floundering on the software and just looking for tech magic solutions to fundamental problems.
Intel always dies when the software gets more complex than really simple drivers, it's their achilles heel.
KNL also had the whole MCDRAM on package for basically HBM bandwidth, but that didn't actually work very well in practice, again due to software issues (you have to pick where you allocate, and using it as an l4 cache was not always effective).
Still I expected them to try harder this time, because the technologies to develop a good GPU, are strategically important in other areas too
I think I read somewhere that they're having problems getting AIB partners for Battlemage. That would be a significant impediment for continuing in the consumer desktop market unless Battlemage can perform better (business-wise) than Alchemist.
They probably will continue investing in GPU even if they give up on Arc, it might just be for the specialized stuff.
Yeah true, plus I bought my a770 at pretty much half price during the whole driver issues and so eventually got a 3070 performing card for like $250, which is an insane deal for me but no way intel made anything on it after all the rnd and production costs
The main reason Intel can't compete is the fact CUDA is both proprietary and the industry standard, if you want to use a library you have to translate it yourself which is kind of inconvenient and no datacentre is going to go for that
The main reason Intel can’t compete is the fact CUDA is both proprietary and the industry standard
Funnily enough this is actually changing because of the AI boom. Would-be buyers can't get Nvidia AI cards so they're buying AMD and Intel and reworking their stacks as needed. It helps that there's also translation layers available now too which translate CUDA and other otherwise vebdor-specific stuff to the open protocols supported by Intel and AMD
I think intel support it (or at least a translation layer) but there's no motivation for Nvidia to standardise to something open-source as the status quo works pretty well