Things would be completely different if only Intel hadn't cancelled the Larrabee GPU.
Intel CEO laments Nvidia's 'extraordinarily lucky' AI dominance, claims it coulda-woulda-shoulda have been Intel::Intel CEO Pat Gelsinger has taken a shot at his main rival in high performance computing, dismissing Nvidia's success in providing GPUs for AI modelling as "extraordinarily lucky." Gels
Having read the article it sounds like Pat is more complaining that Intel would have been positioned to milk the AI cow if the previous CEOs weren't fucking idiots.
He's right. Intel dumped their Tick Tock fast development cycle and let their product lines languish. Real stupid way to generate short term investor gains.
Well Bob Swan did dump $20 billion on a single stock buyback that only elevated the share price for like 2 weeks and Brian Krzanitch saw delay after delay on 14nm which resulted in turning their two year lead into a two year defecit. There's definitely truth to Pat's words.
I think it's shorter to name the things Intel did strike gold on than the boats they missed despite all kinda of advantages.
They missed the entire mobile market processor explosion, inevitable as it was since 2009 with Android launch(or 2007 if you want to say they should have seen what Apple was doing and thought they could compete, which if you're an exec at Intel you should have).
They missed cloud computing.
Classic "big company hires/keeps overpaid check drawers instead of those with finger on the pulse". Intel deserves the very little innovation, success and relevancy theyve had post 86.
Bought my first AMD computer this year, an and 6800 Ryzen 7 with an on proc 680m gpu that is equivalent of ~ Nvidia 2050 discrete card. Game over for Intel.
Bought my first AMD computer this year, an and 6800 Ryzen 7 with an on proc 680m gpu that is equivalent of ~ Nvidia 2050 discrete card. Game over for Intel.
While the rest of your post is logical, this is insane cope. No one is buying integrated graphics for gaming. 2050s are a joke in terms of power - you're talking about a 2 year old budget mobile gpu.... If anything this is basically a "I need to do some photoshop but don't want a dedicated gpu on my laptop" type card. Intel has never given a fuck about mobile graphics. Their offerings have always been "serviceable, but get a real gpu if you want one". Laptops are arguably better with ARM so there's competition there....
Intel is still selling their bread and butter and still has a huge stranglehold on their core market. Claiming "game over" because of an off case of an offshoot of one of their secondary markets is hugely overreacting.
Plenty of people game on integrated graphics and a 2050 is damn powerful for an integrated card. Not everyone is a hyper nerd who builds their own PC or pays big bucks for a good gaming laptop. There are tons of casual players using integrated GPUs
Of course they still have a stranglehold on computer CPUs, the point is they haven't done anything with that for 30 years. Their bread and butter is an ever lower margin game in PCs that have largely peaked; they missed mobile phones and are set to miss whatever comes next with ARM.
My purchase of a laptop without their proc isn't even a drop in the ocean, but the point is that the only market they do have they can't even dominate any more, just keep shipping manually faster CPUs once a year, just like they have been for the past 15.
Not really. It did for one generation, against AMD's ancient Vega iGPU, if you ignore that ~30% of games were buggy or straight up don't even work on Intel iGPUs.
I'm becoming tired of Pat's whining every other week. You don't hear Lisa Su crying about Nvidia (or Intel for that matter) just being lucky, spouting this woe is me bullshit. You hear her, and Jensen, shutting up unless they have something meaningful to show.
I've noticed this in a lot of Intel slides and presentations recently - they talk more about their competitor's products than they do their own!
Perhaps if intel had branched out into GPUs earlier than they did, while they still had mountains of money to do so, they could've leveraged the AI boom.
But what did they do? They spaffed billions up the wall on stupid acquisitions like fucking McAfee Antivirus.
Rather than swallowing their pride and using their manufacturing and CPU design skills to make ARM mobile CPUs, they engaged in the stupid decision to pay phone manufacturers to use x86 CPUs that simply weren't efficient enough for mobile use. Needless to say, they all flopped hard.
They let their foundries and design teams rot to the extent that AMD, a dying company, was able to surpass them.
Now they're struggling to release CPUs without them being 6 months to a year late. Sometimes they don't release at all (where is desktop Meteor Lake, Intel?); that's their bread and butter product, FFS! If Intel can't do CPUs, what can they do?!
Intel deserves everything they're getting right now. Them getting left behind in AI is a problem entirely of their own making.
Not really. ATI were always “G is for graphics” and built video games cards. They never really saw the potential (nor did they have the resources anyway) for GPGPU, which is why NVIDIA had a huge first-player advantage (CUDA is 16 years old, 2 years before AMD acquired ATI). When AMD bought them it was already very late.
Then AMD wanted to build cards for people to buy while NVIDIA was more than happy selling overpriced cards to crypto miners.
OpenCL was an ambitious project that was too big and too open for what was capable from the Khronos group. Vulkan was too late.
Intel could have done it but IIRC the CEO at that time (can’t remember the name) didn’t want to diversify their products after Itanium was a failure. They just doubled down on CPU.
They never really saw the potential (nor did they have the resources anyway) for GPGPU
Maybe ATI, which ended in 2010.
AMD launched ROCm in 2016, after the first AI boom of 2012, but before GANs and transformers exploded. In recent years they're better positioned in than Intel ever was.
Disagree. GCN cards were incredibly compute focused.
Shit, AMD even invented HBM memory because they saw the value in ridiculously high bandwidth, dense, energy efficient memory in data centre applications. HBM is still used today in the enterprise market.
AMD's problem was that they had no money at the time and couldn't build out their software ecosystem like Nvidia could - they had to bank on just getting the ball rolling and open sourcing their efforts in the hope that others would contribute, which didn't happen to the extent that they'd have liked, especially when Nvidia with their mountains of cash could just pump out CUDA and flood universities with free GPUs to get them hooked in the Nvidia software stack.
Amd dropped the ball when it came to software and has now separated their GPU architecture so that they only have enterprise cards for data science. NVIDIA got in early and made CUDA default among all product lineups so that consumer cards could be used as entry-level cards by hobbyists. While it would’ve been nice to see more competition, the only company taking this space seriously has been nvidia.