Do new CPUs threaten Nvidia's future?

Nvidia's Chief Scientist David Kirk is immediately dismissive about the danger posed by integrated CPU-GPU chips

It's worrying times for Nvidia. The graphics industry is apparently poised to transition to ray-tracing technology. Fusion-style processors threaten the existence of traditional CPUs and 3D chips. And Intel is getting serious about PC graphics for the first time.

And yet Nvidia's chief scientist David Kirk reckons it's all going to plan.

There used to be a pleasing symmetry among the four key players in PC technology. AMD took on Intel for processing prowess, while ATI and Nvidia duked it out for graphics grunt. Everyone knew their place. Then AMD snaffled up ATI at the end of 2006 for a few billion greenbacks and the balance in the universe was upset.

Since then, shockwaves from the AMD/ATI deal have spread far and wide. It completely torched what had hitherto been an extremely productive alliance between AMD and Nvidia. According to one Nvidia insider, the daily contact he had with AMD dried up literally overnight following the ATI purchase.

Fusion reaction

More recently, the strategic implications of the AMD/ATI tie up have become rather ominous. Intel and AMD both have plans for CPUs with integrated graphics, sometimes known as fusion processors. In the case of Intel, such a chip should be on sale before the end of the year in the form of Nehalem.

Granted, AMD has suffered all manner of woes following its acquisition of ATI, not least slow and buggy processors. But it's actually Nvidia that now looks most vulnerable.

In part, that's because Nvidia has no x86 kit currently on its books. What's more, even if it wanted to make PC-compatible CPUs, it lacks the necessary x86 license. And isn't the general trend supposed to be a gradual convergence of CPU and GPU technology towards a single chip containing a massively multi-core array of floating point fun? That would surely leave Nvidia out in the cold.

But even if Nvidia can carve out a long term strategy that doesn't include CPU production, its core graphics competency is under threat.

Momentum appears to be building for a new approach to graphics rendering known as ray tracing. It's proponents claim it produces more accurate and realistic graphics than the raster-based technology that currently dominates the graphics chip industry.

Ray tracing revolution

The key worry for Nvidia is that ray tracing might just present an industry-wide inflection point that allows Intel to enter the graphics market on an equal footing. That's certainly what Intel seems to be banking on with Larrabee, a multi-core chip that's thought to be highly optimised for ray tracing.

Suffice to say, therefore, that the easy domination Nvidia currently enjoys in the graphics market is hardly a given for the future. To find out exactly how the green-tinged graphics goliath plans to face up to these challenges, TechRadar crossed swords with none other than Dr David Kirk, Nvidia's chief scientist since 1997. If there's a man alive who has a better grasp of where Nvidia is heading, well, he wasn't available for interview!

Kirk is immediately dismissive about the danger posed by upcoming integrated CPU-GPU chips.

"Integrated graphics has traditionally been a low cost play," Kirk told us. Intel first began integrating graphics into its motherboard chipsets because it could be done almost for free. As Kirk says, the integrated option is essentially "the best graphics that no money can buy".

Not exactly a money spinning market segment, therefore. Yes, AMD's Fusion CPU is likely to raise the bar for integrated graphics performance, as Kirk concedes. But it nevertheless won't come close to what one might describe as acceptable gaming graphics performance. And if there's one thing Kirk is confident about, it's that consumers continue to demand high performance graphics.

TOPICS
Contributor

Technology and cars. Increasingly the twain shall meet. Which is handy, because Jeremy (Twitter) is addicted to both. Long-time tech journalist, former editor of iCar magazine and incumbent car guru for T3 magazine, Jeremy reckons in-car technology is about to go thermonuclear. No, not exploding cars. That would be silly. And dangerous. But rather an explosive period of unprecedented innovation. Enjoy the ride.

Latest in CPU
AMD Ryzen AI
New leak suggests AMD's working on an Arm-based processor to rival Qualcomm's Snapdragon X series
AMD Ryzen 9950X3D chip next to its packaging on a pink table
Asus' AI Cache Boost promises to "pump up" your AMD Ryzen 9000 processor's AI performance
An AMD Ryzen processor slotted into a motherboard
Future AMD-powered gaming handhelds and notebooks could miss out on a key feature – and it might be a deal breaker for gamers
John Loeffler holding the Ryzen 7 7800X3D
Great news! The best gaming CPU ever made is finally available for its original launch price again
The Ryzen AI Max+ 395 could power the latest generation of powerful mini PCs
The AMD Ryzen AI Max+ 395 dominates as the "most powerful" APU on the market, but its competition is questionable
Intel Lunar Lake concept
Intel's Panther Lake processors won't arrive until Q1 2026 - corroborates previous delay rumors despite former Intel CEO's promise of 2025 launch
Latest in News
DeepSeek
Deepseek’s new AI is smarter, faster, cheaper, and a real rival to OpenAI's models
Open AI
OpenAI unveiled image generation for 4o – here's everything you need to know about the ChatGPT upgrade
Apple WWDC 2025 announced
Apple just announced WWDC 2025 starts on June 9, and we'll all be watching the opening event
Hornet swings their weapon in mid air
Hollow Knight: Silksong gets new Steam metadata changes, convincing everyone and their mother that the game is finally releasing this year
OpenAI logo
OpenAI just launched a free ChatGPT bible that will help you master the AI chatbot and Sora
An aerial view of an Instavolt Superhub for charging electric vehicles
Forget gas stations – EV charging Superhubs are using solar power to solve the most annoying thing about electric motoring