Do new CPUs threaten Nvidia's future?
Chief scientist David Kirk talks ray tracing, Fusion and Intel
It's worrying times for Nvidia. The graphics industry is apparently poised to transition to ray-tracing technology. Fusion-style processors threaten the existence of traditional CPUs and 3D chips. And Intel is getting serious about PC graphics for the first time.
And yet Nvidia's chief scientist David Kirk reckons it's all going to plan.
There used to be a pleasing symmetry among the four key players in PC technology. AMD took on Intel for processing prowess, while ATI and Nvidia duked it out for graphics grunt. Everyone knew their place. Then AMD snaffled up ATI at the end of 2006 for a few billion greenbacks and the balance in the universe was upset.
Since then, shockwaves from the AMD/ATI deal have spread far and wide. It completely torched what had hitherto been an extremely productive alliance between AMD and Nvidia. According to one Nvidia insider, the daily contact he had with AMD dried up literally overnight following the ATI purchase.
Fusion reaction
More recently, the strategic implications of the AMD/ATI tie up have become rather ominous. Intel and AMD both have plans for CPUs with integrated graphics, sometimes known as fusion processors. In the case of Intel, such a chip should be on sale before the end of the year in the form of Nehalem.
Granted, AMD has suffered all manner of woes following its acquisition of ATI, not least slow and buggy processors. But it's actually Nvidia that now looks most vulnerable.
Get the best Black Friday deals direct to your inbox, plus news, reviews, and more.
Sign up to be the first to know about unmissable Black Friday deals on top tech, plus get all your favorite TechRadar content.
In part, that's because Nvidia has no x86 kit currently on its books. What's more, even if it wanted to make PC-compatible CPUs, it lacks the necessary x86 license. And isn't the general trend supposed to be a gradual convergence of CPU and GPU technology towards a single chip containing a massively multi-core array of floating point fun? That would surely leave Nvidia out in the cold.
But even if Nvidia can carve out a long term strategy that doesn't include CPU production, its core graphics competency is under threat.
Momentum appears to be building for a new approach to graphics rendering known as ray tracing. It's proponents claim it produces more accurate and realistic graphics than the raster-based technology that currently dominates the graphics chip industry.
Ray tracing revolution
The key worry for Nvidia is that ray tracing might just present an industry-wide inflection point that allows Intel to enter the graphics market on an equal footing. That's certainly what Intel seems to be banking on with Larrabee, a multi-core chip that's thought to be highly optimised for ray tracing.
Suffice to say, therefore, that the easy domination Nvidia currently enjoys in the graphics market is hardly a given for the future. To find out exactly how the green-tinged graphics goliath plans to face up to these challenges, TechRadar crossed swords with none other than Dr David Kirk, Nvidia's chief scientist since 1997. If there's a man alive who has a better grasp of where Nvidia is heading, well, he wasn't available for interview!
Kirk is immediately dismissive about the danger posed by upcoming integrated CPU-GPU chips.
"Integrated graphics has traditionally been a low cost play," Kirk told us. Intel first began integrating graphics into its motherboard chipsets because it could be done almost for free. As Kirk says, the integrated option is essentially "the best graphics that no money can buy".
Not exactly a money spinning market segment, therefore. Yes, AMD's Fusion CPU is likely to raise the bar for integrated graphics performance, as Kirk concedes. But it nevertheless won't come close to what one might describe as acceptable gaming graphics performance. And if there's one thing Kirk is confident about, it's that consumers continue to demand high performance graphics.
Technology and cars. Increasingly the twain shall meet. Which is handy, because Jeremy (Twitter) is addicted to both. Long-time tech journalist, former editor of iCar magazine and incumbent car guru for T3 magazine, Jeremy reckons in-car technology is about to go thermonuclear. No, not exploding cars. That would be silly. And dangerous. But rather an explosive period of unprecedented innovation. Enjoy the ride.