Print 84 comment(s) - last by dwalton.. on Jul 16 at 4:09 PM

Intel says CUDA will be nothing but a footnote in computer history

Intel and NVIDIA compete in many different ways. The most notable place we see competition between the two companies is in chipset manufacturing. Intel and NVIDIA also compete in the integrated graphics market where Intel’s integrated graphics chips lead the market.

NVIDIA started competing with Intel in the data processing arena with the CUDA programming language. Intel’s Pat Gelsinger, co-general manager of Intel’s Digital Enterprise Group, told Custom PC that NVIDIA’s CUDA programming model would be nothing more than an interesting footnote in the annals of computing history.

According to Gelsinger, programmers simply don’t have enough time to learn how to program for new architectures like CUDA. Gelsinger told Custom PC, “The problem that we’ve seen over and over and over again in the computing industry is that there’s a cool new idea, and it promises a 10x or 20x performance improvements, but you’ve just got to go through this little orifice called a new programming model. Those orifices have always been insurmountable as long as the general purpose computing models evolve into the future.”

The Sony Cell architecture illustrates the point according to Gelsinger. The Cell architecture promised huge performance gains compared to normal architectures, but the architecture still isn’t supported widely by developers.

Intel’s Larrabee graphics chip will be entirely based on Intel Architecture x86 cores says Gelsinger. The reason for this is so that developers can program for the graphics processor without having to learn a new language. Larrabee will have full support for APIs like DX and OpenGL.

NVIDIA’s CUDA architecture is what makes it possible to process complex physics calculations on the GPU, enabling PhysX on the GPU rather than CPU.

Comments     Threshold

This article is over a month old, voting and posting comments is disabled

RE: pwnd
By nafhan on 7/2/2008 2:43:16 PM , Rating: 1
Just to comment on your "in order processing is a huge bottleneck" comment.
That's only true if you have a single in order core. With multiple in order cores working on seperate threads, it's not as much of an issue. For the most part, GPU's have been in order, and I think Intel's Atom processor is in order as well. In order processors provide significant transistor and power savings over their out of order counterparts.

RE: pwnd
By Elementalism on 7/2/2008 2:51:28 PM , Rating: 2
afaik Itanium is also in order.

RE: pwnd
By omnicronx on 7/2/2008 3:06:47 PM , Rating: 3
Just to comment on your "in order processing is a huge bottleneck" comment.
That's only true if you have a single in order core. With multiple in order cores working on seperate threads, it's not as much of an issue
I really can not agree with you, you brought up the atom so I will use it as an example. The current Intel 1.6GHZ atom can barely compete with the old celeron 1.2ghz, in which the celeron actually beats it out on most tests. Although this is a single core processor, Intel had to bring back a new implementation of hyperthreading just to bring up the performance to somewhat of a respectable level. So even with multiple threads being excecuted at the same time, performance was at least 1/3 below what a 3 year old out of order processor can do.

I also understand about the power savings with in order processing, but really in a GPGPU who cares? Intel is trying to come out and say they have the holy grail of GPUS that can do just about anything from laptop to desktop to high end GPGPU computing, when in reality they have come up with a unified architecture between all three with one small hiccup, its seems to be far less efficient in two of those fields. (desktop and GPGPU markets)

The way I see it, an all in one solution has never been as good as a standalone product that curtails to the certain area or market. I do give the nod in the fact it seems they have found a way to unify their archecture along all of its lines, but in the end will this be better for Intel or for the consumer, my guess is the later, but what do I know ;)
In the end, only time will tell.

RE: pwnd
By encia on 7/2/2008 11:29:28 PM , Rating: 2
Run Swiftshader 2.01 on Intel Atom (or Intel Core 2 Quad@3Ghz) vs ATI Radeon 9600. Tell me one which is faster in DX9b.

RE: pwnd
By sa365 on 7/3/2008 8:40:46 AM , Rating: 2
Who now thinks AMD's purchase of ATI was a bad move?

"Google fired a shot heard 'round the world, and now a second American company has answered the call to defend the rights of the Chinese people." -- Rep. Christopher H. Smith (R-N.J.)
Related Articles
GeForce 8 To Get Software PhysX Engine
February 15, 2008, 10:33 AM

Most Popular ArticlesAre you ready for this ? HyperDrive Aircraft
September 24, 2016, 9:29 AM
Leaked – Samsung S8 is a Dream and a Dream 2
September 25, 2016, 8:00 AM
Inspiron Laptops & 2-in-1 PCs
September 25, 2016, 9:00 AM
Snapchat’s New Sunglasses are a Spectacle – No Pun Intended
September 24, 2016, 9:02 AM
Walmart may get "Robot Shopping Carts?"
September 17, 2016, 6:01 AM

Copyright 2016 DailyTech LLC. - RSS Feed | Advertise | About Us | Ethics | FAQ | Terms, Conditions & Privacy Information | Kristopher Kubicki