backtop


Print 37 comment(s) - last by Spivonious.. on Jan 25 at 9:36 AM

Ashwood memory architecture allows for much faster memory speeds

Chipmakers realized long ago that extracting more performance from computer processors could be accomplished in ways other than simply reducing the size of the manufacturing process to squeeze more transistors onto a die.

One of the ways chipmakers improved performance was by building multi-core CPUs, like Intel's Penryn processors, that allow for parallel execution of data. Memory chips haven’t been able to keep up with the performance increases we are seeing in processors making for a bottleneck in the performance of computer systems and other devices.

In order to tackle this problem, a cryptographer named Joseph Ashwood has developed a new memory architecture that allows for multi-core memory.

Ashwood dubbed his memory architecture the Ashwood Architecture. According to EETimes the Ashwood architecture integrates smart controller circuitry next to the memory array on a single chip. This provides parallel access to the memory array for hundreds of concurrent processes leading to increased throughput and lower average access times.

Ashwood says, “My design borrows extensively from today's modern multicore CPUs. As far as concurrency goes, my memory architecture shares some features with Fibre Channel.”

Ashwood says his architecture can hit 16Gbytes per second compared to the DDR2 limit of 12 Gbytes per second. The hallmark of the Ashwood architecture is that the larger the number of bit cells in the memory the better the performance.

Ashwood does admit to a couple downsides to his design. The first is that his design is paper only, though it was independently verified by researchers from Carnegie Mellon University. No design was tested of the architecture at the electrical signal level.

The second drawback is that parallel access overhead of the architecture slows down access time to individual memory cells. However, Ashwood says that the parallel nature of his architecture more than makes up for any slowdowns by executing more commands at the same time.

Ashwood has filed a patent on his architecture that is still pending; until the patent is granted the intricate details of his architecture remain unknown.





Comments     Threshold


This article is over a month old, voting and posting comments is disabled

RE: Bandwidth?
By mathew7 on 1/18/2008 7:49:49 AM , Rating: 2
You don't need 64-bit to reach the theoretical bandwidth. The bits are changed at the same speed. The difference is only in overhead, because in 64-bit mode you have not only larger registers, but also more. But you can obtain a bus bottleneck if you do only sums on a large set of data even on 32-bit systems (which can be done using SSE). Of couse, if your processor can process those "bits" fast enough.
And even DDR2 had a slow start after DDR, which also had a slow start after "SD-RAM".
Although "latency" is used in memory relation as the time between command and start of data transfer, the real latency also contains the transfer time (I don't think the CPU cache lets you retrieve bytes before the whole line is fetched). And this is why DDR3 needs higher frequencies to shine: it's higher "start latency" needs to be compensated by burst time.


"If you look at the last five years, if you look at what major innovations have occurred in computing technology, every single one of them came from AMD. Not a single innovation came from Intel." -- AMD CEO Hector Ruiz in 2007
Related Articles
Engineers Explain 45nm Delays, Errata
January 16, 2008, 10:32 AM













botimage
Copyright 2015 DailyTech LLC. - RSS Feed | Advertise | About Us | Ethics | FAQ | Terms, Conditions & Privacy Information | Kristopher Kubicki