To: Joe NYC who wrote (59089 ) 10/18/2001 11:22:10 AM From: dale_laroy Read Replies (1) | Respond to of 275872 >The next generation of AMD processors will have on chip memory controller, and I don't know if it is possible to have onchip memory controllse plus off-chip L3. And off-chip L3 would not have the performance of onchip increase. Off-chip L3 would need some kind of lookup mechanism, which causes additional delay in memory access.< This is not entirely true. If one does not maintain any open pages the cache tag lookup of a L3 cache controller would certainly add delay, but it is possible to design a L3 cache controller using fast enough circuitry to match the delay of a typical open page hit detection circuit. >It would be interesting if someone could calculate how much of L3 would you need to equal performance of say additional 256K of L2.< This would depend upon the architecture of the L3 cache. For example, if one were to integrate the L3 cache into the DRAM itself, it would be possible to use very large cache lines by defining each open DRAM page as a cache line. Such an architecture could exceed the performance of 512KB of integrated L2 cache with 256KB of integrated L2 cache and just 256KB of off chip L3 cache. Building upon the Virtual Channel Memory architecture, it would be possible to integrate as much as one page of SRAM cache for every 64-bits of DRAM. If the page size of the integrated L3 SRAM cache is 4KB to match the granularity of the PMMU of the x86 architecture, 16MB of 16-way associative L3 cache could be controlled with a L3 cache controller the size and complexity of the controller for the 256KB of L2 cache on the Athlon. When you consider how long the Athlon requires to complete it's lookup in the 16-way set associative integrated L2 cache, it becomes obvious that the L3 cache controller could potentially add as little as one FSB cycle to the access, versus a no open page architecture, and even a single open page architecture adds at least on FSB cycle.