Performance trade-offs for microprocessor cache memories

Design trade-offs for integrated microprocessors caches are examined. A model of cache utilization is introduced to evaluate the effects on cache performance of varying the block size. By considering the overhead cost of sorting address tags and replacement information along with data, it is found that large block sizes lead to more cost-effective cache designs than predicted by previous studies. When the overhead cost is high, caches that fetch only partial blocks on a miss perform better than similar caches that fetch entire blocks. This study indicates that lessons from mainframe and minicomputer design practice should be critically examined to benefit the design of microprocessors.<<ETX>>

[1]  Irving L. Traiger,et al.  Evaluation Techniques for Storage Hierarchies , 1970, IBM Syst. J..

[2]  Leonard Jay Shustek,et al.  Analysis and performance of computer instruction sets , 1978 .

[3]  Alan Jay Smith,et al.  Experimental evaluation of on-chip microprocessor cache memories , 1984, ISCA 1984.

[4]  Alan Jay Smith,et al.  Cache Memories , 1982, CSUR.

[5]  M. Shoji,et al.  A pipelined 32b microprocessor with 13Kb of cache memory , 1987, 1987 IEEE International Solid-State Circuits Conference. Digest of Technical Papers.

[6]  F. Fox,et al.  A 32b CMOS microprocessor with on-chip instruction and data caching and memory management , 1987, 1987 IEEE International Solid-State Circuits Conference. Digest of Technical Papers.

[7]  Robert O. Winder,et al.  Cache-based Computer Systems , 1973, Computer.

[8]  Mahadev Satyanarayanan,et al.  Design Trade-Offs in VAX-11 Translation Buffer Organization , 1981, Computer.

[9]  John S. Liptay,et al.  Structural Aspects of the System/360 Model 85 II: The Cache , 1968, IBM Syst. J..

[10]  William D. Strecker Cache memories for PDP-11 family computers , 1976, ISCA.