- 1.C. J. Conti. Concepts of buffer storage,/EEE Computer Group News, 2 (March 1969).Google Scholar
- 2.R. M. Meade. How a cache memory enhances a computer's performance, Electronics (Jan. 1972}.Google Scholar
- 3.K. R. Kaplan and R. O. Winder. Cache-based computer systems, IEEE Computer (March 1973).Google Scholar
- 4.j. Bell, D. Casasent, and C. G. Bell. An investigation of alternative cache organizations. IEEE Transactions on Computers, C- 23 (April 1974).Google Scholar
- 5.j. H. Kroeger and R. M. Meade (of Cogar Corporation, Woppingers Fail, NY). Cache buffer memory specification.Google Scholar
- 6.A. V. Pohm, O. P. Agrawal, and R. N. Monroe. The cost end performance tradeoffs of buffered memories. Proceedings of the IEE~, 63 (Aug. 1973).Google Scholar
- 7.A. J. Smith. Sequential program prefetching in memory hierachies. IEEE Computer (Dec 1978).Google Scholar
- 8.G. H. Tools. instruction iookahead and execution traffic considerations for the ~ cache design (Development division internal paper), Control Data-Canada, 1975.Google Scholar
Index Terms
- Lockup-free instruction fetch/prefetch cache organization
Recommendations
Lockup-free instruction fetch/prefetch cache organization
ISCA '81: Proceedings of the 8th annual symposium on Computer ArchitectureIn the past decade, there has been much literature describing various cache organizations that exploit general programming idiosyncrasies to obtain maximum hit rate (the probability that a requested datum is now resident in the cache). Little, if any, ...
Improving direct-mapped cache performance by the addition of a small fully-associative cache and prefetch buffers
Special Issue: Proceedings of the 17th annual international symposium on Computer ArchitectureProjections of computer technology forecast processors with peak performance of 1,000 MIPS in the relatively near future. These processors could easily lose half or more of their performance in the memory hierarchy if the hierarchy design is based on ...
Temporal instruction fetch streaming
MICRO 41: Proceedings of the 41st annual IEEE/ACM International Symposium on MicroarchitectureL1 instruction-cache misses pose a critical performance bottleneck in commercial server workloads. Cache access latency constraints preclude L1 instruction caches large enough to capture the application, library, and OS instruction working sets of these ...
Comments