Abstract
As the big data era is coming, it brings new challenges to the massive data processing. A combination of GPU and CPU on chip is the trend to release the pressure of large scale computing. We found that there are different memory access characteristics between GPU and CPU. The most important one is that the programs of GPU include a large number of threads, which lead to higher access frequency in cache than the CPU programs. Although the LRU policy favors the programs with high memory access frequency, the programs of GPU can’t get the corresponding performance boost even more cache resources are provided. So LRU policy is not suitable for heterogeneous multi-core processor.
Based on the different characteristics of GPU and CPU programs on memory access, this paper proposes an LLC dynamic replacement policy–DIPP (Dynamic Insertion / Promotion Policy) for heterogeneous multi-core processors. The core idea of the replacement policy is to reduce the miss rate of the program and enhance the overall system performance by limiting the cache resources that GPU can acquire and reducing the thread interferences between programs.
Experiments compare the DIPP replacement policy with LRU and we conduct a classified discussion according to the program results of GPU. Friendly programs enhance 23.29% on the average performance (using arithmetic mean). Large working sets programs can improve 13.95%, compute-intensive programs enhance 9.66% and stream class programs improve 3.8%.
This is a preview of subscription content, log in via an institution.
Buying options
Tax calculation will be finalised at checkout
Purchases are for personal use only
Learn about institutional subscriptionsPreview
Unable to display preview. Download preview PDF.
References
Fan, Z., Qiu, F., Kaufman, A., Yoakum-Stover, S.: GPU cluster for high performance computing. In: SC 2004: Proceedings of the 2004 ACM/IEEE Conference on Supercomputing. IEEE Computer Society, Washington, DC (2004)
Dally, W.J., Hanrahan, P., Erez, M., Knight, T.J., et al.: Merrimac: Supercomputing with Streams. In: SC 2003: Proceedings of Supercomputing Conference 2003, vol. 5270, pp. 35–42 (2003)
Hennessy, J.L., Patterson, D.A.: Computer Architecture: A Quantitative Approach, 4th edn. Morgan Kaufmann (2006)
Qureshi, M.K., Jaleel, A., Patt, Y.N., et al.: Adaptive insertion policies for high performance caching. ACM SIGARCH Computer Architecture News 35(2), 381–391 (2007)
Jaleel, A., Theobald, K.B., Steely Jr., S.C., Emer, J.: High performance cache replacement using re-reference interval prediction (RRIP). In: ISCA-32, pp. 60–71 (2010)
Qureshi, M.K., Patt, Y.N.: Utility-based cache partitioning:A low-overhead, high-per-formance, runtime mechanism to partition shared caches. In: MICRO-39, pp. 423–432 (2006)
Xie, Y., Loh, G.H.: Scalable shared-cache management by containing thrashing workloads. In: Patt, Y.N., Foglia, P., Duesterwald, E., Faraboschi, P., Martorell, X. (eds.) HiPEAC 2010. LNCS, vol. 5952, pp. 262–276. Springer, Heidelberg (2010)
Xie, Y., Loh, G.H.: PIPP: promotion/insertion pseudopartitioning of multi-core shared caches. In: ISCA-31, pp. 174–183 (2009)
MacSim, http://code.google.com/p/macsim/
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2015 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Yang, Z., Zuocheng, X., Xiao, M. (2015). DIPP—An LLC Replacement Policy for On-chip Dynamic Heterogeneous Multi-core Architecture. In: Wang, H., et al. Intelligent Computation in Big Data Era. ICYCSEE 2015. Communications in Computer and Information Science, vol 503. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-662-46248-5_47
Download citation
DOI: https://doi.org/10.1007/978-3-662-46248-5_47
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-662-46247-8
Online ISBN: 978-3-662-46248-5
eBook Packages: Computer ScienceComputer Science (R0)