Managing shared last-level cache in a heterogeneous multicore processor

@article{Mekkat2013ManagingSL,
  title={Managing shared last-level cache in a heterogeneous multicore processor},
  author={Vineeth Mekkat and Anup Holey and P. Yew and Antonia Zhai},
  journal={Proceedings of the 22nd International Conference on Parallel Architectures and Compilation Techniques},
  year={2013},
  pages={225-234}
}
  • Vineeth Mekkat, Anup Holey, +1 author Antonia Zhai
  • Published 2013
  • Computer Science
  • Proceedings of the 22nd International Conference on Parallel Architectures and Compilation Techniques
  • Heterogeneous multicore processors that integrate CPU cores and data-parallel accelerators such as GPU cores onto the same die raise several new issues for sharing various on-chip resources. The shared last-level cache (LLC) is one of the most important shared resources due to its impact on performance. Accesses to the shared LLC in heterogeneous multicore processors can be dominated by the GPU due to the significantly higher number of threads supported. Under current cache management policies… CONTINUE READING
    76 Citations
    Exploiting Dynamic Reuse Probability to Manage Shared Last-level Caches in CPU-GPU Heterogeneous Processors
    • 5
    • Highly Influenced
    • PDF
    Set variation-aware shared LLC management for CPU-GPU heterogeneous architecture
    • 1
    A Sample-Based Dynamic CPU and GPU LLC Bypassing Method for Heterogeneous CPU-GPU Architectures
    • Xin Wang, Wei Zhang
    • Computer Science
    • 2017 IEEE Trustcom/BigDataSE/ICESS
    • 2017
    Evaluating the effect of last-level cache sharing on integrated GPU-CPU systems with heterogeneous applications
    • 16
    • PDF
    Adaptive Cache Management for Energy-Efficient GPU Computing
    • 121
    • PDF
    Heterogeneity Aware Shared DRAM Cache for Integrated Heterogeneous Architectures
    Using Criticality of GPU Accesses in Memory Management for CPU-GPU Heterogeneous Multi-Core Processors
    • 5
    • Highly Influenced
    • PDF
    Cache locking vs. partitioning for real-time computing on integrated CPU-GPU processors
    • Xin Wang, Wei Zhang
    • Computer Science
    • 2016 IEEE 35th International Performance Computing and Communications Conference (IPCCC)
    • 2016
    • 2
    Reducing Cache Contention On Gpus

    References

    SHOWING 1-10 OF 10 REFERENCES
    Scalable Shared-Cache Management by Containing Thrashing Workloads
    • 34
    • Highly Influential
    PIPP: promotion/insertion pseudo-partitioning of multi-core shared caches
    • 304
    • Highly Influential
    • PDF
    Utility-Based Cache Partitioning: A Low-Overhead, High-Performance, Runtime Mechanism to Partition Shared Caches
    • 915
    • Highly Influential
    • PDF
    Sampling Dead Block Prediction for Last-Level Caches
    • 154
    • Highly Influential
    • PDF
    Adaptive insertion policies for high performance caching
    • 619
    • Highly Influential
    • PDF
    Multi2Sim: A simulation framework for CPU-GPU computing
    • 397
    • Highly Influential
    • PDF
    High performance cache replacement using re-reference interval prediction (RRIP)
    • 563
    • Highly Influential
    • PDF
    Runtime cache bypassing
    • IEEE Transactions on Computers
    • 1999
    Runtime cache bypassing
    • IEEE Transactions on Computers
    • 1999
    AMD Fusion Family of APUs: Enabling a Superior, Immersive PC Experience
    • AMD White Paper
    • 2010