Energy-efficient partitioning of hybrid caches in multi-core architecture
@inproceedings{Lee2014EnergyefficientPO, title={Energy-efficient partitioning of hybrid caches in multi-core architecture}, author={Dongwook Lee and Kiyoung Choi}, booktitle={VLSI-SoC}, year={2014} }
This chapter presents a technique for reducing energy consumed by hybrid caches that have both SRAM and STT-RAM (SpinTransfer Torque RAM) in multi-core architecture. It is based on dynamic way partitioning of the SRAM cache as well as the STT-RAM cache. Each core is allocated with a specific number of ways consisting of SRAM ways and STT-RAM ways. Then a cache miss fills the corresponding block in the SRAM or STT-RAM region based on an existing technique called read-write aware region-based…Â
Figures and Tables from this paper
8 Citations
Restricting writes for energy-efficient hybrid cache in multi-core architectures
- Computer Science2016 IFIP/IEEE International Conference on Very Large Scale Integration (VLSI-SoC)
- 2016
An STt-RAM/SRAM hybrid cache architecture is proposed that reduces the number of writes and energy consumption of the STT-RAM region in the LLC by considering the existence of private blocks and makes changes in the conventional MESI protocol by adding new states to deal with the dataless entries.
Improving the Performance of Hybrid Caches Using Partitioned Victim Caching
- Computer ScienceACM Trans. Embed. Comput. Syst.
- 2021
Non-Volatile Memory technologies are coming as a viable option on account of the high density and low-leakage power over the conventional SRAM counterpart. However, the increased write latency…
Reuse-Distance-Aware Write-Intensity Prediction of Dataless Entries for Energy-Efficient Hybrid Caches
- Computer ScienceIEEE Transactions on Very Large Scale Integration (VLSI) Systems
- 2018
This paper presents a data allocation policy that reduces the number of writes and energy consumption of the STT-RAM region in the last-level cache by considering the existence of private blocks and presents a predictor that helps to redirect the write backs from L1 of dataless entries directly to SRAM region, depending on the predicted reuse-distance-aware write intensity.
Towards a dynamic associativity enabled write prediction based hybrid cache
- Computer Science2016 20th International Symposium on VLSI Design and Test (VDAT)
- 2016
A block prediction technique for hybrid cache that is based on existing prediction mechanism that is compared with the baseline SRAM and STT-RAM caches and shows significant reduction of the energy with slight improvement of performance.
Write Variation Aware Cache Partitioning for Improved Lifetime in Non-volatile Caches
- Computer Science2019 32nd International Conference on VLSI Design and 2019 18th International Conference on Embedded Systems (VLSID)
- 2019
A hardware-based cache partitioning technique to intelligently partition the shared last level cache among concurrently running applications by assigning heavily written cache ways to less written partitions so that writes are evenly distributed across the ways within the cache sets.
Per-Operation Reusability Based Allocation and Migration Policy for Hybrid Cache
- Computer ScienceIEEE Transactions on Computers
- 2020
This paper proposes an energy-efficient per-operation reusability-based allocation and migration policy (ORAM) with a unified LRU replacement policy, and presents a migration policy, victim and target cache line selection scheme, to resolve memory type inconsistency between replacement policy and the allocation policy.
Architectural Techniques for Improving the Power Consumption of NoC-Based CMPs: A Case Study of Cache and Network Layer
- Computer Science
- 2017
This work presents a survey of power-saving techniques for efficient NoC designs with a focus on the cache and router components, such as the buffer and crossbar, to compile a quick reference guide for engineers and researchers.
Reuse locality aware cache partitioning for last-level cache
- Computer ScienceComput. Electr. Eng.
- 2019
References
SHOWING 1-10 OF 19 REFERENCES
Energy-efficient partitioning of hybrid caches in multi-core architecture
- Computer Science2014 22nd International Conference on Very Large Scale Integration (VLSI-SoC)
- 2014
Simulation results show that the proposed technique improves the performance of the multi-core architecture and significantly reduces energy consumption in the hybrid caches compared to the state-of-the-art migration-based hybrid cache management.
STT-RAM based energy-efficiency hybrid cache for CMPs
- Computer Science2011 IEEE/IFIP 19th International Conference on VLSI and System-on-Chip
- 2011
This paper proposes to integrate SRAM with STT-RAM to construct a novel hybrid cache architecture for CMPs and proposes dedicated microarchitectural mechanisms to make the hybrid cache robust to workloads with different write patterns.
MAC: migration-aware compilation for STT-RAM based hybrid cache in embedded systems
- Computer ScienceISLPED '12
- 2012
A Migration-Aware Compilation (MAC) approach to improve the energy efficiency and performance of STT-RAM based hybrid cache by re-arranging data layout, the data access pattern in memory blocks is changed such that the number of migrations is reduced without any hardware modification.
Hybrid cache architecture with disparate memory technologies
- Computer ScienceISCA '09
- 2009
This paper discusses and evaluates two types of hybrid cache architectures: inter cache Level HCA (LHCA), in which the levels in a cache hierarchy can be made of disparate memory technologies; and intra cache level or cache Region based H CA (RHCA), where a single level of cache can be partitioned into multiple regions, each of a different memory technology.
OAP: An obstruction-aware cache management policy for STT-RAM last-level caches
- Computer Science2013 Design, Automation & Test in Europe Conference & Exhibition (DATE)
- 2013
An obstruction-aware cache management policy called OAP is proposed, which monitors the cache to periodically detect LLC-obstruction processes, and manage the cache accesses from different processes to minimize the write overhead.
Cooperative partitioning: Energy-efficient cache partitioning for high-performance CMPs
- Computer ScienceIEEE International Symposium on High-Performance Comp Architecture
- 2012
Cooperative Partitioning is presented, a runtime partitioning scheme that reduces both dynamic and static energy while maintaining high performance and maintains high performance while transferring ways five times faster than an existing state-of-the-art technique.
Writeback-aware partitioning and replacement for last-level caches in phase change main memory systems
- Computer ScienceTACO
- 2012
This paper proposes two writeback-aware schemes to manage the LLC for PCM main memory systems and shows that WCP and WQB result in 21% improvement in throughput, 49% reduction in PCM writes, and 14% reduced in energy over a state-of-the-art cache partitioning scheme.
A novel architecture of the 3D stacked MRAM L2 cache for CMPs
- Computer Science2009 IEEE 15th International Symposium on High Performance Computer Architecture
- 2009
This paper stacks MRAM-based L2 caches directly atop CMPs and compares it against SRAM counterparts in terms of performance and energy, and proposes two architectural techniques: read-preemptive write buffer and SRAM-MRAM hybrid L2 cache.
Write intensity prediction for energy-efficient non-volatile caches
- Computer ScienceInternational Symposium on Low Power Electronics and Design (ISLPED)
- 2013
A block placement strategy driven by write intensity prediction for SRAM/STT-RAM hybrid caches is proposed, which reduces write energy consumption by 55% on average compared to the existing hybrid cache architecture.
Static and dynamic co-optimizations for blocks mapping in hybrid caches
- Computer ScienceISLPED '12
- 2012
A combined static and dynamic scheme is proposed to optimize the block placement for endurance and energy-efficiency in a hybrid SRAM and STT-RAM cache that uses the compiler to provide static hints to guide initial data placement, and the hardware to correct the hints based on the run-time cache behavior.