Abstract
When hardware cache coherence scales to many cores on chip, over saturated traffic of the shared memory system may offset the benefit from massive hardware concurrency. In this article, we investigate the cost of a write-update protocol in terms of on-chip memory network traffic and its adverse effects on the system performance based on a multithreaded many-core architecture with distributed caches. We discuss possible software and hardware solutions to alleviate the network pressure. We find that in the context of massive concurrency, by introducing a write-merging buffer with 0.46% area overhead to each core, applications with good locality and concurrency are boosted up by 18.74% in performance on average. Other applications also benefit from this addition and even achieve a throughput increase of 5.93%. In addition, this improvement indicates that higher levels of concurrency per core can be exploited without impacting performance, thus tolerating latency better and giving higher processor efficiencies compared to other solutions.
| Original language | English |
|---|---|
| Article number | 103 |
| Journal | Transactions on Embedded Computing Systems |
| Volume | 13 |
| Issue number | 3s |
| DOIs | |
| Publication status | Published - Mar 2014 |
| Externally published | Yes |
Bibliographical note
Publisher Copyright:© 2014 ACM.
Keywords
- C.4.0 [performance of systems]: design studies
- Design
- Distributed cache
- Experimentation
- Hardware coherence
- Many-core system
- Massive parallelism
- On-chip memory network
- Performance
- Write combination