Full metadata record
DC FieldValueLanguage
dc.contributor.authorKuo, Hsien-Kaien_US
dc.contributor.authorLai, Bo-Cheng Charlesen_US
dc.contributor.authorJou, Jing-Yangen_US
dc.date.accessioned2015-07-21T11:21:09Z-
dc.date.available2015-07-21T11:21:09Z-
dc.date.issued2014-11-01en_US
dc.identifier.issn1084-4309en_US
dc.identifier.urihttp://dx.doi.org/10.1145/2676550en_US
dc.identifier.urihttp://hdl.handle.net/11536/123923-
dc.description.abstractDeploying the Shared Last-Level Cache (SLLC) is an effective way to alleviate the memory bottleneck in modern throughput processors, such as GPGPUs. A commonly used scheduling policy of throughput processors is to render the maximum possible thread-level parallelism. However, this greedy policy usually causes serious cache contention on the SLLC and significantly degrades the system performance. It is therefore a critical performance factor that the thread scheduling of a throughput processor performs a careful trade-off between the thread-level parallelism and cache contention. This article characterizes and analyzes the performance impact of cache contention in the SLLC of throughput processors. Based on the analyses and findings of cache contention and its performance pitfalls, this article formally formulates the aggregate working-set-size-constrained thread scheduling problem that constrains the aggregate working-set size on concurrent threads. With a proof to be NP-hard, this article has integrated a series of algorithms to minimize the cache contention and enhance the overall system performance on GPGPUs. The simulation results on NVIDIA\'s Fermi architecture have shown that the proposed thread scheduling scheme achieves up to 61.6% execution time enhancement over a widely used thread clustering scheme. When compared to the state-of-the-art technique that exploits the data reuse of applications, the improvement on execution time can reach 47.4%. Notably, the execution time improvement of the proposed thread scheduling scheme is only 2.6% from an exhaustive searching scheme.en_US
dc.language.isoen_USen_US
dc.subjectAlgorithmsen_US
dc.subjectPerformanceen_US
dc.subjectThroughput processorsen_US
dc.subjectthread-level parallelismen_US
dc.subjectcache contentionen_US
dc.subjectshared last-level cacheen_US
dc.subjectthread schedulingen_US
dc.titleReducing Contention in Shared Last-Level Cache for Throughput Processorsen_US
dc.typeArticleen_US
dc.identifier.doi10.1145/2676550en_US
dc.identifier.journalACM TRANSACTIONS ON DESIGN AUTOMATION OF ELECTRONIC SYSTEMSen_US
dc.citation.volume20en_US
dc.contributor.department交大名義發表zh_TW
dc.contributor.departmentNational Chiao Tung Universityen_US
dc.identifier.wosnumberWOS:000345523400012en_US
dc.citation.woscount0en_US
Appears in Collections:Articles


Files in This Item:

  1. 000345523400012.pdf

If it is a zip file, please download the file and unzip it, then open index.html in a browser to view the full text content.