Cache Impacts of Datatype Acceleration

Lisa K. Wu; Martha Allen Kim; Stephen A. Edwards

Cache Impacts of Datatype Acceleration
Wu, Lisa K.
Kim, Martha Allen
Edwards, Stephen A.
Computer Science
Permanent URL:
Book/Journal Title:
Computer Architecture Letters
Hardware acceleration is a widely accepted solution for performance and energy efficient computation because it removes unnecessary hardware for general computation while delivering exceptional performance via specialized control paths and execution units. The spectrum of accelerators available today ranges from coarse-grain off-load engines such as GPUs to fine-grain instruction set extensions such as SSE. This research explores the benefits and challenges of managing memory at the data-structure level and exposing those operations directly to the ISA. We call these instructions. Abstract Datatype Instructions (ADIs). This paper quantifies the performance and energy impact of ADIs on the instruction and data cache hierarchies. For instruction fetch, our measurements indicate that ADIs can result in 21-48% and 16-27% reductions in instruction fetch time and energy respectively. For data delivery, we observe a 22-40% reduction in total data read/write time and 9-30% in total data read/write energy.
Computer science
Publisher DOI:
Item views:
text | xml
Suggested Citation:
Lisa K. Wu, Martha Allen Kim, Stephen A. Edwards, 2011, Cache Impacts of Datatype Acceleration, Columbia University Academic Commons, http://hdl.handle.net/10022/AC:P:12744.

In Partnership with the Center for Digital Research and Scholarship at Columbia University Libraries/Information Services | Terms of Use | Copyright