4.4 Article

Processing data where it makes sense: Enabling in-memory computation

期刊

MICROPROCESSORS AND MICROSYSTEMS
卷 67, 期 -, 页码 28-41

出版社

ELSEVIER SCIENCE BV
DOI: 10.1016/j.micpro.2019.01.009

关键词

Data movement; Main memory; Processing-in-memory; 3D-Stacked memory; Near-data processing

资金

  1. Alibaba
  2. Google
  3. Huawei
  4. Intel
  5. Microsoft
  6. NVIDIA
  7. Samsung
  8. Seagate
  9. VMware
  10. Intel Science and Technology Center for Cloud Computing
  11. Semiconductor Research Corporation
  12. Data Storage Systems Center at Carnegie Mellon University
  13. NSF grants
  14. NSF CAREER Award
  15. Intel Faculty Honor Program Award

向作者/读者索取更多资源

Today's systems are overwhelmingly designed to move data to computation. This design choice goes directly against at least three key trends in systems that cause performance, scalability and energy bottlenecks: (1) data access from memory is already a key bottleneck as applications become more data-intensive and memory bandwidth and energy do not scale well, (2) energy consumption is a key constraint in especially mobile and server systems, (3) data movement is very expensive in terms of bandwidth, energy and latency, much more so than computation. These trends are especially severely-felt in the data-intensive server and energy-constrained mobile systems of today. At the same time, conventional memory technology is facing many scaling challenges in terms of reliability, energy, and performance. As a result, memory system architects are open to organizing memory in different ways and making it more intelligent, at the expense of higher cost. The emergence of 3D-stacked memory plus logic as well as the adoption of error correcting codes inside DRAM chips, and the necessity for designing new solutions to serious reliability and security issues, such as the RowHammer phenomenon, are an evidence of this trend. In this work, we discuss some recent research that aims to practically enable computation close to data. After motivating trends in applications as well as technology, we discuss at least two promising directions for processing-in-memory (PIM): (1) performing massively-parallel bulk operations in memory by exploiting the analog operational properties of DRAM, with low-cost changes, (2) exploiting the logic layer in 3D-stacked memory technology to accelerate important data-intensive applications. In both approaches, we describe and tackle relevant cross-layer research, design, and adoption challenges in devices, architecture, systems, and programming models. Our focus is on the development of in-memory processing designs that can be adopted in real computing platforms at low cost. (C) 2019 Published by Elsevier B.V.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.4
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据