Bandwidth-Efficient On-Chip Interconnect Designs for GPGPUs Conference Paper uri icon

abstract

  • 2015 ACM. Modern computational workloads require abundant thread level parallelism (TLP), necessitating highly-parallel, many-core accelerators such as General Purpose Graphics Processing Units (GPGPUs). GPGPUs place a heavy demand on the on-chIP interconnect between the many cores and a few memory controllers (MCs). Thus, traffic is highly asymmetric, impacting on-chIP resource utilization and system performance. Here, we analyze the communication demands of typical GPGPU applications, and propose efficient Network-on-ChIP (NoC) designs to meet those demands. We show that the proposed schemes improve performance by up to 64.7%. Compared to the best of class prior work, our VC monopolizing and partitioning schemes improve performance by 25%.

name of conference

  • Proceedings of the 52nd Annual Design Automation Conference

published proceedings

  • 2015 52ND ACM/EDAC/IEEE DESIGN AUTOMATION CONFERENCE (DAC)

author list (cited authors)

  • Jang, H., Kim, J., Gratz, P., Yum, K. H., & Kim, E. J.

citation count

  • 53

complete list of authors

  • Jang, Hyunjun||Kim, Jinchun||Gratz, Paul||Yum, Ki Hwan||Kim, Eun Jung

publication date

  • June 2015