GPU-accelerated language and communication support by FPGA

Taisuke Boku, Toshihiro Hanawa, Hitoshi Murai, Masahiro Nakao, Yohei Miki, Hideharu Amano, Masayuki Umemura

Research output: Chapter in Book/Report/Conference proceedingChapter


Although the GPU is one of the most successfully used accelerating devices for HPC, there are several issues when it is used for large-scale parallel systems. To describe real applications on GPU-ready parallel systems, we need to combine different paradigms of programming such as CUDA/OpenCL, MPI, and OpenMP for advanced platforms. In the hardware configuration, inter-GPU communication through PCIe channel and support by CPU are required which causes large overhead to be a bottleneck of total parallel processing performance. In our project to be described in this chapter, we developed an FPGA-based platform to reduce the latency of inter-GPU communication and also a PGAS language for distributed-memory programming with accelerating devices such as GPU. Through this work, a new approach to compensate the hardware and software weakness of parallel GPU computing is provided. Moreover, FPGA technology for computation and communication acceleration is described upon astrophysical problem where GPU or CPU computation is not sufficient on performance.

Original languageEnglish
Title of host publicationAdvanced Software Technologies for Post-Peta Scale Computing
Subtitle of host publicationThe Japanese Post-Peta CREST Research Project
PublisherSpringer Singapore
Number of pages17
ISBN (Electronic)9789811319242
ISBN (Print)9789811319235
Publication statusPublished - 2018 Dec 6

ASJC Scopus subject areas

  • Computer Science(all)


Dive into the research topics of 'GPU-accelerated language and communication support by FPGA'. Together they form a unique fingerprint.

Cite this