skip to main content


Search for: All records

Award ID contains: 1763503

Note: When clicking on a Digital Object Identifier (DOI) number, you will be taken to an external site maintained by the publisher. Some full text articles may not yet be available without a charge during the embargo (administrative interval).
What is a DOI Number?

Some links on this page may take you to non-federal websites. Their policies may differ from this site.

  1. Free, publicly-accessible full text available November 13, 2024
  2. Free, publicly-accessible full text available November 12, 2024
  3. Free, publicly-accessible full text available September 26, 2024
  4. Free, publicly-accessible full text available August 30, 2024
  5. Free, publicly-accessible full text available August 17, 2024
  6. Free, publicly-accessible full text available June 1, 2024
  7. Free, publicly-accessible full text available May 1, 2024
  8. With the ever-growing popularity of Graph Neural Networks (GNNs), efficient GNN inference is gaining tremendous attention. Field-Programmable Gate Arrays (FPGAs) are a promising execution platform due to their fine-grained parallelism, low power consumption, reconfigurability, and concurrent execution. Even better, High-Level Synthesis (HLS) tools help bridge the gap between the non-trivial FPGA development efforts and rapid emergence of new GNN models. To enable investigation into how effectively modern HLS tools can accelerate GNN inference, we present GNNHLS, a benchmark suite containing a software stack for data generation and baseline deployment and FPGA implementations of 6 well-tuned GNN HLS kernels. 
    more » « less