Skip to main navigation Skip to search Skip to main content

CU.POKer: Placing DNNs on WSE With Optimal Kernel Sizing and Efficient Protocol Optimization

  • Bentian Jiang*
  • , Jingsong Chen
  • , Jinwei Liu
  • , Lixin Liu
  • , Fangzhou Wang
  • , Xiaopeng Zhang
  • , Evangeline F.Y. Young
  • *Corresponding author for this work

Research output: Contribution to journalJournal articlepeer-review

1 Citation (Scopus)

Abstract

The tremendous growth in deep learning (DL) applications has created an exponential demand for computing power, which leads to the rise of AI-specific hardware. Targeted toward accelerating computation-intensive DL applications, AI hardware, including but not limited to GPGPU, TPU, ASICs, etc., have been adopted ubiquitously. As a result, domain-specific CAD tools play more and more important roles and have been deeply involved in both the design and compilation stages of modern AI hardware. Recently, ISPD 2020 contest introduced a special challenge targeting at the physical mapping of neural network workloads onto the largest commercial DL accelerator, CS-1 wafer-scale engine (WSE). In this article, we proposed CU.POKer, a high-performance engine fully customized for WSE's deep neural network workload placement challenge. A provably optimal placeable kernel candidate searching scheme and a data-flow-aware placement tool are developed accordingly to ensure the state-of-the-art (SOTA) quality on the real industrial benchmarks. Experimental results on ISPD 2020 contest evaluation suites demonstrated the superiority of our proposed framework over not only the SOTA placer but also the conventional heuristics used in general floorplanning.

Original languageEnglish
Pages (from-to)1888-1901
Number of pages14
JournalIEEE Transactions on Computer-Aided Design of Integrated Circuits and Systems
Volume41
Issue number6
Early online date12 Jul 2021
DOIs
Publication statusPublished - 1 Jun 2022

UN SDGs

This output contributes to the following UN Sustainable Development Goals (SDGs)

  1. SDG 9 - Industry, Innovation, and Infrastructure
    SDG 9 Industry, Innovation, and Infrastructure

User-Defined Keywords

  • AI chip compilation
  • deep learning (DL) accelerator
  • neural network workload placement
  • wafer-scale engine (WSE)

Fingerprint

Dive into the research topics of 'CU.POKer: Placing DNNs on WSE With Optimal Kernel Sizing and Efficient Protocol Optimization'. Together they form a unique fingerprint.

Cite this