Tensor Processing Unit: Google's LSI designed for neural network pending

This session has NOT been accepted yet. If you would like to see it in the conference, please post it in social networks, as those numbers will be tallied and used as part of the selection criteria

Abstract

This session takes a detailed look at the architecture of Google’s Tensor Processing Unit (TPU), the LSI designed for neural network processing. We’ll cover quantization, CISC design, as well as the systolic array matrix unit that is the heart of the device. You’ll learn how a minimalistic design philosophy and a tight focus on neural network inference use-cases enabled Google to build and deploy such a high-performing device in just 15 months.

Session Information
Material Level Intermediate
Starts On N/A
Room
Session Duration 30 min
Spoken Language Japanese
Slide Language English
Photo Release Allow
Recording Release Allow
Materials Release Allow