Skip to main content

and
  1. No Access

    Chapter

    Conclusion

    The use of deep neural networks (DNNs) has recently seen explosive growth. They are currently widely used for many artificial intelligence (AI) applications including computer vision, speech recognition, and r...

    Vivienne Sze, Yu-Hsin Chen, Tien-Ju Yang in Efficient Processing of Deep Neural Networ… (2020)

  2. No Access

    Chapter

    Kernel Computation

    The fundamental computation of both CONV and FC layers described in Chapter 2 are multiply-and-accumulate (MAC) operations. Because there are negligible dependencies between these operations and the accumulati...

    Vivienne Sze, Yu-Hsin Chen, Tien-Ju Yang in Efficient Processing of Deep Neural Networ… (2020)

  3. No Access

    Chapter

    Introduction

    Deep neural networks (DNNs) are currently the foundation for many modern artificial intelligence (AI) applications [5]. Since the breakthrough application of DNNs to speech recognition [6] and image recognition1 ...

    Vivienne Sze, Yu-Hsin Chen, Tien-Ju Yang in Efficient Processing of Deep Neural Networ… (2020)

  4. No Access

    Book

  5. No Access

    Chapter

    Overview of Deep Neural Networks

    Deep Neural Networks (DNNs) come in a wide variety of shapes and sizes depending on the application.1 The popular shapes and sizes are also evolving rapidly to improve accuracy and efficiency. In all cases, the i...

    Vivienne Sze, Yu-Hsin Chen, Tien-Ju Yang in Efficient Processing of Deep Neural Networ… (2020)

  6. No Access

    Chapter

    Operation Map** on Specialized Hardware

    In Chapter 5, we discussed various key design considerations and techniques for the implementation of specialized DNN hardware. Also introduced was the notion of the map** of the computation for a particular wo...

    Vivienne Sze, Yu-Hsin Chen, Tien-Ju Yang in Efficient Processing of Deep Neural Networ… (2020)

  7. No Access

    Chapter

    Exploiting Sparsity

    A salient characteristic of the data used in DNN computations is that it is (or can be made to be) sparse. By saying that the data is sparse, we are referring to the fact that there are many repeated values in...

    Vivienne Sze, Yu-Hsin Chen, Tien-Ju Yang in Efficient Processing of Deep Neural Networ… (2020)

  8. No Access

    Chapter

    Advanced Technologies

    As highlighted throughout the previous chapters, data movement dominates energy consumption. The energy is consumed both in the access to the memory as well as the transfer of the data. The associated physical...

    Vivienne Sze, Yu-Hsin Chen, Tien-Ju Yang in Efficient Processing of Deep Neural Networ… (2020)

  9. No Access

    Chapter

    Key Metrics and Design Objectives

    Over the past few years, there has been a significant amount of research on efficient processing of DNNs. Accordingly, it is important to discuss the key metrics that one should consider when comparing and eva...

    Vivienne Sze, Yu-Hsin Chen, Tien-Ju Yang in Efficient Processing of Deep Neural Networ… (2020)

  10. No Access

    Chapter

    Designing DNN Accelerators

    In Chapter 4, we discussed how DNN processing can undergo transforms to leverage optimized libraries or reduce the number of operations, specifically multiplications, in order to achieve higher performance (i....

    Vivienne Sze, Yu-Hsin Chen, Tien-Ju Yang in Efficient Processing of Deep Neural Networ… (2020)

  11. No Access

    Chapter

    Reducing Precision

    As highlighted in the previous chapters, data movement dominates energy consumption and can affect the throughput for memory-bound systems. One way to address this issue to reduce the number of bits (bit width...

    Vivienne Sze, Yu-Hsin Chen, Tien-Ju Yang in Efficient Processing of Deep Neural Networ… (2020)

  12. No Access

    Chapter

    Designing Efficient DNN Models

    The previous two chapters discussed the use of DNN model and hardware co-design approaches, such as reducing precision (Chapter 7) and exploiting sparsity (Chapter 8), to reduce storage requirements, data move...

    Vivienne Sze, Yu-Hsin Chen, Tien-Ju Yang in Efficient Processing of Deep Neural Networ… (2020)

  13. Chapter and Conference Paper

    NetAdapt: Platform-Aware Neural Network Adaptation for Mobile Applications

    This work proposes an algorithm, called NetAdapt, that automatically adapts a pre-trained deep neural network to a mobile platform given a resource budget. While many existing algorithms simplify networks based o...

    Tien-Ju Yang, Andrew Howard, Bo Chen, **ao Zhang, Alec Go in Computer Vision – ECCV 2018 (2018)