Professional Documents
Culture Documents
Full download Accelerators for Convolutional Neural Networks Arslan Munir file pdf all chapter on 2024
Full download Accelerators for Convolutional Neural Networks Arslan Munir file pdf all chapter on 2024
https://ebookmass.com/product/graph-neural-networks-for-
efficient-learning-of-mechanical-properties-of-polycrystals-
jonathan-m-hestroffer/
https://ebookmass.com/product/physics-informed-neural-networks-m-
raissi-p-perdikaris-g-e-karniadakis/
https://ebookmass.com/product/analysis-and-visualization-of-
discrete-data-using-neural-networks-koji-koyamada/
https://ebookmass.com/product/building-computer-vision-
applications-using-artificial-neural-networks-2nd-edition-
shamshad-ansari/
Quantum-like Networks: An Approach to Neural Behavior
through their Mathematics and Logic 1st Edition Stephen
A. Selesnick
https://ebookmass.com/product/quantum-like-networks-an-approach-
to-neural-behavior-through-their-mathematics-and-logic-1st-
edition-stephen-a-selesnick/
https://ebookmass.com/product/ai-applications-to-communications-
and-information-technologies-the-role-of-ultra-deep-neural-
networks-daniel-minoli/
https://ebookmass.com/product/python-deep-learning-understand-
how-deep-neural-networks-work-and-apply-them-to-real-world-
tasks-3rd-edition-vasilev/
https://ebookmass.com/product/networks-for-social-impact-
michelle-shumate/
https://ebookmass.com/product/startup-program-design-a-practical-
guide-for-creating-corporate-accelerators-and-incubators-paolo-
lombardi/
Accelerators for Convolutional Neural Networks
IEEE Press
445 Hoes Lane
Piscataway, NJ 08854
Arslan Munir
Kansas State University
USA
Joonho Kong
Kyungpook National University
South Korea
No part of this publication may be reproduced, stored in a retrieval system, or transmitted in any
form or by any means, electronic, mechanical, photocopying, recording, scanning, or otherwise,
except as permitted under Section 107 or 108 of the 1976 United States Copyright Act, without
either the prior written permission of the Publisher, or authorization through payment of the
appropriate per-copy fee to the Copyright Clearance Center, Inc., 222 Rosewood Drive, Danvers,
MA 01923, (978) 750-8400, fax (978) 750-4470, or on the web at www.copyright.com. Requests to
the Publisher for permission should be addressed to the Permissions Department, John Wiley &
Sons, Inc., 111 River Street, Hoboken, NJ 07030, (201) 748-6011, fax (201) 748-6008, or online at
http://www.wiley.com/go/permission.
Trademarks: Wiley and the Wiley logo are trademarks or registered trademarks of John Wiley &
Sons, Inc. and/or its affiliates in the United States and other countries and may not be used
without written permission. All other trademarks are the property of their respective owners.
John Wiley & Sons, Inc. is not associated with any product or vendor mentioned in this book.
Limit of Liability/Disclaimer of Warranty: While the publisher and author have used their best
efforts in preparing this book, they make no representations or warranties with respect to the
accuracy or completeness of the contents of this book and specifically disclaim any implied
warranties of merchantability or fitness for a particular purpose. No warranty may be created or
extended by sales representatives or written sales materials. The advice and strategies contained
herein may not be suitable for your situation. You should consult with a professional where
appropriate. Further, readers should be aware that websites listed in this work may have
changed or disappeared between when this work was written and when it is read. Neither the
publisher nor authors shall be liable for any loss of profit or any other commercial damages,
including but not limited to special, incidental, consequential, or other damages.
For general information on our other products and services or for technical support, please
contact our Customer Care Department within the United States at (800) 762-2974, outside the
United States at (317) 572-3993 or fax (317) 572-4002.
Wiley also publishes its books in a variety of electronic formats. Some content that appears in
print may not be available in electronic formats. For more information about Wiley products,
visit our web site at www.wiley.com.
Joonho Kong dedicates this book to his wife Jiyeon, children Eunseo and Eunyu,
and his parents for their continuous support.
Mahmood Azhar Qureshi dedicates this book to his parents, siblings, and his wife
Kiran, all of whom provided continuous support throughout his academic and
professional career.
vii
Contents
Part I Overview 1
1 Introduction 3
1.1 History and Applications 5
1.2 Pitfalls of High-Accuracy DNNs/CNNs 6
1.2.1 Compute and Energy Bottleneck 6
1.2.2 Sparsity Considerations 9
1.3 Chapter Summary 11
2.3.3 GoogleNet 27
2.3.4 SqueezeNet 27
2.3.5 Binary Neural Networks 29
2.3.6 EfficientNet 29
2.4 Popular CNN Datasets 30
2.4.1 MNIST Dataset 30
2.4.2 CIFAR 30
2.4.3 ImageNet 31
2.5 CNN Processing Hardware 31
2.5.1 Temporal Architectures 32
2.5.2 Spatial Architectures 34
2.5.3 Near-Memory Processing 36
2.6 Chapter Summary 37
16 Conclusions 259
References 265
Index 285
xiii
Preface
weights and activations in CNNs are zero and result in ineffectual computations)
to deliver high effective throughput. The book then presents a sparse CNN
accelerator that performs in situ decompression and convolution of sparse input
feature maps. Afterwards, the book discusses a sparse CNN accelerator, which
has the capability to actively skip a huge number of ineffective computations (i.e.,
computations involving zero weights and/or activations), while only favoring
effective computations (nonzero weights and nonzero activations) to drastically
improve the hardware utilization. The book then presents another sparse CNN
accelerator that uses sparse binary mask representation to actively lookahead
into sparse computations, and dynamically schedule its computational threads
to maximize the thread utilization and throughput. The fifth part of the book
targets hardware/software co-design and co-scheduling for CNN acceleration.
The book discusses hardware/software co-design and co-scheduling that can
lead to better optimization and utilization of the available hardware resources
for CNN acceleration. The book summarizes recent works on hardware/software
co-design and scheduling. The book then presents a technique that utilizes
software, algorithm, and hardware co-design to reduce the response time of
CNN inferences. Afterwards, the book discusses a CPU-accelerator co-scheduling
technique, which co-utilizes the CPU and CNN accelerators to expedite the CNN
inference. The book also provides directions for future research and development
for CNN accelerators.
This is the first book on the subject of accelerators for CNNs that introduces
readers to advances and state-of-the-art research in design of CNN accelerators.
This book can serve as a good reference for students, researchers, and practi-
tioners working in the area of hardware design, computer architecture, and AI
acceleration.
Part I
Overview
Another random document with
no related content on Scribd:
prohibition against accepting unsolicited donations from donors in
such states who approach us with offers to donate.
Please check the Project Gutenberg web pages for current donation
methods and addresses. Donations are accepted in a number of
other ways including checks, online payments and credit card
donations. To donate, please visit: www.gutenberg.org/donate.
Most people start at our website which has the main PG search
facility: www.gutenberg.org.