Professional Documents
Culture Documents
Cryptocurrency JStor Paper
Cryptocurrency JStor Paper
DOI:10.1145/ 3024928
such as the DAO vulnerability in June
Article development led by
queue.acm.org
2016—have captured the public imagi-
nation and the eye of the popular press.
In the meantime, academics have been
Expert-curated guides to busy, delivering new results in main-
the best of CS research. taining anonymity, ensuring usability,
detecting errors, and reasoning about
Research
decentralized markets, all through the
lens of these modern cryptocurrency
systems. It is a pleasure having two
academic experts deliver the latest up-
for Practice:
dates from the burgeoning body of aca-
demic research on this subject.
Next, Song Han provides an over-
view of hardware trends related to an-
Cryptocurrencies,
other long-studied academic problem
that has recently seen an explosion in
popularity: deep learning. Fueled by
large amounts of training data and in-
Blockchains, and
expensive parallel and scale-out com-
pute, deep-learning-model architec-
tures have seen a massive resurgence
Smart Contracts;
of interest based on their excellent
performance on traditionally diffi-
cult tasks such as image recognition.
Hardware for
These deep networks are compute-
intensive to train and evaluate, and
many of the best minds in computer
Deep Learning
systems (for example, the team that
developed MapReduce) and AI are
working to improve them. As a result,
Song has provided a fantastic over-
view of recent advances devoted to
using hardware and hardware-aware
techniques to compress networks,
improve their performance, and re-
duce their often large amounts of en-
ergy consumption.
As always, our goal in this column is
OUR FOURTH INSTALLMENT of Research for Practice covers to allow our readers to become experts
in the latest topics in computer science
two of the hottest topics in computer science research research in a weekend afternoon’s
and practice: cryptocurrencies and deep learning. worth of reading. To facilitate this
First, Arvind Narayanan and Andrew Miller, co- process, as always, we have provided
open access to the ACM Digital Library
authors of the increasingly popular open access Bitcoin for the relevant citations from these
textbook, provide an overview of ongoing research in selections so you can read the research
results in full. Please enjoy!
cryptocurrencies. This is a topic with a long history —Peter Bailis
in the academic literature that has recently come to
prominence with the rise of Bitcoin, blockchains, and Peter Bailis is an assistant professor of computer science
at Stanford University. His research in the Future Data
similar implementations of advanced, decentralized Systems group (futuredata.stanford.edu/) focuses on
the design and implementation of next-generation data-
protocols. These developments—and colorful exploits intensive systems.
using and applying cryptocurrencies: curacy. On top of the compression al- on Architectural Support for Programming
privacy, security, and interfacing with gorithm, EIE is a hardware accelerator Languages and Operating Systems, 2014.
http://pages.saclay.inria.fr/olivier.temam/files/
the real world. These will be fertile ar- that works directly on the compressed
eval/CDSWWCT14.pdf.
eas of research and development in model and solves the problem of ir-
the years to come. regular computation patterns (sparsity Recent state-of-the-art CNNs and
and indirection) brought about by the DNNs are characterized by their large
Arvind Narayanan is an assistant professor of computer compression algorithm. EIE efficiently sizes. With layers of thousands of neu-
science at Princeton, where he leads a research team
investigating the security, anonymity, and stability parallelizes the compressed model rons and millions of synapses, they
of cryptocurrencies as well as novel applications onto multiple processing elements and place a special emphasis on interac-
of blockchains. He also leads the Princeton Web
Transparency and Accountability Project, to uncover how proposes an efficient way of partition- tions with memory. DianNao is an
companies collect and use our personal information.
ing and load balancing both the storage accelerator for large-scale CNNs and
Andrew Miller is an assistant professor in Electrical and the computation. This achieves a DNNs, with a special emphasis on
and Computer Engineering at the University of Illinois
at Urbana-Champaign. He is an associate director of the
speedup of 189/13 times and an energy the impact of memory on accelerator
Initiative for Cryptocurrencies and Contracts (IC3) at efficiency improvement of 24,000/3,400 design, performance, and energy. It
Cornell and an advisor to the Zcash project.
times over a modern CPU/GPU. takes advantage of dedicated storage,
which is key for achieving good perfor-
Hardware For Deep Optimized Dataflow mance and power. By carefully exploit-
Learning ing the locality properties of neural
By Song Han network models, and by introducing
Chen, Y.-H., Emer, J., Sze, V.
Deep neural networks Eyeriss: A spatial architecture for energy-
storage structures custom designed
(DNNs) have evolved to efficient dataflow for convolutional neural to take advantage of these proper-
a state-of-the-art tech- networks. In Proceedings of the International ties, DianNao shows it is possible to
nique for machine-learning tasks Symposium on Computer Architecture, design a machine-learning accelera-
ranging from computer vision to 2016. https://www.researchgate.net/ tor capable of high performance in a
publication/301891800_Eyeriss_A_Spatial_
speech recognition to natural lan- Architecture_for_Energy-Efficient_Dataflow_ very small footprint. It is possible to
guage processing. Deep-learning for_Convolutional_Neural_Networks. achieve a speedup of 117.87 times and
algorithms, however, are both com- an energy reduction of 21.08 times
putationally and memory intensive, Deep-learning algorithms are memory over a 128-bit 2GHz SIMD (single in-
making them power-hungry to deploy intensive, and accessing memory con- struction, multiple data) core with a
on embedded systems. Running deep- sumes energy more than two orders of normal cache hierarchy.
learning algorithms in real time at magnitude more than ALU (arithmetic
subwatt power consumption would be logic unit) operations. Thus, it’s criti- Looking Forward
ideal in embedded devices, but gener- cal to develop dataflow that can reduce Specialized hardware will be a key
al-purpose hardware is not providing memory reference. Eyeriss presents a solution to make deep-learning al-
satisfying energy efficiency to deploy novel dataflow called RS (row-station- gorithms faster and more energy ef-
such a DNN. The three papers pre- ary) that minimizes data-movement ficient. Reducing memory footprint
sented here suggest ways to solve this energy consumption on a spatial ar- is the most critical issue. The papers
problem with specialized hardware. chitecture. This is realized by exploit- presented here demonstrate three
ing local data reuse of filter weights and ways to solve this problem: optimize
The Compressed Model feature map pixels (that is, activations) in both algorithm and hardware and
the high-dimensional convolutions, and accelerate the compressed model;
by minimizing data movement of partial use an optimized dataflow to sched-
Han, S., Liu, X., Mao, H., Pu, J., Pedram, A.,
Horowitz, M.A., Dally, W.J. sum accumulations. Unlike dataflows ule the data movements; and design
EIE: Efficient inference engine on compressed used in existing designs, which reduce dedicated memory buffers for the
deep neural network. In Proceedings of only certain types of data movement, the weights, input activations, and out-
the International Symposium on Computer proposed RS dataflow can adapt to dif- put activations. We can look forward
Architecture, 2016.
ferent CNN (convolutional neural net- to seeing more artificial intelligence
https://arxiv.org/pdf/1602.01528v2.pdf.
work) shape configurations and reduce applications benefit from such hard-
This work is a combination of algo- all types of data movement through ware optimizations, putting AI every-
rithm optimization and hardware spe- maximum use of PE (processing engine) where, in every device in our lives.
cialization. EIE (efficient inference local storage, direct inter-PE communi-
engine) starts with a deep-learning- cation, and spatial parallelism. Song Han is a Ph.D. student at Stanford University,
Stanford, CA. He proposed deep compression that can
model compression algorithm that first compress state-of-the art CNNs by 10–49 times and
prunes neural networks by 9–13 times Small-Footprint Accelerator designed EIE (efficient inference engine), a hardware
architecture that does inference directly on the
without hurting accuracy, which leads compressed sparse model.
to both computation saving and mem-
ory saving; next, using pruning plus Chen, T., Wang, J., Du, Z., Wu, C.,
Sun, N., Chen, Y., Temam, O.
weight sharing and Huffman coding, DianNao: A small-footprint high-throughput
EIE further compresses the network accelerator for ubiquitous machine-learning. In Copyright held by owner(s)/authors.
35–49 times, again without hurting ac- Proceedings of the International Conference Publication rights licensed to ACM. $15.00