73 research outputs found
Gibbs Sampling with Low-Power Spiking Digital Neurons
Restricted Boltzmann Machines and Deep Belief Networks have been successfully
used in a wide variety of applications including image classification and
speech recognition. Inference and learning in these algorithms uses a Markov
Chain Monte Carlo procedure called Gibbs sampling. A sigmoidal function forms
the kernel of this sampler which can be realized from the firing statistics of
noisy integrate-and-fire neurons on a neuromorphic VLSI substrate. This paper
demonstrates such an implementation on an array of digital spiking neurons with
stochastic leak and threshold properties for inference tasks and presents some
key performance metrics for such a hardware-based sampler in both the
generative and discriminative contexts.Comment: Accepted at ISCAS 201
Anatomy of a cortical simulator
Insights into brain’s high-level computational principles will lead to novel cognitive systems, computing architectures, programming paradigms, and numerous practical applications. An important step towards this end is the study of large networks of cortical spiking neurons. We have built a cortical simulator, C2, incorporating several algorithmic enhancements to optimize the simulation scale and time, through: computationally efficient simulation of neurons in a clock-driven and synapses in an event-driven fashion; memory efficient representation of simulation state; and communication efficient message exchanges. Using phenomenological, single-compartment models of spiking neurons and synapses with spike-timing dependent plasticity, we represented a rat-scale cortical model (55 million neurons, 442 billion synapses) in 8TB memory of a 32,768processor BlueGene/L. With 1 millisecond resolution for neuronal dynamics and 1-20 milliseconds axonal delays, C2 can simulate 1 second of model time in 9 seconds per Hertz of average neuronal firing rate. In summary, by combining state-of-the-art hardware with innovative algorithms and software design, we simultaneously achieved unprecedented time-to-solution on an unprecedented problem size. 1
Convolutional Networks for Fast, Energy-Efficient Neuromorphic Computing
Deep networks are now able to achieve human-level performance on a broad
spectrum of recognition tasks. Independently, neuromorphic computing has now
demonstrated unprecedented energy-efficiency through a new chip architecture
based on spiking neurons, low precision synapses, and a scalable communication
network. Here, we demonstrate that neuromorphic computing, despite its novel
architectural primitives, can implement deep convolution networks that i)
approach state-of-the-art classification accuracy across 8 standard datasets,
encompassing vision and speech, ii) perform inference while preserving the
hardware's underlying energy-efficiency and high throughput, running on the
aforementioned datasets at between 1200 and 2600 frames per second and using
between 25 and 275 mW (effectively > 6000 frames / sec / W) and iii) can be
specified and trained using backpropagation with the same ease-of-use as
contemporary deep learning. For the first time, the algorithmic power of deep
learning can be merged with the efficiency of neuromorphic processors, bringing
the promise of embedded, intelligent, brain-inspired computing one step closer.Comment: 7 pages, 6 figure
Efficient and Effective Methods for Mixed Precision Neural Network Quantization for Faster, Energy-efficient Inference
For efficient neural network inference, it is desirable to achieve
state-of-the-art accuracy with the simplest networks requiring the least
computation, memory, and power. Quantizing networks to lower precision is a
powerful technique for simplifying networks. As each layer of a network may
have different sensitivity to quantization, mixed precision quantization
methods selectively tune the precision of individual layers to achieve a
minimum drop in task performance (e.g., accuracy). To estimate the impact of
layer precision choice on task performance, two methods are introduced: i)
Entropy Approximation Guided Layer selection (EAGL) is fast and uses the
entropy of the weight distribution, and ii) Accuracy-aware Layer Precision
Selection (ALPS) is straightforward and relies on single epoch fine-tuning
after layer precision reduction. Using EAGL and ALPS for layer precision
selection, full-precision accuracy is recovered with a mix of 4-bit and 2-bit
layers for ResNet-50, ResNet-101 and BERT-base transformer networks,
demonstrating enhanced performance across the entire accuracy-throughput
frontier. The techniques demonstrate better performance than existing
techniques in several commensurate comparisons. Notably, this is accomplished
with significantly lesser computational time required to reach a solution
Implementation of Olfactory Bulb Glomerular-Layer Computations in a Digital Neurosynaptic Core
We present a biomimetic system that captures essential functional properties of the glomerular layer of the mammalian olfactory bulb, specifically including its capacity to decorrelate similar odor representations without foreknowledge of the statistical distributions of analyte features. Our system is based on a digital neuromorphic chip consisting of 256 leaky-integrate-and-fire neurons, 1024 × 256 crossbar synapses, and address-event representation communication circuits. The neural circuits configured in the chip reflect established connections among mitral cells, periglomerular cells, external tufted cells, and superficial short-axon cells within the olfactory bulb, and accept input from convergent sets of sensors configured as olfactory sensory neurons. This configuration generates functional transformations comparable to those observed in the glomerular layer of the mammalian olfactory bulb. Our circuits, consuming only 45 pJ of active power per spike with a power supply of 0.85 V, can be used as the first stage of processing in low-power artificial chemical sensing devices inspired by natural olfactory systems
Scaling, stability and synchronization in mouse-sized (and larger) cortical simulations
A Conceptual Cortical Surface Atlas
Volumetric, slice-based, 3-D atlases are invaluable tools for understanding complex cortical convolutions. We present a simple scheme to convert a slice-based atlas to a conceptual surface atlas that is easier to visualize and understand. The key idea is to unfold each slice into a one-dimensional vector, and concatenate a succession of these vectors – while maintaining as much spatial contiguity as possible – into a 2-D matrix. We illustrate our methodology using a coronal slice-based atlas of the Rhesus Monkey cortex. The conceptual surface-based atlases provide a useful complement to slice-based atlases for the purposes of indexing and browsing
- …
