49 research outputs found
Convolutional Networks for Fast, Energy-Efficient Neuromorphic Computing
Deep networks are now able to achieve human-level performance on a broad
spectrum of recognition tasks. Independently, neuromorphic computing has now
demonstrated unprecedented energy-efficiency through a new chip architecture
based on spiking neurons, low precision synapses, and a scalable communication
network. Here, we demonstrate that neuromorphic computing, despite its novel
architectural primitives, can implement deep convolution networks that i)
approach state-of-the-art classification accuracy across 8 standard datasets,
encompassing vision and speech, ii) perform inference while preserving the
hardware's underlying energy-efficiency and high throughput, running on the
aforementioned datasets at between 1200 and 2600 frames per second and using
between 25 and 275 mW (effectively > 6000 frames / sec / W) and iii) can be
specified and trained using backpropagation with the same ease-of-use as
contemporary deep learning. For the first time, the algorithmic power of deep
learning can be merged with the efficiency of neuromorphic processors, bringing
the promise of embedded, intelligent, brain-inspired computing one step closer.Comment: 7 pages, 6 figure
On relay placemement for deterministic line network
We consider a unicast communication problem where, a source transmits information to a destination through a wireless network with the help of k relays positioned on a line. We adopt the linear deterministic model to capture the wireless signal interactions and study the optimal placement of the relays so that the capacity from the source to the destination in the deterministic network is maximized. Analytical results are provided for a number of special cases, and the insights gained are used to provide a heuristic framework for designing large relay networks.
Analyzing the impact of system architecture on the scalability of OLTP engines for high-contention workloads
Vispark: GPU-accelerated distributed visual computing using spark
With the growing need of big-data processing in diverse application domains, MapReduce (e.g., Hadoop) has become one of the standard computing paradigms for large-scale computing on a cluster system. Despite its popularity, the current MapReduce framework suffers from inflexibility and inefficiency inherent to its programming model and system architecture. In order to address these problems, we propose Vispark, a novel extension of Spark for GPU-accelerated MapReduce processing on array-based scientific computing and image processing tasks. Vispark provides an easy-to-use, Python-like high-level language syntax and a novel data abstraction for MapReduce programming on a GPU cluster system. Vispark introduces a programming abstraction for accessing neighbor data in the mapper function, which greatly simplifies many image processing tasks using MapReduce by reducing memory footprints and bypassing the reduce stage. Vispark provides socket-based halo communication that synchronizes between data partitions transparently from the users, which is necessary for many scientific computing problems in distributed systems. Vispark also provides domain-specific functions and language supports specifically designed for high-performance computing and image processing applications. We demonstrate the performance of our prototype system on several visual computing tasks, such as image processing, volume rendering, K-means clustering, and heat transfer simulation.clos
Investigating Guided Extensive Reading And Vocabulary Knowledge Performance Among Remedial Esl Learners In A Public University In Malaysia
Penyelidikan menyokong pembacaan ekstensif, yang tertumpu pada pembelajaran kebetulan (incidental learning), sebagai wadah utama bagi perkembangan pengetahuan kosa kata bahasa kedua/asing.
Research supports extensive reading, which draws on incidental learning, as a primary tool for second/foreign language vocabulary knowledge development
Joint timing synchronization and channel estimation based on ZCZ sequence set in SC-MIMO-FDE system
Sequence Alignment Through the Looking Glass
AbstractRapid advances in sequencing technologies are producing genomic data on an unprecedented scale. The first, and often one of the most time consuming, step of genomic data analysis is sequence alignment, where sequenced reads must be aligned to a reference genome. Several years of research on alignment algorithms has led to the development of several state-of-the-art sequence aligners that can map tens of thousands of reads per second.In this work, we answer the question “How do sequence aligners utilize modern processors?” We examine four state-of-the-art aligners running on an Intel processor and identify that all aligners leave the processor substantially underutilized. We perform an in-depth microarchitectural analysis to explore the interaction between aligner software and processor hardware. We identify bottlenecks that lead to processor underutilization and discuss the implications of our analysis on next-generation sequence aligner design.</jats:p
