Neuromorphic Computing: Challenges, Opportunities Including Materials, Algorithms, Devices & Ethics

This new research paper titled "2022 roadmap on neuromorphic computing and engineering" is from numerous researchers at Technical University of Denmark, Instituto de Microelectrónica de Sevilla, CSIC, University of Seville, and many others. Partial Abstract: "The aim of this roadmap is to present a snapshot of the present state of neuromorphic technology and provide an opinion on the chall... » read more

Novel Spintronic Neuro-mimetic Device Emulating the LIF Neuron Dynamics w/High Energy Efficiency & Compact Footprints

New technical paper titled "Noise resilient leaky integrate-and-fire neurons based on multi-domain spintronic devices" from researchers at Purdue University. Abstract "The capability of emulating neural functionalities efficiently in hardware is crucial for building neuromorphic computing systems. While various types of neuro-mimetic devices have been investigated, it remains challenging to... » read more

Always-On Sub-Microwatt Spiking Neural Network Based on Spike-Driven Clock- and Power-Gating for an Ultra-Low-Power Intelligent Device

Abstract: "This paper presents a novel spiking neural network (SNN) classifier architecture for enabling always-on artificial intelligent (AI) functions, such as keyword spotting (KWS) and visual wake-up, in ultra-low-power internet-of-things (IoT) devices. Such always-on hardware tends to dominate the power efficiency of an IoT device and therefore it is paramount to minimize its power diss... » read more

An Event-Driven and Fully Synthesizable Architecture for Spiking Neural Networks

Abstract:  "The development of brain-inspired neuromorphic computing architectures as a paradigm for Artificial Intelligence (AI) at the edge is a candidate solution that can meet strict energy and cost reduction constraints in the Internet of Things (IoT) application areas. Toward this goal, we present μBrain: the first digital yet fully event-driven without clock architecture, with co-lo... » read more

Benchmarking Highly Parallel Hardware for Spiking Neural Networks in Robotics

Abstract: "Animal brains still outperform even the most performant machines with significantly lower speed. Nonetheless, impressive progress has been made in robotics in the areas of vision, motion- and path planning in the last decades. Brain-inspired Spiking Neural Networks (SNN) and the parallel hardware necessary to exploit their full potential have promising features for robotic applica... » read more

Easier And Faster Ways To Train AI

Training an AI model takes an extraordinary amount of effort and data. Leveraging existing training can save time and money, accelerating the release of new products that use the model. But there are a few ways this can be done, most notably through transfer and incremental learning, and each of them has its applications and tradeoffs. Transfer learning and incremental learning both take pre... » read more

11 Ways To Reduce AI Energy Consumption

As the machine-learning industry evolves, the focus has expanded from merely solving the problem to solving the problem better. “Better” often has meant accuracy or speed, but as data-center energy budgets explode and machine learning moves to the edge, energy consumption has taken its place alongside accuracy and speed as a critical issue. There are a number of approaches to neural netw... » read more

Edge-Inference Architectures Proliferate

First part of two parts. The second part will dive into basic architectural characteristics. The last year has seen a vast array of announcements of new machine-learning (ML) architectures for edge inference. Unburdened by the need to support training, but tasked with low latency, the devices exhibit extremely varied approaches to ML inference. “Architecture is changing both in the comp... » read more

Neural Networks Without Matrix Math

The challenge of speeding up AI systems typically means adding more processing elements and pruning the algorithms, but those approaches aren't the only path forward. Almost all commercial machine learning applications depend on artificial neural networks, which are trained using large datasets with a back-propagation algorithm. The network first analyzes a training example, typically assign... » read more

Zeroing In On Biological Computing

Artificial spiking neural networks need to replicate both excitatory and inhibitory biological neurons in order to emulate the neural activation patterns seen in biological brains. Doing this with CMOS-based designs is challenging because of the large circuit footprint required. However, researchers at HP Labs observed that one biologically plausible model, the Hodgkins-Huxley model, is math... » read more

← Older posts