In Memory Computing With Emerging Non Volatile Memory For Efficient Processing Of Deep Neural Networks PDF Download

Are you looking for read ebook online? Search for your book and save it on your Kindle device, PC, phones or tablets. Download In Memory Computing With Emerging Non Volatile Memory For Efficient Processing Of Deep Neural Networks PDF full book. Access full book title In Memory Computing With Emerging Non Volatile Memory For Efficient Processing Of Deep Neural Networks.

Efficient Processing of Deep Neural Networks

Efficient Processing of Deep Neural Networks
Author: Vivienne Sze
Publisher: Springer Nature
Total Pages: 254
Release: 2022-05-31
Genre: Technology & Engineering
ISBN: 3031017668

Download Efficient Processing of Deep Neural Networks Book in PDF, ePub and Kindle

This book provides a structured treatment of the key principles and techniques for enabling efficient processing of deep neural networks (DNNs). DNNs are currently widely used for many artificial intelligence (AI) applications, including computer vision, speech recognition, and robotics. While DNNs deliver state-of-the-art accuracy on many AI tasks, it comes at the cost of high computational complexity. Therefore, techniques that enable efficient processing of deep neural networks to improve key metrics—such as energy-efficiency, throughput, and latency—without sacrificing accuracy or increasing hardware costs are critical to enabling the wide deployment of DNNs in AI systems. The book includes background on DNN processing; a description and taxonomy of hardware architectural approaches for designing DNN accelerators; key metrics for evaluating and comparing different designs; features of DNN processing that are amenable to hardware/algorithm co-design to improve energy efficiency and throughput; and opportunities for applying new technologies. Readers will find a structured introduction to the field as well as formalization and organization of key concepts from contemporary work that provide insights that may spark new ideas.


Energy Efficient Hardware Implementation of Neural Networks Using Emerging Non-Volatile Memory Devices

Energy Efficient Hardware Implementation of Neural Networks Using Emerging Non-Volatile Memory Devices
Author: Sangheon Oh
Publisher:
Total Pages: 0
Release: 2023
Genre:
ISBN:

Download Energy Efficient Hardware Implementation of Neural Networks Using Emerging Non-Volatile Memory Devices Book in PDF, ePub and Kindle

Deep learning based on neural networks emerged as a robust solution to various complex problems such as speech recognition and visual recognition. Deep learning relies on a great amount of iterative computation on a huge dataset. As we need to transfer a large amount of data and program between the CPU and the memory unit, the data transfer rate through a bus becomes a limiting factor for computing speed, which is known as Von Neumann bottleneck. Moreover, the data transfer between memory and computation spends a large amount of energy and cause significant delay. To overcome the limitation of Von Neumann bottleneck, neuromorphic computing with emerging nonvolatile memory (eNVM) devices has been proposed to perform iterative calculations in memory without transferring data to a processor. This dissertation presents energy efficient hardware implementation of neuromorphic computing applications using phase change memory (PCM), subquantum conductive bridge random access memory (CBRAM), Ag-based CBRAM, and CuOx-based resistive random access memory (RRAM). Although substantial progress has been made towards in-memory computing with synaptic devices, compact nanodevices implementing non-linear activation functions for efficient full-hardware implementation of deep neural networks is still missing. Since DNNs need to have a very large number of activations to achieve high accuracy, it is critical to develop energy and area efficient implementations of activation functions, which can be integrated on the periphery of the synaptic arrays. In this dissertation, we demonstrate a Mott activation neuron that implements the rectified linear unit function in the analogue domain. The integration of Mott activation neurons with a CBRAM crossbar array is also demonstrated in this dissertation.


Neuro-inspired Computing Using Emerging Non-Volatile Memories

Neuro-inspired Computing Using Emerging Non-Volatile Memories
Author: Yuhan Shi
Publisher:
Total Pages: 0
Release: 2023
Genre:
ISBN:

Download Neuro-inspired Computing Using Emerging Non-Volatile Memories Book in PDF, ePub and Kindle

Data movement between separate processing and memory units in traditional von Neumann computing systems is costly in terms of time and energy. The problem is aggravated by the recent explosive growth in data intensive applications related to artificial intelligence. In-memory computing has been proposed as an alternative approach where computational tasks can be performed directly in memory without shuttling back and forth between the processing and memory units. Memory is at the heart of in-memory computing. Technology scaling of mainstream memory technologies, such as static random-access memory (SRAM) and Dynamic random-access memory (DRAM), is increasingly constrained by fundamental technology limits. The recent research progress of various emerging nonvolatile memory (eNVM) device technologies, such as resistive random-access memory (RRAM), phase-change memory (PCM), conductive bridging random-access memory (CBRAM), ferroelectric random-access memory (FeRAM) and spin-transfer torque magnetoresistive random-access memory (STT-MRAM), have drawn tremendous attentions owing to its high speed, low cost, excellent scalability, enhanced storage density. Moreover, an eNVM based crossbar array can perform in-memory matrix vector multiplications in analog manner with high energy efficiency and provide potential opportunities for accelerating computation in various fields such as deep learning, scientific computing and computer vision. This dissertation presents research work on demonstrating a wide range of emerging memory device technologies (CBRAM, RRAM and STT-MRAM) for implementing neuro-inspired in-memory computing in several real-world applications using software and hardware co-design approach. Chapter 1 presents low energy subquantum CBRAM devices and a network pruning technique to reduce network-level energy consumption by hundreds to thousands fold. We showed low energy (10×-100× less than conventional memory technologies) and gradual switching characteristics of CBRAM as synaptic devices. We developed a network pruning algorithm that can be employed during spiking neural network (SNN) training to further reduce the energy by 10×. Using a 512 Kbit subquantum CBRAM array, we experimentally demonstrated high recognition accuracy on the MNIST dataset for digital implementation of unsupervised learning. Chapter 2 presents the details of SNN pruning algorithm that used in Chapter1. The pruning algorithms exploits the features of network weights and prune weights during the training based on neurons' spiking characteristics, leading significant energy saving when implemented in eNVM based in-memory computing hardware. Chapter 3 presents a benchmarking analysis for the potential use of STT-MRAM in in-memory computing against SRAM at deeply scaled technology nodes (14nm and 7nm). A C++ based benchmarking platform is developed and uses LeNet-5, a popular convolutional neural network model (CNN). The platform maps STT-MRAM based in-memory computing architectures to LeNet-5 and can estimate inference accuracy, energy, latency, and area accurately for proposed architectures at different technology nodes compared against SRAM. Chapter 4 presents an adaptive quantization technique that compensates the accuracy loss due to limited conductance levels of PCM based synaptic devices and enables high-accuracy SNN unsupervised learning with low-precision PCM devices. The proposed adaptive quantization technique uses software and hardware co-design approach by designing software algorithms with consideration of real synaptic device characteristics and hardware limitations. Chapter 5 presents a real-world neural engineering application using in-memory computing. It presents an interface between eNVM based crossbar with neural electrodes to implement a real-time and high-energy efficient in-memory spike sorting system. A real-time hardware demonstration is performed using CuOx based eNVM crossbar to sort spike data in different brain regions recorded from multi-electrode arrays in animal experiments, which further extend the eNVM memory technologies for neural engineering applications. Chapter 6 presents a real-world deep learning application using in-memory computing. We demonstrated a direct integration of Ag-based conductive bridge random access memory (Ag-CBRAM) crossbar arrays with Mott-ReLU activation neurons for scalable, energy and area efficient hardware implementation of DNNs. Chapter 7 is the conclusion of this dissertation. The future directions of in-memory computing system based on eNVM technologies are discussed.


In-Memory Computing Hardware Accelerators for Data-Intensive Applications

In-Memory Computing Hardware Accelerators for Data-Intensive Applications
Author: Baker Mohammad
Publisher: Springer Nature
Total Pages: 145
Release: 2023-10-27
Genre: Technology & Engineering
ISBN: 303134233X

Download In-Memory Computing Hardware Accelerators for Data-Intensive Applications Book in PDF, ePub and Kindle

This book describes the state-of-the-art of technology and research on In-Memory Computing Hardware Accelerators for Data-Intensive Applications. The authors discuss how processing-centric computing has become insufficient to meet target requirements and how Memory-centric computing may be better suited for the needs of current applications. This reveals for readers how current and emerging memory technologies are causing a shift in the computing paradigm. The authors do deep-dive discussions on volatile and non-volatile memory technologies, covering their basic memory cell structures, operations, different computational memory designs and the challenges associated with them. Specific case studies and potential applications are provided along with their current status and commercial availability in the market.


In-/Near-Memory Computing

In-/Near-Memory Computing
Author: Daichi Fujiki
Publisher: Springer Nature
Total Pages: 124
Release: 2022-05-31
Genre: Technology & Engineering
ISBN: 3031017722

Download In-/Near-Memory Computing Book in PDF, ePub and Kindle

This book provides a structured introduction of the key concepts and techniques that enable in-/near-memory computing. For decades, processing-in-memory or near-memory computing has been attracting growing interest due to its potential to break the memory wall. Near-memory computing moves compute logic near the memory, and thereby reduces data movement. Recent work has also shown that certain memories can morph themselves into compute units by exploiting the physical properties of the memory cells, enabling in-situ computing in the memory array. While in- and near-memory computing can circumvent overheads related to data movement, it comes at the cost of restricted flexibility of data representation and computation, design challenges of compute capable memories, and difficulty in system and software integration. Therefore, wide deployment of in-/near-memory computing cannot be accomplished without techniques that enable efficient mapping of data-intensive applications to such devices, without sacrificing accuracy or increasing hardware costs excessively. This book describes various memory substrates amenable to in- and near-memory computing, architectural approaches for designing efficient and reliable computing devices, and opportunities for in-/near-memory acceleration of different classes of applications.


Non-Volatile In-Memory Computing by Spintronics

Non-Volatile In-Memory Computing by Spintronics
Author: Hao Yu
Publisher: Springer Nature
Total Pages: 147
Release: 2022-05-31
Genre: Technology & Engineering
ISBN: 3031020324

Download Non-Volatile In-Memory Computing by Spintronics Book in PDF, ePub and Kindle

Exa-scale computing needs to re-examine the existing hardware platform that can support intensive data-oriented computing. Since the main bottleneck is from memory, we aim to develop an energy-efficient in-memory computing platform in this book. First, the models of spin-transfer torque magnetic tunnel junction and racetrack memory are presented. Next, we show that the spintronics could be a candidate for future data-oriented computing for storage, logic, and interconnect. As a result, by utilizing spintronics, in-memory-based computing has been applied for data encryption and machine learning. The implementations of in-memory AES, Simon cipher, as well as interconnect are explained in details. In addition, in-memory-based machine learning and face recognition are also illustrated in this book.


Processing-in-Memory for AI

Processing-in-Memory for AI
Author: Joo-Young Kim
Publisher: Springer Nature
Total Pages: 168
Release: 2022-07-09
Genre: Technology & Engineering
ISBN: 3030987817

Download Processing-in-Memory for AI Book in PDF, ePub and Kindle

This book provides a comprehensive introduction to processing-in-memory (PIM) technology, from its architectures to circuits implementations on multiple memory types and describes how it can be a viable computer architecture in the era of AI and big data. The authors summarize the challenges of AI hardware systems, processing-in-memory (PIM) constraints and approaches to derive system-level requirements for a practical and feasible PIM solution. The presentation focuses on feasible PIM solutions that can be implemented and used in real systems, including architectures, circuits, and implementation cases for each major memory type (SRAM, DRAM, and ReRAM).


Machine Learning and Non-volatile Memories

Machine Learning and Non-volatile Memories
Author: Rino Micheloni
Publisher: Springer Nature
Total Pages: 178
Release: 2022-05-25
Genre: Technology & Engineering
ISBN: 303103841X

Download Machine Learning and Non-volatile Memories Book in PDF, ePub and Kindle

This book presents the basics of both NAND flash storage and machine learning, detailing the storage problems the latter can help to solve. At a first sight, machine learning and non-volatile memories seem very far away from each other. Machine learning implies mathematics, algorithms and a lot of computation; non-volatile memories are solid-state devices used to store information, having the amazing capability of retaining the information even without power supply. This book will help the reader understand how these two worlds can work together, bringing a lot of value to each other. In particular, the book covers two main fields of application: analog neural networks (NNs) and solid-state drives (SSDs). After reviewing the basics of machine learning in Chapter 1, Chapter 2 shows how neural networks can mimic the human brain; to accomplish this result, neural networks have to perform a specific computation called vector-by-matrix (VbM) multiplication, which is particularly power hungry. In the digital domain, VbM is implemented by means of logic gates which dictate both the area occupation and the power consumption; the combination of the two poses serious challenges to the hardware scalability, thus limiting the size of the neural network itself, especially in terms of the number of processable inputs and outputs. Non-volatile memories (phase change memories in Chapter 3, resistive memories in Chapter 4, and 3D flash memories in Chapter 5 and Chapter 6) enable the analog implementation of the VbM (also called “neuromorphic architecture”), which can easily beat the equivalent digital implementation in terms of both speed and energy consumption. SSDs and flash memories are strictly coupled together; as 3D flash scales, there is a significant amount of work that has to be done in order to optimize the overall performances of SSDs. Machine learning has emerged as a viable solution in many stages of this process. After introducing the main flash reliability issues, Chapter 7 shows both supervised and un-supervised machine learning techniques that can be applied to NAND. In addition, Chapter 7 deals with algorithms and techniques for a pro-active reliability management of SSDs. Last but not least, the last section of Chapter 7 discusses the next challenge for machine learning in the context of the so-called computational storage. No doubt that machine learning and non-volatile memories can help each other, but we are just at the beginning of the journey; this book helps researchers understand the basics of each field by providing real application examples, hopefully, providing a good starting point for the next level of development.


Emerging Non-volatile Memory Technologies

Emerging Non-volatile Memory Technologies
Author: Wen Siang Lew
Publisher: Springer Nature
Total Pages: 439
Release: 2021-01-09
Genre: Science
ISBN: 9811569126

Download Emerging Non-volatile Memory Technologies Book in PDF, ePub and Kindle

This book offers a balanced and comprehensive guide to the core principles, fundamental properties, experimental approaches, and state-of-the-art applications of two major groups of emerging non-volatile memory technologies, i.e. spintronics-based devices as well as resistive switching devices, also known as Resistive Random Access Memory (RRAM). The first section presents different types of spintronic-based devices, i.e. magnetic tunnel junction (MTJ), domain wall, and skyrmion memory devices. This section describes how their developments have led to various promising applications, such as microwave oscillators, detectors, magnetic logic, and neuromorphic engineered systems. In the second half of the book, the underlying device physics supported by different experimental observations and modelling of RRAM devices are presented with memory array level implementation. An insight into RRAM desired properties as synaptic element in neuromorphic computing platforms from material and algorithms viewpoint is also discussed with specific example in automatic sound classification framework.