Gnawali et al., 2019 - Google Patents
Reliability enhancements in memristive neural network architecturesGnawali et al., 2019
View PDF- Document ID
- 4586534792772466931
- Author
- Gnawali K
- Paudel B
- Mozaffari S
- Tragoudas S
- Publication year
- Publication venue
- IEEE Transactions on Nanotechnology
External Links
Snippet
Memristive Crossbar Arrays (MCAs) are widely used in designing fast and compact neuromorphic systems. However, such systems require on-chip implementation of the backpropagation algorithm to accommodate process variations. This paper proposes a low …
- 230000001537 neural 0 title abstract description 38
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06N—COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computer systems based on biological models
- G06N3/02—Computer systems based on biological models using neural network models
- G06N3/06—Physical realisation, i.e. hardware implementation of neural networks, neurons or parts of neurons
- G06N3/063—Physical realisation, i.e. hardware implementation of neural networks, neurons or parts of neurons using electronic means
- G06N3/0635—Physical realisation, i.e. hardware implementation of neural networks, neurons or parts of neurons using electronic means using analogue means
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06N—COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computer systems based on biological models
- G06N3/02—Computer systems based on biological models using neural network models
- G06N3/04—Architectures, e.g. interconnection topology
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F17/00—Digital computing or data processing equipment or methods, specially adapted for specific functions
- G06F17/50—Computer-aided design
- G06F17/5009—Computer-aided design using simulation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06N—COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N99/00—Subject matter not provided for in other groups of this subclass
- G06N99/005—Learning machines, i.e. computer in which a programme is changed according to experience gained by the machine itself during a complete run
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06N—COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computer systems based on biological models
- G06N3/12—Computer systems based on biological models using genetic models
- G06N3/126—Genetic algorithms, i.e. information processing using digital simulations of the genetic system
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F7/00—Methods or arrangements for processing data by operating upon the order or content of the data handled
- G06F7/58—Random or pseudo-random number generators
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F17/00—Digital computing or data processing equipment or methods, specially adapted for specific functions
- G06F17/10—Complex mathematical operations
-
- G—PHYSICS
- G11—INFORMATION STORAGE
- G11C—STATIC STORES
- G11C11/00—Digital stores characterised by the use of particular electric or magnetic storage elements; Storage elements therefor
- G11C11/56—Digital stores characterised by the use of particular electric or magnetic storage elements; Storage elements therefor using storage elements with more than two stable states represented by steps, e.g. of voltage, current, phase, frequency
Similar Documents
Publication | Publication Date | Title |
---|---|---|
Kendall et al. | Training end-to-end analog neural networks with equilibrium propagation | |
Kaiser et al. | Hardware-aware in situ learning based on stochastic magnetic tunnel junctions | |
Marinella et al. | Multiscale co-design analysis of energy, latency, area, and accuracy of a ReRAM analog neural training accelerator | |
Gokmen et al. | Training deep convolutional neural networks with resistive cross-point devices | |
Hu et al. | Dot-product engine for neuromorphic computing: Programming 1T1M crossbar to accelerate matrix-vector multiplication | |
Kadetotad et al. | Parallel architecture with resistive crosspoint array for dictionary learning acceleration | |
Kim et al. | Input voltage mapping optimized for resistive memory-based deep neural network hardware | |
Roy et al. | TxSim: Modeling training of deep neural networks on resistive crossbar systems | |
Charan et al. | Accurate inference with inaccurate rram devices: A joint algorithm-design solution | |
Fouda et al. | Mask technique for fast and efficient training of binary resistive crossbar arrays | |
Gnawali et al. | Reliability enhancements in memristive neural network architectures | |
Yakopcic et al. | Flexible memristor based neuromorphic system for implementing multi-layer neural network algorithms | |
Yan et al. | Training memristor-based multilayer neuromorphic networks with SGD, momentum and adaptive learning rates | |
US20240005162A1 (en) | Error-triggered learning of multi-layer memristive spiking neural networks | |
Rzeszut et al. | Multi-state MRAM cells for hardware neuromorphic computing | |
Ansari et al. | PHAX: Physical characteristics aware ex-situ training framework for inverter-based memristive neuromorphic circuits | |
Djeffal et al. | An approach based on neural computation to simulate the nanoscale CMOS circuits: Application to the simulation of CMOS inverter | |
Vahdat et al. | Reliability enhancement of inverter-based memristor crossbar neural networks using mathematical analysis of circuit non-idealities | |
Antolini et al. | Combined HW/SW drift and variability mitigation for PCM-based analog in-memory computing for neural network applications | |
Ansari et al. | OCTAN: An on-chip training algorithm for memristive neuromorphic circuits | |
Greenberg-Toledo et al. | Supporting the momentum training algorithm using a memristor-based synapse | |
Pagliarini et al. | A probabilistic synapse with strained MTJs for spiking neural networks | |
Zhang et al. | Neural network training with stochastic hardware models and software abstractions | |
Cao et al. | Parasitic-aware modeling and neural network training scheme for energy-efficient processing-in-memory with resistive crossbar array | |
Mozaffari et al. | An aging resilient neural network architecture |