Yakopcic et al., 2017 - Google Patents
Memristor crossbar based implementation of a multilayer perceptronYakopcic et al., 2017
- Document ID
- 1849244547817729583
- Author
- Yakopcic C
- Taha T
- Publication year
- Publication venue
- 2017 IEEE National Aerospace and Electronics Conference (NAECON)
External Links
Snippet
This paper describes a memristor-based neuromorphic system that can be used for ex-situ training of various multi-layer neural network algorithms. This system is based on an analog neuron circuit that is capable of performing an accurate dot product calculation. The …
- 230000001537 neural 0 abstract description 32
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06N—COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computer systems based on biological models
- G06N3/02—Computer systems based on biological models using neural network models
- G06N3/06—Physical realisation, i.e. hardware implementation of neural networks, neurons or parts of neurons
- G06N3/063—Physical realisation, i.e. hardware implementation of neural networks, neurons or parts of neurons using electronic means
- G06N3/0635—Physical realisation, i.e. hardware implementation of neural networks, neurons or parts of neurons using electronic means using analogue means
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06N—COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computer systems based on biological models
- G06N3/02—Computer systems based on biological models using neural network models
- G06N3/04—Architectures, e.g. interconnection topology
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06N—COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computer systems based on biological models
- G06N3/02—Computer systems based on biological models using neural network models
- G06N3/08—Learning methods
- G06N3/082—Learning methods modifying the architecture, e.g. adding or deleting nodes or connections, pruning
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06N—COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N99/00—Subject matter not provided for in other groups of this subclass
- G06N99/005—Learning machines, i.e. computer in which a programme is changed according to experience gained by the machine itself during a complete run
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06N—COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computer systems based on biological models
- G06N3/12—Computer systems based on biological models using genetic models
- G06N3/126—Genetic algorithms, i.e. information processing using digital simulations of the genetic system
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F7/00—Methods or arrangements for processing data by operating upon the order or content of the data handled
- G06F7/58—Random or pseudo-random number generators
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06N—COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N99/00—Subject matter not provided for in other groups of this subclass
- G06N99/002—Quantum computers, i.e. information processing by using quantum superposition, coherence, decoherence, entanglement, nonlocality, teleportation
Similar Documents
Publication | Publication Date | Title |
---|---|---|
Yakopcic et al. | Memristor crossbar deep network implementation based on a convolutional neural network | |
Yakopcic et al. | Extremely parallel memristor crossbar architecture for convolutional neural network implementation | |
Yang et al. | Research progress on memristor: From synapses to computing systems | |
Nandakumar et al. | Mixed-precision architecture based on computational memory for training deep neural networks | |
US9715655B2 (en) | Method and apparatus for performing close-loop programming of resistive memory devices in crossbar array based hardware circuits and systems | |
Wijesinghe et al. | An all-memristor deep spiking neural computing system: A step toward realizing the low-power stochastic brain | |
US11348002B2 (en) | Training of artificial neural networks | |
Wang et al. | Integration and co-design of memristive devices and algorithms for artificial intelligence | |
Chen et al. | Mitigating effects of non-ideal synaptic device characteristics for on-chip learning | |
Hasan et al. | Enabling back propagation training of memristor crossbar neuromorphic processors | |
Narayanan et al. | Toward on-chip acceleration of the backpropagation algorithm using nonvolatile memory | |
Hasan et al. | On-chip training of memristor based deep neural networks | |
Kadetotad et al. | Parallel architecture with resistive crosspoint array for dictionary learning acceleration | |
Xia et al. | Fault-tolerant training enabled by on-line fault detection for RRAM-based neural computing systems | |
Kim et al. | Input voltage mapping optimized for resistive memory-based deep neural network hardware | |
US11386319B2 (en) | Training of artificial neural networks | |
Hu et al. | Dot-product engine as computing memory to accelerate machine learning algorithms | |
Yakopcic et al. | Flexible memristor based neuromorphic system for implementing multi-layer neural network algorithms | |
Hasan et al. | A fast training method for memristor crossbar based multi-layer neural networks | |
Yakopcic et al. | Memristor crossbar based implementation of a multilayer perceptron | |
Sun et al. | Unary coding and variation-aware optimal mapping scheme for reliable ReRAM-based neuromorphic computing | |
Nandakumar et al. | Mixed-precision training of deep neural networks using computational memory | |
Bennett et al. | Contrasting advantages of learning with random weights and backpropagation in non-volatile memory neural networks | |
Cho et al. | An on-chip learning neuromorphic autoencoder with current-mode transposable memory read and virtual lookup table | |
Yeo et al. | A hardware and energy-efficient online learning neural network with an RRAM crossbar array and stochastic neurons |