ReLU, Sparseness, and the Encoding of Optic Flow in Neural Networks
<p>Sample optic flow fields generated with different simulated camera self-motion through different visual environments. (<b>a</b>) Backward translation at 1 m/s at −180° azimuth and 45° elevation away from a frontoparallel wall positioned 4 m in front of the camera. (<b>b</b>) Combination of backward translation at 1 m/s (−180° azimuth, 45° elevation) with respect to a ground plane and 5°/s yaw rotation. The camera is 10 m above the ground plane and is oriented 30° downward. (<b>c</b>) Forward translation at 1 m/s at 0° azimuth and 0° elevation through a 3D dot cloud.</p> "> Figure 2
<p>Neural network activation functions examined in the present article. (<b>a</b>) The rectified linear unit (ReLU) and leaky ReLU activation functions. (<b>b</b>) The Gaussian error linear unit (GELU) and Mish activation functions. The y-axis shows the output of the neuron after applying an activation function to the net input indicated on the x-axis.</p> "> Figure 3
<p>Overview of the CNN and MLP network architecture. The CNN architecture begins with one or more convolutional and max pooling layer stacks. The max pooling layers that reduce the spatial resolution of the optic flow signal. The representation in the final max pooling layer is flattened into a 1D vector, which is passed through one or more densely connected layers. As described in the main text, we created CNN and MLP variants that apply one of the following activation functions in both the convolutional and dense layers: ReLU, leaky ReLU, GELU, or Mish. We schematize where in the network the choice of one of these activation functions is applied with <Act fun>. The output layer contains five neurons, corresponding to the parameters that describe the camera’s self-motion: the azimuth and elevation of observer translation, along with the pitch, yaw, and roll components of observer rotation. The network is trained to minimize a cosine loss function of the translation azimuth angle due to its circularity. Mean squared error (MSE) is used for the other variables. The MLP differs from the CNN in the lack of convolutional and max pooling stages (shown in teal).</p> "> Figure 4
<p>Test accuracy of the neural networks on the TR360 dataset. (<b>a</b>,<b>c</b>) MSE of network estimates of translational (T) and rotational (R) self-motion from optic flow (<b>b</b>,<b>d</b>) mean absolute error (MAE) of networks estimates of the T and R self-motion from optic flow. (<b>e</b>–<b>h</b>) Scatter plots depict the estimate (y-axis) corresponding to each true translational azimuth label (x-axis; “heading_x”) produced by each CNN variant. Each red diagonal line coincides with estimates that match the true label (no error). (<b>i</b>–<b>l</b>) Same format as the row above, but for the MLPs. In the depicted coordinate system, ±180° both refer to straight-backward self-motion.</p> "> Figure 5
<p>Test accuracy on the TR360Cloud optic flow dataset achieved by the CNN and MLP models trained on a different dataset (TR360). Same format and conventions as <a href="#sensors-24-07453-f004" class="html-fig">Figure 4</a>.</p> "> Figure 6
<p>Accuracy of self-motion estimates when noise is added to the TR360 test optic flow samples. (<b>a</b>–<b>c</b>) Example optic flow fields with 0%, 30%, and 60% noise, respectively. (<b>d</b>,<b>e</b>) MAE in estimating translational and rotational self-motion parameters, respectively, when the optic flow contains different proportions of noise (x-axis).</p> "> Figure 7
<p>Accuracy of self-motion estimates when motion vectors are removed from the TR360 test optic flow samples. (<b>a</b>–<b>c</b>) Example optic flow fields with 0%, 30%, and 60% sparseness, respectively. (<b>d</b>,<b>e</b>) MAE in estimating translational and rotational self-motion parameters, respectively. The x-axis indicates the degree of sparseness within each optic flow sample.</p> "> Figure 8
<p>Accuracy of self-motion estimates when the optic flow contains the motion due to an independently moving object. (<b>a</b>–<b>c</b>) Example optic flow fields with a Size 1 (1 × 1 pixels), Size 6 (6 × 6 pixels), and Size 12 (12 × 12 pixels) region of motion induced by the moving object, respectively. (<b>d</b>,<b>e</b>) MAE in estimating translational and rotational self-motion parameters, respectively. The x-axis indicates the size of the moving object in the optic flow field.</p> "> Figure 9
<p>(<b>a</b>–<b>h</b>) The population (red) and lifetime (blue) sparseness in each layer of the 8 models. Both metrics range between 0 (dense code) and 1 (very sparse code). The red and blue dashed lines indicate the mean population and lifetime sparseness across the network, respectively.</p> "> Figure 10
<p>(<b>a</b>–<b>d</b>) The relationship between population sparseness (x-axis) and MAE obtained when estimating the translational self-motion parameters on the TR360Cloud dataset (y-axis). Plot markers correspond to values obtained from the top 3 networks within each model type. Red lines show the regression curves fitted to the data.</p> "> Figure 11
<p>The Sparseness Index (<span class="html-italic">S</span>) computed on the weights in the early, middle, or final third of the CNNs (<b>a</b>,<b>c</b>) and MLPs (<b>b</b>,<b>d</b>). Solid line demarcates that the analysis includes negative network weights. Dashed line demarcates that the analysis includes only non-negative network weights.</p> "> Figure 12
<p>The distribution of translation azimuth angles that yield maximal activation in individual neurons within each dense hidden layer. Each histogram corresponds to the preferences of units in a single model layer, and the histograms associated with the same model are stacked vertically. Histograms assigned smaller “Dense layer” integer labels (top-left panel) correspond to layers earlier in the network, while those with larger integer labels correspond to layers deeper in the network. The x-axis in each histogram corresponds to the preferred translation azimuth angle (0–360°). The y-axis indicates the number of units that possess a particular azimuth angle (bin width: 30°). The schematic atop the 3rd column shows the coordinate system (top-down view).</p> "> Figure 13
<p>The translation elevation angle preference of individual neurons within each model dense hidden layer. Same format as <a href="#sensors-24-07453-f012" class="html-fig">Figure 12</a>. The schematic atop the 3rd column shows the coordinate system (side view).</p> "> Figure 14
<p>The distribution of preferred rotation azimuth angles of individual neurons within each model dense hidden layer. Same format as <a href="#sensors-24-07453-f012" class="html-fig">Figure 12</a>.</p> "> Figure 15
<p>The distribution of preferred rotation elevation angles of individual neurons within each model dense hidden layer. Same format as <a href="#sensors-24-07453-f012" class="html-fig">Figure 12</a>.</p> ">
Abstract
:1. Introduction
2. Materials and Methods
2.1. Optic Flow Datasets
Robustness Tests
2.2. Neural Networks
2.2.1. Architecture
2.2.2. Model Variants
2.2.3. Training Protocol
2.3. Analyses
2.3.1. Population and Lifetime Sparseness Metrics
2.3.2. Weight Sparseness Index
2.3.3. Translation and Rotation Tuning Preferences
2.4. Software Accessibility
3. Results
3.1. Accuracy of Self-Motion Estimation
Generalization: TR360Cloud
3.2. Robustness
3.2.1. Noise
3.2.2. Sparse Optic Flow
3.2.3. Independently Moving Objects
3.3. Sparseness in the Neural Network Encoding of Optic Flow
3.3.1. Population and Lifetime Sparseness
3.3.2. Dead Neurons
3.3.3. Sparseness in Network Weights
3.4. Optic Flow Tuning
3.4.1. Translation Preferences
3.4.2. Rotation Preferences
4. Discussion
5. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Acknowledgments
Conflicts of Interest
Abbreviations
CNN | Convolutional neural network |
MLP | Multi-layer perceptron neural network |
ReLU | Rectified linear unit activation function |
GELU | Gaussian error linear unit activation function |
MSE | Mean squared error |
MAE | Mean absolute error |
LLM | Large language model |
References
- Gibson, J.J. The Perception of the Visual World; Houghton Mifflin: Boston, MA, USA, 1950. [Google Scholar]
- Zufferey, J.C.; Beyeler, A.; Floreano, D. Optic flow to control small UAVs. In Workshop on Visual Guidance Systems for Small Autonomous Aerial Vehicles; EPFL: Lausanne, Switzerland, 2008. [Google Scholar]
- Srinivasan, M. Honeybees as a model for the study of visually guided flight, navigation, and biologically inspired robotics. Physiol. Rev. 2011, 91, 413–460. [Google Scholar] [CrossRef] [PubMed]
- De Croon, G.; Ho, H.; De Wagter, C.; Van Kampen, E.; Remes, B.; Chu, Q. Optic-flow based slope estimation for autonomous landing. Int. J. Micro Air Veh. 2013, 5, 287–297. [Google Scholar] [CrossRef]
- De Croon, G.; De Wagter, C.; Seidl, T. Enhancing optical-flow-based control by learning visual appearance cues for flying robots. Nat. Mach. Intell. 2021, 3, 33–41. [Google Scholar] [CrossRef]
- Escobar-Alvarez, H.; Johnson, N.; Hebble, T.; Klingebiel, K.; Quintero, S.; Regenstein, J.; Browning, N. R-ADVANCE: Rapid Adaptive Prediction for Vision-based Autonomous Navigation, Control, and Evasion. J. Field Robot. 2018, 35, 91–100. [Google Scholar] [CrossRef]
- Zhang, J.; Singh, S. Visual-lidar odometry and mapping: Low-drift, robust, and fast. In Proceedings of the 2015 IEEE International Conference on Robotics and Automation (ICRA), Seattle, WA, USA, 26–30 May 2015. [Google Scholar]
- Costante, G.; Mancini, M.; Valigi, P.; Ciarfuglia, T. Exploring representation learning with cnns for frame-to-frame ego-motion estimation. IEEE Robot. Autom. Lett. 2015, 1, 18–25. [Google Scholar] [CrossRef]
- Costante, G.; Ciarfuglia, T. LS-VO: Learning Dense Optical Subspace for Robust Visual Odometry Estimation. arXiv 2017, arXiv:1709.06019. [Google Scholar] [CrossRef]
- Kouris, A.; Bouganis, C.S. Learning to fly by myself: A self-supervised cnn-based approach for autonomous navigation. In Proceedings of the 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Madrid, Spain, 1–5 October 2018; pp. 1–9. [Google Scholar]
- Kashyap, H.; Fowlkes, C.; Krichmar, J. Sparse Representations for Object and Ego-motion Estimation in Dynamic Scenes. arXiv 2019, arXiv:1903.03731. [Google Scholar] [CrossRef]
- Mineault, P.; Bakhtiari, S.; Richards, B.; Pack, C. Your head is there to move you around: Goal-driven models of the primate dorsal pathway. Adv. Neural Inf. Process. Syst. 2021, 34, 28757–28771. [Google Scholar]
- Zhao, B.; Huang, Y.; Wei, H.; Hu, X. Ego-Motion Estimation Using Recurrent Convolutional Neural Networks through Optical Flow Learning. Electronics 2021, 10, 222. [Google Scholar] [CrossRef]
- Xu, Y.; Croon, G. CNN-based Ego-Motion Estimation for Fast MAV Maneuvers. arXiv 2021, arXiv:2101.01841v2. [Google Scholar]
- Layton, O.; Powell, N.; Steinmetz, S.; Fajen, B. Estimating curvilinear self-motion from optic flow with a biologically inspired neural system. Bioinspir. Biomim. 2022, 17, 046013. [Google Scholar] [CrossRef]
- Maus, N.; Layton, O. Estimating heading from optic flow: Comparing deep learning network and human performance. Neural Netw. 2022, 154, 383–396. [Google Scholar] [CrossRef] [PubMed]
- Layton, O.; Steinmetz, S. Accuracy optimized neural networks do not effectively model optic flow tuning in brain area MSTd. bioRxiv 2024. [Google Scholar] [CrossRef] [PubMed]
- Lu, L.; Shin, Y.; Su, Y.; Karniadakis, G. Dying ReLU and Initialization: Theory and Numerical Examples. arXiv 2019, arXiv:1903.06733v3. [Google Scholar] [CrossRef]
- Maas, A.; Hannun, A.; Ng, A. Rectifier Nonlinearities Improve Neural Network Acoustic Models. In Proceedings of the 30th International Conference on Machine Learning, Atlanta, GA, USA, 17–19 June 2013; Volume 28. [Google Scholar]
- Glorot, X.; Bordes, A.; Bengio, Y. Deep sparse rectifier neural networks. In Proceedings of the 14th International Conference on Artificial Intelligence and Statistics, Fort Lauderdale, FL, USA, 11–13 April 2011; Volume 15, pp. 315–323. [Google Scholar]
- Doi, E.; Balcan, D.; Lewicki, M. Robust coding over noisy overcomplete channels. IEEE Trans. Image Process 2007, 16, 442–452. [Google Scholar] [CrossRef]
- Ranzato, M.; Poultney, C.; Chopra, S.; Cun, Y. Efficient learning of sparse representations with an energy-based model. In Proceedings of the 19th International Conference on Neural Information Processing Systems (NIPS’06), Vancouver, BC, Canada, 4–7 December 2006. [Google Scholar]
- Ranzato, M.; Boureau, Y.L.; Cun, Y. Sparse feature learning for deep belief networks. In Proceedings of the 20th International Conference on Neural Information Processing Systems (NIPS’07), Vancouver, BC, Canada, 3–6 December 2007. [Google Scholar]
- Beyeler, M.; Rounds, E.; Carlson, K.; Dutt, N.; Krichmar, J. Neural correlates of sparse coding and dimensionality reduction. PLoS Comput. Biol. 2019, 15, e1006908. [Google Scholar] [CrossRef]
- Shi, S.; Chu, X. Speeding up Convolutional Neural Networks By Exploiting the Sparsity of Rectifier Units. arXiv 2017, arXiv:1704.07724v2. [Google Scholar]
- Kurtz, M.; Kopinsky, J.; Gelashvili, R.; Matveev, A.; Carr, J.; Goin, M.; Leiserson, W.; Moore, S.; Shavit, N.; Alistarh, D. Inducing and exploiting activation sparsity for fast inference on deep neural networks. Int. Conf. Mach. Learn. 2020, 119, 5533–5543. [Google Scholar]
- Hendrycks, D.; Gimpel, K. Gaussian Error Linear Units (GELUs). arXiv 2016, arXiv:1606.08415v5. [Google Scholar]
- Misra, D. Mish: A self regularized non-monotonic activation function. arXiv 2019, arXiv:1908.08681. [Google Scholar]
- Liu, Z.; Mao, H.; Wu, C.Y.; Feichtenhofer, C.; Darrell, T.; Xie, S. A convnet for the 2020s. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022. [Google Scholar]
- Layton, O.; Steinmetz, S. Accuracy optimized neural networks do not effectively model optic flow tuning in brain area MSTd. Front. Neurosci. 2024, 18, 1441285. [Google Scholar] [CrossRef] [PubMed]
- Raudies, F.; Neumann, H. Modeling heading and path perception from optic flow in the case of independently moving objects. Front. Behav. Neurosci. 2013, 7, 23. [Google Scholar] [CrossRef] [PubMed]
- Beyeler, M.; Dutt, N.; Krichmar, J. 3D Visual Response Properties of MSTd Emerge from an Efficient, Sparse Population Code. J. Neurosci. 2016, 36, 8399–8415. [Google Scholar] [CrossRef] [PubMed]
- Longuet-Higgins, H.; Prazdny, K. The interpretation of a moving retinal image. Proc. R. Soc. Lond. B 1980, 208, 385–397. [Google Scholar]
- Royden, C.; Hildreth, E. Human heading judgments in the presence of moving objects. Percept. Psychophys. 1996, 58, 836–856. [Google Scholar] [CrossRef]
- Layton, O.; Mingolla, E.; Browning, N. A motion pooling model of visually guided navigation explains human behavior in the presence of independently moving objects. J. Vis. 2012, 12, 20. [Google Scholar] [CrossRef]
- Glorot, X.; Bengio, Y. Understanding the difficulty of training deep feedforward neural networks. In Proceedings of the Thirteenth International Conference on Artificial Intelligence and Statistics, JMLR Workshop and Conference Proceedings, Chia Laguna Resort, Sardinia, Italy, 13–15 May 2010. [Google Scholar]
- Abadi, M.; Agarwal, A.; Barham, P.; Brevdo, E.; Chen, Z.; Citro, C.; Corrado, G.; Davis, A.; Dean, J.; Devin, M. Tensorflow: Large-scale machine learning on heterogeneous distributed systems. arXiv 2016, arXiv:1603.04467. [Google Scholar]
- Vinje, W.; Gallant, J. Sparse coding and decorrelation in primary visual cortex during natural vision. Science 2000, 287, 1273–1276. [Google Scholar] [CrossRef]
- Quiroga, R.; Reddy, L.; Koch, C.; Fried, I. Decoding visual inputs from multiple neurons in the human temporal lobe. J. Neurophysiol. 2007, 98, 1997–2007. [Google Scholar] [CrossRef]
- Georgopoulos, A.; Schwartz, A.; Kettner, R. Neuronal population coding of movement direction. Science 1986, 233, 1416–1419. [Google Scholar] [CrossRef]
- Harris, C.; Millman, K.; Van Der Walt, S.; Gommers, R.; Virtanen, P.; Cournapeau, D.; Wieser, E.; Taylor, J.; Berg, S.; Smith, N. Array programming with NumPy. Nature 2020, 585, 357–362. [Google Scholar] [CrossRef] [PubMed]
- Virtanen, P.; Gommers, R.; Oliphant, T.; Haberland, M.; Reddy, T.; Cournapeau, D.; Burovski, E.; Peterson, P.; Weckesser, W.; Bright, J.; et al. SciPy 1.0: Fundamental algorithms for scientific computing in Python. Nat. Methods 2020, 17, 261–272. [Google Scholar] [CrossRef] [PubMed]
- McKinney, W. Data structures for statistical computing in python. In Proceedings of the 9th Python in Science Conference, Austin, TX, USA, 28 June–3 July 2010; Volume 445. [Google Scholar]
- Waskom, M. seaborn: Statistical data visualization. J. Open Source Softw. 2021, 6, 3021. [Google Scholar] [CrossRef]
- Layton, O.; Fajen, B. Competitive dynamics in MSTd: A mechanism for robust heading perception based on optic flow. PLoS Comput. Biol. 2016, 12, e1004942. [Google Scholar] [CrossRef]
- Layton, O.; Fajen, B. Possible role for recurrent interactions between expansion and contraction cells in MSTd during self-motion perception in dynamic environments. J. Vis. 2017, 17, 5. [Google Scholar] [CrossRef]
- Krizhevsky, A.; Sutskever, I.; Hinton, G. Imagenet classification with deep convolutional neural networks. Adv. Neural Inf. Process. Syst. 2012, 25, 1097–1105. [Google Scholar] [CrossRef]
- Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks for Large-Scale Image Recognition. arXiv 2014, arXiv:1409.1556v6. [Google Scholar]
- He, K.; Zhang, X.; Ren, S.; Sun, J. Delving deep into rectifiers: Surpassing human-level performance on imagenet classification. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015. [Google Scholar]
- Radford, A.; Narasimhan, K.; Salimans, T.; Sutskever, I. Improving Language Understanding by Generative Pre-Training. 2018. Available online: https://cdn.openai.com/research-covers/language-unsupervised/language_understanding_paper.pdf (accessed on 12 September 2024).
- Radford, A.; Wu, J.; Child, R.; Luan, D.; Amodei, D.; Sutskever, I. Language models are unsupervised multitask learners. OpenAI Blog 2019, 1, 9. [Google Scholar]
- Mirzadeh, I.; Alizadeh, K.; Mehta, S.; Mundo, C.; Tuzel, O.; Samei, G.; Rastegari, M.; Farajtabar, M. ReLU Strikes Back: Exploiting Activation Sparsity in Large Language Models. arXiv 2023, arXiv:2310.04564v1. [Google Scholar]
- Zhang, Z.; Song, Y.; Yu, G.; Han, X.; Lin, Y.; Xiao, C.; Song, C.; Liu, Z.; Mi, Z.; Sun, M. ReLU Wins: Discovering Efficient Activation Functions for Sparse LLMs. arXiv 2024, arXiv:2402.03804v1. [Google Scholar]
Dataset | Description | Size (Num Samples N) | Independent Variables |
---|---|---|---|
TR360 | Simulated self-motion toward either a frontoparallel plane (wall) or above a ground plane. T and R direction is uniform random: TR elevation [−180, 180]°, TR azimuth [−90, 90]°. | total: 12,060 (6030 frontoparallel + 6030 ground) train: 6030 validation: 3015 test: 3015 | T speed: [0.5, 1.0, 1.5] m/s R speed: [0, 5, 10] °/s Frontoparallel plane depth: [2, 4, 8, 16, 32] m |
TR360Cloud | Same as TR360, except self-motion is simulated through a 3D cloud of dots. Depth of each dot is uniform random: [2, 32] m. | 3015 test | T speed: [0.5, 1.0, 1.5] m/s R speed: [0, 5, 10] °/s |
TestProtocolT | Diagnostic set of optic flow patterns used to evaluate neural tuning to specific T directions | 514 (512 combinations of T azimuth and elevation & ±90° vertical) | T azimuth: [0, ±11.25, ±22.5, …, ±180]° T elevation: [0, ±11.25, ±22.5, …, ±90]° |
TestProtocolR | Diagnostic set of optic flow patterns used to evaluate neural tuning to specific R directions | 514 (512 combinations of R azimuth and elevation & ±90° vertical) | R azimuth: [0, ±11.25, ±22.5, …, ±180]° R elevation: [0, ±11.25, ±22.5, …, ±90]° |
Hyperparameter | Value Range |
---|---|
Number of convolution and max pooling stacks | [1, 3] |
Number of dense layers | [1, 5] |
Number of convolutional filters | [2, 300] |
Number of dense units | [2, 10,000] |
Convolutional unit filter size | [2, 15] |
Max pooling window size | [2, 4] |
Max pooling stride length | [1, 3] |
Learning rate | [, , , ] * |
Leaky ReLU activation function | [0.01, 0.1, 1, 2] * |
Hyperparameter | CNN_RELU | CNN_LEAKY_RELU | CNN_GELU | CNN_MISH |
---|---|---|---|---|
Number of convolution and max pooling stacks | 1 | 1 | 2 | 3 |
Number of dense layers | 5 | 5 | 3 | 5 |
Number of convolutional filters | [157] | [105] | [104, 213] | [105, 112, 279] |
Number of dense units | [2997, 7566, 5979, 6709, 2631] | [2073, 6418, 5792, 1020, 8613] | [6305, 7621, 2990] | [4796, 3330, 6632, 411, 5166] |
Convolutional unit filter size | [2] | [2] | [2, 2] | [2, 2, 2] |
Max pooling window size | [2] | [2] | [4, 4] | [3, 3, 3] |
Max pooling stride length | [3] | [2] | [1, 2] | [1, 3, 1] |
Learning rate |
Hyperparameter | MLP_RELU | MLP_LEAKY_RELU | MLP_GELU | MLP_MISH |
---|---|---|---|---|
Number of dense layers | 5 | 5 | 5 | 5 |
Number of dense units | [2958, 6244, 3234, 5067, 2651] | [7628, 4705, 8044, 6970, 2027] | [6783, 502, 6201, 5015, 9342] | [4371, 5412, 3814, 6625, 3240] |
Learning rate |
Model | Number of Parameters |
---|---|
CNN_RELU | 137,480,586 |
CNN_LEAKY_RELU | 75,904,186 |
CNN_GELU | 104,532,948 |
CNN_MISH | 48,476,092 |
MLP_RELU | 69,846,657 |
MLP_LEAKY_RELU | 147,374,365 |
MLP_GELU | 87,565,173 |
MLP_MISH | 93,013,120 |
Network Model | Number of Dead Neurons | Percentage of Neurons Dead |
---|---|---|
CNN_RELU | [0/157, 20/2997, 121/7566, 258/5979, 323/6709, 179/2631] | [0.0%, 0.67%, 1.60%, 4.32%, 4.81%, 6.80%] |
CNN_LEAKY_RELU | [0/105, 0/2073, 0/6418, 0/5792, 0/1020, 0/8613] | [0.0%, 0.0%, 0.0%, 0.0%, 0.0%, 0.0%] |
CNN_GELU | [0/104, 0/213, 0/6305, 0/7621, 0/2990] | [0.0%, 0.0%, 0.0%, 0.0%, 0.0%] |
CNN_MISH | [0/105, 0/112, 0/279, 0/4796, 0/3330, 0/6632, 0/411, 0/5166] | [0.0%, 0.0%, 0.0%, 0.0%, 0.0%, 0.0%, 0.0%, 0.0%] |
MLP_RELU | [224/2958, 5903/6244, 3144/3234, 4975/5067, 2505/2651] | [7.57%, 94.54%, 97.22%, 98.18%, 94.49%] |
MLP_LEAKY_RELU | [0/7628, 0/4705, 0/8044, 0/6970, 0/2027] | [0.0%, 0.0%, 0.0%, 0.0%, 0.0%] |
MLP_GELU | [0/6783, 0/502, 0/6201, 0/5015, 0/9342] | [0.0%, 0.0%, 0.0%, 0.0%, 0.0%] |
MLP_MISH | [0/4371, 0/5412, 0/3814, 0/6625, 0/3240] | [0.0%, 0.0%, 0.0%, 0.0%, 0.0%] |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Layton, O.W.; Peng, S.; Steinmetz, S.T. ReLU, Sparseness, and the Encoding of Optic Flow in Neural Networks. Sensors 2024, 24, 7453. https://doi.org/10.3390/s24237453
Layton OW, Peng S, Steinmetz ST. ReLU, Sparseness, and the Encoding of Optic Flow in Neural Networks. Sensors. 2024; 24(23):7453. https://doi.org/10.3390/s24237453
Chicago/Turabian StyleLayton, Oliver W., Siyuan Peng, and Scott T. Steinmetz. 2024. "ReLU, Sparseness, and the Encoding of Optic Flow in Neural Networks" Sensors 24, no. 23: 7453. https://doi.org/10.3390/s24237453
APA StyleLayton, O. W., Peng, S., & Steinmetz, S. T. (2024). ReLU, Sparseness, and the Encoding of Optic Flow in Neural Networks. Sensors, 24(23), 7453. https://doi.org/10.3390/s24237453