[go: up one dir, main page]
More Web Proxy on the site http://driver.im/
You seem to have javascript disabled. Please note that many of the page functionalities won't work as expected without javascript enabled.
 
 
Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

Article Types

Countries / Regions

Search Results (8)

Search Parameters:
Keywords = auditory perception inspired

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
15 pages, 2932 KiB  
Article
The Role of Sensory Cues in Collective Dynamics: A Study of Three-Dimensional Vicsek Models
by Poorendra Ramlall and Subhradeep Roy
Appl. Sci. 2025, 15(3), 1556; https://doi.org/10.3390/app15031556 - 4 Feb 2025
Viewed by 544
Abstract
This study presents a three-dimensional collective motion model that integrates auditory and visual sensing modalities, inspired by organisms like bats that rely on these senses for navigation. Most existing models of collective motion consider vision-based sensing, likely reflecting an inherent human bias towards [...] Read more.
This study presents a three-dimensional collective motion model that integrates auditory and visual sensing modalities, inspired by organisms like bats that rely on these senses for navigation. Most existing models of collective motion consider vision-based sensing, likely reflecting an inherent human bias towards visual perception. However, many organisms utilize multiple sensory modalities, and this study explores how the integration of these distinct sensory inputs influences group behavior. We investigate a generalized scenario of three-dimensional motion, an area not previously explored for combining sensory information. Through numerical simulations, we investigate the combined impact of auditory and visual sensing on group behavior, contrasting these effects with those observed when relying solely on vision or audition. The results demonstrate that composite sensing allows particles to interact with more neighbors, thereby gaining more information. This interaction allows the formation of a single, large, perfectly aligned group using a narrow sensing region, achievable by taking advantage of the mechanics of both auditory and visual sensing. Our findings demonstrate the importance of integrating multiple sensory modalities in shaping emergent group behavior, with potential applications in both biological studies and the development of robotic swarms. Full article
Show Figures

Figure 1

Figure 1
<p>Schematic explaining the implementation of sensing modalities, where both the field of vision and the acoustic beam are modeled as spherical cones in three dimensions. (<b>a</b>) In the auditory mode, the orange has the blue particle as its neighbor, but not the yellow, since orange resides within the acoustic coverage of the blue and thus can hear it. (<b>b</b>) In the visual mode, the orange has the yellow particle as its neighbor, but not the blue, since the yellow resides within its field of vision. (<b>c</b>) In the composite mode, the orange particle can ‘see’ the yellow and ‘hear’ the blue, making both yellow and blue its neighbors.</p>
Full article ">Figure 2
<p>Snapshots of the 3D Vicsek model implementing (<b>a</b>) auditory, (<b>b</b>) visual, and (<b>c</b>) composite sensing modalities, with parameters <math display="inline"><semantics> <mrow> <mi>ρ</mi> <mo>=</mo> <mn>10</mn> </mrow> </semantics></math>, <math display="inline"><semantics> <mrow> <mi>η</mi> <mo>=</mo> <mn>0.2</mn> </mrow> </semantics></math>, and <math display="inline"><semantics> <mrow> <mi>ϕ</mi> <mo>=</mo> <mn>6</mn> <mi>π</mi> <mo>/</mo> <mn>15</mn> </mrow> </semantics></math>, at time step <math display="inline"><semantics> <mrow> <mi>t</mi> <mo>=</mo> <mn>1500</mn> </mrow> </semantics></math>.</p>
Full article ">Figure 3
<p>Monte-Carlo simulations of polarization over 20 iterations for five different pairs of parameters comparing three modalities at <math display="inline"><semantics> <mrow> <mi>ρ</mi> <mo>=</mo> <mn>2</mn> </mrow> </semantics></math>, <math display="inline"><semantics> <mrow> <mi>L</mi> <mo>=</mo> <mn>5</mn> </mrow> </semantics></math>, and <math display="inline"><semantics> <mrow> <mi>R</mi> <mo>=</mo> <mn>0.5</mn> </mrow> </semantics></math>.</p>
Full article ">Figure 4
<p>Results of the order parameter analysis comparing three modalities. Mean polarization (<b>top row</b>), mean cohesion (<b>middle row</b>), and mean largest cluster size (<b>bottom row</b>) are calculated for auditory (<b>left column</b>), visual (<b>middle column</b>), and composite sensing (<b>right column</b>) modes.</p>
Full article ">
16 pages, 5334 KiB  
Article
An Auditory Convolutional Neural Network for Underwater Acoustic Target Timbre Feature Extraction and Recognition
by Junshuai Ni, Fang Ji, Shaoqing Lu and Weijia Feng
Remote Sens. 2024, 16(16), 3074; https://doi.org/10.3390/rs16163074 - 21 Aug 2024
Cited by 1 | Viewed by 1055
Abstract
In order to extract the line-spectrum features of underwater acoustic targets in complex environments, an auditory convolutional neural network (ACNN) with the ability of frequency component perception, timbre perception and critical information perception is proposed in this paper inspired by the human auditory [...] Read more.
In order to extract the line-spectrum features of underwater acoustic targets in complex environments, an auditory convolutional neural network (ACNN) with the ability of frequency component perception, timbre perception and critical information perception is proposed in this paper inspired by the human auditory perception mechanism. This model first uses a gammatone filter bank that mimics the cochlear basilar membrane excitation response to decompose the input time-domain signal into a number of sub-bands, which guides the network to perceive the line-spectrum frequency information of the underwater acoustic target. A sequence of convolution layers is then used to filter out interfering noise and enhance the line-spectrum components of each sub-band by simulating the process of calculating the energy distribution features, after which the improved channel attention module is connected to select line spectra that are more critical for recognition, and in this module, a new global pooling method is proposed and applied in order to better extract the intrinsic properties. Finally, the sub-band information is fused using a combination layer and a single-channel convolution layer to generate a vector with the same dimensions as the input signal at the output layer. A decision module with a Softmax classifier is added behind the auditory neural network and used to recognize the five classes of vessel targets in the ShipsEar dataset, achieving a recognition accuracy of 99.8%, which is improved by 2.7% compared to the last proposed DRACNN method, and there are different degrees of improvement over the other eight compared methods. The visualization results show that the model can significantly suppress the interfering noise intensity and selectively enhance the radiated noise line-spectrum energy of underwater acoustic targets. Full article
(This article belongs to the Topic AI and Data-Driven Advancements in Industry 4.0)
Show Figures

Figure 1

Figure 1
<p>Time-frequency line-spectra diagrams. (<b>a</b>) Time-frequency diagram of the small boat. (<b>b</b>) Time-frequency diagram of the test vessel in a stationary state with only auxiliary machinery operation. (<b>c</b>) Time-frequency diagram of the fishing vessel with shaft system failure. (<b>d</b>) Time-frequency diagram of the motor boat-radiated noise when its propeller is rotating at a high speed.</p>
Full article ">Figure 2
<p>ACNN model structure.</p>
Full article ">Figure 3
<p>The frequency magnitude responses of the gammatone filter bank.</p>
Full article ">Figure 4
<p>Channel attention mechanism.</p>
Full article ">Figure 5
<p>Structure of the global pooling layer.</p>
Full article ">Figure 6
<p>ACNN_DRACNN model structure.</p>
Full article ">Figure 7
<p>Training curves of ACNN_DRACNN model.</p>
Full article ">Figure 8
<p>Confusion matrix for recognizing results when batch size is 64.</p>
Full article ">Figure 9
<p>Cost function value of the model on validation dataset.</p>
Full article ">Figure 10
<p>Recognition accuracy of the model on the validation dataset.</p>
Full article ">Figure 11
<p>Power spectrum of input data and output data. (<b>a</b>) Sample of category A. (<b>b</b>) Sample of category B. (<b>c</b>) Sample of category C. (<b>d</b>) Sample of category D. (<b>e</b>) Sample of category E.</p>
Full article ">Figure 12
<p>Data visualization by TSNE. (<b>a</b>) Raw signals of ShipsEar dataset. (<b>b</b>) Output data of the ACNN model.</p>
Full article ">
18 pages, 3363 KiB  
Article
A Concert-Based Study on Melodic Contour Identification among Varied Hearing Profiles—A Preliminary Report
by Razvan Paisa, Jesper Andersen, Francesco Ganis, Lone M. Percy-Smith and Stefania Serafin
J. Clin. Med. 2024, 13(11), 3142; https://doi.org/10.3390/jcm13113142 - 27 May 2024
Viewed by 771
Abstract
Background: This study investigated how different hearing profiles influenced melodic contour identification (MCI) in a real-world concert setting with a live band including drums, bass, and a lead instrument. We aimed to determine the impact of various auditory assistive technologies on music [...] Read more.
Background: This study investigated how different hearing profiles influenced melodic contour identification (MCI) in a real-world concert setting with a live band including drums, bass, and a lead instrument. We aimed to determine the impact of various auditory assistive technologies on music perception in an ecologically valid environment. Methods: The study involved 43 participants with varying hearing capabilities: normal hearing, bilateral hearing aids, bimodal hearing, single-sided cochlear implants, and bilateral cochlear implants. Participants were exposed to melodies played on a piano or accordion, with and without an electric bass as a masker, accompanied by a basic drum rhythm. Bayesian logistic mixed-effects models were utilized to analyze the data. Results: The introduction of an electric bass as a masker did not significantly affect MCI performance for any hearing group when melodies were played on the piano, contrary to its effect on accordion melodies and previous studies. Greater challenges were observed with accordion melodies, especially when accompanied by an electric bass. Conclusions: MCI performance among hearing aid users was comparable to other hearing-impaired profiles, challenging the hypothesis that they would outperform cochlear implant users. A cohort of short melodies inspired by Western music styles was developed for future contour identification tasks. Full article
(This article belongs to the Special Issue Advances in the Diagnosis, Treatment, and Prognosis of Hearing Loss)
Show Figures

Figure 1

Figure 1
<p>Distribution of hearing profiles for eligible participants; N = 43 (18 NH, 6 Bilateral HA, and 19 CI users), NH = normal hearing, bilateral CI = both ears are implanted, bimodal = CI + hearing aid, SSCI = single-side CI, Bilateral HA = hearing aids in both ears.</p>
Full article ">Figure 2
<p>Average demographic information self-reported by participants: “HA use” refers to hearing aid usage before obtaining a cochlear implant, “Music Experience” refers to formal music education or practice prior to implantation (if relevant); NH = normal hearing, Bilateral CI = both ears are implanted, Bimodal = CI + hearing aid, SSCI = single-side CI, Bilateral HA = hearing aid in both ears.</p>
Full article ">Figure 3
<p>Melody nr. 47 with undulating contour played on accordion and bass; the playing tempo was 70 BPM, but to increase resolution the scores have been written at 140 BPM.</p>
Full article ">Figure 4
<p>Spectrogram (10—20 kHz) (<b>top</b>) and waveform (<b>bottom</b>) of an accordion (<b>left</b>) and a grand piano (<b>right</b>) and playing a single D4 note.</p>
Full article ">Figure 5
<p>Detailed distribution of correct answers for normal hearing users, separated by contours and instruments played. Normal hearing performance is shown in each plot as the transparent section, while the performance of hearing impaired groups is shown in solid color and numerical value.</p>
Full article ">Figure 6
<p>Probability distribution that describes the performance difference from the reference level (0 = Normal hearing) for each hearing profile—orange areas indicate a credibility interval of 95%, purple line indicates the median.</p>
Full article ">Figure 7
<p>Pairwise probabilities of profile performance differences, with respect to the normal hearing reference level. The mark indicates an “0” equal chance of profile A performing better than profile B.</p>
Full article ">Figure 8
<p>Probability distribution that describes the difference with the reference level (Piano) for each instrument combination—orange areas indicate a credibility interval of 95%, purple line indicates the median.</p>
Full article ">Figure 9
<p>Probability distribution that describes the difference from the reference level (bilateral HA) for each instrument combination, fitted with data excluding the normal hearing population—orange areas indicate a credibility interval of 95%, purple line indicates the median.</p>
Full article ">
18 pages, 12292 KiB  
Article
Correlations among Firing Rates of Tactile, Thermal, Gustatory, Olfactory, and Auditory Sensations Mimicked by Artificial Hybrid Fluid (HF) Rubber Mechanoreceptors
by Kunio Shimada
Sensors 2023, 23(10), 4593; https://doi.org/10.3390/s23104593 - 9 May 2023
Cited by 1 | Viewed by 1751
Abstract
In order to advance the development of sensors fabricated with monofunctional sensation systems capable of a versatile response to tactile, thermal, gustatory, olfactory, and auditory sensations, mechanoreceptors fabricated as a single platform with an electric circuit require investigation. In addition, it is essential [...] Read more.
In order to advance the development of sensors fabricated with monofunctional sensation systems capable of a versatile response to tactile, thermal, gustatory, olfactory, and auditory sensations, mechanoreceptors fabricated as a single platform with an electric circuit require investigation. In addition, it is essential to resolve the complicated structure of the sensor. In order to realize the single platform, our proposed hybrid fluid (HF) rubber mechanoreceptors of free nerve endings, Merkel cells, Krause end bulbs, Meissner corpuscles, Ruffini endings, and Pacinian corpuscles mimicking the bio-inspired five senses are useful enough to facilitate the fabrication process for the resolution of the complicated structure. This study used electrochemical impedance spectroscopy (EIS) to elucidate the intrinsic structure of the single platform and the physical mechanisms of the firing rate such as slow adaption (SA) and fast adaption (FA), which were induced from the structure and involved the capacitance, inductance, reactance, etc. of the HF rubber mechanoreceptors. In addition, the relations among the firing rates of the various sensations were clarified. The adaption of the firing rate in the thermal sensation is the opposite of that in the tactile sensation. The firing rates in the gustation, olfaction, and auditory sensations at frequencies of less than 1 kHz have the same adaption as in the tactile sensation. The present findings are useful not only in the field of neurophysiology, to research the biochemical reactions of neurons and brain perceptions of stimuli, but also in the field of sensors, to advance salient developments in sensors mimicking bio-inspired sensations. Full article
(This article belongs to the Special Issue Applications of Flexible Tactile Sensors in Intelligent Systems)
Show Figures

Figure 1

Figure 1
<p>EIS results of HF rubber mechanoreceptors: (<b>a</b>) dissipation factor; (<b>b</b>) relation between reactance and resistance; (<b>c</b>) phase.</p>
Full article ">Figure 2
<p>Electric circuits of HF rubber mechanoreceptors: (<b>a</b>) approximated parallel circuit of <span class="html-italic">R<sub>2</sub></span>, <span class="html-italic">L<sub>2</sub></span>, <span class="html-italic">C<sub>2,1</sub></span>, and <span class="html-italic">C<sub>2,2</sub></span>; (<b>b</b>) approximated parallel circuit of <span class="html-italic">R<sub>2</sub></span>, <span class="html-italic">C<sub>2,1</sub></span>, and <span class="html-italic">C<sub>2,2</sub></span>; (<b>c</b>) approximated parallel circuit of <span class="html-italic">L<sub>2</sub></span>, <span class="html-italic">C<sub>2,1</sub></span>, and <span class="html-italic">C<sub>2,2</sub></span>; (<b>d</b>) primary electric circuit of HF rubber mechanoreceptors.</p>
Full article ">Figure 3
<p>Relation between the behavior of ions, particles and electrons, and the electric circuit.</p>
Full article ">Figure 4
<p>Relation between the voltage and the electric current in the inductor.</p>
Full article ">Figure 5
<p>Illustration summarizing the relations: (<b>a</b>) among EIS, the electric circuit of the substance, and the firing rate; (<b>b</b>) among EIS, the electric circuit of the substance, the changes in voltage of the sensor, and the firing rate on FA and SA, respectively.</p>
Full article ">Figure 6
<p>Difference in voltage in response to touching a heater.</p>
Full article ">Figure 7
<p>Typical cases of the relation of the equivalent rate to the application of force.</p>
Full article ">Figure 8
<p>Typical cases of the equivalent firing rates and voltages related to gustatory sensation: (<b>a</b>) Type B, saltiness; (<b>b</b>) Type C, saltiness; (<b>c</b>) Type D, saltiness; (<b>d</b>) Type D, sourness; (<b>e</b>) Type E, sweetness.</p>
Full article ">Figure 8 Cont.
<p>Typical cases of the equivalent firing rates and voltages related to gustatory sensation: (<b>a</b>) Type B, saltiness; (<b>b</b>) Type C, saltiness; (<b>c</b>) Type D, saltiness; (<b>d</b>) Type D, sourness; (<b>e</b>) Type E, sweetness.</p>
Full article ">Figure 9
<p>Typical results of <span class="html-italic">V</span>–<span class="html-italic">I</span> curves for gustatory sensation: (<b>a</b>) Type B; (<b>b</b>) Type D.</p>
Full article ">Figure 10
<p>Detailed behavior of ions and electrons in the liquids in Cases F and G: (<b>a</b>) ionic transfer in Case F; (<b>b</b>) electron transfer in Case F; (<b>c</b>) ionic transfer in Case G; (<b>d</b>) electron transfer in Case G.</p>
Full article ">Figure 11
<p>Olfactory sensation: (<b>a</b>) initial responsive voltage; (<b>b</b>) <span class="html-italic">V</span>–<span class="html-italic">I</span> curves.</p>
Full article ">Figure 12
<p>Typical cases of equivalent firing rate and voltage with respect to olfactory sensation: (<b>a</b>) Type B; (<b>b</b>) Type D.</p>
Full article ">Figure 13
<p>Comparison of voltage in response to the vibration of the HF rubber mechanoreceptors.</p>
Full article ">Figure 14
<p>Equivalent firing rates for the application of sound for auditory sensation: (<b>a</b>) Type A; (<b>b</b>) Type B; (<b>c</b>) Type C; (<b>d</b>) Type D; (<b>e</b>) Type E.</p>
Full article ">Figure 14 Cont.
<p>Equivalent firing rates for the application of sound for auditory sensation: (<b>a</b>) Type A; (<b>b</b>) Type B; (<b>c</b>) Type C; (<b>d</b>) Type D; (<b>e</b>) Type E.</p>
Full article ">Figure A1
<p>A schematic of our mechanoreceptor. Following the arrows, a schematic of a virtual biomedical mechanoreceptor, an image of the mechanoreceptor fabricated with HF rubber, a schematic of the intrinsic fabrication, the electric structure, the equivalent electric circuit, and an illustration of free nerve endings (Type A) [<a href="#B24-sensors-23-04593" class="html-bibr">24</a>].</p>
Full article ">Figure A2
<p>Intrinsic conditions of the HF rubber [<a href="#B7-sensors-23-04593" class="html-bibr">7</a>].</p>
Full article ">Figure A3
<p>EIS results of the HF rubber mechanoreceptors: (<b>a</b>) impedance; (<b>b</b>) capacitance; (<b>c</b>) inductance; (<b>d</b>) resistance [<a href="#B24-sensors-23-04593" class="html-bibr">24</a>].</p>
Full article ">Figure A4
<p>Relations among the fabricated mechanoreceptors, capacitance, and inductance [<a href="#B24-sensors-23-04593" class="html-bibr">24</a>].</p>
Full article ">Figure A5
<p>Physical model of the electric field and voltage in the equivalent electric circuits of the mechanoreceptors [<a href="#B34-sensors-23-04593" class="html-bibr">34</a>].</p>
Full article ">
16 pages, 5113 KiB  
Article
Origami-Inspired Structure with Pneumatic-Induced Variable Stiffness for Multi-DOF Force-Sensing
by Wenchao Yue, Jiaming Qi, Xiao Song, Shicheng Fan, Giancarlo Fortino, Chia-Hung Chen, Chenjie Xu and Hongliang Ren
Sensors 2022, 22(14), 5370; https://doi.org/10.3390/s22145370 - 19 Jul 2022
Cited by 11 | Viewed by 4013
Abstract
With the emerging need for human–machine interactions, multi-modal sensory interaction is gradually pursued rather than satisfying common perception forms (visual or auditory), so developing flexible, adaptive, and stiffness-variable force-sensing devices is the key to further promoting human–machine fusion. However, current sensor sensitivity is [...] Read more.
With the emerging need for human–machine interactions, multi-modal sensory interaction is gradually pursued rather than satisfying common perception forms (visual or auditory), so developing flexible, adaptive, and stiffness-variable force-sensing devices is the key to further promoting human–machine fusion. However, current sensor sensitivity is fixed and nonadjustable after fabrication, limiting further development. To solve this problem, we propose an origami-inspired structure to achieve multiple degrees of freedom (DoFs) motions with variable stiffness for force-sensing, which combines the ductility and flexibility of origami structures. In combination with the pneumatic actuation, the structure can achieve and adapt the compression, pitch, roll, diagonal, and array motions (five motion modes), which significantly increase the force adaptability and sensing diversity. To achieve closed-loop control and avoid excessive gas injection, the ultra-flexible microfiber sensor is designed and seamlessly embedded with an approximately linear sensitivity of ∼0.35 Ω/kPa at a relative pressure of 0–100 kPa, and an exponential sensitivity at a relative pressure of 100–350 kPa, which can render this device capable of working under various conditions. The final calibration experiment demonstrates that the pre-pressure value can affect the sensor’s sensitivity. With the increasing pre-pressure of 65–95 kPa, the average sensitivity curve shifts rightwards around 9 N intervals, which highly increases the force-sensing capability towards the range of 0–2 N. When the pre-pressure is at the relatively extreme air pressure of 100 kPa, the force sensitivity value is around 11.6 Ω/N. Therefore, our proposed design (which has a low fabrication cost, high integration level, and a suitable sensing range) shows great potential for applications in flexible force-sensing development. Full article
(This article belongs to the Special Issue Advances in Tactile Sensing and Robotic Grasping)
Show Figures

Figure 1

Figure 1
<p>(<b>a</b>) The potential application scenarios of the wearable device for multiple DOF force-sensing: the foldable structure can be fixed to the fixed strap for easy wearing on the human wrist. The foldable structure in an uninflated state is easy to wear and can be adjusted to a fixed position on the wrist for wearing comfort; the foldable structure in an inflated state can be fixed on the wrist, and the sensitivity of the sensor can be variable by adjusting the pre-pressure value. (<b>b</b>) This core origami model is made of a parallel waterbomb-like unit. Based on the advantages of soft materials, the structure has multiple pseudo-DOFs compared to the rigid ones, allowing for various basic motion modes. (<b>c</b>) The folding pattern of this origami structure is shown, where solid red lines represent the mountain folds, and solid blue lines represent the valley folds. The essential size parameters <math display="inline"><semantics> <msub> <mi>L</mi> <mn>1</mn> </msub> </semantics></math>, <math display="inline"><semantics> <msub> <mi>L</mi> <mn>2</mn> </msub> </semantics></math>, and <math display="inline"><semantics> <msub> <mi>L</mi> <mn>3</mn> </msub> </semantics></math> of this structure are designed as 60, 60, and 50 mm respectively. The diamond cutouts release the fixed constraints from the middle creases and stress concentrations from the folding process.</p>
Full article ">Figure 2
<p>(<b>a</b>) Stretching test of liquid metal microfiber from 100% (<b>left</b>) to 400% (<b>right</b>), and the digital multimeter is utilized to measure the change of the resistance value; (<b>b</b>) shows the basic demonstration of the sensor layout.</p>
Full article ">Figure 3
<p>The passively linear motion of our origami structure. (<b>a</b>,<b>b</b>) show that this origami structure generates the linear motion when normal pressure is manually applied.</p>
Full article ">Figure 4
<p>The fabrication process of the pressure-aware foldable structure can be mainly divided into the preparation process and layered the curing process. The preparation process includes I. The molding fabrication is by 3D printing and II. The Eco-Flex mixing solution preparation. After the preparation, the layered curing processes follow the steps: 01. Molding the outermost layer molding. 02. Attaching the middle layer. 03. Molding the innermost layer. 04. Mounting the proposed microfiber sensor. 05. De-molding the model of the sandwich structure. 06. Folding and PDMS, sealing the structure. To accelerate curing and compactness between layers, the molding procedure is processed in the vacuum oven with a temperature of 70 <math display="inline"><semantics> <msup> <mrow/> <mo>∘</mo> </msup> </semantics></math>C to realize accelerated curing and bubble removal.</p>
Full article ">Figure 5
<p>Origami kinematic model: we model our system as two rigid plates connected with four legs. When this foldable structure generates the pitch and roll motion, the top surface will form a certain angle relative to the bottom plane.</p>
Full article ">Figure 6
<p>The five motion modes (linear, pitch, roll, diagonal, and array motion) of our origami structure are demonstrated from (<b>a</b>–<b>e</b>) based on the simulation results in ANSYS Workbench. Here, one vertex of origami structure is selected as the tracking objective, which indicates that the maximum displacement along the <span class="html-italic">Z</span> direction is around 30 mm, and the maximum pitch or roll angle is around <math display="inline"><semantics> <msup> <mn>40</mn> <mo>∘</mo> </msup> </semantics></math>.</p>
Full article ">Figure 7
<p>(<b>a</b>) Adaptation test setup. The experiment setup for the origami structure’s adaptation test. (<b>b</b>) The origami structure’s linear demonstration results. (<b>c</b>) The origami structure’s pitch or roll demonstration results. (<b>d</b>) The origami structure’s diagonal demonstration results.</p>
Full article ">Figure 8
<p>Comparison of simulation results on the (<b>a</b>) displacement chart and (<b>b</b>) stress chart for the microfiber sensor between the common sandwich structure with the upper layer and our proposed dual-layer structure without the upper layer under external pressure.</p>
Full article ">Figure 9
<p>(<b>a</b>) The basic test setup system for the Instron tensile tester, which includes the testing indenter by 3D printing and our proposed microfiber sensor; (<b>b</b>) the working schematic of the Instron tester for the cycle load testing.</p>
Full article ">Figure 10
<p>(<b>a</b>–<b>d</b>) indicate the working stability test curve: under–over 10 cycles for loads of 10 N, 20 N, 100 N, and 200 N respectively, the output signals exhibit excellent repeatability and stability; (<b>e</b>) refers to the calibration curve to describe the relationship between the resistance value and the pressure applied, and the measurement range is divided into two zones: approximate linear sensitivity zone (0–100 kPa) and exponential sensitivity zone (over 100 kPa).</p>
Full article ">Figure 11
<p>(<b>a</b>) Represents the air path diagram, where the whole system is divided into the air pump part, the digital valve controlled by the ESP32 MCU, and the foldable prototype. (<b>b</b>) There is a basic experimental setup for the pneumatic actuation of our prototype, and the valve is controlled by ESP32-MCU to realize the switch on intake or exhaust state; (<b>c</b>,<b>d</b>) indicate the two working states: the release state when exhausting and the actuation state when intaking. (<b>e</b>) The real-time resistance acquisition curve of our microfiber sensor can precisely record the air pressure change inside the foldable structure when intaking air with a constant pressure of 100.524 kPa. The foldable structure’s state is changed from I. Rest State; II. Actuation State; III. Release State; and finally back to the IV. Rest State.</p>
Full article ">Figure 12
<p>(<b>a</b>) Indicates the setup of the calibration experiment, and the force sensor is fixed into the suitable clamps (fabricated by 3D printing) to identify the load value; (<b>b</b>) represents the offset relationship curves between the microfiber sensor’s resistance and adding force when the pre-pressure is at the approximate linear range (65–95 kPa); (<b>c</b>) represents the relationship curves between the microfiber sensor’s resistance and adding force when the pre-pressure is at the exponential zone (100 kPa).</p>
Full article ">
28 pages, 19143 KiB  
Article
Behavioral Outcomes and Neural Network Modeling of a Novel, Putative, Recategorization Sound Therapy
by Mithila Durai, Zohreh Doborjeh, Philip J. Sanders, Dunja Vajsakovic, Anne Wendt and Grant D. Searchfield
Brain Sci. 2021, 11(5), 554; https://doi.org/10.3390/brainsci11050554 - 27 Apr 2021
Cited by 6 | Viewed by 3694
Abstract
The mechanisms underlying sound’s effect on tinnitus perception are unclear. Tinnitus activity appears to conflict with perceptual expectations of “real” sound, resulting in it being a salient signal. Attention diverted towards tinnitus during the later stages of object processing potentially disrupts high-order auditory [...] Read more.
The mechanisms underlying sound’s effect on tinnitus perception are unclear. Tinnitus activity appears to conflict with perceptual expectations of “real” sound, resulting in it being a salient signal. Attention diverted towards tinnitus during the later stages of object processing potentially disrupts high-order auditory streaming, and its uncertain nature results in negative psychological responses. This study investigated the benefits and neurophysiological basis of passive perceptual training and informational counseling to recategorize phantom perception as a more real auditory object. Specifically, it examined underlying psychoacoustic correlates of tinnitus and the neural activities associated with tinnitus auditory streaming and how malleable these are to change with targeted intervention. Eighteen participants (8 females, 10 males, mean age = 61.6 years) completed the study. The study consisted of 2 parts: (1) An acute exposure over 30 min to a sound that matched the person’s tinnitus (Tinnitus Avatar) that was cross-faded to a selected nature sound (Cicadas, Fan, Water Sound/Rain, Birds, Water and Bird). (2) A chronic exposure for 3 months to the same “morphed” sound. A brain-inspired spiking neural network (SNN) architecture was used to model and compare differences between electroencephalography (EEG) patterns recorded prior to morphing sound presentation, during, after (3-month), and post-follow-up. Results showed that the tinnitus avatar generated was a good match to an individual’s tinnitus as rated on likeness scales and was not rated as unpleasant. The five environmental sounds selected for this study were also rated as being appropriate matches to individuals’ tinnitus and largely pleasant to listen to. There was a significant reduction in the Tinnitus Functional Index score and subscales of intrusiveness of the tinnitus signal and ability to concentrate with the tinnitus trial end compared to baseline. There was a significant decrease in how strong the tinnitus signal was rated as well as ratings of how easy it was to ignore the tinnitus signal on severity rating scales. Qualitative analysis found that the environmental sound interacted with the tinnitus in a positive way, but participants did not experience change in severity, however, characteristics of tinnitus, including pitch and uniformity of sound, were reported to change. The results indicate the feasibility of the computational SNN method and preliminary evidence that the sound exposure may change activation of neural tinnitus networks and greater bilateral hemispheric involvement as the sound morphs over time into natural environmental sound; particularly relating to attention and discriminatory judgments (dorsal attention network, precentral gyrus, ventral anterior network). This is the first study that attempts to recategorize tinnitus using passive auditory training to a sound that morphs from resembling the person’s tinnitus to a natural sound. These findings will be used to design future-controlled trials to elucidate whether the approach used differs in effect and mechanism from conventional Broadband Noise (BBN) sound therapy. Full article
(This article belongs to the Special Issue Neurorehabilitation of Sensory Disorders)
Show Figures

Figure 1

Figure 1
<p>Consolidated Standards of Reporting Trials (CONSORT) reporting of the study.</p>
Full article ">Figure 2
<p>The study protocol: Diagram of data collection across (<b>a</b>) 18 chronic tinnitus patients that underwent auditory training through (<b>b</b>) depicting parameters (pitch, bandwidth, and spatial location) used to determine tinnitus avatar; and conceptualization of gradual morphing of high-pitched tinnitus from its current perceptual characteristics (spectral change is illustrated in this example) to that matching a real target auditory object (a humming bird is used in this example); (<b>c</b>) spatiotemporal brain data measured before, during, and after the auditory training; (<b>d</b>) illustration of the SNN-based methodology, containing: EEG encoding into spike sequences and computational modeling of data into a 3D space of artificial neurons.</p>
Full article ">Figure 3
<p>Average hearing thresholds (dB HL) of participants across frequencies for the Right and Left ear. Error bars represent ± 1 SD.</p>
Full article ">Figure 4
<p>(<b>a</b>) Changes in TFI Total scores across time; (<b>b</b>) changes in TFI Intrusiveness scores across time; (<b>c</b>) changes in TFI Concentration scores across time. Error bars represent ±1 SD.</p>
Full article ">Figure 5
<p>Changes in TFI Concentration scores across time by Intervention sound type. Error bars represent ± 1 SD.</p>
Full article ">Figure 6
<p>(<b>a</b>) Changes in TSNS Strong scores across time; (<b>b</b>) changes in TSNS Ignore scores across time. Error bars represent ±1 SD.</p>
Full article ">Figure 7
<p>Changes in DASS Stress scores across time by Tinnitus Location. Error bars represent ± 1 SD.</p>
Full article ">Figure 8
<p>(<b>a</b>) Changes in Positive Emotionality PANAS scores across time; (<b>b</b>) changes in Positive Emotionality PANAS scores across time by intervention sound type; and (<b>c</b>) changes in Negative Emotionality PANAS scores across time by intervention sound type. Error bars represent ±1 SD.3.3.5 Psychoacoustic tinnitus characteristic changes.</p>
Full article ">Figure 9
<p>Tinnitus pitch match for participants in Hz at baseline and at 3 months following the end of the feasibility trial.</p>
Full article ">Figure 10
<p>Dynamic visualization of the evolution of neuronal connectivity and spiking activity in an SNN model of 1471 spiking neurons with Talairach-based coordinates. It shows differences between the connectivity in the trained SNN models of Pre auditory training (baseline) from the first 10 min of auditory training-denoted as ‘Sound 1′. The green lines are increase connections, while the red lines are decrease connection changes. 5% of relative changes are presented.</p>
Full article ">Figure 11
<p>Dynamic visualization of the evolution of neuronal connectivity and spiking activity in an SNN model of 1471 spiking neurons with Talairach-based coordinates. It shows differences between the connectivity in the trained SNN models of Pre auditory training (baseline) from the middle of 10 min of auditory training-denoted as ‘Sound 2′. The green lines are increase connections, while the red lines are decrease connection changes. 5% of relative changes are presented.</p>
Full article ">Figure 12
<p>Dynamic visualization of the evolution of neuronal connectivity and spiking activity in an SNN model of 1471 spiking neurons with Talairach-based coordinates. It shows differences between the connectivity in the trained SNN models of Pre auditory training (baseline) from the last 10 min of auditory training-denoted as ‘Sound 3′. The green lines are increase connections, while the red lines are decrease connection changes. 5% of relative changes are presented.</p>
Full article ">Figure 13
<p>Dynamic visualization of the evolution of neuronal connectivity and spiking activity in an SNN model of 1471 spiking neurons with Talairach-based coordinates. It shows differences between the connectivity in the trained SNN models of Pre auditory training (baseline) after 10 min of stopping the auditory training-denoted as ‘Post.’ The green lines are increase connections, while the red lines are decrease connection changes. 5% of relative changes are presented.</p>
Full article ">Figure 14
<p>Dynamic visualization of the evolution of neuronal connectivity and spiking activity in an SNN model of 1471 spiking neurons with Talairach-based coordinates. It shows differences between the connectivity in the trained SNN models of pre-auditory training (baseline) from after 3-month post. The green lines increase connections, while the red lines are decrease connection changes. 5% of relative changes are presented.</p>
Full article ">
14 pages, 3443 KiB  
Article
Bio-Inspired Modality Fusion for Active Speaker Detection
by Gustavo Assunção, Nuno Gonçalves and Paulo Menezes
Appl. Sci. 2021, 11(8), 3397; https://doi.org/10.3390/app11083397 - 10 Apr 2021
Cited by 1 | Viewed by 2102
Abstract
Human beings have developed fantastic abilities to integrate information from various sensory sources exploring their inherent complementarity. Perceptual capabilities are therefore heightened, enabling, for instance, the well-known "cocktail party" and McGurk effects, i.e., speech disambiguation from a panoply of sound signals. This fusion [...] Read more.
Human beings have developed fantastic abilities to integrate information from various sensory sources exploring their inherent complementarity. Perceptual capabilities are therefore heightened, enabling, for instance, the well-known "cocktail party" and McGurk effects, i.e., speech disambiguation from a panoply of sound signals. This fusion ability is also key in refining the perception of sound source location, as in distinguishing whose voice is being heard in a group conversation. Furthermore, neuroscience has successfully identified the superior colliculus region in the brain as the one responsible for this modality fusion, with a handful of biological models having been proposed to approach its underlying neurophysiological process. Deriving inspiration from one of these models, this paper presents a methodology for effectively fusing correlated auditory and visual information for active speaker detection. Such an ability can have a wide range of applications, from teleconferencing systems to social robotics. The detection approach initially routes auditory and visual information through two specialized neural network structures. The resulting embeddings are fused via a novel layer based on the superior colliculus, whose topological structure emulates spatial neuron cross-mapping of unimodal perceptual fields. The validation process employed two publicly available datasets, with achieved results confirming and greatly surpassing initial expectations. Full article
(This article belongs to the Special Issue Computer Vision for Mobile Robotics)
Show Figures

Figure 1

Figure 1
<p>Exemplary output of an multi-speaker active speaker detection (ASD) system where an intervening speaker is denoted in green and a passive speaker is encircled in red.</p>
Full article ">Figure 2
<p>Overall superior colliculus fusion (SCF) layer structure, where the synapses of the central darker neuron (representing neuron <math display="inline"><semantics> <mrow> <mi>i</mi> <mi>j</mi> </mrow> </semantics></math> in a <math display="inline"><semantics> <mrow> <mi>N</mi> <mo>×</mo> <mi>M</mi> </mrow> </semantics></math> neural area) are modelled after a Mexican Hat wavelet. Each neuron is laterally linked to its area <span class="html-italic">s</span> neighbors by excitatory <math display="inline"><semantics> <mrow> <mi>e</mi> <mi>x</mi> </mrow> </semantics></math> and inhibitory <math display="inline"><semantics> <mrow> <mi>i</mi> <mi>n</mi> </mrow> </semantics></math> synapses (red arrows). The described spatially conditional connection between uni-modal neurons (visual/auditory) and their multi-modal counterparts is shown on the right.</p>
Full article ">Figure 3
<p>Structure overview of the proposed model. From a video, the image sequence of a face is extracted as well as the corresponding audio. A narrowband spectrogram is generated from the audio and progressed through VGGVox. The image sequence is progressed through the RN-LSTM without any preprocessing. The obtained audio and visual embeddings are then classified either separately or merged by concatenation or the SCF layer.</p>
Full article ">Figure 4
<p>Performance of the trained full model on unseen data: (<b>a</b>) Single speaker scenario. (<b>b</b>) Multiple speaker scenario.</p>
Full article ">
12 pages, 8465 KiB  
Article
A Deep Convolutional Neural Network Inspired by Auditory Perception for Underwater Acoustic Target Recognition
by Honghui Yang, Junhao Li, Sheng Shen and Guanghui Xu
Sensors 2019, 19(5), 1104; https://doi.org/10.3390/s19051104 - 4 Mar 2019
Cited by 77 | Viewed by 5702
Abstract
Underwater acoustic target recognition (UATR) using ship-radiated noise faces big challenges due to the complex marine environment. In this paper, inspired by neural mechanisms of auditory perception, a new end-to-end deep neural network named auditory perception inspired Deep Convolutional Neural Network (ADCNN) is [...] Read more.
Underwater acoustic target recognition (UATR) using ship-radiated noise faces big challenges due to the complex marine environment. In this paper, inspired by neural mechanisms of auditory perception, a new end-to-end deep neural network named auditory perception inspired Deep Convolutional Neural Network (ADCNN) is proposed for UATR. In the ADCNN model, inspired by the frequency component perception neural mechanism, a bank of multi-scale deep convolution filters are designed to decompose raw time domain signal into signals with different frequency components. Inspired by the plasticity neural mechanism, the parameters of the deep convolution filters are initialized randomly, and the is n learned and optimized for UATR. The n, max-pooling layers and fully connected layers extract features from each decomposed signal. Finally, in fusion layers, features from each decomposed signal are merged and deep feature representations are extracted to classify underwater acoustic targets. The ADCNN model simulates the deep acoustic information processing structure of the auditory system. Experimental results show that the proposed model can decompose, model and classify ship-radiated noise signals efficiently. It achieves a classification accuracy of 81.96%, which is the highest in the contrast experiments. The experimental results show that auditory perception inspired deep learning method has encouraging potential to improve the classification performance of UATR. Full article
(This article belongs to the Special Issue Intelligent Sensor Signal in Machine Learning)
Show Figures

Figure 1

Figure 1
<p>The architecture of ADCNN.</p>
Full article ">Figure 2
<p>Spectrogram of recordings. (<b>a</b>) Cargo recording; (<b>b</b>) Passenger ship recording; (<b>c</b>) Tanker recording; (<b>d</b>) Environment noise recording.</p>
Full article ">Figure 3
<p>Visualization of the output of each filter. (<b>a</b>) Testing sample of Cargo class; (<b>b</b>) Testing sample of Passenger ship class.</p>
Full article ">Figure 4
<p>Result of t-SNE feature visualization. (<b>a</b>–<b>e</b>) Feature groups of deep filter sub-networks; (<b>f</b>) Features of layer-1; (<b>g</b>) Features of layer-2.</p>
Full article ">Figure 5
<p>ROC curves of the proposed model and its competitors. (<b>a</b>) Cargo class is positive class; (<b>b</b>) Passenger ship class is positive class; (<b>c</b>) Tanker class is positive class; (<b>d</b>) Environment noise class is positive class.</p>
Full article ">
Back to TopTop