[go: up one dir, main page]
More Web Proxy on the site http://driver.im/
You seem to have javascript disabled. Please note that many of the page functionalities won't work as expected without javascript enabled.
 
 
Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (987)

Search Parameters:
Keywords = ImageStream

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
19 pages, 8290 KiB  
Article
Multi-Scale Contrastive Learning with Hierarchical Knowledge Synergy for Visible-Infrared Person Re-Identification
by Yongheng Qian and Su-Kit Tang
Sensors 2025, 25(1), 192; https://doi.org/10.3390/s25010192 - 1 Jan 2025
Viewed by 236
Abstract
Visible-infrared person re-identification (VI-ReID) is a challenging cross-modality retrieval task to match a person across different spectral camera views. Most existing works focus on learning shared feature representations from the final embedding space of advanced networks to alleviate modality differences between visible and [...] Read more.
Visible-infrared person re-identification (VI-ReID) is a challenging cross-modality retrieval task to match a person across different spectral camera views. Most existing works focus on learning shared feature representations from the final embedding space of advanced networks to alleviate modality differences between visible and infrared images. However, exclusively relying on high-level semantic information from the network’s final layers can restrict shared feature representations and overlook the benefits of low-level details. Different from these methods, we propose a multi-scale contrastive learning network (MCLNet) with hierarchical knowledge synergy for VI-ReID. MCLNet is a novel two-stream contrastive deep supervision framework designed to train low-level details and high-level semantic representations simultaneously. MCLNet utilizes supervised contrastive learning (SCL) at each intermediate layer to strengthen visual representations and enhance cross-modality feature learning. Furthermore, a hierarchical knowledge synergy (HKS) strategy for pairwise knowledge matching promotes explicit information interaction across multi-scale features and improves information consistency. Extensive experiments on three benchmarks demonstrate the effectiveness of MCLNet. Full article
Show Figures

Figure 1

Figure 1
<p>(<b>a</b>) The traditional supervised learning paradigm only imposes supervision on the last layer of a neural network. (<b>b</b>) Deep supervised learning involves training the last and intermediate layers concurrently. (<b>c</b>) Grad-CAM [<a href="#B21-sensors-25-00192" class="html-bibr">21</a>] visualization of attention maps at different feature extraction stages. Deeper red colors signify higher weights.</p>
Full article ">Figure 2
<p>Illustration of the proposed MCLNet framework. MCLNet first decouples input IR and RGB images into modality-specific and modality-shared features. It then applies a generalized mean pooling (GeM) layer to generate a feature vector, followed by a batch normalization (BN) layer for identity inference. Meanwhile, the projection head maps multi-scale low-level features to the embedding space, where circles represent logits from the final layer and intermediate layers. Here, supervised contrastive learning (SCL) jointly supervises high-level semantics and low-level details while introducing a hierarchical knowledge synergy (HKS) strategy, using pairwise knowledge matching to enhance information consistency across supervised branches.</p>
Full article ">Figure 3
<p>Analysis of trade-off coefficient <math display="inline"><semantics> <msub> <mi>λ</mi> <mrow> <mi>M</mi> <mi>R</mi> </mrow> </msub> </semantics></math>. Re-identification rates at Rank-1 (%) and mAP (%).</p>
Full article ">Figure 4
<p>The distribution of intra-person and inter-person similarities on the two search modes of the SYSU-MM01 dataset.</p>
Full article ">Figure 5
<p>Visualization results of different modality images of randomly selected two identities. Grad-CAM [<a href="#B21-sensors-25-00192" class="html-bibr">21</a>] visualization of the attention maps of B, B + H, B + S, and MCLNet methods are performed, respectively. Deeper red colors signify higher weights.</p>
Full article ">Figure 6
<p>Top-10 retrieved results of some example queries with the MCLNet on SYSU-MM01 and RegDB. The green and red bounding boxes indicate query results matching the same identity and different identities from the gallery, respectively.</p>
Full article ">
33 pages, 3827 KiB  
Review
Distinguishing Reality from AI: Approaches for Detecting Synthetic Content
by David Ghiurău and Daniela Elena Popescu
Computers 2025, 14(1), 1; https://doi.org/10.3390/computers14010001 - 24 Dec 2024
Viewed by 336
Abstract
The advancement of artificial intelligence (AI) technologies, including generative pre-trained transformers (GPTs) and generative models for text, image, audio, and video creation, has revolutionized content generation, creating unprecedented opportunities and critical challenges. This paper systematically examines the characteristics, methodologies, and challenges associated with [...] Read more.
The advancement of artificial intelligence (AI) technologies, including generative pre-trained transformers (GPTs) and generative models for text, image, audio, and video creation, has revolutionized content generation, creating unprecedented opportunities and critical challenges. This paper systematically examines the characteristics, methodologies, and challenges associated with detecting the synthetic content across multiple modalities, to safeguard digital authenticity and integrity. Key detection approaches reviewed include stylometric analysis, watermarking, pixel prediction techniques, dual-stream networks, machine learning models, blockchain, and hybrid approaches, highlighting their strengths and limitations, as well as their detection accuracy, independent accuracy of 80% for stylometric analysis and up to 92% using multiple modalities in hybrid approaches. The effectiveness of these techniques is explored in diverse contexts, from identifying deepfakes and synthetic media to detecting AI-generated scientific texts. Ethical concerns, such as privacy violations, algorithmic bias, false positives, and overreliance on automated systems, are also critically discussed. Furthermore, the paper addresses legal and regulatory frameworks, including intellectual property challenges and emerging legislation, emphasizing the need for robust governance to mitigate misuse. Real-world examples of detection systems are analyzed to provide practical insights into implementation challenges. Future directions include developing generalizable and adaptive detection models, hybrid approaches, fostering collaboration between stakeholders, and integrating ethical safeguards. By presenting a comprehensive overview of AIGC detection, this paper aims to inform stakeholders, researchers, policymakers, and practitioners on addressing the dual-edged implications of AI-driven content creation. Full article
Show Figures

Figure 1

Figure 1
<p>Interest over time for artificial intelligence according to search engines [<a href="#B6-computers-14-00001" class="html-bibr">6</a>].</p>
Full article ">Figure 2
<p>Prisma flow chart with the total number of studies and reports included.</p>
Full article ">Figure 3
<p>Audio waveform analyzed for specific mismatches in tonality, rhythm, and fluency.</p>
Full article ">Figure 4
<p>Dall-E generated image of an apple.</p>
Full article ">Figure 5
<p>Analyzed AIGC image using Sight Engine.</p>
Full article ">Figure 6
<p>Video frame extraction for authenticity analysis.</p>
Full article ">Figure 7
<p>Accuracy of identifying generated and manipulated content [<a href="#B30-computers-14-00001" class="html-bibr">30</a>,<a href="#B31-computers-14-00001" class="html-bibr">31</a>].</p>
Full article ">Figure 8
<p>The process of watermarking an image [<a href="#B35-computers-14-00001" class="html-bibr">35</a>].</p>
Full article ">Figure 9
<p>Regularization technique flow with specific components.</p>
Full article ">Figure 10
<p>News submission on a blockchain ledger with crowdsourcing consensus.</p>
Full article ">
29 pages, 5462 KiB  
Article
Phytochemical Profile and In Vitro Cytotoxic, Genotoxic, and Antigenotoxic Evaluation of Cistus monspeliensis L. Leaf Extract
by Ghanya Al-Naqeb, Gianluca Zorzi, Amanda Oldani, Alberto Azzalin, Linda Avesani, Flavia Guzzo, Alessia Pascale, Rachele De Giuseppe and Hellas Cena
Int. J. Mol. Sci. 2024, 25(24), 13707; https://doi.org/10.3390/ijms252413707 - 22 Dec 2024
Viewed by 373
Abstract
Cistus monspeliensis L. (C. monspeliensis) is used in Italian folk medicine. This study was performed to determine genotoxic and antigenotoxic effects of C. monspeliensis leaf extract against mitomycin C (MMC) using an in vitro cytokinesis-block micronucleus assay (CBMN) in the Chinese [...] Read more.
Cistus monspeliensis L. (C. monspeliensis) is used in Italian folk medicine. This study was performed to determine genotoxic and antigenotoxic effects of C. monspeliensis leaf extract against mitomycin C (MMC) using an in vitro cytokinesis-block micronucleus assay (CBMN) in the Chinese Hamster Ovarian K1 (CHO-K1) cell line. The phytochemical composition of C. monspeliensis extract was evaluated using an untargeted metabolomic approach by employing UPLC-PDA-ESI/MS. The automated in vitro CBMN assay was carried out using image analysis systems with a widefield fluorescence microscope and the ImageStreamX imaging flow cytometer. The phytochemical profile of C. monspeliensis extract showed, as the most abundant metabolites, punicalagin, myricetin, gallocathechin, and a labdane-type diterpene. C. monspeliensis, at the tested concentrations of 50, 100, and 200 μg/mL, did not induce significant micronuclei frequency, thus indicating the absence of a genotoxic potential. When testing the C. monspeliensis extract for antigenotoxicity in the presence of MMC, we observed a hormetic concentration-dependent effect, where low concentrations resulted in a significant protective effect against MMC-induced micronuclei frequency, and higher concentrations resulted in no effect. In conclusion, our findings demonstrate that C. monspeliensis extract is not genotoxic and, at low concentration, exhibits an antigenotoxic effect. In relation to this final point, C. monspeliensis may act as a potential chemo-preventive against genotoxic agents. Full article
Show Figures

Figure 1

Figure 1
<p>Base peak chromatogram (BPC) of diluted (1:10 <span class="html-italic">V</span>/<span class="html-italic">V</span>) methanolic leaf extract of <span class="html-italic">Cistus monspeliensis</span> in negative ionization mode.</p>
Full article ">Figure 2
<p>(<b>A</b>) Viability (% of DMSO control) of CHO-K1 cells after 24 h incubation with different concentrations of <span class="html-italic">C. monspeliensis</span> methanolic extract (9.4–600 µg/mL). Data represent mean ± standard deviation of three independent experiments. (<b>B</b>) A nonlinear regression of log-transformed concentration values (curve fit) was applied to determine the IC<sub>50</sub> value. The percentage of viable cells upon treatment was calculated using this equation: T/C × 100, where T stands for test sample and C for control. Statistical analysis was performed using GraphPad Prism Ver.7. ANOVA followed by Tukey multiple comparison post-test. Different symbols indicate significant differences from DMSO control (*** <span class="html-italic">p</span> = 0.0008; **** <span class="html-italic">p</span> &lt; 0.0001).</p>
Full article ">Figure 3
<p>(<b>A</b>): Percentage of cytotoxicity (%) CN (blue) and % CBPI (black) in CHO-K1 cells (<b>B</b>): % of binucleated cells in CHO-K1 cells after 24 h incubation with different concentrations of MMC, followed by 24 h incubation with 3 μg/mL of cytochalasin B. Graphs represent data collected from three independent experiments. One-way ANOVA, Tukey’s multiple comparisons test using GraphPad Prism 7 software was applied to calculate statistical significance in comparison with NC. (**** <span class="html-italic">p</span> &lt; 0.0001).</p>
Full article ">Figure 4
<p>(<b>A</b>) Percentage of cytotoxicity (%) CN (blue) and CBPI (black) in CHO-K1 cells; (<b>B</b>) % of binucleated cells in CHO-K1 cells after 24 h incubation with different concentrations of <span class="html-italic">C. monspeliensis</span> extract, followed by 24 h incubation with 3 μg/mL of cytochalasin B. Graphs represent data collected from three independent experiments. One-way ANOVA, Tukey’s multiple comparisons test using GraphPad Prism 7 software was applied to calculate statistical significance in comparison with NC.</p>
Full article ">Figure 5
<p>(<b>A</b>): Micronuclei frequency in CHO-K1cells after 24 h incubation with three different concentrations of <span class="html-italic">C. monspeliensis</span> extract, followed by 24 h incubation with 3 μg/mL of cytochalasin B. Graphs represent data collected from three independent experiments. One-way ANOVA, Tukey’s multiple comparisons test using GraphPad Prism 7 software was applied to calculate statistical significance in comparison with NC (**** <span class="html-italic">p</span> &lt; 0.0001). Micronuclei frequency (%) = (binucleated cells with MN/cells × 100). (<b>B</b>): Representative microscopic images of micronuclei formation in binucleated CHO-K1 cells with 40× objective after 24 h incubation with NC, MMC and <span class="html-italic">C. monspeliensis</span> at 50 and 200 μg/mL. CHO-K1 cell DNA was stained with bisbenzimide (Hoechst dye no. 33258). The white arrows showed the micronuclei. The white line in the image shows the scale bar = 50 µm.</p>
Full article ">Figure 6
<p>(<b>A</b>): Percentage of cytotoxicity (%) CN (blue) and CBPI (red) in CHO-K1 cells. (<b>B</b>) % of binucleated cells in CHO-K1 cells after 24 h incubation with different concentrations of <span class="html-italic">C. monspeliensis</span> extract in the presence of 0.025 μg/mL MMC followed by 24 h incubation with 3 μg/mL of cytochalasin B. Graphs represent data collected from three independent experiments. One-way ANOVA, Tukey’s multiple comparisons test using GraphPad Prism 7 software was applied to calculate statistical significance in comparison with MMC control. ** <span class="html-italic">p</span> = 0.0020, *** <span class="html-italic">p</span> = 0.0001, **** <span class="html-italic">p</span> &lt; 0.0001.</p>
Full article ">Figure 7
<p>(<b>A</b>) Micronuclei frequency in CHO-K1 cells after 24 h incubation with three different concentrations of <span class="html-italic">C. monspeliensis</span> extract in the presence of MMC at 0.025 μg/mL, followed by 24 h incubation with 3 μg/mL of cytochalasin B. Graphs represent data collected from three independent experiments. One-way ANOVA, Tukey’s multiple comparisons test using GraphPad Prism 7 software was applied to calculate statistical significance in comparison with MMC control. * <span class="html-italic">p</span> &lt; 0.0492, *** <span class="html-italic">p</span> = 0.0001. (<b>B</b>) Representative microscopic images of for micronuclei formation in binucleated CHO-K1 cells with 40× objective after 24 h incubation with 0.025 μg/mL MMC alone or MMC + <span class="html-italic">C. monspeliensis</span> at 5, 100, and 200 μg/mL. CHO-K1 cell DNA was stained with bisbenzimide (Hoechst dye no. 33258). The white arrows showed the micronuclei. The white line in the image shows the scale bar = 50 µm.</p>
Full article ">Figure 8
<p>Micronuclei frequency (normalized with NC) per 2000 binucleated CHO-K1cells after 24 h incubation with three different concentrations (50, 100, and 200 μg/mL) of <span class="html-italic">C. monspeliensis</span> extract, followed by 24 h incubation with 3 μg/mL of cytochalasin B. Graphs represent data collected from three independent experiments. One-way ANOVA, Tukey’s multiple comparisons test using GraphPad Prism 7 software was applied to calculate statistical significance in comparison with NC control. Micronuclei frequency (%) = (binucleated cells with MN/binucleated cells × 100). * <span class="html-italic">p</span> = 0.0192, *** <span class="html-italic">p</span> = 0.0001, **** <span class="html-italic">p</span> &lt; 0.0001.</p>
Full article ">Figure 9
<p>Representative images captured by the ImageStreamX, with 40× objective, that show bright field (<b>a</b>) image of single cells, (<b>b</b>): Ch05, binucleated cells with micronuclei with or without micronuclei stained with Draq5 of NC, C. monspeliensis extract at tested concentrations of 50 and 200 μg/mL in presence or absence of MMC at 0.025 μg/mL, (<b>c</b>) represents side scatter (SSC) image of each cell.</p>
Full article ">Figure 10
<p>Representative steps for micronuclei analysis using the CellProfiler software (version number 4.2.6): (<b>a</b>) Raw image obtained from widefield microscope acquisition. (<b>b</b>) Deconvolved image obtained from Leica Lightning software tool. (<b>c</b>) Processed image (image crop, smoothing filter, noise-reduction filter) obtained from CellProfiler. (<b>d</b>) Nuclei segmentation obtained from CellProfiler. (<b>e</b>) Nuclei splitting into classes (mononucleated and binucleated cells) obtained from CellProfiler. (<b>f</b>) Definition of cell boundaries (secondary objects), expanding nuclei for a specific distance, obtained from CellProfiler. (<b>g</b>) Definition of cell cytoplasm (tertiary objects) obtained from CellProfiler. (<b>h</b>) Micronuclei segmentation, obtained from CellProfiler. (<b>i</b>) Filtered micronuclei assigned to the nucleus they belong to, obtained from CellProfiler. (<b>j</b>) Outlines of mononucleated and binucleated cells overlayed to (<b>c</b>).</p>
Full article ">Figure 11
<p>Representative steps of micronuclei analysis using the IDEAS Software (version number 6.2). (<b>a</b>) During the analysis, first, selection was based on the Gradient_RMS parameter to confirm in-focus events in the brightfield channel (BF, Ch01). (<b>b</b>) A dot plot of Draq5 lobe count versus aspect ratio was created: the reported gates include all cells with two Draq5-stained nuclei, two lobes, single lobe, and cells with more than two nuclei. (<b>c</b>) A histogram of BF Aspect Ratio, displaying the gate used to discriminate single cells: all events with Aspect Ratio higher than 0.5 were considered for defining bi-nucleated cells in the histogram reported in (<b>d</b>) that used Draq5 Area for this purpose. (<b>e</b>) Dot plot of Draq5 Width versus Homogeneity parameters, where the blue gate includes cells with more uniform distribution of Draq5 stain: the majority of the events of interest showed Homogeneity greater than 10 and Width greater than 15. (<b>f</b>) Previous Homog/Width population was considered in a dot plot of Draq5 Aspect Ratio intensity versus Draq5-Compactness to finally define the gate that encompasses the acceptable BNC population (yellow gate). Each sample was checked to better define each single gate, evaluating each single event. (<b>g</b>) A histogram of our specific spot count feature (micronuclei_Count_A) generated over the masks described in Materials and Methods and in <a href="#app1-ijms-25-13707" class="html-app">Supplementary Data (2)</a>. The linear gates over each bar display the number of BNCs without micronuclei (0 micronuclei) and, respectively, with 1, 2, and 3 micronuclei. The normalized frequency represents the percentage of each type of cell among the total number of cells in the BNC population. BF = brightfield channel, BN = binucleated cells, BNC = binucleated cells with micronuclei.</p>
Full article ">Figure 12
<p>Mask staining (blue shadows over the images): (<b>a</b>) Ch01, brightfield (BF) image with the default mask applied that stains all the picture; (<b>b</b>) Ch05, binucleated cells with micronuclei (BNC) with a single micronucleus stained with Draq5 with the nuclear mask applied (micronuclei _MaskB_Step3, see <a href="#app1-ijms-25-13707" class="html-app">Supplementary Data</a>); (<b>c</b>) Ch05, BNC with a single micronucleus stained with Draq5 with the complete micronuclei mask applied (micronuclei _MaskA_Step3 and Not micronuclei _MaskB_Step3, see <a href="#app1-ijms-25-13707" class="html-app">Supplementary Data</a>).</p>
Full article ">
11 pages, 1974 KiB  
Article
Nuclear Factor-κB Signaling Regulates the Nociceptin Receptor but Not Nociceptin Itself
by Lan Zhang, Ulrike M. Stamer, Robin Moolan-Vadackumchery and Frank Stüber
Cells 2024, 13(24), 2111; https://doi.org/10.3390/cells13242111 - 20 Dec 2024
Viewed by 354
Abstract
The nociceptin receptor (NOP) and nociceptin are involved in the pathways of pain and inflammation. The potent role of nuclear factor-κB (NFκB) in the modulation of tumor necrosis factor-α (TNF-α) and interleukin (IL)-1β on the nociceptin system in human THP-1 cells under inflammatory [...] Read more.
The nociceptin receptor (NOP) and nociceptin are involved in the pathways of pain and inflammation. The potent role of nuclear factor-κB (NFκB) in the modulation of tumor necrosis factor-α (TNF-α) and interleukin (IL)-1β on the nociceptin system in human THP-1 cells under inflammatory conditions were investigated. Cells were stimulated without/with phorbol-myristate-acetate (PMA), TNF-α, IL-1β, or PMA combined with individual cytokines. To examine NFκB’s contribution to the regulation of the nociceptin system, PMA-stimulated cells were treated with NFκB inhibitor BAY 11-7082, JSH-23, or anacardic acid before culturing with TNF-α or IL-1β. NOP and prepronociceptin (ppNOC) mRNA were quantified by RT-qPCR; cell membrane NOP and intracellular nociceptin protein levels were measured by flow cytometry. Phosphorylation and localization of NFκB/p65 were determined using ImageStream. PMA + TNF-α decreased NOP mRNA compared to stimulation with PMA alone, while PMA + IL-1β did not. BAY 11-7082 and JSH-23 reversed the repression of NOP by PMA + TNF-α. TNF-α and IL-1β attenuated PMA’s upregulating effects on ppNOC. None of the inhibitors preserved the upregulation of ppNOC in PMA + TNF-α and PMA + IL-1β cultures. TNF-α strongly mediated the nuclear translocation of NFκB/p65 in PMA-treated cells, while IL-1β did not. Proinflammatory cytokines suppressed NOP and ppNOC mRNA in PMA-induced human THP-1 cells. NFκB signaling seems to be an important regulator controlling the transcription of NOP. These findings suggest that the nociceptin system may play an anti-inflammatory role during immune responses. Full article
Show Figures

Figure 1

Figure 1
<p>Effects of PMA on NOP (<b>A</b>) and nociceptin (<b>B</b>) mRNA and protein levels in THP-1 cells. Cells were treated with PMA 5 ng/mL or without PMA (controls) for 24 h. Median with interquartile range and 10–90 percentiles; mean “+”; mRNA and protein data are representative of six and twelve independent experiments, respectively. Wilcoxon tests * <span class="html-italic">p</span> &lt; 0.05; ** <span class="html-italic">p</span> &lt; 0.005; *** <span class="html-italic">p</span> &lt; 0.001, compared to the respective controls.</p>
Full article ">Figure 2
<p>Dose-dependent effects of cytokines and NFκB inhibitors. THP-1 cells were cultured without/with PMA 5 ng/mL and without/with various concentrations of individual cytokines for 24 h. mRNA expression of <span class="html-italic">NOP</span> (<b>A</b>) and <span class="html-italic">ppNOC</span> (<b>B</b>) is presented as mRNA ratio related to the respective PMA samples. (<b>C</b>) Dose-dependent inhibitory effects of different NFκB inhibitors on <span class="html-italic">IL1B</span> mRNA levels induced by LPS. Cells were pre-treated without/with various concentrations of BAY 11-7082 (BAY), JSH-23 (JSH), or anarcadic acid (AA) for 1 h prior to culturing without/with LPS 100 ng/mL for 6 h. <span class="html-italic">IL1B</span> mRNA levels are presented as mRNA ratio related to an untreated group. Data are from two independent experiments and measures are expressed in mean with range. One-sample t test, * <span class="html-italic">p</span> &lt; 0.05; ** <span class="html-italic">p</span> &lt; 0.005, compared to the PMA group (<span class="html-italic">NOP</span> and <span class="html-italic">ppNOC</span>) and compared to the LPS group (<span class="html-italic">IL1B</span>).</p>
Full article ">Figure 3
<p>Effects of TNF-α and IL-1β on <span class="html-italic">NOP</span> and <span class="html-italic">ppNOC</span> mRNA expression. THP-1 cells were cultured without/with PMA 5 ng/mL in the presence or absence of TNF-α 10 ng/mL or IL-1β 10 ng/mL for 24 h. Quantitative PCR analysis of <span class="html-italic">NOP</span> (<b>A</b>) and <span class="html-italic">ppNOC</span> mRNA expression (<b>B</b>). Fold change values in mRNA levels are normalized against the respective PMA groups. Data are from six independent experiments and are presented as median with interquartile range and 10–90 percentiles; mean “+”. Wilcoxon test with post hoc test. ** <span class="html-italic">p</span> &lt; 0.005.</p>
Full article ">Figure 4
<p>Effects of NFκB inhibitors on <span class="html-italic">NOP</span> and <span class="html-italic">ppNOC</span> mRNA expression. THP-1 cells were stimulated without/with PMA 5 ng/mL for 24 h, followed by treatment without/with one of the NFκB inhibitors (BAY 11-7082 (BAY, 100 nM), JSH-23 (JSH, 100 nM), and anacardic acid (AA, 100 nM)) for 1 h before exposure to TNF-α 10 ng/mL or IL-1β 10 ng/mL for 6 h (BAY and JSH) and for 12 h (AA). <span class="html-italic">NOP</span> (<b>A</b>–<b>C</b>) and <span class="html-italic">ppNOC</span> mRNA levels (<b>D</b>–<b>F</b>) are fold-change related to the corresponding PMA groups (controls). Median with interquartile range and 10–90 percentiles, experiment using BAY or AA, <span class="html-italic">n</span> = 4; experiment using JSH, <span class="html-italic">n</span> = 7. Mann–Whitney U test with post hoc test. * <span class="html-italic">p</span> &lt; 0.05; ** <span class="html-italic">p</span> &lt; 0.005.</p>
Full article ">Figure 5
<p>NFκB/p65 phosphorylation and nuclear translocation. THP-1 cells were treated without/with PMA 5 ng/mL for 24 h, followed by incubation with TNF-α 10 ng/mL or IL-1β 10 ng/mL or without additional cytokines for 1 h. (<b>A</b>) Composite images of representative cells (60×) in real time showing level of NFκB/p65 protein and its localization. From left to right in each panel, brightfield images of each cell, followed by the nuclear (red) and NFκB/p65 (green) images, and the merged image of the nucleus with NFκB/p65. The scale bar is 7 μm. (<b>B</b>) NFκB/p65 translocation into the nucleus, measured using similarity scores (SS). Histogram displaying varying SS in the cells of each group: untreated (white), TNF-α (yellow), IL-1β (green), PMA (purple), PMA + TNF-α (red) and PMA+ IL-1β (blue). (<b>C</b>) Median SS (MSS) for NFκB/p65 of the individual groups. Data are from three independent experiments and measures are expressed in median with interquartile range.</p>
Full article ">
30 pages, 13159 KiB  
Article
GLMAFuse: A Dual-Stream Infrared and Visible Image Fusion Framework Integrating Local and Global Features with Multi-Scale Attention
by Fu Li, Yanghai Gu, Ming Zhao, Deji Chen and Quan Wang
Electronics 2024, 13(24), 5002; https://doi.org/10.3390/electronics13245002 - 19 Dec 2024
Viewed by 434
Abstract
Integrating infrared and visible-light images facilitates a more comprehensive understanding of scenes by amalgamating dual-sensor data derived from identical environments. Traditional CNN-based fusion techniques are predominantly confined to local feature emphasis due to their inherently limited receptive fields. Conversely, Transformer-based models tend to [...] Read more.
Integrating infrared and visible-light images facilitates a more comprehensive understanding of scenes by amalgamating dual-sensor data derived from identical environments. Traditional CNN-based fusion techniques are predominantly confined to local feature emphasis due to their inherently limited receptive fields. Conversely, Transformer-based models tend to prioritize global information, which can lead to a deficiency in feature diversity and detail retention. Furthermore, methods reliant on single-scale feature extraction are inadequate for capturing extensive scene information. To address these limitations, this study presents GLMAFuse, an innovative dual-stream encoder–decoder network, which utilizes a multi-scale attention mechanism to harmoniously integrate global and local features. This framework is designed to maximize the extraction of multi-scale features from source images while effectively synthesizing local and global information across all layers. We introduce the global-aware and local embedding (GALE) module to adeptly capture and merge global structural attributes and localized details from infrared and visible imagery via a parallel dual-branch architecture. Additionally, the multi-scale attention fusion (MSAF) module is engineered to optimize attention weights at the channel level, facilitating an enhanced synergy between high-frequency edge details and global backgrounds. This promotes effective interaction and fusion of dual-modal features. Extensive evaluations using standard datasets demonstrate that GLMAFuse surpasses the existing leading methods in both qualitative and quantitative assessments, highlighting its superior capability in infrared and visible image fusion. On the TNO and MSRS datasets, our method achieves outstanding performance across multiple metrics, including EN (7.15, 6.75), SD (46.72, 47.55), SF (12.79, 12.56), MI (2.21, 3.22), SCD (1.75, 1.80), VIF (0.79, 1.08), Qbaf (0.58, 0.71), and SSIM (0.99, 1.00). These results underscore its exceptional proficiency in infrared and visible image fusion. Full article
(This article belongs to the Special Issue Artificial Intelligence Innovations in Image Processing)
Show Figures

Figure 1

Figure 1
<p>The qualitative fusion results based on CNN, AE, GAN, Transformer frameworks, and GLMAFuse.</p>
Full article ">Figure 2
<p>Deep model based on feature-level fusion.</p>
Full article ">Figure 3
<p>The framework of the proposed GLMAFuse for IVIF (where <math display="inline"><semantics> <mrow> <msub> <mrow> <mi mathvariant="bold">I</mi> </mrow> <mrow> <mi mathvariant="bold">F</mi> </mrow> </msub> </mrow> </semantics></math> means fused image, <math display="inline"><semantics> <mrow> <msub> <mrow> <mi mathvariant="bold">I</mi> </mrow> <mrow> <mi mathvariant="bold">i</mi> <mi mathvariant="bold">r</mi> </mrow> </msub> </mrow> </semantics></math> means infrared image, and <math display="inline"><semantics> <mrow> <msub> <mrow> <mi mathvariant="bold">I</mi> </mrow> <mrow> <mi mathvariant="bold">v</mi> <mi mathvariant="bold">i</mi> <mi mathvariant="bold">s</mi> </mrow> </msub> </mrow> </semantics></math> means visible image).</p>
Full article ">Figure 4
<p>Overview of the global-aware and local embedding module.</p>
Full article ">Figure 5
<p>Overview of the dual-modal interactive residual fusion block.</p>
Full article ">Figure 6
<p>Comparative analysis of visual fusion methods on the TNO dataset. Subfigure (<b>a</b>) displays images of rural scene, while subfigure (<b>b</b>) shows images of urban scene.</p>
Full article ">Figure 7
<p>Cumulative distribution of 8 metrics from the TNO dataset. The point on the curve <math display="inline"><semantics> <mrow> <mo>(</mo> <mi mathvariant="normal">x</mi> <mo>,</mo> <mi mathvariant="normal">y</mi> <mo>)</mo> </mrow> </semantics></math> indicates that there are <math display="inline"><semantics> <mrow> <mo>(</mo> <mn>100</mn> <mo>×</mo> <mi mathvariant="normal">x</mi> <mo>)</mo> <mi mathvariant="normal">%</mi> </mrow> </semantics></math> image pairs with metric values not exceeding y.</p>
Full article ">Figure 8
<p>Qualitative comparison results on the MSRS dataset. Subfigure (<b>a</b>) displays images of nighttime road scene, while subfigure (<b>b</b>) shows images of daytime road scene.</p>
Full article ">Figure 9
<p>Cumulative distribution of 8 metrics from the MSRS dataset. The point on the curve <math display="inline"><semantics> <mrow> <mo>(</mo> <mi mathvariant="normal">x</mi> <mo>,</mo> <mi mathvariant="normal">y</mi> <mo>)</mo> </mrow> </semantics></math> indicates that there are <math display="inline"><semantics> <mrow> <mo>(</mo> <mn>100</mn> <mo>×</mo> <mi mathvariant="normal">x</mi> <mo>)</mo> <mi mathvariant="normal">%</mi> </mrow> </semantics></math> of image pairs with metric values not exceeding y.</p>
Full article ">Figure 10
<p>Qualitative results of the ablation experiment of GALE on the Roadscene dataset.</p>
Full article ">Figure 11
<p>Qualitative results of the ablation experiment of MSAF on the Roadscene dataset.</p>
Full article ">
22 pages, 7963 KiB  
Article
WTSM-SiameseNet: A Wood-Texture-Similarity-Matching Method Based on Siamese Networks
by Yizhuo Zhang, Guanlei Wu, Shen Shi and Huiling Yu
Information 2024, 15(12), 808; https://doi.org/10.3390/info15120808 - 16 Dec 2024
Viewed by 364
Abstract
In tasks such as wood defect repair and the production of high-end wooden furniture, ensuring the consistency of the texture in repaired or jointed areas is crucial. This paper proposes the WTSM-SiameseNet model for wood-texture-similarity matching and introduces several improvements to address the [...] Read more.
In tasks such as wood defect repair and the production of high-end wooden furniture, ensuring the consistency of the texture in repaired or jointed areas is crucial. This paper proposes the WTSM-SiameseNet model for wood-texture-similarity matching and introduces several improvements to address the issues present in traditional methods. First, to address the issue that fixed receptive fields cannot adapt to textures of different sizes, a multi-receptive field fusion feature extraction network was designed. This allows the model to autonomously select the optimal receptive field, enhancing its flexibility and accuracy when handling wood textures at different scales. Secondly, the interdependencies between layers in traditional serial attention mechanisms limit performance. To address this, a concurrent attention mechanism was designed, which reduces interlayer interference by using a dual-stream parallel structure that enhances the ability to capture features. Furthermore, to overcome the issues of existing feature fusion methods that disrupt spatial structure and lack interpretability, this study proposes a feature fusion method based on feature correlation. This approach not only preserves the spatial structure of texture features but also improves the interpretability and stability of the fused features and the model. Finally, by introducing depthwise separable convolutions, the issue of a large number of model parameters is addressed, significantly improving training efficiency while maintaining model performance. Experiments were conducted using a wood texture similarity dataset consisting of 7588 image pairs. The results show that WTSM-SiameseNet achieved an accuracy of 96.67% on the test set, representing a 12.91% improvement in accuracy and a 14.21% improvement in precision compared to the pre-improved SiameseNet. Compared to CS-SiameseNet, accuracy increased by 2.86%, and precision improved by 6.58%. Full article
Show Figures

Figure 1

Figure 1
<p>Diagram of the SiameseNet architecture.</p>
Full article ">Figure 2
<p>Diagram of the WTSM-SiameseNet architecture.</p>
Full article ">Figure 3
<p>Diagram of the MRF-Resnet architecture.</p>
Full article ">Figure 4
<p>Multi-scale receptive field fusion.</p>
Full article ">Figure 5
<p>Concurrent attention.</p>
Full article ">Figure 6
<p>CBAM attention.</p>
Full article ">Figure 7
<p>Texture feature aggregation and matching module.</p>
Full article ">Figure 8
<p>Sample dataset.</p>
Full article ">Figure 9
<p>Training loss.</p>
Full article ">Figure 10
<p>Wood-texture-similarity matching example.</p>
Full article ">
24 pages, 4616 KiB  
Article
Estimating a 3D Human Skeleton from a Single RGB Image by Fusing Predicted Depths from Multiple Virtual Viewpoints
by Wen-Nung Lie and Veasna Vann
Sensors 2024, 24(24), 8017; https://doi.org/10.3390/s24248017 - 15 Dec 2024
Viewed by 706
Abstract
In computer vision, accurately estimating a 3D human skeleton from a single RGB image remains a challenging task. Inspired by the advantages of multi-view approaches, we propose a method of predicting enhanced 2D skeletons (specifically, predicting the joints’ relative depths) from multiple virtual [...] Read more.
In computer vision, accurately estimating a 3D human skeleton from a single RGB image remains a challenging task. Inspired by the advantages of multi-view approaches, we propose a method of predicting enhanced 2D skeletons (specifically, predicting the joints’ relative depths) from multiple virtual viewpoints based on a single real-view image. By fusing these virtual-viewpoint skeletons, we can then estimate the final 3D human skeleton more accurately. Our network consists of two stages. The first stage is composed of a two-stream network: the Real-Net stream predicts 2D image coordinates and the relative depth for each joint from the real viewpoint, while the Virtual-Net stream estimates the relative depths in virtual viewpoints for the same joints. Our network’s second stage consists of a depth-denoising module, a cropped-to-original coordinate transform (COCT) module, and a fusion module. The goal of the fusion module is to fuse skeleton information from the real and virtual viewpoints so that it can undergo feature embedding, 2D-to-3D lifting, and regression to an accurate 3D skeleton. The experimental results demonstrate that our single-view method can achieve a performance of 45.7 mm on average per-joint position error, which is superior to that achieved in several other prior studies of the same kind and is comparable to that of other sequence-based methods that accept tens of consecutive frames as the input. Full article
Show Figures

Figure 1

Figure 1
<p>(<b>a</b>) Multi-view geometry; (<b>b</b>) our setup with multiple virtual viewpoints (the blue camera is real, the other <span class="html-italic">N</span> (here, <span class="html-italic">N</span> = 7) cameras are virtual, and the two green cameras are selected after experiments (<a href="#sec4dot2dot1-sensors-24-08017" class="html-sec">Section 4.2.1</a>)); (<b>c</b>) geometry for depth error analysis.</p>
Full article ">Figure 1 Cont.
<p>(<b>a</b>) Multi-view geometry; (<b>b</b>) our setup with multiple virtual viewpoints (the blue camera is real, the other <span class="html-italic">N</span> (here, <span class="html-italic">N</span> = 7) cameras are virtual, and the two green cameras are selected after experiments (<a href="#sec4dot2dot1-sensors-24-08017" class="html-sec">Section 4.2.1</a>)); (<b>c</b>) geometry for depth error analysis.</p>
Full article ">Figure 2
<p>(<b>a</b>) Overall architecture of our proposed two-stream method. (<b>b</b>) Detailed architecture of the first-stage network, including the “real” stream (Real-Net) and virtual stream (Virtual-Net). (<b>c</b>) Detailed architecture of the fusion module (FM) in the second stage. <span class="html-italic">N</span> denotes the number of virtual viewpoints, <span class="html-italic">J</span> denotes the number of joints, and <span class="html-italic">D</span> denotes the dimension of the embeddings.</p>
Full article ">Figure 2 Cont.
<p>(<b>a</b>) Overall architecture of our proposed two-stream method. (<b>b</b>) Detailed architecture of the first-stage network, including the “real” stream (Real-Net) and virtual stream (Virtual-Net). (<b>c</b>) Detailed architecture of the fusion module (FM) in the second stage. <span class="html-italic">N</span> denotes the number of virtual viewpoints, <span class="html-italic">J</span> denotes the number of joints, and <span class="html-italic">D</span> denotes the dimension of the embeddings.</p>
Full article ">Figure 3
<p>Global context information of humans (P1–P3) with the same 3D pose captured from different viewpoints (with horizontal viewing angles of −α, 0, and β, respectively) by the camera.</p>
Full article ">Figure 4
<p>The architecture of the fusion network in the fusion module, where <span class="html-italic">N</span> is the total number of virtual viewpoints: (<b>a</b>) DenseFC network; (<b>b</b>) GCN.</p>
Full article ">Figure 5
<p>Illustration of the bone vector connections in our system.</p>
Full article ">Figure 6
<p>(<b>a</b>) Error distribution across different actions, where the dotted red line refers to the overall MPJPE value of 45.7 mm; (<b>b</b>) average MPJPE of each joint.</p>
Full article ">Figure 6 Cont.
<p>(<b>a</b>) Error distribution across different actions, where the dotted red line refers to the overall MPJPE value of 45.7 mm; (<b>b</b>) average MPJPE of each joint.</p>
Full article ">Figure 7
<p>Visualized results on the Human3.6M dataset: (<b>a</b>) successful predictions; (<b>b</b>) failed predictions on some joints.</p>
Full article ">Figure 8
<p>Qualitative results of the in-the-wild scenarios: (<b>a</b>) successful cases; (<b>b</b>) failed cases.</p>
Full article ">
14 pages, 2007 KiB  
Article
Hardware-Assisted Low-Latency NPU Virtualization Method for Multi-Sensor AI Systems
by Jong-Hwan Jean and Dong-Sun Kim
Sensors 2024, 24(24), 8012; https://doi.org/10.3390/s24248012 - 15 Dec 2024
Viewed by 855
Abstract
Recently, AI systems such as autonomous driving and smart homes have become integral to daily life. Intelligent multi-sensors, once limited to single data types, now process complex text and image data, demanding faster and more accurate processing. While integrating NPUs and sensors has [...] Read more.
Recently, AI systems such as autonomous driving and smart homes have become integral to daily life. Intelligent multi-sensors, once limited to single data types, now process complex text and image data, demanding faster and more accurate processing. While integrating NPUs and sensors has improved processing speed and accuracy, challenges like low resource utilization and long memory latency remain. This study proposes a method to reduce processing time and improve resource utilization by virtualizing NPUs to simultaneously handle multiple deep-learning models, leveraging a hardware scheduler and data prefetching techniques. Experiments with 30,000 SA resources showed that the hardware scheduler reduced memory cycles by over 10% across all models, with reductions of 30% for NCF and 70% for DLRM. The hardware scheduler effectively minimized memory latency and idle NPU resources in resource-constrained environments with frequent context switching. This approach is particularly valuable for real-time applications like autonomous driving, enabling smooth transitions between tasks such as object detection and route planning. It also enhances multitasking in smart homes by reducing latency when managing diverse data streams. The proposed system is well suited for resource-constrained environments that demand efficient multitasking and low-latency processing. Full article
(This article belongs to the Section Intelligent Sensors)
Show Figures

Figure 1

Figure 1
<p>Examples of real-life applications of multi-sensor AI in various fields.</p>
Full article ">Figure 2
<p>NPU virtualization operation flow. The symbol ‘#’ represents the number of cores.</p>
Full article ">Figure 3
<p>TPU v4 architecture.</p>
Full article ">Figure 4
<p>Hardware-assisted NPU virtualization system.</p>
Full article ">Figure 5
<p>Comparison before and after the hardware scheduler when the burst size was changed.</p>
Full article ">Figure 6
<p>Comparison before and after hardware scheduler application when the number of available SA changed.</p>
Full article ">Figure 7
<p>Difference in memory cycles with and without a hardware scheduler.</p>
Full article ">
25 pages, 2229 KiB  
Article
MIRA-CAP: Memory-Integrated Retrieval-Augmented Captioning for State-of-the-Art Image and Video Captioning
by Sabina Umirzakova, Shakhnoza Muksimova, Sevara Mardieva, Murodjon Sultanov Baxtiyarovich and Young-Im Cho
Sensors 2024, 24(24), 8013; https://doi.org/10.3390/s24248013 - 15 Dec 2024
Viewed by 533
Abstract
Generating accurate and contextually rich captions for images and videos is essential for various applications, from assistive technology to content recommendation. However, challenges such as maintaining temporal coherence in videos, reducing noise in large-scale datasets, and enabling real-time captioning remain significant. We introduce [...] Read more.
Generating accurate and contextually rich captions for images and videos is essential for various applications, from assistive technology to content recommendation. However, challenges such as maintaining temporal coherence in videos, reducing noise in large-scale datasets, and enabling real-time captioning remain significant. We introduce MIRA-CAP (Memory-Integrated Retrieval-Augmented Captioning), a novel framework designed to address these issues through three core innovations: a cross-modal memory bank, adaptive dataset pruning, and a streaming decoder. The cross-modal memory bank retrieves relevant context from prior frames, enhancing temporal consistency and narrative flow. The adaptive pruning mechanism filters noisy data, which improves alignment and generalization. The streaming decoder allows for real-time captioning by generating captions incrementally, without requiring access to the full video sequence. Evaluated across standard datasets like MS COCO, YouCook2, ActivityNet, and Flickr30k, MIRA-CAP achieves state-of-the-art results, with high scores on CIDEr, SPICE, and Polos metrics, underscoring its alignment with human judgment and its effectiveness in handling complex visual and temporal structures. This work demonstrates that MIRA-CAP offers a robust, scalable solution for both static and dynamic captioning tasks, advancing the capabilities of vision–language models in real-world applications. Full article
(This article belongs to the Section Sensing and Imaging)
Show Figures

Figure 1

Figure 1
<p>This figure presents the architecture of the MIRA-CAP model, illustrating the sequential flow of components involved in generating contextually rich and temporally coherent captions for images and videos. The diagram begins with an input image on the left, representing the visual data fed into the model. The processing pipeline is divided into multiple stages, each corresponding to a key component of the MIRA-CAP framework.</p>
Full article ">Figure 2
<p>Illustrate the Adaptive Dataset Pruning process in detail.</p>
Full article ">Figure 3
<p>This figure illustrates the effectiveness of the MIRA-CAP model in generating accurate, contextually rich captions for various scenes, highlighting the model’s ability to identify objects, actions, and environments in diverse visual contexts. Each image is accompanied by a caption generated by MIRA-CAP, with key descriptive terms highlighted in red to showcase elements that contribute to semantic richness and specificity.</p>
Full article ">Figure 4
<p>The <b>top row</b> shows the original input images, while the <b>bottom row</b> overlays attention heatmaps (using Grad-CAM) to highlight salient areas.</p>
Full article ">Figure 5
<p>Illustrates examples of limitations encountered by MIRA-CAP in generating captions for complex visual scenes. The <b>top image</b> demonstrates a scenario where the model misidentifies background objects and fails to capture the primary action accurately. Instead of focusing on the fish seller’s preparation of the fish, the caption incorrectly highlights the shopping activity, leading to a misalignment of context. The <b>bottom image</b> showcases a case where the model struggles to describe the environment and activity appropriately. While the man is standing in a trench beside a concrete structure, the generated caption inaccurately describes the scene as a man inspecting a garden. These examples emphasize the challenges MIRA-CAP faces in crowded environments, misaligned object relationships, and failure to distinguish primary actions from secondary elements.</p>
Full article ">
18 pages, 25764 KiB  
Article
Evaluating Landsat- and Sentinel-2-Derived Burn Indices to Map Burn Scars in Chyulu Hills, Kenya
by Mary C. Henry and John K. Maingi
Fire 2024, 7(12), 472; https://doi.org/10.3390/fire7120472 - 11 Dec 2024
Viewed by 570
Abstract
Chyulu Hills, Kenya, serves as one of the region’s water towers by supplying groundwater to surrounding streams and springs in southern Kenya. In a semiarid region, this water is crucial to the survival of local people, farms, and wildlife. The Chyulu Hills is [...] Read more.
Chyulu Hills, Kenya, serves as one of the region’s water towers by supplying groundwater to surrounding streams and springs in southern Kenya. In a semiarid region, this water is crucial to the survival of local people, farms, and wildlife. The Chyulu Hills is also very prone to fires, and large areas of the range burn each year during the dry season. Currently, there are no detailed fire records or burn scar maps to track the burn history. Mapping burn scars using remote sensing is a cost-effective approach to monitor fire activity over time. However, it is not clear whether spectral burn indices developed elsewhere can be directly applied here when Chyulu Hills contains mostly grassland and bushland vegetation. Additionally, burn scars are usually no longer detectable after an intervening rainy season. In this study, we calculated the Differenced Normalized Burn Ratio (dNBR) and two versions of the Relative Differenced Normalized Burn Ratio (RdNBR) using Landsat Operational Land Imager (OLI) and Sentinel-2 MultiSpectral Instrument (MSI) data to determine which index, threshold values, instrument, and Sentinel near-infrared (NIR) band work best to map burn scars in Chyulu Hills, Kenya. The results indicate that the Relative Differenced Normalized Burn Ratio from Landsat OLI had the highest accuracy for mapping burn scars while also minimizing false positives (commission error). While mapping burn scars, it became clear that adjusting the threshold value for an index resulted in tradeoffs between false positives and false negatives. While none were perfect, this is an important consideration going forward. Given the length of the Landsat archive, there is potential to expand this work to additional years. Full article
(This article belongs to the Special Issue Fire in Savanna Landscapes, Volume II)
Show Figures

Figure 1

Figure 1
<p>Location of Chyulu Hills, Kenya, in East Africa. Protected areas are shown in hatch-filled areas with labels in legend. Study area falls within three counties, Kajiado, Makueni, and Taita Taveta, as shown in map. Elevation is also shown in map, with higher elevations in white. Major roads include Mombasa Road to east of study area.</p>
Full article ">Figure 2
<p>Flow chart showing methods used in this study. OLI = Operational Land Imager, OLI2 = Operational Land Imager 2; MSI = MultiSpectral Instrument; BOA = Bottom of Atmosphere Reflectance; NBR = Normalized Burn Ratio; dNBR = Differenced Normalized Burn Ratio; RdNBR = Relative Differenced Normalized Burn Ratio; RdNBR2 = Relative Differenced Normalized Burn Ratio alternate calculation. Boxes with bold outline indicate inputs to final analysis.</p>
Full article ">Figure 3
<p>Mapped burn scars for the 2021 fire season in Chyulu Hills, Kenya. Yellow shows areas mapped as burned using Landsat RdNBR with a threshold of 0.23. Clouds and cloud shadows are masked out and shown in black. Purple boundaries indicate protected areas in the Chyulu Hills.</p>
Full article ">Figure 4
<p>Mapped burn scars for the 2021 fire season in Chyulu Hills, Kenya. Yellow shows areas mapped as burned using Sentinel-2 RdNBR with a threshold of 0.22. Clouds and cloud shadows are masked out and shown in black. Purple boundaries indicate protected areas in the Chyulu Hills.</p>
Full article ">
16 pages, 4755 KiB  
Article
Experimental and Mathematical Modelling Investigation of Plasma Electrolytic Oxidation (PEO) for Surface Hardening of 20Ch Steel
by Kuat Kombayev, Fuad Khoshnaw, Gulzhaz Uazyrkhanova and Gulzhaz Moldabayeva
Materials 2024, 17(24), 6043; https://doi.org/10.3390/ma17246043 - 10 Dec 2024
Viewed by 602
Abstract
This study aimed to develop an alternative surface hardening technique for low-carbon steel alloy type 20Ch using plasma electrolytic oxidation (PEO). The surface hardening of 20Ch alloy steel samples was achieved through PEO in a Na2CO3 electrolyte solution. Optimal processing [...] Read more.
This study aimed to develop an alternative surface hardening technique for low-carbon steel alloy type 20Ch using plasma electrolytic oxidation (PEO). The surface hardening of 20Ch alloy steel samples was achieved through PEO in a Na2CO3 electrolyte solution. Optimal processing parameters were determined experimentally by measuring voltage and applied current. Quenching was performed in the electrolyte stream, and plasma was ionised through excitation. A mathematical model based on thermal conductivity equations and regression analysis was developed to relate the key parameters of the hardening process. The results from both the experimental and mathematical models demonstrated that PEO significantly reduces hardening time compared to traditional methods. The microstructural images revealed the transformation of the coarse-grained pearlite–ferrite structure into quenched martensite. Vickers microhardness tests indicated a substantial increase in surface hardness after PEO treatment, compared to the untreated samples. The major advantages of PEO include lower energy consumption, high quenching rates, and the ability to perform localised surface treatments. These benefits contribute to overall cost reduction, making PEO a promising surface hardening method for various industrial applications. Full article
Show Figures

Figure 1

Figure 1
<p>Experimental setup for plasma electrolyte oxidation. 1—power supply; 2—working bath; 3—conical nozzle; 4—reservoir; 5—clamping mechanism; 6—table; 7—pump; 8—filter; 9—hood; 10—manometer; 11—thermometer; 12—ball valve; 13, 14—high-pressure hoses.</p>
Full article ">Figure 2
<p>PEO process of a sample in a working bath. 1—working bath; 2—tube for return of electrolyte; 3—anode plate; 4—conical nozzle; 5—cathode—sample/detail; 6—clamping mechanism; 7—holes on the anode; 8—nozzle for electrolyte supply.</p>
Full article ">Figure 3
<p>Sample preparation for the microstructure and cross-section analysis.</p>
Full article ">Figure 4
<p>Microstructure of steel 20Ch in its initial state; optical microscopy.</p>
Full article ">Figure 5
<p>Volt–temperature characteristics of PEO, in the time interval of one cycle.</p>
Full article ">Figure 6
<p>A mathematical model estimation of the temperature profile dependence on the quenching time of the modelled steels in the electrolyte.</p>
Full article ">Figure 7
<p>Microstructure of the surface layer of steel 20Ch after PEO, U = 200 V, t = 4 min.</p>
Full article ">Figure 8
<p>EDS analysis, using SEM, after PEO.</p>
Full article ">Figure 9
<p>An illustration of the electrolytic-plasma treatment of the workpiece surface.</p>
Full article ">Figure 10
<p>(<b>a</b>) The surface of the as-received alloy before PEO, (<b>b</b>) the surface of the alloy after PEO treatment, and (<b>c</b>) a cross-section of the alloy showing the high-hardness region (surface layer), medium-hardness region (intermediate zone), and low-hardness region (subsurface zone).</p>
Full article ">Figure 11
<p>Variations in microhardness were measured on 20Ch steel: (<b>a</b>) Vickers hardness values for the surfaces before and after PEO, and (<b>b</b>) microhardness values across the cross-sectional depths after PEO.</p>
Full article ">
18 pages, 5749 KiB  
Article
Multivariantism of Auditory Perceptions as a Significant Element of the Auditory Scene Analysis Concept
by Adam Rosiński
Arts 2024, 13(6), 180; https://doi.org/10.3390/arts13060180 - 9 Dec 2024
Viewed by 450
Abstract
The concept of auditory scene analysis, popularized in scientific experiments by A. S. Bregman, the primary architect of the perceptual streaming theory, and his research team, along with more recent analyses by subsequent researchers, highlights a specific scientific gap that has not been [...] Read more.
The concept of auditory scene analysis, popularized in scientific experiments by A. S. Bregman, the primary architect of the perceptual streaming theory, and his research team, along with more recent analyses by subsequent researchers, highlights a specific scientific gap that has not been thoroughly explored in previous studies. This article seeks to expand on this concept by introducing the author’s observation of the multivariant nature of auditory perception. This notion suggests that listeners focusing on different components of an auditory image (such as a musical piece) may perceive the same sounds but interpret them as distinct sound structures. Notably, even the same listener may perceive various structures (different mental figures) when re-listening to the same piece, depending on which musical elements they focus on. The thesis of multivariantism was examined and confirmed through the analysis of selected classical music pieces, providing concrete evidence of different interpretations of the same sound stimuli. To enhance clarity and understanding, the introduction to multivariantism was supplemented with graphic examples from the visual arts, which were then related to musical art through score excerpts from the works of composers such as C. Saint-Saëns, F. Liszt, and F. Mendelssohn Bartholdy. Full article
(This article belongs to the Special Issue Applied Musicology and Ethnomusicology)
Show Figures

Figure 1

Figure 1
<p>The field presented in (<b>a</b>) can be interpreted by the receiver in two different ways, as shown in (<b>b</b>,<b>c</b>). Source: own research.</p>
Full article ">Figure 2
<p>The field presented in (<b>a</b>) can be interpreted by the receiver in a few different ways, as shown in (<b>b</b>–<b>g</b>). Source: own research.</p>
Full article ">Figure 3
<p>The extended field presented in (<b>a</b>) can be interpreted as very many perceptual variants (<b>a</b>–<b>ad</b>), depending on the element that the viewer focuses his attention on in a given moment. Source: own research.</p>
Full article ">Figure 4
<p>Johann Sebastian Bach, <span class="html-italic">Toccata and fugue in D minor</span> (<span class="html-italic">Dorian</span>), BWV 538, <span class="html-italic">Toccata</span>, a fragment of measures 3–4, mental representation for variant No. 1. Source: own research.</p>
Full article ">Figure 5
<p>Johann Sebastian Bach, <span class="html-italic">Toccata and fugue in D minor</span> (<span class="html-italic">Dorian</span>), BWV 538, <span class="html-italic">Toccata</span>, a fragment of measures 3–4, mental representation for the variant No. 2. Source: own research.</p>
Full article ">Figure 6
<p>Johann Sebastian Bach, <span class="html-italic">Toccata and fugue in D minor</span> (<span class="html-italic">Dorian</span>), BWV 538, <span class="html-italic">Toccata</span>, a fragment of measures 3–4, mental representation for variant No. 3. Source: own research.</p>
Full article ">Figure 7
<p>Johann Sebastian Bach, <span class="html-italic">Toccata and fugue in D minor</span> (<span class="html-italic">Dorian</span>), BWV 538, <span class="html-italic">Toccata</span>, a fragment of measures 3–4, mental representation for variant No. 4. Source: own research.</p>
Full article ">Figure 8
<p>Johann Sebastian Bach, <span class="html-italic">Toccata and fugue in D minor</span> (<span class="html-italic">Dorian</span>), BWV 538, <span class="html-italic">Toccata</span>, a fragment of measures 3–4, mental representation for variant No. 5. Source: own research.</p>
Full article ">Figure 9
<p>Johann Sebastian Bach, <span class="html-italic">Toccata and fugue in D minor</span> (<span class="html-italic">Dorian</span>), BWV 538, <span class="html-italic">Toccata</span>, a fragment of measures 3–4, mental representation for variant No. 6. Source: own research.</p>
Full article ">Figure 10
<p>Johann Sebastian Bach, <span class="html-italic">Toccata and fugue in D minor</span> (<span class="html-italic">Dorian</span>), BWV 538, <span class="html-italic">Toccata</span>, a fragment of measures 3–4, mental representation for variant No. 7. Source: own research.</p>
Full article ">Figure 11
<p>Camille Saint-Saëns, <span class="html-italic">Pour l’indépendance des doigts</span>, op. 52, no. 2, <span class="html-italic">Andantino malinconico</span>, m. 1–14. Source: <a href="https://imslp.org/wiki/Special:IMSLPImageHandler/27025%2Fcy28" target="_blank">https://imslp.org/wiki/Special:IMSLPImageHandler/27025%2Fcy28</a> (accessed on 23 August 2023).</p>
Full article ">Figure 12
<p>Camille Saint-Saëns, <span class="html-italic">Prélude et fugue, F Minor</span>, op. 52, no. 3, <span class="html-italic">Fugue, Animato</span>, m. 1–4. Source: <a href="https://imslp.org/wiki/Special:IMSLPImageHandler/27025%2Fcy28" target="_blank">https://imslp.org/wiki/Special:IMSLPImageHandler/27025%2Fcy28</a> (accessed on 23 August 2023).</p>
Full article ">Figure 13
<p>Camille Saint-Saëns, <span class="html-italic">Prélude et fugue, A Major</span>, op. 52, no. 5, <span class="html-italic">Allegro moderato</span>, m. 1–5. Source: <a href="https://imslp.org/wiki/Special:IMSLPImageHandler/27025%2Fcy28" target="_blank">https://imslp.org/wiki/Special:IMSLPImageHandler/27025%2Fcy28</a> (accessed on 23 August 2023).</p>
Full article ">Figure 14
<p>Camille Saint-Saëns, <span class="html-italic">Prélude et fugue, A Major</span>, op. 52, no. 5, <span class="html-italic">Allegro moderato</span>, m. 9–10. Source: <a href="https://imslp.org/wiki/Special:IMSLPImageHandler/27025%2Fcy28" target="_blank">https://imslp.org/wiki/Special:IMSLPImageHandler/27025%2Fcy28</a> (accessed on 23 August 2023).</p>
Full article ">Figure 15
<p>Franz Liszt, <span class="html-italic">Réminiscences de Don Juan</span>, S. 418, <span class="html-italic">Andantino</span>, m. 61–68. Source: <a href="https://ks15.imslp.org/files/imglnks/usimg/e/e5/IMSLP95879-PMLP08541-Liszt_Klavierwerke_Peters_Sauer_Band_8_14_Don_Juan_Phantasie_scan.pdf" target="_blank">https://ks15.imslp.org/files/imglnks/usimg/e/e5/IMSLP95879-PMLP08541-Liszt_Klavierwerke_Peters_Sauer_Band_8_14_Don_Juan_Phantasie_scan.pdf</a> (accessed on 6 December 2024).</p>
Full article ">Figure 16
<p>Felix Mendelssohn Bartholdy, <span class="html-italic">Sechs Lieder ohne Worte</span>, op. 19, no. 1, <span class="html-italic">Andante con moto</span>, m. 3–19. Source: <a href="https://imslp.org/wiki/Special:ImagefromIndex/52241/cy28" target="_blank">https://imslp.org/wiki/Special:ImagefromIndex/52241/cy28</a> (accessed on 2 November 2023).</p>
Full article ">
16 pages, 8072 KiB  
Article
Research on a Panoramic Image Stitching Method for Images of Corn Ears, Based on Video Streaming
by Yi Huangfu, Hongming Chen, Zhonghao Huang, Wenfeng Li, Jie Shi and Linlin Yang
Agronomy 2024, 14(12), 2884; https://doi.org/10.3390/agronomy14122884 - 3 Dec 2024
Viewed by 539
Abstract
Background: Corn is the main grain crop grown in China, and the ear shape index of corn is an important parameter for breeding new varieties, including ear length, diameter, row number of ears, row number of grains per ear, and so on. Objective: [...] Read more.
Background: Corn is the main grain crop grown in China, and the ear shape index of corn is an important parameter for breeding new varieties, including ear length, diameter, row number of ears, row number of grains per ear, and so on. Objective: In order to solve the problem of limited field of view associated with computer detection of the corn ear shape index against a complex background, this paper proposes a panoramic splicing method for corn ears against a complex background, which can splice 10 corn ear panoramic images at the same time, to improve information collection efficiency, display comprehensive information, and support data analysis, so as to realize automatic corn seed examination. Methods: A summary of corn ear panoramic stitching methods under complex backgrounds is presented as follows: 1. a perceptual hash algorithm and histogram equalization were used to extract video frames; 2. the U-Net image segmentation model based on transfer learning was used to predict corn labels; 3. a mask preprocessing algorithm was designed; 4. a corn ear splicing positioning algorithm was designed; 5. an algorithm for irregular surface expansion was designed; 6. an image stitching method based on template matching was adopted to assemble the video frames. Results: The experimental results showed that the proposed corn ear panoramic stitching method could effectively solve the problems of virtual stitching, obvious stitching seams, and too-high similarity between multiple images. The success rate of stitching was as high as 100%, and the speed of single-corn-ear panoramic stitching was about 9.4 s, indicating important reference value for corn breeding and disease and insect detection. Discussions: Although the experimental results demonstrated the significant advantages of the panoramic splicing method for corn ear images proposed in this paper in terms of improving information collection efficiency and automating corn assessment, the method still faces certain challenges. Future research will focus on the following points: 1. addressing the issue of environmental interference caused by diseases, pests, and plant nutritional status on the measurement of corn ear parameters in order to enhance the stability and accuracy of the algorithm; 2. expanding the dataset for the U-Net model to include a wider range of corn ears with complex backgrounds, different growth stages, and various environmental conditions to improve the model’s segmentation recognition rate and precision. Recently, our panoramic splicing algorithm has been deployed in practical applications with satisfactory results. We plan to continue optimizing the algorithm and more broadly promote its use in fields such as corn breeding and pest and disease detection in an effort to advance the development of agricultural automation technology. Full article
(This article belongs to the Section Precision and Digital Agriculture)
Show Figures

Figure 1

Figure 1
<p>Algorithm flow chart.</p>
Full article ">Figure 2
<p>Fruit Spike Parameter Measuring Device (The image displays the fruiting spike parameter collection device along with its various components).</p>
Full article ">Figure 3
<p>Rotating Visualizer model.</p>
Full article ">Figure 4
<p>Partial dataset of corn ears ((<b>a</b>) shows white corn ears, while (<b>b</b>) shows yellow corn ears).</p>
Full article ">Figure 5
<p>Histogram equalization comparison (left figure (<b>a</b>) is the original image, right figure (<b>b</b>) is the histogram equalization figure).</p>
Full article ">Figure 6
<p>Prediction results of U-Net model (left figure (<b>a</b>) is the original image of the ear, and right figure (<b>b</b>) is the result of U-Net prediction of the ear).</p>
Full article ">Figure 7
<p>Fruit ear mapping results (left figure (<b>a</b>) is the result image processed by the ear positioning algorithm, right figure (<b>b</b>) is the spike mask map to be treated).</p>
Full article ">Figure 8
<p>Flow chart of ear mask preprocessing algorithm.</p>
Full article ">Figure 9
<p>Fruit ear mask preprocessing results (left figure (<b>a</b>) is the original image of corn ear mask, right figure (<b>b</b>) is the mask result after ear mask preprocessing algorithm).</p>
Full article ">Figure 10
<p>Fixed-point plot of mask.</p>
Full article ">Figure 11
<p>Surface unwrapping diagram of the mesh optimization method (left figure (<b>a</b>) is the input spike mask and the original spike map, middle figure (<b>b</b>) is the output mesh map, representing the area to be unwrapped, and right figure (<b>c</b>) is the surface unwrapping result diagram).</p>
Full article ">Figure 12
<p>Splicing results of corn ears (figure (<b>a</b>) on the left shows the mosaic result after processing with the median filter algorithm, while figure (<b>b</b>) on the right presents the original mosaic without any treatment).</p>
Full article ">Figure 13
<p>Similarity comparison (Figure (<b>a</b>,<b>b</b>) are sequential frames before histogram equalization, and Figure (<b>c</b>,<b>d</b>) are images after histogram equalization).</p>
Full article ">Figure 14
<p>Histogram comparison.</p>
Full article ">Figure 15
<p>Comparison of image stitching results before and after video frame (figure (<b>a</b>) on the left is the panoramic image stitching result before video frame deduplication, and figure (<b>b</b>) on the right is the panoramic image stitching result after video frame deduplication).</p>
Full article ">Figure 16
<p>Comparison of the effects of different interpolation methods (Figure (<b>a</b>) is the original figure, Figure (<b>b</b>) is the grid optimization diagram using the proximity interpolation method, Figure (<b>c</b>) is the grid optimization diagram using the linear interpolation method, and Figure (<b>d</b>) is the grid optimization diagram using the cubic interpolation method).</p>
Full article ">Figure 17
<p>Binarization results of panoramic image stitching of corn ears.</p>
Full article ">
18 pages, 44302 KiB  
Article
DuSiamIE: A Lightweight Multidimensional Infrared-Enhanced RGBT Tracking Algorithm for Edge Device Deployment
by Jiao Li, Haochen Wu, Yuzhou Gu, Junyu Lu and Xuecheng Sun
Electronics 2024, 13(23), 4721; https://doi.org/10.3390/electronics13234721 - 29 Nov 2024
Viewed by 443
Abstract
Advancements in deep learning and infrared sensors have facilitated the integration of RGB-thermal (RGBT) tracking technology in computer vision. However, contemporary RGBT tracking methods handle complex image data, resulting in inference procedures with a large number of floating-point operations and parameters, which limits [...] Read more.
Advancements in deep learning and infrared sensors have facilitated the integration of RGB-thermal (RGBT) tracking technology in computer vision. However, contemporary RGBT tracking methods handle complex image data, resulting in inference procedures with a large number of floating-point operations and parameters, which limits their performance on general-purpose processors. We present a lightweight Siamese dual-stream infrared-enhanced RGBT tracking algorithm, called DuSiamIE.It is implemented on the low-power NVIDIA Jetson Nano to assess its practicality for edge-device applications in resource-limited settings. Our algorithm replaces the conventional backbone network with a modified MobileNetV3 and incorporates light-aware and infrared feature enhancement modules to extract and integrate multimodal information. Finally, NVIDIA TensorRT is used to improve the inference speed of the algorithm on edge devices. We validated our algorithm on two public RGBT tracking datasets. On the GTOT dataset, DuSiamIE achieved a precision (PR) of 83.4% and a success rate (SR) of 66.8%, with a tracking speed of 40.3 frames per second (FPS). On the RGBT234 dataset, the algorithm achieved a PR of 75.3% and an SR of 52.6%, with a tracking speed of 34.7 FPS. Compared with other algorithms, DuSiamIE exhibits a slight loss in accuracy but significantly outperforms them in speed on resource-constrained edge devices. It is the only algorithm among those tested that can perform real-time tracking on such devices. Full article
(This article belongs to the Special Issue Machine Learning and Deep Learning Based Pattern Recognition)
Show Figures

Figure 1

Figure 1
<p>Examples of poor lighting conditions (<b>left</b>) and exposure to shade (<b>right</b>).</p>
Full article ">Figure 2
<p>The DuSiamIE network: feature extraction via a Siamese network, modal fusion with the LIFA module, infrared enhancement using the MIFE module, and tracking output after classification and regression.</p>
Full article ">Figure 3
<p>LIFA module.</p>
Full article ">Figure 4
<p>MIFE module: (<b>a</b>) MIFE’s network structure. (<b>b</b>) Structure of the mechanism of self-attention in different dimensions.</p>
Full article ">Figure 5
<p>NVIDIA TensorRT optimization process.</p>
Full article ">Figure 6
<p>Comparison of speeds of various tracking methods. (<b>a</b>) PR and speed based on GTOT. (<b>b</b>) SR and speed based on GTOT.</p>
Full article ">Figure 7
<p>Visual comparison of the tracker proposed in this article with other trackers on three video sequences of GTOT: (<b>a</b>) Torabi1, (<b>b</b>) Quarrying, and (<b>c</b>) Fastcar2.</p>
Full article ">Figure 8
<p>Visual comparison of the tracker proposed in this article with other trackers on three video sequences of RGBT234: (<b>a</b>) dog11, (<b>b</b>) electric bicycle in front car, and (<b>c</b>) flower2.</p>
Full article ">
24 pages, 15074 KiB  
Article
A Symmetric Reversible Audio Information Hiding Algorithm Using Matrix Embedding Within Image Carriers
by Yongqiang Tuo, Guodong Li and Kaiyue Hou
Symmetry 2024, 16(12), 1586; https://doi.org/10.3390/sym16121586 - 27 Nov 2024
Viewed by 536
Abstract
To address the vulnerability of existing hiding algorithms to differential attacks and the limitations of single chaotic systems, such as small key space and low security, a novel algorithm combining audio encryption with information hiding is proposed. First, the original audio is divided [...] Read more.
To address the vulnerability of existing hiding algorithms to differential attacks and the limitations of single chaotic systems, such as small key space and low security, a novel algorithm combining audio encryption with information hiding is proposed. First, the original audio is divided into blocks to enhance efficiency. A “one-time pad” mechanism is achieved by associating the key with the plaintext, and a new multidimensional sine-coupled chaotic map is designed, which, in conjunction with multiple chaotic systems, generates the key stream. Next, the block-processed audio signals are matrix-converted and then encrypted using cyclic remainder scrambling, an improved Josephus scrambling, XOR diffusion, and bit diffusion. This results in an encrypted audio information matrix. Finally, the GHM multiwavelet transform is used to select embedding channels, and the least significant bit (LSB) method is employed to hide the information within the carrier image. The algorithm is symmetric, and decryption involves simply reversing the encryption process on the stego image. Experimental results demonstrate that the Structural Similarity Index (SSIM) between the carrier image and the stego image is 0.992540, the Peak Signal-to-Noise Ratio (PSNR) is 49.659404 dB, and the Mean Squared Error (MSE) is 0.708044. These metrics indicate high statistical similarity and indistinguishability in visual appearance. The key space of the encryption algorithm is approximately 2850, which effectively resists brute-force attacks. The energy distribution of the encrypted audio approximates noise, with information entropy close to 8, uniform histograms, high scrambling degree, strong resistance to differential attacks, and robustness against noise and cropping attacks. Full article
(This article belongs to the Special Issue Algebraic Systems, Models and Applications)
Show Figures

Figure 1

Figure 1
<p>Tent map phase diagram (<b>left</b>) and bifurcation and Lyapunov exponent plots of Tent chaos map (<b>right</b>).</p>
Full article ">Figure 2
<p>(<b>a</b>) Trajectories; (<b>b</b>) bifurcation diagrams; (<b>c</b>) Lyapunov exponents for multidimensional sine-coupled chaotic maps.</p>
Full article ">Figure 3
<p>Attractors in the <span class="html-italic">x</span>−<span class="html-italic">z</span> plane of the unified chaotic system and maximum Lyapunov exponent plots for varying parameters. (<b>a</b>) x-z Plane of the Unified Chaotic System with Parameter α= 0; (<b>b</b>) x-z Plane of the Unified Chaotic System with Parameter α = 0.6; (<b>c</b>) x-z Plane of the Unified Chaotic System with Parameter α = 0.8; (<b>d</b>) x-z Plane of the Unified Chaotic System with Parameter α = 1; (<b>e</b>) Maximum Lyapunov Exponent of the Unified Chaotic System with Varying Parameter α.</p>
Full article ">Figure 4
<p>Illustration of the enhanced Josephus permutation algorithm.</p>
Full article ">Figure 5
<p>Illustration of the enhanced Josephus permutation algorithm.</p>
Full article ">Figure 6
<p>Schematic diagram of the bit diffusion algorithm.</p>
Full article ">Figure 7
<p>Flowchart of audio encryption and steganography algorithms.</p>
Full article ">Figure 8
<p>Simulation experiment results. (<b>a</b>) Original audio time series; (<b>b</b>) Encrypted audio time series; (<b>c</b>) Carrier image; (<b>d</b>) Audio image; (<b>e</b>) Encrypted audio image; (<b>f</b>) Decrypted audio time Series.</p>
Full article ">Figure 9
<p>Original carrier image vs. carrier image with embedded encrypted audio. (<b>a</b>) Carrier image 1; (<b>b</b>) Carrier image 2; (<b>c</b>) Carrier image 3; (<b>d</b>) Carrier image 4; (<b>e</b>) Stego image 1; (<b>f</b>) Stego image 2; (<b>g</b>) Stego image 3; (<b>h</b>) Stego image 4.</p>
Full article ">Figure 10
<p>Histogram contrast: original carrier image vs. stego carrier with embedded data. (<b>a</b>) Histogram comparison of carrier image 1 before and after embedding; (<b>b</b>) Histogram comparison of carrier image 2 before and after embedding; (<b>c</b>) Histogram comparison of carrier image 3 before and after embedding; (<b>d</b>) Histogram comparison of carrier image 4 before and after embedding.</p>
Full article ">Figure 11
<p>Visualization of decryption outcomes using incorrect key.</p>
Full article ">Figure 12
<p>Spectrograms of original audio (<b>top</b>) and encrypted audio (<b>bottom</b>).</p>
Full article ">Figure 13
<p>Scatter plot of adjacent signal amplitudes for plaintext and encrypted audio. (<b>a</b>) Correlation plot of adjacent signal amplitudes for original audio 1; (<b>b</b>) Correlation plot of adjacent signal amplitudes for original audio 2; (<b>c</b>) Correlation plot of adjacent signal amplitudes for original audio 3; (<b>d</b>) Correlation plot of adjacent signal amplitudes for encrypted audio 1; (<b>e</b>) Correlation plot of adjacent signal amplitudes for encrypted audio 2; (<b>f</b>) Correlation plot of adjacent signal amplitudes for encrypted audio 3.</p>
Full article ">Figure 14
<p>Histograms of original audio (<b>left</b>) and encrypted audio (<b>right</b>).</p>
Full article ">Figure 15
<p>Scatter plot of permuted index distribution.</p>
Full article ">Figure 16
<p>Impact of noise attacks on hidden information robustness in steganography. (<b>a</b>) Stego image with 1% salt and pepper noise; (<b>b</b>) Stego image with 5% salt and pepper noise; (<b>c</b>) Stego image with 10% salt and pepper noise; (<b>d</b>) Decrypted audio from 1% salt and pepper noise; (<b>e</b>) Decrypted audio from 5% salt and pepper noise; (<b>f</b>) Decrypted audio from 10% salt and pepper noise.</p>
Full article ">Figure 17
<p>Results of cropping attack analysis on information hiding systems. (<b>a</b>) Stego image after 1% cropping attack; (<b>b</b>) Stego image after 5% cropping attack; (<b>c</b>) Stego image after 10% cropping attack; (<b>d</b>) Decrypted audio after 1% cropping attack; (<b>e</b>) Decrypted audio after 5% cropping attack; (<b>f</b>) Decrypted audio after 10% cropping attack.</p>
Full article ">
Back to TopTop