[go: up one dir, main page]
More Web Proxy on the site http://driver.im/
You seem to have javascript disabled. Please note that many of the page functionalities won't work as expected without javascript enabled.
 
 
Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (3,155)

Search Parameters:
Keywords = fine-scale

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
13 pages, 2327 KiB  
Article
Reconstruction of High-Resolution Solar Spectral Irradiance Based on Residual Channel Attention Networks
by Peng Zhang, Jianwen Weng, Qing Kang and Jianjun Li
Remote Sens. 2024, 16(24), 4698; https://doi.org/10.3390/rs16244698 - 17 Dec 2024
Viewed by 147
Abstract
The accurate measurement of high-resolution solar spectral irradiance (SSI) and its variations at the top of the atmosphere is crucial for solar physics, the Earth’s climate, and the in-orbit calibration of optical satellites. However, existing space-based solar spectral irradiance instruments achieve high-precision SSI [...] Read more.
The accurate measurement of high-resolution solar spectral irradiance (SSI) and its variations at the top of the atmosphere is crucial for solar physics, the Earth’s climate, and the in-orbit calibration of optical satellites. However, existing space-based solar spectral irradiance instruments achieve high-precision SSI measurements at the cost of spectral resolution, which falls short of meeting the requirements for identifying fine solar spectral features. Therefore, this paper proposes a new method for reconstructing high-resolution solar spectral irradiance based on a residual channel attention network. This method considers the stability of SSI spectral features and employs residual channel attention blocks to enhance the expression ability of key features, achieving the high-accuracy reconstruction of spectral features. Additionally, to address the issue of excessively large output features from the residual channel attention blocks, a scaling coefficient adjustment network block is introduced to achieve the high-accuracy reconstruction of spectral absolute values. Finally, the proposed method is validated using the measured SSI dataset from SCIAMACHY on Envisat-1 and the simulated dataset from TSIS-1 SIM. The validation results show that, compared to existing scaling coefficient adjustment algorithms, the proposed method achieves single-spectrum super-resolution reconstruction without relying on external data, with a Mean Absolute Percentage Error (MAPE) of 0.0302% for the reconstructed spectra based on the dataset. The proposed method achieves higher-resolution reconstruction results while ensuring the accuracy of SSI. Full article
Show Figures

Figure 1

Figure 1
<p>(<b>a</b>) SCIAMACHY spectra and convolutional solar spectra; (<b>b</b>) The response function at 443 nm of TSIS-1 SIM.</p>
Full article ">Figure 2
<p>The network architecture of our work.</p>
Full article ">Figure 3
<p>Detailed description of the SFL module.</p>
Full article ">Figure 4
<p>Comparison of relative deviations of reconstruction results between our model and the model after removing the residual skip connection.</p>
Full article ">Figure 5
<p>A squeeze-and-excitation block.</p>
Full article ">Figure 6
<p>Comparison of relative deviations of reconstruction results between our model and ResNets.</p>
Full article ">Figure 7
<p>Training error.</p>
Full article ">Figure 8
<p>The 0.1 nm resolution reconstruction of solar spectral irradiance.</p>
Full article ">
14 pages, 1758 KiB  
Article
Unsupervised Temporal Adaptation in Skeleton-Based Human Action Recognition
by Haitao Tian and Pierre Payeur
Algorithms 2024, 17(12), 581; https://doi.org/10.3390/a17120581 - 16 Dec 2024
Viewed by 168
Abstract
With deep learning approaches, the fundamental assumption of data availability can be severely compromised when a model trained on a source domain is transposed to a target application domain where data are unlabeled, making supervised fine-tuning mostly impossible. To overcome this limitation, the [...] Read more.
With deep learning approaches, the fundamental assumption of data availability can be severely compromised when a model trained on a source domain is transposed to a target application domain where data are unlabeled, making supervised fine-tuning mostly impossible. To overcome this limitation, the present work introduces an unsupervised temporal-domain adaptation framework for human action recognition from skeleton-based data that combines Contrastive Prototype Learning (CPL) and Temporal Adaptation Modeling (TAM), with the aim of transferring the knowledge learned from a source domain to an unlabeled target domain. The CPL strategy, inspired by recent success in contrastive learning applied to skeleton data, learns a compact temporal representation from the source domain, from which the TAM strategy leverages the capacity for self-training to adapt the representation to a target application domain using pseudo-labels. The research demonstrates that simultaneously solving CPL and TAM effectively enables the training of a generalizable human action recognition model that is adaptive to both domains and overcomes the requirement of a large volume of labeled skeleton data in the target domain. Experiments are conducted on multiple large-scale human action recognition datasets such as NTU RGB+D, PKU MMD, and Northwestern–UCLA to comprehensively evaluate the effectiveness of the proposed method. Full article
(This article belongs to the Special Issue Machine Learning Models and Algorithms for Image Processing)
Show Figures

Figure 1

Figure 1
<p>Unsupervised domain adaptation (UDA) framework combining the proposed CPL and TAM strategies. In CPL, the training data (labeled) from the source domain supports supervised representation learning. The learned backbone is reused in TAM for refining over data samples (initially unlabeled) from the target domain. Network embedding is denoted with dotted lines if it is updated during training and with solid lines otherwise.</p>
Full article ">Figure 2
<p>Tensor flow of the training pipeline. Green symbols relate to the source domain, while red ones relate to the target domain.</p>
Full article ">Figure 3
<p>T-SNE visualization on action prototypes on the feature space of ST-GCN (upper row: “vanilla without CPL”; bottom row: “with CPL”). Action prototypes are distinguished by colors where twenty actions are randomly selected among fifty common actions for clarity. The left column represents action clusters of the source domain (NTU RGB+D), and the right column shows clusters of the target domain (PKU-MMD), respectively.</p>
Full article ">Figure 4
<p>Performance (Top1 Acc) of TAM with varying hyperparameter <math display="inline"><semantics> <mrow> <mi mathvariant="script">T</mi> </mrow> </semantics></math>.</p>
Full article ">Figure 5
<p>Performance (Top1 Acc) of TAM with varying hyperparameter <math display="inline"><semantics> <mrow> <mi>λ</mi> </mrow> </semantics></math>.</p>
Full article ">
33 pages, 8586 KiB  
Article
Decoding Land Use Conflicts: Spatiotemporal Analysis and Constraint Diagnosis from the Perspectives of Production–Living–Ecological Functions
by Yong Liu, Rui Xu, Jixin Yang, Xinpeng Xie and Xufeng Cui
Land 2024, 13(12), 2187; https://doi.org/10.3390/land13122187 - 14 Dec 2024
Viewed by 628
Abstract
Exploring the intensity and constraint factors of land use conflicts provides essential insights for efficient land use planning. Currently, China’s spatial development is gradually transitioning towards the coordinated development of production, living, and ecological functions (PLEFs). Previous studies have typically focused on land [...] Read more.
Exploring the intensity and constraint factors of land use conflicts provides essential insights for efficient land use planning. Currently, China’s spatial development is gradually transitioning towards the coordinated development of production, living, and ecological functions (PLEFs). Previous studies have typically focused on land use conflicts from a micro perspective, examining conflicts between production, living, and ecological land uses at a fine scale. There is limited research from a macro perspective that conducts a theoretical analysis based on the production, living, and ecological functions of land use conflicts themselves. In addition, existing studies primarily analyze the influencing factors of land use conflicts, with limited literature directly addressing the constraint factors of land use conflicts. This study focuses on 12 prefecture-level cities in Hubei Province, China, using data from 2010 to 2020. It categorizes land use conflicts at the macro level into production perspective, living perspective, and ecological perspective conflicts. For each of these conflict perspectives, different pressure, state, and response indicators are introduced. This approach leads to the development of a theoretical framework for analyzing land use conflicts at the macro level. On this basis, a spatiotemporal evolution analysis of land use conflicts was conducted. Additionally, using a constraint factor diagnosis model, the study analyzed the constraint factors of land use conflicts at the macro level across cities, leading to the following research conclusions: (1) the land use conflicts from the production and living perspectives in the 12 prefecture-level cities of Hubei showed an upward trend from 2010 to 2020, while the land use conflicts from the ecological perspective exhibited a downward trend; (2) during the study period, Wuhan exhibited the highest intensity of land use conflicts from both the production and living perspectives, while Ezhou experienced the highest intensity of land use conflicts from the ecological perspective for most of the study period; (3) the main constraining factors of land use conflicts from the production perspective in the 12 prefecture-level cities of Hubei are population density, average land GDP, and effective irrigation rate. The primary constraining factors of land use conflicts from the living perspective are population density, urbanization rate, and average land real estate development investment. The main constraining factors of land use conflicts from the ecological perspective are population density, average land fertilizer input, and effective irrigation rate. This study constructs a new theoretical framework for land use conflict assessment at the macro level, providing a novel approach for studying land use conflicts at the macro scale. Full article
(This article belongs to the Special Issue Land Resource Assessment)
Show Figures

Figure 1

Figure 1
<p>The map of Hubei.</p>
Full article ">Figure 2
<p>PSR theoretical analysis framework for land use conflicts from production–living–ecological perspectives.</p>
Full article ">Figure 3
<p>The spatiotemporal evolution of three perspectives of land use conflicts in Hubei from 2010 to 2020.</p>
Full article ">Figure 4
<p>Level of land use conflicts from a production perspective.</p>
Full article ">Figure 5
<p>Level of land use conflicts from a living perspective.</p>
Full article ">Figure 6
<p>Level of land use conflicts from an ecological perspective.</p>
Full article ">Figure 7
<p>Diagnosis of constraint degrees on land use conflicts in Hubei from 2010 to 2020.</p>
Full article ">
17 pages, 5303 KiB  
Article
Carbon Soil Mapping in a Sustainable-Managed Farm in Northeast Italy: Geochemical and Geophysical Applications
by Gian Marco Salani, Enzo Rizzo, Valentina Brombin, Giacomo Fornasari, Aaron Sobbe and Gianluca Bianchini
Environments 2024, 11(12), 289; https://doi.org/10.3390/environments11120289 - 14 Dec 2024
Viewed by 394
Abstract
Recently, there has been increasing interest in organic carbon (OC) certification of soil as an incentive for farmers to adopt sustainable agricultural practices. In this context, this pilot project combines geochemical and geophysical methods to map the distribution of OC contents in agricultural [...] Read more.
Recently, there has been increasing interest in organic carbon (OC) certification of soil as an incentive for farmers to adopt sustainable agricultural practices. In this context, this pilot project combines geochemical and geophysical methods to map the distribution of OC contents in agricultural fields, allowing us to detect variations in time and space. Here we demonstrated a relationship between soil OC contents estimated in the laboratory and the apparent electrical conductivity (ECa) measured in the field. Specifically, geochemical elemental analyses were used to evaluate the OC content and relative isotopic signature in collected soil samples from a hazelnut orchard in the Emilia–Romagna region of Northeastern Italy, while the geophysical Electromagnetic Induction (EMI) method enabled the in situ mapping of the ECa distribution in the same soil field. According to the results, geochemical and geophysical data were found to be reciprocally related, as both the organic matter and soil moisture were mainly incorporated into the fine sediments (i.e., clay) of the soil. Therefore, such a relation was used to create a map of the OC content distribution in the investigated field, which could be used to monitor the soil C sequestration on small-scale farmland and eventually develop precision agricultural services. In the future, this method could be used by farmers and regional and/or national policymakers to periodically certify the farm’s soil conditions and verify the effectiveness of carbon sequestration. These measures would enable farmers to pursue Common Agricultural Policy (CAP) incentives for the reduction of CO2 emissions. Full article
Show Figures

Figure 1

Figure 1
<p>(<b>a</b>) Location of the sampling area (MB), in the Northeast sector of the municipality of Ferrara in the Emilia–Romagna region (Northeastern Italy); (<b>b</b>) the hazel orchard–grassland field before the geochemical and geophysical investigation of 19 October 2021; (<b>c</b>) soil sampling locations represented by light blue dots; (<b>d</b>) at each location, a sample was collected and mixed with five aliquots of soil per square probed at a depth of 0–30 cm; (<b>e</b>) geophysical measurements were indicated with red dots and georeferenced with an internal GPR; and (<b>f</b>) a Profiler EMP-400 (GSSI) was used to acquire the Hp and Hs electromagnetic fields at different positions.</p>
Full article ">Figure 2
<p>Elemental and isotopic composition of the total carbon (TC), organic carbon (OC), and inorganic carbon (IC) fractions of the soil samples.</p>
Full article ">Figure 3
<p>Boxplots of the (<b>a</b>) LOI 105 °C, (<b>b</b>) LOI 550 °C, (<b>c</b>) LOI 1000 °C, (<b>d</b>) TC, (<b>e</b>) OC, (<b>f</b>) IC, (<b>g</b>) δ<sup>13</sup>C<sub>TC</sub>, and (<b>h</b>) δ¹³C<sub>OC</sub> of the samples divided into three classes based on their aspect in the field and OC/IC ratio (see the text for details). In each box plot, the black line represents the median. Letters below the box plots represent the results of the Tukey post hoc test. Different letters denote significant differences between classes. The one-way ANOVA results are also reported (** <span class="html-italic">p</span> &lt; 0.001; *** <span class="html-italic">p</span> &lt; 0.0001).</p>
Full article ">Figure 4
<p>Spatial variability and distribution of the ECa values obtained from the EMI acquisition field survey using three different frequencies: (<b>a</b>) 16, (<b>b</b>) 14, and (<b>c</b>) 10 kHz.</p>
Full article ">Figure 5
<p>The elemental TC contents and δ¹³C<sub>TC</sub> of MB samples and average elemental TC contents and δ¹³C<sub>TC</sub> recognized as deposits from the paleochannel and levee of the easternmost Padanian plain soils, as studied by Natali et al. [<a href="#B36-environments-11-00289" class="html-bibr">36</a>] and Salani et al. [<a href="#B37-environments-11-00289" class="html-bibr">37</a>].</p>
Full article ">Figure 6
<p>OC/IC (in logarithmic scale) versus (<b>a</b>) δ<sup>13</sup>C<sub>TC</sub> shows a strong negative correlation; the insets reproduce the relationships between OC/IC, (<b>b</b>) δ<sup>13</sup>C<sub>IC</sub>, and (<b>c</b>) δ<sup>13</sup>C<sub>OC</sub>.</p>
Full article ">Figure 7
<p>Principal Component Analysis (PCA) for δ<sup>13</sup>CTC, OC, IC, TC, and ECa (measured at 10 kHz), clustered in Class I (green dots and dash-dotted line ellipse), Class II (yellow triangles and solid line ellipse), and Class III (red squares and dashed line ellipse).</p>
Full article ">Figure 8
<p>Linear regression graphics used to observe the relationships between the ECa measured at 10 kHz and (<b>a</b>) OC, (<b>b</b>) OC/IC, and (<b>c</b>) δ<sup>13</sup>C<sub>TC</sub>. The data are represented as green dots, yellow triangles, and red squares, for Class I, Class II, and Class III, respectively. The regression line (in black) and relative equation, R<sup>2</sup> value, and 95% confidence intervals (the red curves) are provided for each plot.</p>
Full article ">Figure 9
<p>Predictive maps realized using ordinary kriging for (<b>a</b>) the OC values, (<b>b</b>) the ECa values measured at 10 kHz, and cokriging to predict (<b>c</b>) a new OC surface, with the OC values and the ECa values at 10 kHz as a covariate variable. The legend values for each map represent a quantile classification.</p>
Full article ">
14 pages, 2919 KiB  
Article
Evaluation of Potential Developmental Precursors to Executive Function in Young Children with Motor Delays: Exploratory Study
by Andrea B. Cunha, Iryna Babik, Regina T. Harbourne, Stacey C. Dusing, Lin-Ya Hsu, Natalie A. Koziol, Sarah Westcott-McCoy, Sandra L. Willett, James A. Bovaird and Michele A. Lobo
Behav. Sci. 2024, 14(12), 1201; https://doi.org/10.3390/bs14121201 - 14 Dec 2024
Viewed by 424
Abstract
This study aimed to explore whether early developmental abilities are related to future executive function (EF) in children with motor delays. Fourteen children with motor delays (Mean age = 10.76, SD = 2.55) were included from a larger study. Object interaction and [...] Read more.
This study aimed to explore whether early developmental abilities are related to future executive function (EF) in children with motor delays. Fourteen children with motor delays (Mean age = 10.76, SD = 2.55) were included from a larger study. Object interaction and developmental outcomes (Bayley-III) were evaluated at baseline and 3, 6, and 12 months post-baseline. Bayley-III and EF assessments (Minnesota Executive Function Scale) were conducted at 36 months post-baseline. Children with high EF demonstrated advanced early bimanual, visual–bimanual, receptive language, expressive language, and fine motor skills compared to children with low EF. Significant positive correlations between later Bayley-III and EF scores were found for cognitive, expressive language, and fine motor scores. These preliminary results suggest that early developmental skills support the emergence of EF in children with motor delays. Full article
(This article belongs to the Special Issue The Role of Early Sensorimotor Experiences in Cognitive Development)
Show Figures

Figure 1

Figure 1
<p>Experimental setup for the Minnesota Executive Function Scale (MEFS) at 36 months post-baseline.</p>
Full article ">Figure 2
<p>Estimated trajectories for the early object interaction skills found to differ between children with low vs. high EF at 36 months post-baseline.</p>
Full article ">Figure 3
<p>Estimated trajectories for the early Bayley-III areas of development found to differ between children with low vs. high executive function at 36 months post-baseline.</p>
Full article ">
19 pages, 5351 KiB  
Article
GMTP: Enhanced Travel Time Prediction with Graph Attention Network and BERT Integration
by Ting Liu and Yuan Liu
AI 2024, 5(4), 2926-2944; https://doi.org/10.3390/ai5040141 - 13 Dec 2024
Viewed by 477
Abstract
(1) Background: Existing Vehicle travel time prediction applications face challenges in modeling complex road network and handling irregular spatiotemporal traffic state propagation. (2) Methods: To address these issues, we propose a Graph Attention-based Multi-Spatiotemporal Features for Travel Time Prediction (GMTP) model, which integrates [...] Read more.
(1) Background: Existing Vehicle travel time prediction applications face challenges in modeling complex road network and handling irregular spatiotemporal traffic state propagation. (2) Methods: To address these issues, we propose a Graph Attention-based Multi-Spatiotemporal Features for Travel Time Prediction (GMTP) model, which integrates an enhanced graph attention network (GATv2) and Bidirectional Encoder Representations from Transformers (BERT) to analyze dynamic correlations across spatial and temporal dimensions. The pre-training process consists of two blocks: the Road Segment Interaction Pattern to Enhance GATv2, which generates road segment representation vectors, and a traffic congestion-aware trajectory encoder by incorporating a shared attention mechanism for high computational efficiency. Additionally, two self-supervised tasks are designed for improved model accuracy and robustness. (3) Results: The fine-tuned model had comparatively optimal performance metrics with significant reductions in Mean Absolute Error (MAE), Mean Absolute Percentage Error (MAPE), and Root Mean Squared Error (RMSE). (4) Conclusions: Ultimately, the integration of this model into travel time prediction, based on two large-scale real-world trajectory datasets, demonstrates enhanced performance and computational efficiency. Full article
Show Figures

Figure 1

Figure 1
<p>The architecture of GMTP. This figure illustrates the overall design: (<b>a</b>) The graph attention network V2 (GATv2) module captures spatial relationships by modeling the road network structure and incorporating Interaction Transfer Frequency for spatial information extraction. (<b>b</b>) The traffic congestion-aware trajectory encoder, equipped with adaptive shared attention, encodes road segment representations into trajectory vectors. Workday and peak hour information are integrated into the trajectory embeddings, and the attention mechanism is dynamically adjusted using transfer time and hybrid matrices for better spatiotemporal fusion. (<b>c</b>) The adaptive masked trajectory reconstruction task applies a random masking strategy to improve trajectory recovery, enhancing the accuracy of the trajectory representations. (<b>d</b>) The contrastive learning module strengthens feature extraction through Time-Frequency Perturbation and road segment masking, enhancing robustness and generalization.</p>
Full article ">Figure 2
<p>The calculation process of multi-head shared attention mechanism (MSA). For <math display="inline"><semantics> <mrow> <msub> <mi>N</mi> <mi>h</mi> </msub> <mo>=</mo> <mn>3</mn> </mrow> </semantics></math>, the attention scores for input <math display="inline"><semantics> <msub> <mi>X</mi> <mi>n</mi> </msub> </semantics></math> are computed. The three independent heads are represented by different colors, and the block structure of the mixing matrix <math display="inline"><semantics> <msub> <mi>H</mi> <mrow> <mi>m</mi> <mi>a</mi> <mi>t</mi> <mi>r</mi> <mi>i</mi> <mi>x</mi> </mrow> </msub> </semantics></math> ensures that the dot products for each head are performed on non-overlapping dimensions. (<b>a</b>) represents a more generalized hybrid matrix <math display="inline"><semantics> <msub> <mi>H</mi> <mrow> <mi>m</mi> <mi>a</mi> <mi>t</mi> <mi>r</mi> <mi>i</mi> <mi>x</mi> </mrow> </msub> </semantics></math> as opposed to simple head concatenation, and the blocks-of-1 represents a ones matrix. (<b>b</b>) involves sharing head projections by learning all entries of the matrix. (<b>c</b>) reduces the number of projections from <math display="inline"><semantics> <msub> <mi>D</mi> <mi>k</mi> </msub> </semantics></math> to <math display="inline"><semantics> <msub> <mover accent="true"> <mi>D</mi> <mo stretchy="false">˜</mo> </mover> <mi>k</mi> </msub> </semantics></math>, allowing heads to share redundant projections, thus improving efficiency.</p>
Full article ">Figure 3
<p>Effectiveness of various algorithm: Comparison of No-GATv2, Node2vec, GAT and GMTP, highlighting differences in embedding initialization and performance with respect to road features. Lower values represent better performance.</p>
Full article ">Figure 4
<p>The impact of adaptive HSSTA. “No-TimeEmb” indicates the absence of time characteristics, “No-TransferMatrix” means transfer times are not considered, and “No-HybridMatrix” denotes the exclusion of the mixing matrix in attention weight calculation. Ignoring these components leads to a significant drop in model performance.</p>
Full article ">Figure 5
<p>Impact of self-supervised tasks. In the “No-Mask” and “No-Contra” settings, the model’s MAPE, Macro-F1, and MR values all increased, highlighting the significance of both trajectory masking and contrastive learning in self-supervised training.</p>
Full article ">Figure 6
<p>Impact of data augmentation strategies. Darker colors in the heatmap represent lower MAPE values, indicating better performance. The combination of Perturb and Mask yields the best results.</p>
Full article ">Figure 7
<p>The impact of hyperparameters. The vertical axis represents Macro-F1. (<b>a</b>) Encoding layer depth (<span class="html-italic">L</span>): Balances learning capacity and overfitting. (<b>b</b>) Embedding dimension (<span class="html-italic">d</span>): Affects representation quality. (<b>c</b>) Batch size (<span class="html-italic">N</span>): Impacts gradient estimation and contrastive learning.</p>
Full article ">Figure 8
<p>Trajectory encoding time comparison. The horizontal axis represents dataset size (in K), and the vertical axis represents encoding cost (in seconds).</p>
Full article ">
16 pages, 7607 KiB  
Article
Airwave Noise Identification from Seismic Data Using YOLOv5
by Zhenghong Liang, Lu Gan, Zhifeng Zhang, Xiuju Huang, Fengli Shen, Guo Chen and Rongjiang Tang
Appl. Sci. 2024, 14(24), 11636; https://doi.org/10.3390/app142411636 - 12 Dec 2024
Viewed by 411
Abstract
Airwave interference presents a major source of noise in seismic exploration, posing significant challenges to the quality control of raw seismic data. With the increasing data volume in 3D seismic exploration, manual identification methods fall short of meeting the demands of high-density 3D [...] Read more.
Airwave interference presents a major source of noise in seismic exploration, posing significant challenges to the quality control of raw seismic data. With the increasing data volume in 3D seismic exploration, manual identification methods fall short of meeting the demands of high-density 3D seismic surveys. This study employs the YOLOv5 model, a widely used tool in object detection, to achieve rapid identification of airwave noise in seismic profiles. Initially, the model was pre-trained on the COCO dataset—a large-scale dataset designed for object detection—and subsequently fine-tuned using a training set specifically labeled for airwave noise data. The fine-tuned model achieved an accuracy and recall rate of approximately 85% on the test dataset, successfully identifying not only the presence of noise but also its location, confidence levels, and range. To evaluate the model’s effectiveness, we applied the YOLOv5 model trained on 2D data to seismic records from two regions: 2D seismic data from Ningqiang, Shanxi, and 3D seismic data from Xiushui, Sichuan. The overall prediction accuracy in both regions exceeded 90%, with the accuracy and recall rates for airwave noise surpassing 83% and 90%, respectively. The evaluation time for single-shot 3D seismic data (over 8000 traces) was less than 2 s, highlighting the model’s exceptional transferability, generalization ability, and efficiency. These results demonstrate that the YOLOv5 model is highly effective for detecting airwave noise in raw seismic data across different regions, marking the first successful attempt at computer recognition of airwaves in seismic exploration. Full article
Show Figures

Figure 1

Figure 1
<p>Typical characteristics of airwave noise (indicated in green). (<b>a</b>) No airwave noise; (<b>b</b>) Weak airwave noise; (<b>c</b>,<b>d</b>) represent airwave noise in three-dimensional exploration, where (<b>c</b>) is closer to the seismic source than (<b>d</b>); (<b>e</b>,<b>f</b>) depict typical airwave noise in two-dimensional seismic exploration.</p>
Full article ">Figure 2
<p>The YOLOv5m network architecture (modified from Zhang, 2022 [<a href="#B44-applsci-14-11636" class="html-bibr">44</a>]). Conv refers to two-dimensional convolution operations, while BN (Batch Normalization) accelerates network convergence during training and reduces sensitivity to parameter initialization. Concat denotes the concatenation operation used to merge features from different sources. SiLU is the activation function, defined as <math display="inline"><semantics> <mrow> <mi>S</mi> <mi>i</mi> <mi>L</mi> <mi>U</mi> <mfenced separators="|"> <mrow> <mi>x</mi> </mrow> </mfenced> <mo>=</mo> <mstyle scriptlevel="0" displaystyle="true"> <mfrac> <mrow> <mi>x</mi> </mrow> <mrow> <mn>1</mn> <mo>+</mo> <msup> <mrow> <mi>e</mi> </mrow> <mrow> <mi>x</mi> </mrow> </msup> </mrow> </mfrac> </mstyle> </mrow> </semantics></math>, which is a smooth, nonlinear activation function. This means that it does not introduce discontinuity at zero like ReLU (Rectified Linear Unit), thereby avoiding gradient instability issues. MaxPool refers to the max pooling layer used for downsampling, which reduces the size of the feature map by selecting the maximum value within local regions of the feature map, while preserving the most significant features.</p>
Full article ">Figure 3
<p>The diagram illustrates the grid division (black grid lines), marked boxes or label (yellow bounding boxes), prior anchor boxes (red bounding boxes), and the non-maximum suppression process. Since only one class is predicted, each predicted unit greater than the IoU threshold (indicated in purple) corresponds to three anchor boxes, which are merged into one by the non-maximum suppression algorithm. (Modified form <a href="https://cloud.tencent.com/developer/article/1118040" target="_blank">https://cloud.tencent.com/developer/article/1118040</a> (accessed on 1 January 2020)).</p>
Full article ">Figure 4
<p>YOLOv5 training and prediction workflow for airwave detection in seismic profiles. The pentagram represents the source location.</p>
Full article ">Figure 5
<p>The evolution curves of loss and precision-recall during the secondary training of YOLOv5m.</p>
Full article ">Figure 6
<p>Statistical distribution of samples in the test dataset: (<b>a</b>) A distribution map of predicted locations of airwave noise, where x and y represent the two spatial dimensions of the image, with the top-left corner of the image set as the coordinate origin. (<b>b</b>) Statistical distribution of widths and heights of predicted airwave noise boxes. (<b>c</b>) The relationship between recall and accuracy at different confidence levels.</p>
Full article ">Figure 7
<p>Randomly selected predictions from the test dataset, where the red boxes indicate airwave noise detected by YOLOv5m, and the numbers in the top-left corner represent the confidence levels.</p>
Full article ">Figure 8
<p>Instances labeled as no interference but predicted as interference; the boxes represent the prediction results, and the numbers indicate the confidence levels.</p>
Full article ">Figure 9
<p>Randomly selected airwave noise prediction results from the Xiushui three-dimensional seismic data. The green boxes indicate YOLOv5m predictions of airwave noise, with the first and second rows representing seismic records corresponding to different seismic sources.</p>
Full article ">Figure 10
<p>The 3D seismic records in the Xiushui area exhibit unilateral airwave noise, primarily caused by the terrain.</p>
Full article ">Figure 11
<p>Airwave predictions of different colormap in test dataset, and the numbers in the top-left corner represent the confidence levels. Up: TWILIGHT. Down: CIVIDIS.</p>
Full article ">
23 pages, 1149 KiB  
Article
MGAFN-ISA: Multi-Granularity Attention Fusion Network for Implicit Sentiment Analysis
by Yifan Huo, Ming Liu, Junhong Zheng and Lili He
Electronics 2024, 13(24), 4905; https://doi.org/10.3390/electronics13244905 (registering DOI) - 12 Dec 2024
Viewed by 314
Abstract
Although significant progress has been made in sentiment analysis tasks based on image–text data, existing methods still have limitations in capturing cross-modal correlations and detailed information. To address these issues, we propose a Multi-Granularity Attention Fusion Network for Implicit Sentiment Analysis (MGAFN-ISA). MGAFN-ISA [...] Read more.
Although significant progress has been made in sentiment analysis tasks based on image–text data, existing methods still have limitations in capturing cross-modal correlations and detailed information. To address these issues, we propose a Multi-Granularity Attention Fusion Network for Implicit Sentiment Analysis (MGAFN-ISA). MGAFN-ISA that leverages neural networks and attention mechanisms to effectively reduce noise interference between different modalities and captures distinct, fine-grained visual and textual features. The model includes two key feature extraction modules: a multi-scale attention fusion-based visual feature extractor and a hierarchical attention mechanism-based textual feature extractor, each designed to extract detailed and discriminative visual and textual representations. Additionally, we introduce an image translator engine to produce accurate and detailed image descriptions, further narrowing the semantic gap between the visual and textual modalities. A bidirectional cross-attention mechanism is also incorporated to utilize correlations between fine-grained local regions across modalities, extracting complementary information from heterogeneous visual and textual data. Finally, we designed an adaptive multimodal classification module that dynamically adjusts the contribution of each modality through an adaptive gating mechanism. Extensive experimental results demonstrate that MGAFN-ISA achieves a significant performance improvement over nine state-of-the-art methods across multiple public datasets, validating the effectiveness and advancement of our proposed approach. Full article
(This article belongs to the Section Artificial Intelligence)
Show Figures

Figure 1

Figure 1
<p>Examples of multimodal sentiment analysis.</p>
Full article ">Figure 2
<p>The overall structure of the proposed MGAFN-ISA method. The visual–text feature extraction utilizes ResNet-101 to extract multi-scale image features, combined with ViT-GPT-2 to generate image descriptions, while BERT is used to extract textual features. The multimodal fine-grained correlation fusion module adopts a bidirectional cross-attention mechanism to capture fine-grained associations between visual and textual modalities. Furthermore, multi-head attention and layer stacking are used to enhance feature interactions. The classification module employs an adaptive gating mechanism to dynamically fuse visual and textual features, achieving accurate classification of multimodal implicit sentiment by integrating the improved focal loss function.</p>
Full article ">Figure 3
<p>The performance of MGAFN-ISA under different proportions of training data on each dataset.</p>
Full article ">
15 pages, 605 KiB  
Article
Towards the Particle Spectrum, Tickled by a Distant Massive Object
by Mehdi Dehghani, Salman A. Nejad and Maryam Mardaani
Astronomy 2024, 3(4), 304-318; https://doi.org/10.3390/astronomy3040019 (registering DOI) - 12 Dec 2024
Viewed by 397
Abstract
To investigate the gravitational effects of massive objects on a typical observer, we studied the dynamics of a test particle following BMS3 geodesics. We constructed the BMS3 framework using the canonical phase space formalism and the corresponding Hamiltonian. We focused on [...] Read more.
To investigate the gravitational effects of massive objects on a typical observer, we studied the dynamics of a test particle following BMS3 geodesics. We constructed the BMS3 framework using the canonical phase space formalism and the corresponding Hamiltonian. We focused on analyzing these effects at fine scales of spacetime, which led us to quantization of the phase space. By deriving and studying the solutions of the quantum equations of motion for the test particle, we obtained its energy spectrum and explored the behavior of its wave function. These findings offer a fresh perspective on gravitational interactions in the context of quantum mechanics, providing an alternative approach to traditional quantum field theory analyses. Full article
Show Figures

Figure 1

Figure 1
<p>The coordinates <span class="html-italic">u</span> and <span class="html-italic">r</span> in spacetime are shown as in [<a href="#B1-astronomy-03-00019" class="html-bibr">1</a>], with the time coordinate <math display="inline"><semantics> <msup> <mi>x</mi> <mn>0</mn> </msup> </semantics></math> pointing upward. The wavy red line represents a massless particle emitted at <math display="inline"><semantics> <mrow> <mi>r</mi> <mo>=</mo> <mn>0</mn> </mrow> </semantics></math> and traveling outward along a light cone generator (<math display="inline"><semantics> <mrow> <mi>u</mi> <mo>=</mo> </mrow> </semantics></math> const) to a non-zero <span class="html-italic">r</span>. The circle of radius <span class="html-italic">r</span> represents a sphere in four-dimensional spacetime.</p>
Full article ">
17 pages, 2272 KiB  
Article
Convolutional Neural Network–Vision Transformer Architecture with Gated Control Mechanism and Multi-Scale Fusion for Enhanced Pulmonary Disease Classification
by Okpala Chibuike and Xiaopeng Yang
Diagnostics 2024, 14(24), 2790; https://doi.org/10.3390/diagnostics14242790 - 12 Dec 2024
Viewed by 422
Abstract
Background/Objectives: Vision Transformers (ViTs) and convolutional neural networks (CNNs) have demonstrated remarkable performances in image classification, especially in the domain of medical imaging analysis. However, ViTs struggle to capture high-frequency components of images, which are critical in identifying fine-grained patterns, while CNNs have [...] Read more.
Background/Objectives: Vision Transformers (ViTs) and convolutional neural networks (CNNs) have demonstrated remarkable performances in image classification, especially in the domain of medical imaging analysis. However, ViTs struggle to capture high-frequency components of images, which are critical in identifying fine-grained patterns, while CNNs have difficulties in capturing long-range dependencies due to their local receptive fields, which makes it difficult to fully capture the spatial relationship across lung regions. Methods: In this paper, we proposed a hybrid architecture that integrates ViTs and CNNs within a modular component block(s) to leverage both local feature extraction and global context capture. In each component block, the CNN is used to extract the local features, which are then passed through the ViT to capture the global dependencies. We implemented a gated attention mechanism that combines the channel-, spatial-, and element-wise attention to selectively emphasize the important features, thereby enhancing overall feature representation. Furthermore, we incorporated a multi-scale fusion module (MSFM) in the proposed framework to fuse the features at different scales for more comprehensive feature representation. Results: Our proposed model achieved an accuracy of 99.50% in the classification of four pulmonary conditions. Conclusions: Through extensive experiments and ablation studies, we demonstrated the effectiveness of our approach in improving the medical image classification performance, while achieving good calibration results. This hybrid approach offers a promising framework for reliable and accurate disease diagnosis in medical imaging. Full article
(This article belongs to the Section Machine Learning and Artificial Intelligence in Diagnostics)
Show Figures

Figure 1

Figure 1
<p>The proposed hybrid architecture.</p>
Full article ">Figure 2
<p>Gated mechanism with attention.</p>
Full article ">Figure 3
<p>Inception-styled multi-scale fusion module proposed in this study.</p>
Full article ">Figure 4
<p>A confusion matrix for the proposed model.</p>
Full article ">Figure 5
<p>Impact of different augmentation methods on original images.</p>
Full article ">Figure 6
<p>Impact of gated mechanism and multi-scale fusion using LIME explainability analysis.</p>
Full article ">
21 pages, 3698 KiB  
Article
Child-Centric Robot Dialogue Systems: Fine-Tuning Large Language Models for Better Utterance Understanding and Interaction
by Da-Young Kim, Hyo Jeong Lym, Hanna Lee, Ye Jun Lee, Juhyun Kim, Min-Gyu Kim and Yunju Baek
Sensors 2024, 24(24), 7939; https://doi.org/10.3390/s24247939 - 12 Dec 2024
Viewed by 317
Abstract
Dialogue systems must understand children’s utterance intentions by considering their unique linguistic characteristics, such as syntactic incompleteness, pronunciation inaccuracies, and creative expressions, to enable natural conversational engagement in child–robot interactions. Even state-of-the-art large language models (LLMs) for language understanding and contextual awareness cannot [...] Read more.
Dialogue systems must understand children’s utterance intentions by considering their unique linguistic characteristics, such as syntactic incompleteness, pronunciation inaccuracies, and creative expressions, to enable natural conversational engagement in child–robot interactions. Even state-of-the-art large language models (LLMs) for language understanding and contextual awareness cannot comprehend children’s intent as accurately as humans because of their distinctive features. An LLM-based dialogue system should acquire the manner by which humans understand children’s speech to enhance its intention reasoning performance in verbal interactions with children. To this end, we propose a fine-tuning methodology that utilizes the LLM–human judgment discrepancy and interactive response data. The former data represent cases in which the LLM and human judgments of the contextual appropriateness of a child’s answer to a robot’s question diverge. The latter data involve robot responses suitable for children’s utterance intentions, generated by the LLM. We developed a fine-tuned dialogue system using these datasets to achieve human-like interpretations of children’s utterances and to respond adaptively. Our system was evaluated through human assessment using the Robotic Social Attributes Scale (RoSAS) and Sensibleness and Specificity Average (SSA) metrics. Consequently, it supports the effective interpretation of children’s utterance intentions and enables natural verbal interactions, even in cases with syntactic incompleteness and mispronunciations. Full article
(This article belongs to the Special Issue Challenges in Human-Robot Interactions for Social Robotics)
Show Figures

Figure 1

Figure 1
<p>Overview of AI home robot service and interaction design from our previous study.</p>
Full article ">Figure 2
<p>Results of Godspeed questionnaire.</p>
Full article ">Figure 3
<p>Process of fine-tuning dataset construction (Q: robot’s question; A: child’s answer; R: interactive response).</p>
Full article ">Figure 4
<p>Example of prompts and response judgment data provided to LLM and humans.</p>
Full article ">Figure 5
<p>Structure of fine-tuning dataset with message roles.</p>
Full article ">Figure 6
<p>Comparison of dialogue systems for child’s utterance with lack of specificity.</p>
Full article ">Figure 7
<p>Comparison of dialogue systems for child’s utterance with subtle affirmative expression.</p>
Full article ">Figure 8
<p>Comparison of dialogue systems for child’s utterance with mispronunciation or misrecognition.</p>
Full article ">Figure 9
<p>Evaluation results for dialogue system.</p>
Full article ">Figure A1
<p>Dialogue system prompts.</p>
Full article ">
22 pages, 15973 KiB  
Article
Three-Dimensional Bone-Image Synthesis with Generative Adversarial Networks
by Christoph Angermann, Johannes Bereiter-Payr, Kerstin Stock, Gerald Degenhart and Markus Haltmeier
J. Imaging 2024, 10(12), 318; https://doi.org/10.3390/jimaging10120318 - 11 Dec 2024
Viewed by 384
Abstract
Medical image processing has been highlighted as an area where deep-learning-based models have the greatest potential. However, in the medical field, in particular, problems of data availability and privacy are hampering research progress and, thus, rapid implementation in clinical routine. The generation of [...] Read more.
Medical image processing has been highlighted as an area where deep-learning-based models have the greatest potential. However, in the medical field, in particular, problems of data availability and privacy are hampering research progress and, thus, rapid implementation in clinical routine. The generation of synthetic data not only ensures privacy but also allows the drawing of new patients with specific characteristics, enabling the development of data-driven models on a much larger scale. This work demonstrates that three-dimensional generative adversarial networks (GANs) can be efficiently trained to generate high-resolution medical volumes with finely detailed voxel-based architectures. In addition, GAN inversion is successfully implemented for the three-dimensional setting and used for extensive research on model interpretability and applications such as image morphing, attribute editing, and style mixing. The results are comprehensively validated on a database of three-dimensional HR-pQCT instances representing the bone micro-architecture of the distal radius. Full article
(This article belongs to the Special Issue Advances in Medical Imaging and Machine Learning)
Show Figures

Figure 1

Figure 1
<p>HR-pQCT bone samples of real patients with isotropic voxel size 60.7 <math display="inline"><semantics> <mi mathvariant="sans-serif">μ</mi> </semantics></math><math display="inline"><semantics> <mi mathvariant="normal">m</mi> </semantics></math>. Volumes are cropped to a region of interest (ROI) with varying numbers of voxels for each scan.</p>
Full article ">Figure 2
<p>Preprocessing. From left to right: The sample is cropped or padded to a constant size of 168 × 576 × 448 voxels. The mirrored volume is used as padding. The samples are considered regarding the discrete cosine basis. Clipping the basis coefficients to range <math display="inline"><semantics> <mrow> <mo stretchy="false">[</mo> <mo>−</mo> <mn>0.001</mn> <mo>,</mo> <mn>0.001</mn> <mo stretchy="false">]</mo> </mrow> </semantics></math> yields the noise volume. The padded regions are replaced by the corresponding noise volume.</p>
Full article ">Figure 3
<p>Exemplary visualization of the progressive growing strategy for the synthesis of 3D bone HR-pQCT data.</p>
Full article ">Figure 4
<p>Ten HR-pQCT volumes sampled from the proposed 3D-ProGAN (<b>first row</b>) and 3D-StyleGAN (<b>second row</b>). Synthesized volumes have spatial size of 32 × 288 × 224.</p>
Full article ">Figure 5
<p><b>First row</b>: samples with weak trabecular bone mineralization (Tb.BMD). <b>Second row</b>: samples with weak cortical bone mineralization (Ct.BMD). From left to right: <math display="inline"><semantics> <mrow> <msub> <mi>x</mi> <mn>1</mn> </msub> <mo>,</mo> <mspace width="4pt"/> <msubsup> <mi>x</mi> <mrow> <mn>1</mn> <mo>,</mo> <mn>2</mn> </mrow> <mrow> <mn>0.25</mn> </mrow> </msubsup> <mo>,</mo> <mspace width="4pt"/> <msubsup> <mi>x</mi> <mrow> <mn>1</mn> <mo>,</mo> <mn>2</mn> </mrow> <mrow> <mn>0.5</mn> </mrow> </msubsup> <mo>,</mo> <mspace width="4pt"/> <msubsup> <mi>x</mi> <mrow> <mn>1</mn> <mo>,</mo> <mn>2</mn> </mrow> <mrow> <mn>0.75</mn> </mrow> </msubsup> <mo>,</mo> <mspace width="4pt"/> <msub> <mi>x</mi> <mn>2</mn> </msub> </mrow> </semantics></math>. The areas marked in red allow the reader to better recognize the low Tb.BMD and the weak Ct.BMD of the examined radii, respectively.</p>
Full article ">Figure 6
<p>An illustration of the style combination based on the 3D-StyleGAN approach. For both examples, the first row denotes the source image (real patient data). The second row contains the target image at the left most position and style mix results where the style of the source is fed to the generator in the first three convolutional layers (<math display="inline"><semantics> <msubsup> <mi>x</mi> <mrow> <mi>s</mi> <mo>,</mo> <mi>t</mi> </mrow> <mn>3</mn> </msubsup> </semantics></math>), in the first seven layers (<math display="inline"><semantics> <msubsup> <mi>x</mi> <mrow> <mi>s</mi> <mo>,</mo> <mi>t</mi> </mrow> <mn>7</mn> </msubsup> </semantics></math>) and in the first twelve layers (<math display="inline"><semantics> <msubsup> <mi>x</mi> <mrow> <mi>s</mi> <mo>,</mo> <mi>t</mi> </mrow> <mn>12</mn> </msubsup> </semantics></math>), from left to right.</p>
Full article ">Figure 7
<p>3D-ProGAN results for attribute editing. For each volumetric sample, the center axial slice is visualized. Left: Existing patient <span class="html-italic">x</span>. Middle: Generated samples <math display="inline"><semantics> <mrow> <msub> <mi>G</mi> <mn>1</mn> </msub> <mrow> <mo stretchy="false">(</mo> <msub> <mi>z</mi> <mi>opt</mi> </msub> <mrow> <mo stretchy="false">(</mo> <mi>x</mi> <mo stretchy="false">)</mo> </mrow> <mo>+</mo> <mi>α</mi> <msub> <mi>n</mi> <mi>k</mi> </msub> <mo stretchy="false">)</mo> </mrow> <mo>,</mo> <mspace width="4pt"/> <mi>k</mi> <mo>=</mo> <mn>1</mn> <mo>,</mo> <mn>2</mn> <mo>,</mo> <mn>3</mn> <mo>,</mo> <mn>4</mn> </mrow> </semantics></math>. Right: difference <math display="inline"><semantics> <mrow> <msub> <mi>G</mi> <mn>1</mn> </msub> <mrow> <mo stretchy="false">(</mo> <msub> <mi>z</mi> <mi>opt</mi> </msub> <mrow> <mo stretchy="false">(</mo> <mi>x</mi> <mo stretchy="false">)</mo> </mrow> <mo>+</mo> <mi>α</mi> <msub> <mi>n</mi> <mi>k</mi> </msub> <mo stretchy="false">)</mo> </mrow> <mo>−</mo> <mi>x</mi> </mrow> </semantics></math>, where red and blue voxels denote positive and negative residuals, respectively.</p>
Full article ">Figure 8
<p>Comparison between computer-based realism scores and the subjective rating by Expert 1 (<b>first row</b>) and Expert 2 (<b>second row</b>) on HR-pQCT images. The horizontal axes denote the expert rating 1–5, while the vertical axes show the calculated realism scores. From left to right: <math display="inline"><semantics> <msub> <mi mathvariant="bold">r</mi> <mi>inc</mi> </msub> </semantics></math>, <math display="inline"><semantics> <msub> <mi mathvariant="bold">r</mi> <mi>res</mi> </msub> </semantics></math>, <math display="inline"><semantics> <msub> <mi mathvariant="bold">r</mi> <mi>vgs</mi> </msub> </semantics></math>.</p>
Full article ">Figure A1
<p>Synthetic HR-pQCT volumes sampled from the proposed 3D-ProGAN approach with varying parameters for the truncated normal distribution. From left to right column: truncation parameter equals <math display="inline"><semantics> <mrow> <mo>{</mo> <mn>2.6</mn> <mo>,</mo> <mn>1.8</mn> <mo>,</mo> <mn>1</mn> <mo>,</mo> <mn>0.2</mn> <mo>}</mo> </mrow> </semantics></math>.</p>
Full article ">Figure A2
<p>Synthetic HR-pQCT volumes sampled from the proposed 3D-StyleGAN approach with varying truncation levels. From left to right column: <math display="inline"><semantics> <mrow> <mi>ψ</mi> <mo>=</mo> <mo>{</mo> <mn>1</mn> <mo>,</mo> <mn>0.7</mn> <mo>,</mo> <mn>0.4</mn> <mo>,</mo> <mn>0.1</mn> <mo>}</mo> </mrow> </semantics></math>.</p>
Full article ">
17 pages, 4703 KiB  
Article
Robotics Classification of Domain Knowledge Based on a Knowledge Graph for Home Service Robot Applications
by Yiqun Wang, Rihui Yao, Keqing Zhao, Peiliang Wu and Wenbai Chen
Appl. Sci. 2024, 14(24), 11553; https://doi.org/10.3390/app142411553 - 11 Dec 2024
Viewed by 349
Abstract
The representation and utilization of environmental information by service robots has become increasingly challenging. In order to solve the problems that the service robot platform has, such as high timeliness requirements for indoor environment recognition tasks and the small scale of indoor scene [...] Read more.
The representation and utilization of environmental information by service robots has become increasingly challenging. In order to solve the problems that the service robot platform has, such as high timeliness requirements for indoor environment recognition tasks and the small scale of indoor scene data, a method and model for rapid classification of household environment domain knowledge is proposed, which can achieve high recognition accuracy by using a small-scale indoor scene and tool dataset. This paper uses a knowledge graph to associate data for home service robots. The application requirements of knowledge graphs for home service robots are analyzed to establish a rule base for the system. A domain ontology of the home environment is constructed for use in the knowledge graph system, and the interior functional areas and functional tools are classified. This designed knowledge graph contributes to the state of the art by improving the accuracy and efficiency of service decision making. The lightweight network MobileNetV3 is used to pre-train the model, and a lightweight convolution method with good feature extraction performance is selected. This proposal adopts a combination of MobileNetV3 and transfer learning, integrating large-scale pre-training with fine-tuning for the home environment to address the challenge of limited data for home robots. The results show that the proposed model achieves higher recognition accuracy and recognition speed than other common methods, meeting the work requirements of service robots. With the Scene15 dataset, the proposed scheme has the highest recognition accuracy of 0.8815 and the fastest recognition speed of 63.11 microseconds per sheet. Full article
(This article belongs to the Special Issue Artificial Intelligence in Complex Networks (2nd Edition))
Show Figures

Figure 1

Figure 1
<p>Home service robot service system.</p>
Full article ">Figure 2
<p>Domain ontology of the home environment.</p>
Full article ">Figure 3
<p>Template for developing the service inference SWRL rule base.</p>
Full article ">Figure 4
<p>The acquisition mechanism of missing attributes of objects.</p>
Full article ">Figure 5
<p>The acquisition mechanism of missing attributes category, physical, and visual attributes of objects.</p>
Full article ">Figure 6
<p>Network structure of MobileNetV3.</p>
Full article ">Figure 7
<p>The proposed transfer learning strategy.</p>
Full article ">Figure 8
<p>The structure of the semantic cognitive framework.</p>
Full article ">Figure 9
<p>Examples of the CIFAR-100 dataset.</p>
Full article ">Figure 10
<p>Examples of the Scene15 dataset.</p>
Full article ">Figure 11
<p>Examples of the UMD part affordance dataset.</p>
Full article ">Figure 12
<p>Loss values and accuracy during training for the classification of indoor functional areas.</p>
Full article ">Figure 13
<p>Loss values and accuracy.</p>
Full article ">
21 pages, 7882 KiB  
Article
Multi-Scale Gross Ecosystem Product (GEP) Valuation for Wetland Ecosystems: A Case Study of Lishui City
by Zhixin Zhu, Keyue Wu, Shuyue Zhou, Zhe Wang and Weiya Chen
Water 2024, 16(24), 3554; https://doi.org/10.3390/w16243554 - 10 Dec 2024
Viewed by 424
Abstract
Traditional gross ecosystem product (GEP) accounting methods often operate at macro scales, failing to reflect the localized and nuanced values of wetland ecosystems. This study addresses these challenges by introducing a fine-grained classification system based on a localized adaptation of international standards. The [...] Read more.
Traditional gross ecosystem product (GEP) accounting methods often operate at macro scales, failing to reflect the localized and nuanced values of wetland ecosystems. This study addresses these challenges by introducing a fine-grained classification system based on a localized adaptation of international standards. The framework integrates high-precision national land surveys and remote sensing quantitative analysis while incorporating fisheries resource models, climate regulation beneficiary mapping, and visitor interpolation to address data scarcity related to human activities. This approach refines the spatial calculation methods for functional quantity accounting at fine scales. The results demonstrate that the refined classification maintains consistency with traditional methods in total value while adapting to multi-scale accounting, filling gaps at small and medium scales and providing a more accurate representation of localized wetland characteristics. Additionally, the study highlights the dominance of cultural services in GEP, emphasizing the need to balance cultural and regulatory services to ensure fairness in decision-making. Finally, a village-scale decision-support model is proposed, offering actionable guidance for wetland management and sustainable development planning. Full article
(This article belongs to the Special Issue Hydro-Economic Models for Sustainable Water Resources Management)
Show Figures

Figure 1

Figure 1
<p>Layout of study area.</p>
Full article ">Figure 2
<p>Wetland classification mapping process.</p>
Full article ">Figure 3
<p>Spatial calculation methods for functional quantities maps based on interpolation optimization. (<b>a</b>) Fishery suitability mapping and adjustment. (<b>b</b>) Wetland climate regulation beneficiaries analysis and (<b>c</b>) wetland tourism distribution.</p>
Full article ">Figure 4
<p>GEP accounting process: In the ecosystem product amout part, the “·” symbol in the figure indicates data calculations based on coefficients or models specified in the standard, while the “+” symbol denotes the spatial calculations or adjustment coefficients added for optimizing functional quantities in the research design.</p>
Full article ">Figure 5
<p>Lishui wetland classification map.</p>
Full article ">Figure 6
<p>Multi-scale wetland waterbody area statistics.</p>
Full article ">Figure 7
<p>Gross ecosystem product result in multi scale.</p>
Full article ">Figure 8
<p>Wetland area vs. GEP value with Z-score selections.</p>
Full article ">Figure 9
<p>Cluster analysis of regulatory vs. cultural service contributions.</p>
Full article ">Figure 10
<p>Village-level wetland development strategy map.</p>
Full article ">
15 pages, 2348 KiB  
Article
Fine Tuning the Glass Transition Temperature and Crystallinity by Varying the Thiophene-Quinoxaline Copolymer Composition
by Xun Pan and Mats R. Andersson
Materials 2024, 17(24), 6031; https://doi.org/10.3390/ma17246031 - 10 Dec 2024
Viewed by 372
Abstract
In recent years, the design and synthesis of high-performing conjugated materials for the application in organic photovoltaics (OPVs) have achieved lab-scale devices with high power conversion efficiency. However, most of the high-performing materials are still synthesised using complex multistep procedures, resulting in high [...] Read more.
In recent years, the design and synthesis of high-performing conjugated materials for the application in organic photovoltaics (OPVs) have achieved lab-scale devices with high power conversion efficiency. However, most of the high-performing materials are still synthesised using complex multistep procedures, resulting in high cost. For the upscaling of OPVs, it is also important to focus on conjugated polymers that can be made via fewer simple synthetic steps. Therefore, an easily synthesised amorphous thiophene−quinoxaline donor polymer, TQ1, has attracted our attention. An analogue, TQ-EH that has the same polymer backbone as TQ1 but with short branched side-chains, was previously reported as a donor polymer with increased crystallinity. We have synthesised copolymers with varied ratios between octyloxy and branched (2-ethylhexyl)oxy-substituted quinoxaline units having the same polymer backbone, with the aim to control the aggregation/crystallisation behaviour of the resulting copolymers. The optical properties, glass transition temperatures and degree of crystallinity of the new copolymers were systematically examined in relation to their copolymer composition, revealing that the composition can be used to fine-tune these properties of conjugated polymers. In addition, multiple sub-Tg transitions were found from some of the polymers, which are not commonly or clearly seen in other conjugated polymers. The new copolymers were tested in photovoltaic devices with a fullerene derivative as the acceptor, achieving slightly higher performances compared to the homopolymers. This work demonstrates that side-chain modification by copolymerisation can fine-tune the properties of conjugated polymers without requiring complex organic synthesis, thereby expanding the number of easily synthesised polymers for future upscaling of OPVs. Full article
Show Figures

Figure 1

Figure 1
<p>Chemical structures of TQ-EH and copolymers with different loadings of octyloxy and (2-ethylhexyl)oxy side-chains.</p>
Full article ">Figure 2
<p>UV−vis spectra of polymers TQ-EH, TQ-O2-EH8, TQ-O4-EH6, TQ-O6-EH4, TQ-O8-EH2 and TQ1 in the 80 °C <span class="html-italic">o</span>-DCB solutions (<b>a</b>), <span class="html-italic">o</span>-DCB solutions at r.t. (<b>b</b>) and solid films (<b>c</b>). (<b>d</b>) The UV−vis spectra of TQ-EH in different conditions with the inserted photo showing the colours of TQ-EH <span class="html-italic">o</span>-DCB solutions at different temperatures.</p>
Full article ">Figure 3
<p>DMTA plots of polymers: (<b>a</b>) TQ-EH; (<b>b</b>) TQ-O2-EH8; (<b>c</b>) TQ-O4-EH6; (<b>d</b>) TQ-O6-EH4; (<b>e</b>) TQ-O8-EH2; (<b>f</b>) TQ1. The DMTA samples of polymers TQ-O4-EH6 and TQ-O6-EH4 broke at ~280 °C.</p>
Full article ">Figure 4
<p>Arrhenius plot of TQ-O2-EH8 showing the linear fits and resulting activation energies for the thermal transitions. The dashed lines represent the linear fit of the scatter plot.</p>
Full article ">Figure 5
<p>Representative <span class="html-italic">J−V</span> curves of polymer:PC<sub>71</sub>BM-based OPVs.</p>
Full article ">Figure 6
<p>AFM height images (0.5 × 0.5 µm<sup>2</sup>) of polymer:PC<sub>71</sub>BM (1:2.5) films. Scale bars are 200 nm.</p>
Full article ">
Back to TopTop