Artificial Intelligence-Based Fault Diagnosis for Steam Traps Using Statistical Time Series Features and a Transformer Encoder-Decoder Model
<p>A five-step process for visualizing steam trap fault diagnosis.</p> "> Figure 2
<p>This figure shows thermal images of the main equipment in the steam system at an aluminum processing plant and a food manufacturing plant. Most of the steam pipes are insulated with heat insulators, while equipment that can be manually set (such as steam traps) is not insulated.</p> "> Figure 3
<p>Receiver Operating Characteristic (ROC) Curve of a Transformer-based fault diagnosis model.</p> "> Figure 4
<p>Comparison of 2D diagnostic projection with statistical and machine learning methods.</p> "> Figure 5
<p>Magnified view of cluster overlapping in Transformer-based 2D diagnostic projection.</p> "> Figure 6
<p>Schematic of steam trap structure with temperature measurement points.</p> ">
Abstract
:1. Introduction
1.1. Challenges in Steam Trap Maintenance
- Manual inspections are time-consuming, labor-intensive, and prone to human error.
- Failures often go undetected for extended periods, leading to excessive energy waste.
- Unexpected failures result in production downtime and increased maintenance costs.
1.2. AI-Based Predictive Maintenance for Steam Trap Monitoring
1.3. Advancements in AI-Based Predictive Maintenance
2. Related Works
2.1. Existing Steam Trap Management Methods
- Heterogeneous sensor data collected from different manufacturers lacks standardization, making integration difficult [8].
- Environmental factors such as insulation layers and varying steam pressure complicate direct fault detection, requiring advanced methods such as infrared thermal imaging-based diagnostics [9].
2.2. Advancement of Predictive Maintenance Technology
- Traditional models, such as PCA and t-SNE, provide dimensionality reduction but may fail to preserve both local and global structures in high-dimensional data, leading to suboptimal representation [13].
- Supervised learning models require extensive labeled datasets, which are not always available due to the cost of manual annotations. This limitation has been highlighted in studies addressing data collection and quality challenges in deep learning [14].
- Transformer-based models have demonstrated superior feature extraction capabilities for time-series data [15], but their application to steam trap diagnostics remains unexplored.
Dimensionality Reduction and Visualization Techniques
2.3. Transformer Encoder–Decoder Model in Fault Diagnosis
- It captures both local and global dependencies: The self-attention mechanism enables the model to effectively preserve both short-term variations and long-term trends in sensor data, improving fault detection accuracy [20].
- It improves computational efficiency: The parallelized processing of Transformer models significantly reduces training and inference time compared to recurrent architectures such as LSTMs, making them well suited for handling large-scale industrial data [21].
- It enables real-time fault diagnosis: The Transformer’s ability to rapidly analyze streaming sensor data makes it ideal for industrial IoT applications, where fast response times are critical for minimizing system downtime [21].
- It enhances fault pattern separation: The model’s robust feature extraction capabilities improve fault-clustering performance, as indicated by metrics such as the Adjusted Rand Index (ARI) and Davies–Bouldin Index (DBI), leading to more interpretable diagnostic insights [22].
3. Our Approaches
3.1. Data Acquisition
3.2. Network Configuration
3.3. AI-Based Steam Trap Fault Diagnosis
3.3.1. Steam Trap Fault Diagnosis Technology
- Sensor data: Inlet and outlet temperatures of the steam trap.
- Trap properties: Application, size, model, manufacturer, and steam trap type.
- Environmental information: Temperature inside the plant, local weather conditions, work schedules, and seasonal patterns reflected through calendar data.
- Sensor malfunction (three types): The sensor is malfunctioning, making accurate measurement difficult (0: Out of range, 1: Frozen, 2: Inlet and outlet reversed).
- Unused: The steam trap is not currently in use.
- Leaked: The steam trap is not properly closed, causing continuous leakage of steam or condensate.
- Blocked (or a valve is closed): The trap is blocked or the valve is closed, preventing condensate discharge.
- Back-pressure: High outlet pressure disrupts the proper operation of the steam trap.
- Normal: The steam trap is operating normally.
3.3.2. Enhanced Two-Dimensional Diagnostic Projection for Steam Trap Fault Diagnosis
- Scatter Plot Representation: Each steam trap is represented as a point, with its position indicating proximity to fault clusters (higher likelihood of failure) or boundaries (diagnostic uncertainty).
- Fault Proximity Analysis: Facilitates quick identification of steam traps at high risk of failure or those with lower diagnostic certainty.
- Reliability Assessment: Provides a clear visual representation for assessing the reliability of diagnostic outcomes.
- Interpretability: Two-dimensional visualizations are easier for maintenance engineers to interpret, offering quick, actionable insights.
- Computational Complexity: Two-dimensional projections reduce computational load without compromising the clarity of fault clustering.
- Scalability: Scalable results are easily integrated into existing industrial monitoring dashboards without additional infrastructure.
4. Experiments
- Data acquisitions.
- Experiment setting.
- Performance evaluation.
4.1. Data Acquisitions
4.2. Experiment Settings
- Area Under Curve (AUC).
- Classification accuracy (CA).
- F-score (F1).
- Precision.
- Recall.
- Within-Cluster Sum of Squares (WCSS): Measures the compactness of clusters by calculating the sum of squared distances between each data point and the centroid of its assigned cluster. Lower values indicate tighter, well-defined clusters.
- Bayesian Information Criterion (BIC): Evaluates the trade-off between model complexity and goodness of fit. Lower values indicate better clustering performance with minimal overfitting.
- Davies–Bouldin Index (DBI): Assesses cluster separation and compactness by comparing intra-cluster dispersion with inter-cluster distances. Lower values indicate better-defined clusters.
- Adjusted Rand Index (ARI): Measures clustering accuracy by comparing the agreement between predicted clusters and ground truth labels. Higher values indicate stronger alignment with actual fault patterns.
- Calinski–Harabasz Index (CHI): Computes the ratio of between-cluster dispersion to within-cluster variance. Higher values indicate well-separated clusters with minimal overlap.
4.3. Performance Evaluation
- : Initial steam temperature at 1 atmosphere of pressure.
- : Discharge temperature at the steam trap for the i-th measurement.
- t: The time point when the predictive maintenance system is introduced.
- : The installation period for the predictive maintenance system.
- n: The total number of data points.
4.4. Computational Efficiency Analysis
4.4.1. Dataset and Experimental Setup
4.4.2. Hardware Specifications
4.4.3. Training and Inference Performance
4.4.4. Scalability and Optimization Strategies
- Mixed-Precision Training: Leveraging FP16 computations using NVIDIA’s Automatic Mixed Precision (AMP) can further reduce memory consumption and improve training speed.
- Model Pruning and Quantization: Reducing model size by removing redundant parameters or converting weights to less precise formats (e.g., INT8) can enhance efficiency, especially for embedded systems.
- Distributed Training: Using multi-GPU training strategies such as model parallelism and data parallelism can scale up training for large industrial datasets.
5. Conclusions
- Data Availability: The model requires a large labeled dataset, which may not always be feasible in industrial environments.
- Energy Consumption Variability: External factors such as production schedules, seasonal changes, and maintenance activities may impact energy savings. Additional studies are needed to analyze these influences.
- Sensor Dependency: The current approach primarily relies on temperature data, limiting applicability to diverse steam trap configurations. Future work should explore multimodal sensor integration (e.g., pressure and acoustic signals).
- Model Comparison: While the proposed Transformer-based model outperformed traditional machine learning methods, this study did not include a direct comparison with self-supervised learning approaches. Future work will explore the integration of self-supervised models, which can leverage unlabeled data to reduce labeling costs and enhance scalability. Comparing these models will provide insights into their relative performance, computational efficiency, and cost-effectiveness.
- Cost–Benefit Analysis: Although this study analyzed the energy-saving potential of the proposed system, a comprehensive cost–benefit analysis was not conducted. Future research will incorporate an economic assessment that considers installation and operational costs alongside energy savings to evaluate the financial feasibility of this AI-driven predictive maintenance system.
Author Contributions
Funding
Data Availability Statement
Conflicts of Interest
References
- Balzamov, D.S.; Balzamova, E.Y.; Bronskaya, V.V.; Oykina, G.I.; Kharitonova, O.S.; Shaikhetdinova, R.S.; Khairullina, L.E. Increasing efficiency of technological steam consumption at oil and gas enterprise. In Proceedings of the IOP Conference Series: Materials Science and Engineering, Krasnoyarsk, Russia, 16–18 April 2020; IOP Publishing: Bristol, UK, 2020; Volume 862. [Google Scholar]
- Roh, J.; Jang, S.; Kim, S.; Cho, H.; Kim, J. Steam trap maintenance-prioritizing model based on big data. ACS Omega 2021, 6, 4408–4416. [Google Scholar] [CrossRef] [PubMed]
- Ucar, A.; Karakose, M.; Kırımça, N. Artificial intelligence for predictive maintenance applications: Key components, trustworthiness, and future trends. Appl. Sci. 2024, 14, 898. [Google Scholar] [CrossRef]
- Davari, N.; Veloso, B.; Ribeiro, R.P.; Pereira, P.M.; Gama, J. Predictive maintenance based on anomaly detection using deep learning for air production unit in the railway industry. In Proceedings of the IEEE 8th International Conference on Data Science and Advanced Analytics (DSAA), Porto, Portugal, 6–9 October 2021. [Google Scholar]
- Fresina, E. Anomaly Detection for Predictive Maintenance: Toward a Generalisable Model Based on Early Identification of Malfunctioning. Ph.D. Dissertation, Tilburg University, Tilburg, The Netherlands, 2021. [Google Scholar]
- He, M.; He, D. Deep learning based approach for bearing fault diagnosis. IEEE Trans. Ind. Appl. 2017, 53, 3057–3065. [Google Scholar] [CrossRef]
- Radle, J. The Importance of Intensive Steam Trap Management. Chem. Eng. 2007, 114, 40. [Google Scholar]
- Saleem, J.; Hammoudeh, M.; Raza, U.; Adebisi, B.; Ande, R. IoT standardisation: Challenges, perspectives and solution. In Proceedings of the 2nd International Conference on Future Networks and Distributed Systems, Amman, Jordan, 26–27 June 2018. [Google Scholar]
- Guan, H.; Xiao, T.; Luo, W.; Gu, J.; He, R.; Xu, P. Automatic fault diagnosis algorithm for hot water pipes based on infrared thermal images. Build. Environ. 2022, 218, 109111. [Google Scholar] [CrossRef]
- Faria, J.R.; Semedo, S.; Cardoso, F.J.; Oliveira, J. Condition Monitoring and Diagnosis of Steam Traps with Wireless Smart Sensors. Int. J. Smart Sens. Intell. Syst. 2014, 7, 1–6. [Google Scholar] [CrossRef]
- Isa, D.; Rajkumar, R. Pipeline defect prediction using support vector machines. Appl. Artif. Intell. 2009, 23, 758–771. [Google Scholar] [CrossRef]
- Zhou, J.; An, Z.; Yang, Z.; Zhang, Y.; Chen, H.; Chen, W.; Luo, Y.; Guo, Y. PT-Informer: A Deep Learning Framework for Nuclear Steam Turbine Fault Diagnosis and Prediction. Machines 2023, 11, 846. [Google Scholar] [CrossRef]
- Van der Maaten, L.; Hinton, G. Visualizing data using t-SNE. J. Mach. Learn. Res. 2008, 9, 2579–2605. [Google Scholar]
- Whang, S.E.; Roh, Y.; Song, H.; Lee, J.G. Data collection and quality challenges in deep learning: A data-centric ai perspective. VLDB J. 2023, 32, 791–813. [Google Scholar] [CrossRef]
- Zhou, H.; Zhang, S.; Peng, J.; Zhang, S.; Li, J.; Xiong, H.; Zhang, W. Informer: Beyond efficient transformer for long sequence time-series forecasting. In Proceedings of the AAAI Conference on Artificial Intelligence, Online, 2–9 February 2021; Volume 35. [Google Scholar]
- Liu, Y.; Yu, W.; Dillon, T.; Rahayu, W.; Li, M. Empowering IoT predictive maintenance solutions with AI: A distributed system for manufacturing plant-wide monitoring. IEEE Trans. Ind. Inform. 2021, 18, 1345–1354. [Google Scholar] [CrossRef]
- Sandu, A.K. AI-Powered Predictive Maintenance for Industrial IoT Systems. Digit. Sustain. Rev. 2022, 2, 1–14. [Google Scholar]
- Kumar, V.; Prakash, M.; Thamburaj, S. Deep Learning-based Predictive Maintenance for Industrial IoT Applications. In Proceedings of the International Conference on Inventive Computation Technologies (ICICT), Lalitpur, Nepal, 24–26 April 2024. [Google Scholar]
- Suthar, A.; Kolhe, K.; Gutte, V.; Patil, D. Predictive Maintenance and Real Time Monitoring using IoT and Cloud Computing. In Proceedings of the 2024 5th International Conference on Image Processing and Capsule Networks (ICIPCN), Dhulikhel, Nepal, 3–4 July 2024. [Google Scholar]
- Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. In Proceedings of the 31st International Conference on Neural Information Processing Systems (NeurIPS), Long Beach, CA, USA, 4–9 December 2017; Curran Associates, Inc.: Red Hook, NY, USA, 2017; Volume 30, pp. 5998–6008. [Google Scholar]
- Li, S.; Jin, X.; Xuan, Y.; Zhou, X.; Chen, W.; Wang, Y.X.; Yan, X. Enhancing the locality and breaking the memory bottleneck of transformer on time series forecasting. Adv. Neural Inf. Process. Syst. 2019, 32. [Google Scholar]
- Wu, H.; Xu, J.; Wang, J.; Long, M. Autoformer: Decomposition transformers with auto-correlation for long-term series forecasting. Adv. Neural Inf. Process. Syst. 2021, 34, 22419–22430. [Google Scholar]
- Haxhibeqiri, J.; De Poorter, E.; Moerman, I.; Hoebeke, J. A survey of LoRaWAN for IoT: From technology to application. Sensors 2018, 18, 3995. [Google Scholar] [CrossRef] [PubMed]
- Andres-Maldonado, P.; Ameigeiras, P.; Prados-Garzon, J.; Ramos-Munoz, J.J.; Lopez-Soler, J.M. Optimized LTE data transmission procedures for IoT: Device side energy consumption analysis. In Proceedings of the 2017 IEEE International Conference on Communications Workshops (ICC Workshops), Paris, France, 21–25 May 2017. [Google Scholar]
- Kim, C.; Joe, I. A Balanced Approach of Rapid Genetic Exploration and Surrogate Exploitation for Hyperparameter Optimization. IEEE Access 2024, 12, 192184–192194. [Google Scholar] [CrossRef]
- Miyawaki Inc. Figure 4.1 Set of Condensate Collection Package. Available online: https://www.miyawaki-inc.com/application/files/2116/4506/6116/Figure_4.1_Set_of_condensate_collection_package.png (accessed on 24 February 2025).
Name | Formula |
---|---|
Maximum | |
Minimum | |
Absolute Average | |
Peak to Peak | |
RMS 1 | |
Mean | |
Standard Deviation | |
Skewness | |
Kurtosis | |
Variance | |
Wave Factor | |
Change Coefficient | |
Median | |
Clearance Factor | |
Impulse Factor | |
Percentile 25 | |
Percentile 50 | |
Percentile 75 | |
Percentile 90 | |
Sum |
Attribute | Example 1 | Example 2 | Attribute | Example 1 | Example 2 |
---|---|---|---|---|---|
site_id | 919 | 919 | outlet_rms | 97.79 | 97.806 |
dev_id | 630 | 630 | outlet_mean | 97.79 | 97.806 |
dev_size | 1 | 1 | outlet_std | 0.123 | 0.218 |
dev_type | 0 | 0 | outlet_skewness | 0.082 | 0.439 |
dev_manufacturer | 0 | 0 | outlet_kurtosis_value | −1.18 | −0.865 |
inlet_max | 154.27 | 152.7 | outlet_variance | 0.015 | 0.047 |
inlet_min | 147.81 | 148.53 | outlet_wave_factor | 1 | 1 |
inlet_absolute_average | 151.744 | 150.49 | outlet_clearance_factor | 0.01 | 0.01 |
inlet_ptp | 6.46 | 4.17 | outlet_impulse_factor | 1.002 | 1.005 |
inlet_rms | 151.752 | 150.495 | outlet_percentile25 | 97.7 | 97.633 |
inlet_mean | 151.744 | 150.49 | outlet_percentile50 | 97.785 | 97.81 |
inlet_std | 1.564 | 1.284 | outlet_percentile75 | 97.91 | 97.975 |
inlet_skewness | −0.755 | −0.013 | outlet_percentile90 | 97.94 | 98.129 |
inlet_kurtosis_value | −0.018 | −1.296 | outlet_sum_value | 2346.96 | 2347.35 |
inlet_variance | 2.445 | 1.648 | outlet_median | 97.785 | 97.81 |
inlet_wave_factor | 1 | 1 | env_temp | 15.907 | 17.545 |
inlet_clearance_factor | 0.007 | 0.007 | env_humi | 23.669 | 28.493 |
inlet_impulse_factor | 1.017 | 1.015 | outdoor_humi | 67.324 | 71.406 |
inlet_percentile25 | 150.72 | 149.222 | outdoor_temp | −1.251 | 1.013 |
inlet_percentile50 | 152.125 | 150.315 | outdoor_feellike | −3.367 | 0.95 |
inlet_percentile75 | 152.888 | 151.538 | outdoor_windspeed | 1.867 | 0.779 |
inlet_percentile90 | 153.187 | 152.048 | year | 2024 | 2024 |
inlet_sum_value | 3641.85 | 3611.75 | month | 1 | 1 |
inlet_median | 152.125 | 150.315 | day | 3 | 10 |
outlet_max | 98.02 | 98.26 | weekday | 3 | 3 |
outlet_min | 97.6 | 97.49 | operation | 1 | 1 |
outlet_absolute_average | 97.79 | 97.806 | pm 1 | 0 | 0 |
outlet_ptp | 0.42 | 0.77 | label | 4 | 4 |
Model | Hyperparameter | Value |
---|---|---|
Random Forest (RF) | Number of Trees | 100 |
Attributes Considered per Split | 5 | |
Gradient Boost (GB) | Number of Trees | 100 |
Learning Rate | 0.1 | |
Max Depth per Tree | 3 | |
Adaptive Boosting (AdaBoost) | Number of Estimators | 50 |
Learning Rate | 1.0 | |
Classification Algorithm | SAMME.R | |
Regression Loss Function | Linear | |
Multi-layer Perceptron (MLP) | Network Structure | 50 × 50 × 50 |
Activation Function | ReLU | |
Optimizer | Adam | |
Transformer (Ours) | Input Dimension | 52 |
Hidden Dimension | 8 | |
Output Dimension | 52 | |
Number of Layers | 4 | |
Attention Heads | 8 | |
Intermediate Size | 32 (Hidden Dim × 4) | |
Dropout Probability | 0.1 | |
Max Sequence Length | 512 | |
Optimizer | Adam (lr = ) | |
Loss Functions | MSE (Reconstruction), BCE (Classification) |
Model | AUC | CA | F1 | Precision | Recall |
---|---|---|---|---|---|
RF | 0.890 | 0.875 | 0.875 | 0.875 | 0.875 |
MLP | 0.902 | 0.895 | 0.898 | 0.900 | 0.899 |
GB | 0.906 | 0.887 | 0.886 | 0.888 | 0.887 |
AdaBoost | 0.915 | 0.901 | 0.902 | 0.901 | 0.902 |
Transformer | 0.927 | 0.932 | 0.938 | 0.916 | 0.914 |
Model | WCSS | BIC | DBI | ARI | CHI |
---|---|---|---|---|---|
Simple stats. 1 | 212,077 | 212,077 | 2.6145 | 0.7671 | 5392 |
PCA | 169,806 | 103,608 | 2.2582 | 0.7479 | 5146 |
t-SNE | 12,083,970 | 217,808 | 1.7633 | 0.8868 | 3579 |
VAE | 7117 | −33,338 | 1.8366 | 0.7421 | 1862 |
Transformer | 19,700 | 74,711 | 0.8808 | 0.9195 | 14,558 |
All Features | 8,638,854 | 161,006 | 4.1308 | 0.9256 | 11,738 |
CPU Specifications | GPU Specifications | ||
---|---|---|---|
Model | Intel Core i9-10920X | Model | NVIDIA GeForce RTX 3090 (x2) |
Architecture | x86_64 (64-bit) | CUDA Cores | 10,496 per GPU |
Cores/Threads | 12 Cores/24 Threads | Memory | 24 GB GDDR6X per GPU |
Base Clock | 3.50 GHz | Memory Bandwidth | 936.2 GB/s |
Max Clock | 4.80 GHz | Base Clock | 1.40 GHz |
L1 Cache | 384 KiB | Boost Clock | 1.70 GHz |
L2 Cache | 12 MiB | TDP | 350 W per GPU |
L3 Cache | 19.3 MiB | CUDA Version | 12.1 |
TDP | 165 W | Driver Version | 530.30.02 |
SIMD Extensions | AVX512, SSE4.2, FMA3 | ||
Virtualization | VT-x Supported |
Device | Train (s) | Inference (s) |
---|---|---|
CPU | ||
GPU |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Kim, C.; Cho, K.; Joe, I. Artificial Intelligence-Based Fault Diagnosis for Steam Traps Using Statistical Time Series Features and a Transformer Encoder-Decoder Model. Electronics 2025, 14, 1010. https://doi.org/10.3390/electronics14051010
Kim C, Cho K, Joe I. Artificial Intelligence-Based Fault Diagnosis for Steam Traps Using Statistical Time Series Features and a Transformer Encoder-Decoder Model. Electronics. 2025; 14(5):1010. https://doi.org/10.3390/electronics14051010
Chicago/Turabian StyleKim, Chul, Kwangjae Cho, and Inwhee Joe. 2025. "Artificial Intelligence-Based Fault Diagnosis for Steam Traps Using Statistical Time Series Features and a Transformer Encoder-Decoder Model" Electronics 14, no. 5: 1010. https://doi.org/10.3390/electronics14051010
APA StyleKim, C., Cho, K., & Joe, I. (2025). Artificial Intelligence-Based Fault Diagnosis for Steam Traps Using Statistical Time Series Features and a Transformer Encoder-Decoder Model. Electronics, 14(5), 1010. https://doi.org/10.3390/electronics14051010