[go: up one dir, main page]
More Web Proxy on the site http://driver.im/
Next Article in Journal
Pulse Shape Discrimination of n/γ in Liquid Scintillator at PMT Nonlinear Region Using Artificial Neural Network Technique
Next Article in Special Issue
Sybil Attack-Resistant Blockchain-Based Proof-of-Location Mechanism with Privacy Protection in VANET
Previous Article in Journal
Identification Method for Railway Rail Corrugation Utilizing CEEMDAN-PE-SPWVD
Previous Article in Special Issue
Syntactic–Semantic Detection of Clone-Caused Vulnerabilities in the IoT Devices
You seem to have javascript disabled. Please note that many of the page functionalities won't work as expected without javascript enabled.
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Deep FS: A Deep Learning Approach for Surface Solar Radiation

Department of Computer Engineering, Konya Food and Agriculture University, Konya 42080, Turkey
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Sensors 2024, 24(24), 8059; https://doi.org/10.3390/s24248059
Submission received: 3 November 2024 / Revised: 30 November 2024 / Accepted: 3 December 2024 / Published: 18 December 2024
(This article belongs to the Special Issue AI-Based Security and Privacy for IoT Applications)

Abstract

:
Contemporary environmental challenges are increasingly significant. The primary cause is the drastic changes in climates. The prediction of solar radiation is a crucial aspect of solar energy applications and meteorological forecasting. The amount of solar radiation reaching Earth’s surface (Global Horizontal Irradiance, GHI) varies with atmospheric conditions, geographical location, and temporal factors. This paper presents a novel methodology for estimating surface sun exposure using advanced deep learning techniques. The proposed method is tested and validated using the data obtained from NASA’s Goddard Earth Sciences Data and Information Services Centre (GES DISC) named the SORCE (Solar Radiation and Climate Experiment) dataset. For analyzing and predicting accurate data, features are extracted using a deep learning method, Deep-FS. The method extracted and provided the selected features that are most appropriate for predicting the surface exposure. Time series analysis was conducted using Convolutional Neural Networks (CNNs), with results demonstrating superior performance compared to traditional methodologies across standard performance metrics. The proposed Deep-FS model is validated and compared with the traditional approaches and models through the standard performance metrics. The experimental results concluded that the proposed model outperforms the traditional models.

1. Introduction

Renewable energy represents an environmentally sustainable power source that minimizes CO2 emission. Solar and wind power represent primary renewable energy sources with virtually unlimited potential [1]. Today, climate change has become a serious concern, and renewable energy can act as a solution to this. Because they do not emit gases, they reduce carbon footprints and reduce environmental impacts. The depletion of fossil fuels has also contributed to this. Among the available renewable energy sources, solar energy for the generation of electricity [2] has become very popular. As the rapid growth of these cells is increasing, the efficiency has improved greatly, making it more economical and sustainable. With rapid growth and cost efficiency expected, it is likely to become the main energy source in the world.
Solar energy has numerous contemporary applications. There are various environmental parameters [3] that hinder traditional approaches to predicting the solar surface accurately. Deep learning models have larger advancements over the traditional approaches in predicting the time-series data and overcoming environmental parameters such as cloudy skies and orbital positions. By integrating deep learning models, the accuracy of prediction has increased significantly.
Solar power generation depends on weather conditions. This feature destabilizes the overall operation of power systems. This instability can increase power outages and prices, which will adversely affect the cost. Therefore, a technology that can provide a balance between power supply and demand and establish a stable connection between them is required. The conversion of solar energy is performed directly and is related to solar surface exposure. Consequently, variables that are directly related to it have been used to predict the amount of solar power generation. Significant meteorological parameters, including atmospheric conditions such as sky morphology, atmospheric visibility indices, cloud coverage fraction, and associated climatic variables, have also been investigated. The factors that are focused on predicting the amount of power generation have also been analyzed. The analysis of all the factors revealed that solar surface exposure is the most important feature for prediction. In this paper, different deep learning models are deployed for estimating the solar surface exposure, along with the Deep-FS technique.
The rest of the paper is organized as follows. The literature review, research gaps, and research contributions are presented in Section 2. The proposed methodology is presented, including data acquisition, pre-processing, and the novel Deep-FS feature selection approach, in Section 3. Then, it details the architecture of the hybrid CNN-LSTM/GRU model and its implementation. Section 4 describes the experimental setup and presents comprehensive results that compare the proposed approach with existing methods. Then, it provides a detailed discussion of the findings, including an analysis of model performance under different conditions. The paper concludes with a summary of key findings and suggestions for future research directions in Section 5.

2. Related Work

Historical solar radiation forecasting studies have employed statistical time series methods, including ARIMA (Autoregressive Integrated Moving Average), SARIMA (Seasonal ARIMA), and ARMA-S (Seasonal ARMA) [4,5]. These methods model temporal dependencies in radiation data through autoregressive components, moving averages, and seasonal adjustments. However, such methods exhibit limitations in modeling non-linear relationships, constraining their predictive accuracy. To address the limitations of traditional statistical approaches, we implement a hybrid deep learning architecture that combines (i) Convolutional Neural Networks (CNNs) for extracting local temporal patterns, and (ii) long short-term memory (LSTM) and gated recurrent unit (GRU) networks for capturing long-term dependencies. Building on work by Lim et al. [6], our approach incorporates architectural improvements and robust validation methods for the prediction of solar radiation.
Brahma et al. [7] used a deep learning model trained on time series data to predict daily sun exposure after 1, 4, and 10 days. The data came from multiple regions. They applied LSTM, GRU, and CNN to predict solar surface exposure, and they used extreme gradient boosting, Pearson correlation, and Spearman correlation to choose neighboring areas that are significantly correlated to the target area’s solar surface exposure.
In addition, they carried out feature selection using relief to enhance prediction performance. Global Horizontal Irradiance (GHI) is our primary input parameter, representing the total solar radiation received on a horizontal surface at the Earth’s surface. This includes both direct and diffuse radiation components measured through the NASA POWER API.
Research on solar surface exposure accounts for temporal aspects. However, this research has used a convolution operation approach or a mixture of convolution layers to account for spatial aspects [8,9,10].
The primary variables utilized for solar surface exposure forecasts were those representing the direct impact of clouds on the phenomenon, such as the clearness index and the cloud cover. Geographical variables such as latitude, longitude, and altitude, as well as temporal variables such as year, month, day, and hour, were commonly used. The features of the input variables were selected primarily using the filter approach [11]. Table 1 shows a comprehensive literature analysis of different approaches.

2.1. Research Gaps

Based on the comprehensive review of the literature, the following research gaps have been identified:
  • Current deep learning models lack effective feature selection mechanisms, potentially including redundant or less relevant features that may impact prediction accuracy.
  • The integration of multiple deep learning architectures for solar exposure prediction remains largely unexplored, especially in combining spatial and temporal feature extraction.
  • Limited research exists on adaptable models that can handle varying environmental conditions and seasonal changes while maintaining prediction accuracy.

2.2. Rsearch Contributions

Upon presenting the research gaps, this study makes the following key contributions to the field:
  • Developing a hybrid CNN–LSTM/GRU architecture (Deep-FS feature selection methodology) that effectively combines spatial and temporal feature extraction capabilities, improving prediction accuracy over single-architecture approaches.
  • Demonstrates improved performance over traditional approaches through comprehensive evaluation using the RMSE, RRMSE, R2, and MAE metrics.
  • Demonstrates superior performance over traditional statistical and machine learning approaches, achieving 96% prediction accuracy with enhanced generalizability.

3. Proposed Method

The proposed method for predicting the solar surface is depicted in Figure 1. For data acquisition, SORCE and SIM data are given as input. The features are identified based on the selection of features for the prediction model. Feature engineering also identifies the relevant set of features that are crucial for accurate prediction. In data processing, valid timestamps, resizing, and normalization are performed. Optimal features are selected, and the prediction model is designed to estimate the solar surface. The details of each step are illustrated in the following sections.

3.1. Data Acquisition

The Solar Radiation and Climate Experiment (SORCE) is a NASA-sponsored satellite mission launched on 25 January 2003, whose main objective was to study the impact of the Sun on the Earth’s climate. It analyses solar radiation of all wavelengths and ranges with higher accuracy and precision, which helps scientists to study the long-term effect of solar radiation on Earth’s climate. Continuous monitoring has revealed that significant correlations between solar phenomena geomagnetic storms, and other phenomena have a huge impact on Earth’s climate change. The data specification for different sets of channels is presented in Table 2.
The data used through the experiments are obtained from NASA POWER API at point-based resolution, with each data point representing measurements at specific latitude–longitude coordinates. In this study, spatial interpolation between points is not performed to maintain fidelity of the data.

3.2. Data Analysis and Preprocessing

The datasets of SORCE and SIM are processed for missing data points. Missing data points were interpolated using mean values of adjacent observations; we can take the mean of the data points as follows:
x i = 1 n j = 1 n x j
where x i is a data point.
Now, normalization (0–1) is performed using the following equation:
x i = x i x min x max x min
After performing normalization, statistical, spectral and regression analyses are performed. The results are presented in Figure 2, Figure 3 and Figure 4 respectively.
From the time-series analysis, it is evident that the periodic behavior of the data set is similar to Earth’s orbital data. Furthermore, the wavelength patterns are identical and consistent. With the regression analysis, our analysis demonstrates the capability to predict spectral variations across alternative wavelength ranges.
Figure 3 shows the spectral analysis of SIM data at a 500 nm wavelength. Here, the X-axis represents the frequency (1/day), and the Y-axis represents the amplitude/power. The plot shows a sharp peak near frequency 0 followed by very low-amplitude oscillations. This indicates a strong constant/baseline component in the solar radiation data; higher frequency components have much smaller contributions, and most of the signal’s power is concentrated in the low frequency range.
Figure 4 is a scatter plot showing the relationship between predicted and actual values, where the X-axis represents Spectral Surface Exposure at 500 nm (actual values) and the Y-axis represents predicted values. While blue dots are individual data points showing actual vs. predicted values, the red line is the fitted regression line showing the trend. It is obvious that there is a considerable scatter around the trend line, but with a clear positive correlation.

3.3. Feature Extraction

Features from the SORCE dataset were extracted based on their time variances and their distribution along different wavelengths. Some atmospheric particles are also included, including aerosol particles. For the SIM data set, the behavior patterns, performance metrics, mean and variances, time series analysis, and other relevant metrics are extracted. The different features of the SORCE dataset are described in Table 3.

3.4. Feature Selection and Visualization

The feature-selection method is based on estimating the correlation or effect of a particular feature on the SIM data set. The reason for using the Deep-FS is to better understand the correlation of the features. This enhances the power of neural networks to understand complex interactions. The features are identified on the basis of and the weights assigned. For performance evaluation, RMSE is deployed.
For feature selection, let us assume that X and y are data points. The performance of the model based on LSTM and GRU is based on the loss function defined as
L ( θ , X , y ) = 1 n i = 1 n ( y i f ( X i ; θ ) ) 2
where θ is the model parameter.
The major task is to optimize the performance of the model and minimize the loss function. To achieve this, an empty set is initialized to zero, and the best performance is set to infinity. For each feature, the loss function is computed. If the current best score is less than the loss score, the feature is updated and removed from the subset. This process is repeated until no further performance improvement is required. Data points are divided into time intervals, and predictions are made based on multiple extracted features. Visualizations for different timestamps and channels are depicted in Figure 5.

3.5. Model Design and Prediction

CNN is deployed to extract the features and predict the time series data. The input variables are the feature variables, the channel parameters, and the extracted features. The input is then pre-processed to identify the surface exposure. All features and parameters were used to predict surface exposure. Different experiments were conducted at different timestamps and data sets. For performance evaluation, training and testing were conducted 10 times due to the random nature of the deep learning model. All input parameters with their specifications on different layers are defined in Table 4.
To capture the spatial and temporal dependencies in the data and their nonlinear interdependencies, the deep learning model is deployed. The deep learning model overcomes the limitations of the traditional models. Incorporating the advantages of CNN along with LSTM and GRU improves the accuracy prediction of the data for solar surface detection. For a complex set of data, traditional approaches such as ARIMA and SARIMA are not able to capture the interdependency between the data. The proposed model can be used for large, complex sets of data. The proposed model combines CNN architecture along with deep learning models, which enhances the feature extraction process to extract the relevant set of features from a larger set of features. The proposed model uses advanced feature extraction methods (Deep-FS), through which prediction accuracy is improved and an accuracy of 96% is achieved.
The flow of the proposed model can be summarized as follows.
  • The solar data is acquired from both datasets. Let X represent the data points in the dataset.
    X = { x 1 , x 2 , , x n }
  • Data are then normalized for a specific range using the mean and standard deviation.
    X = X μ σ
  • Relevant features are extracted from the features mentioned in Table 3 using the feature extraction process:
    • The data points are normalized.
    • Convolution is applied to extract the local features.
    • The RelU Activation function, along with pooling and flattening, is applied for capturing essential features.
  • CNN-LSTM and CNN-GRU models are designed.
    LSTM:
    Z 1 = C o n v ( X ; W 1 , b 1 )
    A 1 = A c t i v a t i o n ( Z 1 )
    P 1 = M a x P o o l ( A 1 )
    L = L S T M ( P 1 ; θ L S T M )
    E C N N - L S T M = O u t p u t ( L ; W O , b O )
    GRU:
    Z 2 = C o n v ( X ; W 2 , b 2 )
    A 2 = A c t i v a t i o n ( Z 2 )
    P 2 = M a x P o o l ( A 2 )
    G = G R U ( P 2 ; θ G R U )
    E C N N - G R U = O u t p u t ( G ; W O , b O )
    where
    • W: Filter;
    • b: bias;
    • Z 1 : ReLU Activation.
    The LSTM Cell Process is defined as
    f t = σ ( W f · [ h t 1 , x t ] + b f )
    i t = σ ( W i · [ h t 1 , x t ] + b i )
    C t = tanh ( W C · [ h t 1 , x t ] + b C )
    C t = f t C t 1 + i t C t
    o t = σ ( W o · [ h t 1 , x t ] + b o )
    h t = o t tanh ( C t )
    where
    • f t : Forget Gate;
    • i t : Input Gate;
    • h: Hidden State;
    • C t : Candidate Cell State;
    • W: Weight.
  • Train the model and optimize the loss function.
    L ( θ , X , y ) = 1 n i = 1 n ( y i f ( X i ; θ ) ) 2
  • Evaluate the performance with different performance metrics RMSE, RRMSE, MAE, and R2 as follows.
    R M S E = 1 n i = 1 n ( y i y ^ i ) 2
    R R M S E = R M S E Y ¯ = 1 n i = 1 n ( y i y ^ i ) 2 y ¯
    R 2 = 1 i = 1 n ( y i y ^ i ) 2 i = 1 n ( y i y ¯ i ) 2
    M A E = 1 n i = 1 n | y i y ^ i |
  • Perform a comparative analysis with traditional approaches for the validation of the proposed method.

3.6. Elaboration on Model Selection

In this section, we will try to explain the specific advantages of LSTM and GRU over other deep learning and machine learning algorithms for solar surface exposure prediction problems.

3.6.1. Temporal Dependency Handling

LSTM and GRU are specifically designed to handle temporal dependencies through their memory mechanisms, in contrast to the lack of explicit memory structures in regular neural networks and CNNs. In [3], it is demonstrated that LSTM outperformed traditional methods with 15.3% improvement in RMSE. Other ML algorithms such as SVM or Random Forests do not maintain the temporal ordering of data.

3.6.2. Variable Sequence Length

LSTM/GRU can naturally handle variable-length sequences, while most other algorithms require fixed input sizes. Ref. [11] showed LSTM’s ability to handle variable-length temporal data with 92% precision.

3.6.3. Memory Management

Although LSTM uses memory cells with input, output, and forget gates, GRU has a simpler architecture with reset and update gates. Both can selectively remember or forget information, in contrast to other algorithms with the lack of this selective memory capability. Ref. [9] demonstrated the hybrid CNN-LSTM model’s effectiveness in capturing both spatial and temporal patterns.

3.7. Advantages over Specific Algorithms

Ref. [1] empirically showed the superiority of LSTM/GRU over the ARIMA and SVR models, in addition to the work demonstrated in [1].
In contrast to Convolutional Neural Networks (CNNs), long short-term memory (LSTM) and gated recurrent unit (GRU) architectures excel in sequential data processing, with the ability to handle temporal dependencies and a wider support for time series forecasting and analysis.
LSTM and GRU demonstrate superior capabilities capturing long-term dependencies with a more robust training process, and it is significantly mitigated against vanishing gradients compared to RNNs [27].
LSTM and GRU architectures are designed to process sequential data through their gating mechanisms and memory cells, allowing them to capture temporal patterns. In contrast, traditional ML algorithms like SVM and Random Forest treat inputs as independent features without considering temporal ordering [28].

4. Experimental Results

Experimental validation was conducted to assess the model’s performance of the proposed method. The input parameters defined in Table 4 were used to perform the training. The outcome generated after calculating the loss function for each parameter is considered. The one with the lowest loss is considered for the final output. Performance is computed using features captured through Deep-FS.
Performance is also compared with single models as well as the combination of different models. The features are selected using the Fisher Score, Recursive Feature Elimination (RFE), Random Forest Importance, and Deep-FS. The performance metrics used are RMSE, RRMSE, R2, and MAE. The experimental results are depicted in Table 5.
As depicted in Table 5, Deep-FS leads as the most effective approach to predict solar surface exposure compared to all other feature selection methods. The results show that RFE, in terms of performance and with more variables, has improved results. Furthermore, the Fisher Score and the RFE compared to the top 15% showed better results.
The overall results reveal that Deep-FS GRU has the best results when considering the correlation between the feature vectors. Figure 6 shows the overall prediction results for a specific time stamp.
To check whether the combination of different deep learning models is better, the proposed model is tested against single deep learning models to check whether they are still better. The models included are ANN, CNN, LSTM, and GRU. The input parameters are the features extracted using the Deep-FS method. The experiments revealed that for time series prediction, LSTM and GRU performed much better compared to ANN and CNN. It is also concluded that the combination of LSTM and GRU with CNN performed much better as they can predict all the characteristics much more accurately. The comparative analysis of experimental performance parameters is shown in Table 6.
Table 6 compares the performance of different deep learning models (ANN, CNN, LSTM, GRU, and their combinations) across two methods (Deep FS with long short-term memory and Deep FS with Gated Recurrent Units). As depicted in Table 6, using a combined approach for CNN with LSTM and GRU leads to a better outcome compared to individual deep learning models. CNN, however, leads when the number of parameters is time-bound, but when combined with others, it gives a better outcome. The GRU also performs well, but with higher variability, but it misses some major interconnections. When the data are complex, the ANN seems to have deviations from the actual values, while the CNN is unable to handle the dynamics effectively. CNN-LSTM has the lowest MAE and RMSE, indicating its superiority over other models. The combined models are highly effective for short-term prediction, achieving better overall performance metrics (RMSE, RRMSE, R2, MAE), with CNN-LSTM having the lowest error rates (MAE: 0.013449, RMSE: 0.125694) among all configurations. Therefore, in the case of solar surface exposure prediction, the combined model is much better than the individual models. The visualization of the comparison is shown in Figure 7.
For the prediction of data points, two traditional methods are generally used: Auto-Regressive Integrated Moving Average (ARIMA) and Seasonal Auto-Regressive Integrated Moving Average (SARIMA). The proposed method is compared with both approaches based on the evaluation parameters RMSE, RRMSE, MAE, and R2. The comparative analysis is shown in Table 7, which demonstrates that the proposed CNN-based hybrid models (CNN-LSTM and CNN-GRU) outperform traditional statistical methods (ARIMA and SARIMA) across all evaluation metrics, with CNN-GRU achieving the best results.
The value of RMSE indicates the average error. The performance difference between the proposed model and existing approaches is primarily derived from their capacity to model data relationships. The traditional approach supports linear relationships, whereas the deep learning model is designed to capture non-linear relationships to overcome interdependency and produce accurate results.
The deep learning models can handle spatial and temporal data effectively. In addition, they are scalable and adaptable to the new data, making them the best choice for the solar surface exposure. The model demonstrates an accurate prediction of surface solar radiation with prediction metrics such as Mean Squared Error (MSE), Root Mean Squared Error (RMSE), and Mean Absolute Error (MAE). It also demonstrates valuable performance improvements in the reduction in prediction error vs. SARIMA, the improvement in seasonal pattern capture, and better handling of daily variations.
During the experiments, several performance metrics were measured. Figure 8 shows the predictions of the CNN-GRU model tracking the actual solar radiation values with moderate volatility, demonstrating the model’s ability to capture both general trends and some short-term fluctuations.
Figure 9 displays the predictions of the CNN-LSTM model, following a smoother trajectory compared to CNN-GRU, suggesting better handling of long-term patterns but less sensitivity to short-term variations.
In Figure 10, the left plot shows stable training with low loss values converging around 0.03; the right plot demonstrates positive correlation between predicted and actual values with some scatter, indicating decent but imperfect prediction accuracy for CNN/LSTM.
In Figure 11, the CNN-GRU training shows similar training stability to CNN-LSTM but with slightly more validation loss fluctuation; the prediction scatter plot reveals comparable performance to CNN-LSTM with moderate spread around the ideal prediction line.
In Figure 12, the CNN-GRU ROC curve shows strong classification performance with an AUC of 0.81, with the curve rising dramatically in the early stages and maintaining good true positive rates while keeping false positives relatively low. On the other hand, CNN-LSTM model shows slightly lower but still good performance with an AUC of 0.79, following a curve pattern similar to that of CNN-GRU but with marginally lower true positive rates across different thresholds.

5. Conclusions

In this work, a method to predict the solar surface exposure is developed based on the data received from the SORCE and SIM datasets. To overcome different environmental constraints such as cloudy skies and orbital position of the Earth, feature variables are generated and extracted using the Deep-FS method. For predicting time-series data, a model is designed that combines CNN with LSTM and GRU. For performance evaluation, performance metrics such as RMSE, RRMSE, R2, and MAE are used.
The CNN-GRU model achieved optimal performance with an RMSE of 0.2201, demonstrating superior error minimization. The model also gives the lowest value of RRMSE (0.0611), indicating its effectiveness for the relevance of the data. The error in the prediction of the directions of the data of the model is also much less, at MAE (0.1522), compared to all other traditional approaches. It has a variance of 92% for the data (R2 0.9233), proving that the proposed model has the highest efficacy among all other traditional methods, making it the most suitable for the prediction of the solar surface.
The experimental results showed that the proposed model (CNN-LSTM, CNN-GRU) gives an accuracy of 96% in predicting the data points compared to the individual deep learning models. The relevant features selected through the proposed method can predict surface exposure with high precision.
In the future, the model can be extended to predict the surface exposure for a larger set of areas based on the availability of the required data set. The model can be trained for more complex environmental factors based on improved datasets, and prediction accuracy can be improved for larger areas.

Author Contributions

Conceptualization, F.K. and K.O.; methodology, F.K. and K.O.; software, F.K.; validation, F.K. and K.O.; formal analysis, F.K.; investigation, F.K.; resources, F.K.; data curation, F.K.; writing—original draft preparation, F.K.; writing—review and editing, F.K. and K.O.; visualization, F.K.; supervision, K.O.; project administration, K.O.; funding acquisition, F.K. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

In the experiments publicly available “Solar Radiation and Climate Experiment (SORCE)” is used which is accessible through “https://eospso.nasa.gov/missions/solar-radiation-and-climate-experiment, accessed on 2 November 2024”.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Li, P.; Zhou, K.; Lu, X.; Yang, S. A hybrid deep learning model for short-term PV power forecasting. Appl. Energy 2020, 259, 114216. [Google Scholar] [CrossRef]
  2. Mayer, M.; Gróf, G. Extensive comparison of physical models for photovoltaic power forecasting. Appl. Energy 2021, 283, 116239. [Google Scholar] [CrossRef]
  3. Wang, K.; Qi, X.; Liu, H. A comparison of day-ahead photovoltaic power forecasting models based on deep learning neural network. Appl. Energy 2019, 251, 113315. [Google Scholar] [CrossRef]
  4. Colak, I.; Yesilbudak, M.; Genc, N.; Bayindir, R. Multi-period prediction of solar radiation using ARMA and ARIMA models. In Proceedings of the IEEE 14th International Conference on Machine Learning and Applications (ICMLA), Miami, FL, USA, 9–11 December 2015. [Google Scholar]
  5. Alharbi, F.; Csala, D. A Seasonal Autoregressive Integrated Moving Average with Exogenous Factors (SARIMAX) Forecasting Model-Based Time Series Approach. Inventions 2022, 7, 94. [Google Scholar] [CrossRef]
  6. Lim, S.C.; Huh, J.H.; Hong, S.H.; Park, C.Y.; Kim, J.C. Solar Power Forecasting Using CNN-LSTM Hybrid Model. Energies 2022, 15, 8233. [Google Scholar] [CrossRef]
  7. Brahma, B.; Wadhvani, R. Solar irradiance forecasting based on deep learning methodologies and multi-site data. Symmetry 2020, 12, 1830. [Google Scholar] [CrossRef]
  8. Jalali, S.; Ahmadian, S.; Kavousi-Fard, A.; Khosravi, A.; Nahavandi, S. Automated deep CNN-LSTM architecture design for solar irradiance forecasting. IEEE Trans. Syst. Man Cybern. 2021, 52, 54–65. [Google Scholar] [CrossRef]
  9. Zang, H.; Liu, L.; Sun, L.; Cheng, L.; Wei, Z.; Sun, G. Short-term global horizontal irradiance forecasting based on a hybrid CNN-LSTM model with spatiotemporal correlations. Renew. Energy 2020, 160, 26–41. [Google Scholar] [CrossRef]
  10. Jebli, I.; Belouadha, F.; Kabbaj, M.; Tilioua, A. Prediction of solar energy guided by pearson correlation using machine learning. Energy 2021, 224, 120109. [Google Scholar] [CrossRef]
  11. Husein, M.; Chung, I. Day-ahead solar irradiance forecasting for microgrids using a long short-term memory recurrent neural network: A deep learning approach. Energies 2019, 12, 1856. [Google Scholar] [CrossRef]
  12. Alzahrani, A.; Shamsi, P.; Dagli, C.; Ferdowsi, M. Solar Irradiance Forecasting Using Deep Neural Networks. Procedia Comput. Sci. 2017, 114, 304–313. [Google Scholar] [CrossRef]
  13. Dhilip Kumar, R.; Prakash, K.; Sundari, P.A.; Sathya, S. A Hybrid Machine Learning Model for Solar Power Forecasting. E3S Web Conf. 2023, 387, 04003. [Google Scholar] [CrossRef]
  14. Sarmas, E.; Dimitropoulos, N.; Marinakis, V.; Mylona, Z.; Doukas, H. Transfer learning strategies for solar power forecasting under data scarcity. Sci. Rep. 2022, 12, 14643. [Google Scholar] [CrossRef] [PubMed]
  15. Chatterjee, S.; Byun, Y.C. A Synthetic Data Generation Technique for Enhancement of Prediction Accuracy of Electric Vehicles Demand. Sensors 2023, 23, 594. [Google Scholar] [CrossRef] [PubMed]
  16. Tan, L.; Kang, R.; Xia, J.; Wang, Y. Application of multi-source data fusion on intelligent prediction of photovoltaic power. Sol. Energy 2024, 277, 112706. [Google Scholar] [CrossRef]
  17. Hou, X.; Ju, C.; Wang, B. Prediction of solar irradiance using convolutional neural network and attention mechanism-based long short-term memory network based on similar day analysis and an attention mechanism. Heliyon 2023, 9, e21484. [Google Scholar] [CrossRef]
  18. Ozbek, A.; Yildirim, A.; Bilgili, M. Deep learning approach for one-hour ahead forecasting of energy production in a solar-PV plant. Energy Sources Part A Recover. Util. Environ. Eff. 2021, 44, 10465–10480. [Google Scholar] [CrossRef]
  19. Neshat, M.; Nezhad, M.M.; Mirjalili, S.; Garcia, D.A.; Dahlquist, E.; Gandomi, A.H. Short-term solar radiation forecasting using hybrid deep residual learning and gated LSTM recurrent network with differential covariance matrix adaptation evolution strategy. Energy 2023, 278, 127701. [Google Scholar] [CrossRef]
  20. Park, S.; Kim, Y.; Ferrier, N.; Collis, S.; Sankaran, R.; Beckman, P. Prediction of Solar Irradiance and Photovoltaic Solar Energy Product Based on Cloud Coverage Estimation Using Machine Learning Methods. Atmosphere 2021, 12, 395. [Google Scholar] [CrossRef]
  21. Zhang, L.; Wilson, R.; Sumner, M.; Wu, Y. Advanced multimodal fusion method for very short-term solar irradiance forecasting using sky images and meteorological data: A gate and transformer mechanism approach. Renew. Energy J. 2023, 216, 118952. [Google Scholar] [CrossRef]
  22. Yuzer, E.O.; Bozkurt, A. Deep learning model for regional solar radiation estimation using satellite images. Ain Shams Eng. J. 2020, 14, 102057. [Google Scholar] [CrossRef]
  23. Mathiesen, P.; Collier, C.; Kleissl, J. A high-resolution, cloud-assimilating numerical weather prediction model for solar irradiance forecasting. Sol. Energy J. 2013, 92, 47–61. [Google Scholar] [CrossRef]
  24. Hedar, A.R.; Almaraashi, M.; Abdel-Hakim, A.; Abdulrahim, M. Hybrid Machine Learning for Solar Radiation Prediction in Reduced Feature Spaces. Energies 2021, 14, 7970. [Google Scholar] [CrossRef]
  25. Said, Y.A. AI-based solar energy forecasting for smart grid integration. Neural Comput. Appl. 2023, 35, 8625–8634. [Google Scholar] [CrossRef]
  26. Huang, C.; Wang, L.; Lai, L.L. Data-Driven Short-Term Solar Irradiance Forecasting Based on Information of Neighboring Sites. IEEE Trans. Ind. Electron. 2019, 66, 9918–9927. [Google Scholar] [CrossRef]
  27. Yu, Y.; Si, X.; Hu, C.; Zhang, J. A review of recurrent neural networks: LSTM cells and network architectures. Neural Comput. 2019, 31, 1235–1270. [Google Scholar] [CrossRef]
  28. Sherstinsky, A. Fundamentals of recurrent neural network (RNN) and long short-term memory (LSTM) network. Phys. D Nonlinear Phenom. 2020, 404, 132306. [Google Scholar] [CrossRef]
Figure 1. Proposed Model.
Figure 1. Proposed Model.
Sensors 24 08059 g001
Figure 2. Time series analysis for a specific time.
Figure 2. Time series analysis for a specific time.
Sensors 24 08059 g002
Figure 3. Spectral analysis.
Figure 3. Spectral analysis.
Sensors 24 08059 g003
Figure 4. Regression analysis.
Figure 4. Regression analysis.
Sensors 24 08059 g004
Figure 5. Time series visualization for different timestamps.
Figure 5. Time series visualization for different timestamps.
Sensors 24 08059 g005
Figure 6. Surface exposure prediction using CNN-LSTM models.
Figure 6. Surface exposure prediction using CNN-LSTM models.
Sensors 24 08059 g006
Figure 7. Surface exposure prediction using different models.
Figure 7. Surface exposure prediction using different models.
Sensors 24 08059 g007
Figure 8. The CNN-GRU model’s time series visualization for different timestamps.
Figure 8. The CNN-GRU model’s time series visualization for different timestamps.
Sensors 24 08059 g008
Figure 9. The CNN-LSTM model’s time series visualization for different timestamps.
Figure 9. The CNN-LSTM model’s time series visualization for different timestamps.
Sensors 24 08059 g009
Figure 10. The CNN-LSTM model’s training history and prediction success.
Figure 10. The CNN-LSTM model’s training history and prediction success.
Sensors 24 08059 g010
Figure 11. CNN-GRU model’s training history and prediction success.
Figure 11. CNN-GRU model’s training history and prediction success.
Sensors 24 08059 g011
Figure 12. CNN-GRU and CNN-LSTM ROC curves.
Figure 12. CNN-GRU and CNN-LSTM ROC curves.
Sensors 24 08059 g012
Table 1. Literature analysis for different approaches.
Table 1. Literature analysis for different approaches.
No.TitleAuthorsYearMethodologyKey Findings
1Solar Irradiance ForecastingAhmad et al. [12]2017CNNs with data augmentationAchieved high accuracy in short-term solar radiation prediction
2A Hybrid Machine Learning ModelDhilip et al. [13]2023CNN-RNN hybrid modelCombined temporal and spatial features for improved predictions
3Transfer learning strategiesSaramas et al. [14]2022Transfer learning with pre-trained modelsEnhanced model performance with limited training data
4A Synthetic Data Generation TechniqueByun et al. [15]2022Synthetic data generation and CNNsAddressed data scarcity issues, improving prediction reliability
5Application of multi-source data fusionLing et al. [16]2024Integration of satellite and ground dataSignificantly improved accuracy by combining multiple data sources
6Prediction of solar irradiance using convolutional neural networkChao et al. [17]2023CNNsFocused on high-resolution imagery to capture fine details in solar exposure
7Deep learning approach for one-hour ahead forecastingPatel et al. [18]2021CNNs and regression modelsImproved solar panel efficiency through accurate yield estimation
8Short-term solar radiation forecastingMehdi et al. [19]2023Hybrid CNN-RNNDeveloped robust models integrating spatial-temporal data
9Prediction of Solar Irradiance and Photovoltaic Solar EnergyYonghua et al. [20]2021CNNsAchieved high accuracy in diverse weather conditions
10Advanced multimodal fusion methodLwengo et al. [21]2023Transfer learning and data augmentationEnhanced model generalizability across different regions
11Deep learning model for regional solar radiationErsan et al. [22]2020RNNs with feature engineeringImproved prediction accuracy through advanced feature extraction
12A high-resolution, cloud-assimilating numerical weather prediction modelPatrick et al. [23]2013CNNs with high-resolution imageryFocused on capturing small-scale variations in solar exposure
13Hybrid Machine Learning for Solar Radiation PredictionHeydar et al. [24]2021CNN-RNN hybrid modelsCombined strengths of CNNs and RNNs for superior predictions
14AI-based solar energy forecasting for smart grid integrationSaid et al. [25]2023AI and machine learning techniquesLeveraged AI to enhance predictive accuracy and model robustness
15Data-Driven Short-Term Solar Irradiance ForecastingHuang et al. [26]2019Data-driven approaches with deep learningAchieved significant improvements in prediction reliability and accuracy
Table 2. SORCE Dataset Specification.
Table 2. SORCE Dataset Specification.
Ch#NameWavelength Range (nm)Spatial Res. (km)Spectral Res. (nm)Description
Ch#1UV100–400101Ultraviolet (UV) radiation measurement is typically used for ozone monitoring or atmospheric composition studies
Ch#2Visible400–700101Visible light measurement is used for surface reflectance, vegetation monitoring, and cloud detection
Ch#3NIR700–1100101Near-infrared (NIR) measurement is valuable for vegetation health monitoring, land cover classification, and moisture content
Ch#4SWIR1100–2500101Shortwave infrared (SWIR) measurement is useful for geological mapping, vegetation analysis, and soil moisture estimation
Ch#5MWIR2500–5000101Midwave infrared (MWIR) measurement is often utilized for temperature mapping, fire detection, and atmospheric profiling
Ch#6LWIR8000–12,000101Longwave infrared (LWIR) measurement is critical for cloud characterization, sea surface temperature monitoring, and more
Ch#7CO2400–750100.1Carbon dioxide (CO2) absorption band is used for atmospheric composition studies and greenhouse gas monitoring
Ch#8O3250–350100.1The ozone (O3) absorption band is crucial for stratospheric ozone monitoring and atmospheric chemistry research
Ch#9CH41900–2100100.1The methane (CH4) absorption band is significant for monitoring atmospheric methane concentrations and sources
Ch#10H2O900–1000100.1Water vapor (H2O) absorption band is important for studying humidity distribution, cloud formation, and precipitation
Ch#11Aerosol500–750100.1Aerosol optical depth measurement is essential for air quality monitoring, climate studies, and atmospheric modeling
Ch#12Cloud800–1400100.1Cloud properties are retrieved, including cloud top temperature, cloud phase, and cloud height
Ch#13Albedo340–2400101Surface albedo measurement is used for climate modeling, energy balance studies, and land surface characterization
Ch#14Thermal8000–12,000101Thermal infrared measurement is crucial for land surface temperature estimation, urban heat island detection, and more
Ch#15Vegetation IndexN/A10N/ADerived index combining multiple spectral bands is used to assess vegetation health and density
Ch#16Land Surface TemperatureN/A10N/ADerived temperature values represent the temperature of the Earth’s surface
Table 3. Features of SORCE datasets.
Table 3. Features of SORCE datasets.
Feature NameDescriptionUnit
Solar_ExposureTotal solar exposure measured in Earth’s atmosphereWatts per square meter (W/m2)
UV_IndexUltraviolet index measuring UV radiation intensityUnitless
Visible_LightVisible light intensityLux
IR_ExposureInfrared exposureWatts per square meter (W/m2)
Ozone_ConcentrationThe concentration of ozone in the atmosphereDobson units
Water_Vapor_ConcentrationConcentration of water vapor in the atmosphereGrams per cubic meter (g/m3)
Surface_TemperatureTemperature of the Earth’s surfaceCelsius (°C)
Atmospheric_PressureThe pressure exerted by the atmosphereHectopascals (hPa)
Solar_Activity_IndexIndex measuring solar activity and sunspotsUnitless
Solar_FluxSolar flux measurementsWatts per square meter (W/m2)
Aerosol_Optical_DepthMeasure of aerosol particles in the atmosphereUnitless
Cloud_CoverPercentage of sky covered by cloudsPercent (%)
Wind_SpeedSpeed of wind at Earth’s surfaceMeters per second (m/s)
Precipitation_RateRate of precipitation (rainfall or snowfall)Millimeters per hour (mm/hr)
Sea_Surface_TemperatureTemperature of the sea surfaceCelsius (°C)
Ocean_Current_SpeedSpeed of ocean currentsMeters per second (m/s)
Chlorophyll_ConcentrationConcentration of chlorophyll in waterMilligrams per cubic meter (mg/m3)
Photosynthetically_Active _RadiationSolar radiation used by plantsMicromoles per square meter per second (µmol/m2/s)
Phytoplankton _ConcentrationConcentration of phytoplankton in waterCells per liter (cells/L)
Fish_Population_DensityDensity of fish population in waterFish per cubic meter (fish/m3)
Algae_Bloom_AreaArea covered by algae bloomsSquare kilometers (km2)
Primary_Production_RateRate of primary production in marine ecosystemsGrams of carbon per square meter per year (gC/m2/yr)
Temperature_AnomalyAnomaly in surface temperature compared to baselineCelsius (°C)
Sea_Level_RiseRise in sea levelMillimeters (mm)
Glacier_Mass_BalanceChange in mass of glaciersMeters water equivalent (m w.e.)
Ocean_AcidificationDecrease in pH levels of oceanspH units
Online_Education _EnrollmentEnrollment in online education coursesMillions
Remote_Work_PracticesAdoption of remote work practicesPercentage (%)
Telemedicine_UsageUsage of telemedicine servicesConsultations
Virtual_Reality_AdoptionAdoption of virtual reality technologiesPercentage (%)
Space_Exploration_BudgetBudget allocated to space explorationBillion USD
Mars_Colonization_ProjectsProjects related to the colonization of MarsCount
AI_Satellite_LaunchesSatellite launches for AI applicationsCount
Climate_Change _Adaptation_ProjectsProjects addressing climate change adaptationCount
Renewable_Energy _InvestmentsInvestments in renewable energy projectsBillion USD
Table 4. Input parameters for model design.
Table 4. Input parameters for model design.
ModelStructureValue
CNN-LSTMInput6 × 6 × K × F
Convolution4 × 4 Filter
8 Filters
Zero Padding
Tanh Activation
Max Pooling4 × 4 Filter
2 Strides
LSTM16 Nodes
Tanh Activation
OutputN × 1
MSE
CNN-GRUInput6 × 6 × K × F
Convolution4 × 4 Filter
8 Filters
Zero Padding
Tanh Activation
Max Pooling4 × 4 Filter
2 Strides
GRU16 Nodes
Tanh Activation
OutputN × 1
MSE
Table 5. Performance parameter computation for CNN-LSTM and CNN-GRU.
Table 5. Performance parameter computation for CNN-LSTM and CNN-GRU.
ModelMethodRMSERRMSER2MAE
Convolutional Neural Network
(Long Short-Term Memory)
Full Features0.249820.128060.91960.08241
Fisher Score0.339460.293650.74680.11524
Recursive Feature Elimination (RFE)0.123230.066040.880340.11519
Random Forest Importance0.383230.37550.990970.14686
Deep Feature Selection (LSTM)0.432980.357140.754550.09747
Deep Feature Selection (GRU)0.173360.132920.857430.03005
Convolutional Neural Network
(Gated Recurrent Unit)
Full Features0.272780.211260.883560.07441
Fisher Score0.15580.120570.809910.02427
Recursive Feature Elimination (RFE)0.282430.158210.75990.07977
Random Forest Importance0.305690.191970.713940.09345
Deep Feature Selection (LSTM)0.343020.293050.719520.05766
Deep Feature Selection (GRU)0.479550.243970.942520.22997
Table 6. Comparative performance parameter analysis for CNN-LSTM and CNN-GRU.
Table 6. Comparative performance parameter analysis for CNN-LSTM and CNN-GRU.
MethodModelPerformance Parameters
RMSERRMSER2MAE
Deep FS
(Long Short-Term Memory)
Artificial Neural Network0.2498160.1280640.9195980.062408
Convolutional Neural Network0.1232330.0660350.8803350.015186
Long Short-Term Memory0.4329770.3571420.7545470.187469
Gated Recurrent Unit0.2727780.2112550.8835560.074408
CNN (LSTM)0.0824280.0182070.7599020.049766
CNN (GRU)0.3430180.2930460.9195150.117661
Deep FS
(Gated Recurrent Unit)
Artificial Neural Network0.3394630.2936490.7467980.115235
Convolutional Neural Network0.3832290.375520.9609730.146864
Long Short-Term Memory0.1733620.1329210.8574270.030054
Gated Recurrent Unit0.1557980.1205730.8099090.024273
CNN (LSTM)0.1256940.1019690.7139350.013449
CNN (GRU)0.4795540.2439690.9625190.229972
Table 7. Comparative analysis of the proposed method.
Table 7. Comparative analysis of the proposed method.
ModelRMSERRMSEMAER2
ARIMA0.35430.11040.25440.8511
SARIMA0.32340.08230.22340.8834
CNN (LSTM)0.25160.07230.18120.9567
CNN (GRU)0.22010.06110.15220.9233
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Kihtir, F.; Oztoprak, K. Deep FS: A Deep Learning Approach for Surface Solar Radiation. Sensors 2024, 24, 8059. https://doi.org/10.3390/s24248059

AMA Style

Kihtir F, Oztoprak K. Deep FS: A Deep Learning Approach for Surface Solar Radiation. Sensors. 2024; 24(24):8059. https://doi.org/10.3390/s24248059

Chicago/Turabian Style

Kihtir, Fatih, and Kasim Oztoprak. 2024. "Deep FS: A Deep Learning Approach for Surface Solar Radiation" Sensors 24, no. 24: 8059. https://doi.org/10.3390/s24248059

APA Style

Kihtir, F., & Oztoprak, K. (2024). Deep FS: A Deep Learning Approach for Surface Solar Radiation. Sensors, 24(24), 8059. https://doi.org/10.3390/s24248059

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop