CN111598225B - Air conditioner cold load prediction method based on self-adaptive deep confidence network - Google Patents
Air conditioner cold load prediction method based on self-adaptive deep confidence network Download PDFInfo
- Publication number
- CN111598225B CN111598225B CN202010415187.7A CN202010415187A CN111598225B CN 111598225 B CN111598225 B CN 111598225B CN 202010415187 A CN202010415187 A CN 202010415187A CN 111598225 B CN111598225 B CN 111598225B
- Authority
- CN
- China
- Prior art keywords
- layer
- training
- network
- cold load
- visual
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
- 238000000034 method Methods 0.000 title claims abstract description 39
- 230000000007 visual effect Effects 0.000 claims abstract description 56
- 238000012549 training Methods 0.000 claims abstract description 54
- 238000009826 distribution Methods 0.000 claims abstract description 27
- 238000011156 evaluation Methods 0.000 claims abstract description 17
- 101001095088 Homo sapiens Melanoma antigen preferentially expressed in tumors Proteins 0.000 claims abstract description 14
- 102100037020 Melanoma antigen preferentially expressed in tumors Human genes 0.000 claims abstract description 14
- ORILYTVJVMAKLC-UHFFFAOYSA-N Adamantane Natural products C1C(C2)CC3CC1CC2C3 ORILYTVJVMAKLC-UHFFFAOYSA-N 0.000 claims abstract description 12
- 238000005265 energy consumption Methods 0.000 claims abstract description 11
- 238000012545 processing Methods 0.000 claims abstract description 11
- 238000005457 optimization Methods 0.000 claims abstract description 10
- 238000013528 artificial neural network Methods 0.000 claims abstract description 9
- 230000002159 abnormal effect Effects 0.000 claims abstract description 8
- 230000007246 mechanism Effects 0.000 claims abstract description 6
- 238000010606 normalization Methods 0.000 claims abstract description 6
- 230000001105 regulatory effect Effects 0.000 claims abstract description 4
- 239000013598 vector Substances 0.000 claims description 17
- 230000004913 activation Effects 0.000 claims description 10
- 230000003044 adaptive effect Effects 0.000 claims description 8
- 238000001816 cooling Methods 0.000 claims description 8
- 239000000284 extract Substances 0.000 claims description 4
- 238000012360 testing method Methods 0.000 claims description 4
- 239000011159 matrix material Substances 0.000 claims description 3
- 230000005855 radiation Effects 0.000 claims description 3
- 238000000638 solvent extraction Methods 0.000 claims description 3
- 230000003213 activating effect Effects 0.000 claims description 2
- 230000006870 function Effects 0.000 description 30
- 230000000694 effects Effects 0.000 description 7
- 238000004458 analytical method Methods 0.000 description 5
- 230000008569 process Effects 0.000 description 5
- 238000004364 calculation method Methods 0.000 description 4
- 238000007726 management method Methods 0.000 description 4
- 238000003066 decision tree Methods 0.000 description 2
- 238000010586 diagram Methods 0.000 description 2
- 238000004134 energy conservation Methods 0.000 description 2
- 238000002474 experimental method Methods 0.000 description 2
- 238000012706 support-vector machine Methods 0.000 description 2
- 238000004378 air conditioning Methods 0.000 description 1
- 230000002547 anomalous effect Effects 0.000 description 1
- 238000013459 approach Methods 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 230000001276 controlling effect Effects 0.000 description 1
- 238000012937 correction Methods 0.000 description 1
- 238000013136 deep learning model Methods 0.000 description 1
- 230000007547 defect Effects 0.000 description 1
- 238000012217 deletion Methods 0.000 description 1
- 230000037430 deletion Effects 0.000 description 1
- 238000013461 design Methods 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 239000005431 greenhouse gas Substances 0.000 description 1
- 238000010438 heat treatment Methods 0.000 description 1
- 230000000977 initiatory effect Effects 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000010248 power generation Methods 0.000 description 1
- 230000009467 reduction Effects 0.000 description 1
- 238000011160 research Methods 0.000 description 1
- 238000009423 ventilation Methods 0.000 description 1
- 239000002699 waste material Substances 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/044—Recurrent networks, e.g. Hopfield networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
- G06N3/084—Backpropagation, e.g. using gradient descent
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
- G06N3/088—Non-supervised learning, e.g. competitive learning
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y04—INFORMATION OR COMMUNICATION TECHNOLOGIES HAVING AN IMPACT ON OTHER TECHNOLOGY AREAS
- Y04S—SYSTEMS INTEGRATING TECHNOLOGIES RELATED TO POWER NETWORK OPERATION, COMMUNICATION OR INFORMATION TECHNOLOGIES FOR IMPROVING THE ELECTRICAL POWER GENERATION, TRANSMISSION, DISTRIBUTION, MANAGEMENT OR USAGE, i.e. SMART GRIDS
- Y04S10/00—Systems supporting electrical power generation, transmission or distribution
- Y04S10/50—Systems or methods supporting the power network operation or management, involving a certain degree of interaction with the load-side end user applications
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Health & Medical Sciences (AREA)
- Computing Systems (AREA)
- Biomedical Technology (AREA)
- Biophysics (AREA)
- Computational Linguistics (AREA)
- Data Mining & Analysis (AREA)
- Evolutionary Computation (AREA)
- Life Sciences & Earth Sciences (AREA)
- Molecular Biology (AREA)
- Artificial Intelligence (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Health & Medical Sciences (AREA)
- Management, Administration, Business Operations System, And Electronic Commerce (AREA)
- Air Conditioning Control Device (AREA)
Abstract
The invention discloses an air conditioner cold load prediction method based on a self-adaptive deep confidence network, which comprises the steps of firstly collecting cold load data, adopting a Lagrange interpolation method to compensate for missing and abnormal energy consumption data, and carrying out normalization processing on the processed energy consumption data; the processed data is processed through independent Gaussian distribution, and the processed CRBM is input to a prediction model; training is carried out through an RBM non-supervision mechanism, an RBM hidden layer of a previous layer is used as a visual layer input of an RBM of a next layer, and network parameters are regulated; then reverse training is carried out through a supervised BP neural network, and network parameters are regulated again; then adopting Adam optimization algorithm to adjust network parameters; finally, selecting parameters and structures of the prediction model, and carrying out structure evaluation selection on the prediction model by adopting a reconstruction error RE; and evaluating the result by adopting a root mean square relative error (RMSPE) and an average absolute percentage error (MAPE) to finish the air conditioner cold load prediction. The method has good prediction precision, universality and applicability.
Description
Technical Field
The invention belongs to the technical field of building air conditioner cold load prediction, and particularly relates to an air conditioner cold load prediction method based on a self-adaptive deep confidence network.
Background
With the increasing urbanization, industry development and population growth, global energy demand continues to rise. It is counted that building operating cold loads account for approximately 40% of global cold loads, with greenhouse gas emissions accounting for 1/3 of the total emissions, indicating that buildings have become the largest energy consumer. For buildings, the heating ventilation and air conditioning system occupies the largest proportion in the total cold load of the building, and the huge energy consumption increases the pressure of the power grid. Therefore, the energy management of the building is particularly important.
The prediction of the building cold load is an important part in the building energy management process, is a key work for realizing building energy conservation, correctly and reasonably predicts the building cold load, can timely and accurately discover some abnormal conditions or potential equipment faults in the building cold load, and is convenient for management personnel to take measures in time so as to avoid excessive waste of energy. Meanwhile, the accurate and reasonable prediction of the building cold load can provide a certain basis for the management personnel to reasonably distribute energy, so that the energy is reasonably and effectively used. And the power generation scheme can be reasonably arranged, so that the supply and demand balance of the power grid is realized, and the power system can stably operate, which is also a very important ring for the power system.
The artificial neural network model, the support vector machine model, the decision tree model and the mixed model in the existing cold load prediction method are widely applied. However, these models suffer from various problems that make the prediction accuracy less than ideal. For example, the learning speed of the artificial neural network algorithm is not high enough, and local optimization and over-fitting phenomena are easy to generate; the support vector machine is difficult to use in a large number of samples, and the problem of multi-classification is difficult to solve; the scheme probability of the decision tree model is easily affected by human factors, and the decision accuracy is reduced; and the problems of high calculation amount and high calculation difficulty of the mixed model and the like. The above prediction algorithm only focuses on the prediction of the self algorithm on the cold load, but cannot fully consider the complex characteristics of the cold load of the building, so that the prediction effect is not ideal.
Disclosure of Invention
Aiming at the defects in the prior art, the invention provides the air conditioner cold load prediction method based on the self-adaptive deep confidence network, which considers complex factors influencing the cold load of the building, realizes the effective prediction of the cold load condition of the building, achieves the technical effects of energy conservation and consumption reduction by controlling the influencing conditions, and solves the problems in the prior art.
The invention adopts the following technical scheme:
the air conditioner cold load prediction method based on the self-adaptive deep confidence network comprises the steps of firstly collecting cold load data, adopting a Lagrange interpolation method to compensate the missing and abnormal energy consumption data, and carrying out normalization processing on the processed energy consumption data; the processed data is processed through independent Gaussian distribution, and the processed CRBM is input to a prediction model; training by using an RBM non-supervision mechanism, taking an RBM hidden layer of a previous layer as a visual layer input of an RBM of a next layer, and adjusting network parameters theta= { w, a, b }; then, reverse training is carried out through a supervised BP neural network, and network parameters theta= { w, a, b } are regulated again; then adopting an Adam optimization algorithm to adjust network parameters theta= { w, a, b }; finally, selecting parameters and structures of the prediction model, and carrying out structure evaluation selection on the prediction model by adopting a reconstruction error RE; and evaluating the result by adopting a root mean square relative error (RMSPE) and an average absolute percentage error (MAPE) to finish the air conditioner cold load prediction.
Specifically, the prediction model adds continuous values of independent Gaussian distribution in the linear unit to simulate real data, and the energy function E (v, h; θ) is as follows:
wherein θ= { w, a, b, σ }, σ i For visual layer v i Standard deviation of corresponding gaussian noise;
the activation probabilities of the update visual layer and the hidden layer are:
wherein N (μ, σ) 2 ) Representing the mean μ and variance σ of the gaussian function 2 ,a i H is the offset of the visual layer j To hide the binary state of cell j, w ij For the weight between the two, n is the number of the predicted cold load of the visual layer, v i Inputting the binary state of i for the visual layer, b j Is the offset of the hidden layer.
Specifically, in the training of the RBM non-supervision mechanism, the RBM converts an input vector from a visual layer to a hidden layer through an activation function, from the hidden layer to the visual layer, and extracts features through training to minimize an internal energy function, so as to obtain a joint configuration energy function E (v, h; theta) between the visual layer and the hidden layer, ensure function distribution standardization through joint distribution of the visual layer and the hidden layer, calculate conditional probability distribution according to the determined visual layer and the hidden layer, and determine an optimal model through maximizing a likelihood function by using a ladder-lifting method, and when a set of training sample sets S= { v is given 1 ,v 2 ,...v n ) And when the objective function is calculated as the maximized lower log likelihood function Ls.
Further, the log-likelihood function Ls under maximization is:
wherein P (v) n ) For the activation probability of each input sample, N is the number of training samples.
Further, the joint distribution between the visual layer and the hidden layer is:
P(v,h|θ)=exp(-E(v,h))/Z(θ)
wherein Z (θ) is a partitioning function, v is a visual layer unit, h is a hidden layer unit, and after the visual layer and the hidden layer are determined, the conditional probability distribution is calculated as follows:
wherein a is i H is the offset of the visual layer j To conceal the binary state of cell j, n is the predicted amount of cooling load for the visual layer, w ij For the weight between the two, m is the number of hidden layers corresponding to the visual layer, v i Inputting the binary state of i for the visual layer, b j Is the offset of the hidden layer.
Specifically, the readjustment network parameter θ= { w, a, b } is specifically:
inputting input vectors containing time, temperature, humidity and solar radiation into a network model, performing unsupervised network training, processing into continuous values through Gaussian distribution CRBM, and performing model training;
and dividing the training samples into f groups of training samples, training the f groups of training samples to adjust network parameters theta= { w, a, b } of the prediction model, and storing the network parameters after reaching the training layer number of the prediction model.
Specifically, the Adam optimization algorithm is adopted to adjust the network parameters θ= { w, a, b } specifically:
initializing the parameter vector, the first moment vector and the second moment vector, then updating each part in a loop and iteration mode to enable the parameter theta to be converged, namely, adding 1 to the time t, correspondingly updating each parameter of the deviation, and finally updating the parameter theta of the prediction model by using the calculated parameter value.
Specifically, the root mean square relative error RMSPE and the mean absolute percentage error MAPE are used for evaluation, the difference between the predicted value and the actual value of the predicted model is compared, the CRBM-DBN network model is evaluated by adopting a reconstruction error, the deviation between the true value and the predicted value forms the reconstruction error, the depth of the predicted model is calculated by calculating the reconstruction error, the number of hidden nodes is calculated, the MAPE and the training time are used as evaluation standards, and the cold load data is predicted by training.
Further, the root mean square relative error RMSPE and the mean absolute percentage error MAPE are specifically:
wherein y is i Andthe actual load value and the load predicted value at the ith moment are respectively;The average value of the air conditioner cold load true value is k, and k is the number of samples of all test sets; the reconstruction error was used for the evaluation as follows:
wherein RE is a reconstruction error, X represents a min-batch matrix, aV i Representing the state value of the current predicted visual element.
Further, the depth of the prediction model is 2, the number of hidden layer nodes is 20, and the prediction step length is 1h.
Compared with the prior art, the invention has at least the following beneficial effects:
the invention provides an air conditioner cold load prediction method based on a self-adaptive deep confidence network, which aims at solving the problem that building cold load data often has nonlinearity and dynamic characteristics, so that the cold load data cannot be reliably and accurately predicted. The algorithm combines the advantages of supervised learning and unsupervised learning, digs high-dimensional features, extracts feature values from bottom to top, and effectively solves a series of problems caused by random initialization of parameters of the traditional neural network. In addition to having a good initiation point, the problems of over-fitting and under-fitting that often occur with neural networks are also effectively solved by pre-training.
Furthermore, the selected data is subjected to deletion and abnormal data compensation processing, so that the problem of data reliability can be effectively solved, and in order to avoid prediction errors caused by different orders of magnitude of parameters of an input layer, the input data is subjected to normalization processing, so that the prediction errors of a network model caused by the data errors can be effectively solved.
Furthermore, the Boltzmann machine is processed by Gaussian distribution, so that the problem that the RBM can only accept binary input and data is easy to lose can be effectively solved.
Furthermore, the deep confidence network optimized by adopting the Adam algorithm has strong self-adaptability and good performance in the aspect of nonlinear problem treatment due to consideration of the second derivative of the objective function, and the problem of slow convergence speed in the process of training network model parameters is effectively solved.
Further, by evaluating the network model by using the reconstruction errors, the size of the reconstruction errors generated by selecting different structures can be clearly seen, so that parameters and structures of the model can be better selected.
In conclusion, the method accurately predicts the building cold load, has prediction precision higher than that of a BP neural network prediction model, has good prediction precision, universality and applicability, is particularly suitable for large public buildings with periodically-changed cold load, and provides cold load prediction data which are more useful for energy-saving planning and energy-saving planning.
The technical scheme of the invention is further described in detail through the drawings and the embodiments.
Drawings
FIG. 1 is an exception data handling diagram;
FIG. 2 is a schematic view of an RBM;
FIG. 3 is a schematic diagram of a CRBM-DBN network prediction;
FIG. 4 is a flow chart of the CRBM-DBN algorithm;
FIG. 5 is a graph of model reconstruction errors, wherein (a) is the reconstruction error at a network depth of 1, (b) is the reconstruction error at a network depth of 2, (c) is the reconstruction error at a network depth of 3, and (d) is the reconstruction error at a network depth of 4;
fig. 6 is a graph of the prediction results, wherein (a) is a graph of the prediction of the cooling load on the rest day and the working day, (b) is an effect of enlarging for 7 to 13 hours, and (c) is an effect of enlarging for 22 to 28 hours.
Detailed Description
The invention discloses an air conditioner cooling load prediction method based on a self-adaptive deep confidence network, which comprises the following steps of:
s1, collecting cold load data (including temperature, humidity, solar load, wind speed, time and the like), adopting a Lagrange interpolation method to compensate the missing and abnormal energy consumption data, and carrying out normalization processing on the processed energy consumption data to be used as cold load energy consumption prediction;
s101, a model aims at a market cold load prediction method, and the market load is 08:00-22:00, recorded for 15 hours. Assuming that the ith data is missing or anomalous during the day, a Lagrangian function is constructed using the remaining 14 known load points during the day:
wherein j is j For the load value at the j-th time point, pj (x) is a polynomial of degree 14,B k for the remaining 14 time pointsCoordinate set, B k E {1, 2..i.1, i+1,..14 }, the load value at the i-th time of the day is y i =L 14 (i)。
Referring to fig. 1, the missing and abnormal data is compensated by using a lagrangian interpolation method, and the original abnormal data and the results before and after the missing data are processed.
S102, in view of the fact that the orders of magnitude of the parameters in the input layer are different, in order to avoid prediction errors generated by the orders, the input parameters are normalized, and a specific formula is as follows:
wherein x' is the normalization result, x is the input vector including temperature, humidity, solar load, time, x min And x max Is the minimum and maximum of the corresponding samples in the input data.
S2, the data processed in the step S1 are firstly processed through independent Gaussian distribution, and the processed CRBM is input into a prediction model to ensure the continuity of the data;
the prediction model adds continuous values of independent Gaussian distribution to the linear unit to simulate real data, so that a traditional limited Boltzmann machine can process continuous input vectors, and an energy function E (v, h; theta) is as follows:
wherein θ= { w, a, b, σ }, σ i For visual layer v i Standard deviation of corresponding gaussian noise, typically σ 2 There are better results when=1.
Thus, the activation probabilities of the updated visual and hidden layers are:
wherein N (μ, σ) 2 ) Representing the mean μ and variance σ of the gaussian function 2 。
S3, training is carried out through an RBM non-supervision mechanism, an RBM hidden layer of a previous layer is used as a visual layer input of an RBM of a next layer, and network parameters theta= { w, a, b } are adjusted for the first time;
referring to fig. 2, the rbm converts an input vector from a visual layer to a hidden layer by activating a function, from the hidden layer to the visual layer, and extracts features by training a minimized internal energy function, where a joint configuration energy function between the visual layer and the hidden layer is:
wherein v is i And h j Is the binary state of the visual layer input i and hidden element j, a i And b j Offset, w, of the visual layer and the hidden layer, respectively ij And m is the number of hidden layers corresponding to the visual layers, and n is the predicted cold load number of the visual layers.
The joint distribution between the visual layer and the hidden layer is:
P(v,h|θ)=exp(-E(v,h;θ))/Z(θ)
where Z (θ) is a partitioning function to ensure that the distribution is normalized, v is the visual layer element, and h is the hidden layer element to ensure that the distribution of the function is normalized.
After the visual layer and the hidden layer are determined, the conditional probability distribution is calculated as follows:
the purpose of RBM training is to determine the best model by maximizing likelihood function using a bench-top approach, when given a set of training sample sets s= { v 1 ,v 2 ,...v n ) When the objective function is the maximum log likelihood function:
wherein P (v) n ) For the activation probability of each input sample, N is the corresponding number of input training samples.
After the training sample set is given, adjusting the network parameters theta= { w, a, b } can adjust the activation probability of the corresponding adjusting hidden layer so as to determine the state of the hidden layer; after the hidden layer is given, the activation probability of the corresponding visual layer can be adjusted by adjusting the network parameters theta= { w, a and b } so as to reconstruct the input data, namely, the activation probability of the hidden layer and the visual layer is maximized by maximizing the log likelihood function, and therefore, the reconstruction of the data is realized.
S4, performing reverse training through a supervised BP neural network, and adjusting network parameters theta= { w, a, b };
referring to fig. 3 and 4, input vectors (including time, temperature, humidity, solar radiation, etc.) are input into a network model, firstly, the model is trained through an unsupervised network, processed into continuous values through gaussian distribution CRBM, and the model is trained according to steps S2 and S3 from bottom to top;
dividing the training samples into f groups of small-batch training samples, training the f groups of training sets to adjust network parameters theta = { w, a, b }, and after the number of training layers of the model is reached, storing the trained network parameters for the next step;
the BP algorithm reversely supervises and finely adjusts network parameters;
s5, after input data are subjected to RBM non-supervision feature training and BP supervised training learning through Gaussian distribution processing, the input data X are converted into another feature space F (Y, theta) through the model, network parameters theta= { w, a, b } are calculated through minimizing errors of F (Y, theta) and X, and a mean square error function is:
where k is the number of test set cooling loads.
Adopting an Adam optimization algorithm to adjust the network parameter theta= { w, a, b };
the Adam optimization algorithm firstly initializes the parameter vector, the first moment vector and the second moment vector, then iteratively updates each part in a circulating way to enable the parameter theta to be converged, namely, t is added with 1, and the corresponding parameter of the updating deviation is updated, and finally the calculated parameter value is used for updating the parameter theta of the model.
The Adam optimization algorithm is adopted to minimize the loss function, the Adam algorithm designs independent self-adaptive learning rate for different parameters by calculating first moment estimation and second moment estimation of gradient, and has high calculation efficiency and lower memory requirement, and the process of updating network parameters is expressed as follows:
m t =β 1 ×m t -1+(1-β 1 )×g t
wherein g t Is the gradient of the mean square error function L (theta) to theta, m t Is a first moment estimate of the gradient, n t Is a second moment estimate of the gradient,and->Respectively to m t And n t Deviation correction of moment estimated exponential decay rate beta 1 Is 0.9, beta 2 A step size eta of 0.99 and a small constant epsilon of 10 with stable numerical value of 0.001 -8 。
S6, selecting parameters and structures of the model, and performing structure evaluation selection on the model by adopting a reconstruction error RE; the results were evaluated using the root mean square relative error RMSPE and the mean absolute percent error MAPE.
The most commonly used evaluation indexes, namely root mean square relative error RMSPE and mean absolute percentage error MAPE, are used for evaluation, and the difference between the predicted value and the actual value of the prediction model is compared, and the calculation formula is as follows.
Wherein y is i Andthe actual load value and the load predicted value at the ith moment are respectively;And k is the number of samples of all test sets, and the smaller the RMSPE and MAPE values are, the better the prediction effect is.
Simple evaluation of the CRBM-DBN (Adam) network model using reconstruction errors (reconstruction error):
wherein RE is a reconstruction error, X represents a min-batch matrix, aV i The state value of the current prediction visual unit is represented, the deviation between the true value and the predicted value forms a reconstruction error, and the depth of the model can be calculated by calculating the reconstruction error.
Referring to fig. 5, the setting of the structural parameters in the deep learning model affects the training process, training time, and prediction result. In order to optimize training to reduce errors, experiments are performed by increasing the number of RBM layers layer by layer, and calculating the reconstruction errors to determine the depth of the network, wherein when the depth of the network is 2 layers, the overall reconstruction errors are reduced and gradually tend to be stable, and the average value of the reconstruction errors is smaller, so that the accuracy is higher. When the network depth is other values, the average value of the reconstruction errors is larger, and the reconstruction errors are always in a fluctuation state. Thus, the network depth of the RBM is determined to be 2 layers, i.e., 2 hidden unit layers and 1 visual unit layer in total.
And for the number of hidden nodes, MAPE and training time are taken as evaluation criteria, and are selected through experiments.
Selection of hidden node number in Table 1
As can be seen from table 1, the model prediction effect is optimal when the number of hidden layer nodes is 20.
Therefore, the depth of the prediction model set by the method is 2, the number of hidden layer nodes is 20, the prediction step length is 1h, and the cold load data is predicted through training.
The predicted daily cooling load curve is generated, after parameters and structures of a network model are obtained, the predicted daily cooling load curve can be obtained, as shown in fig. 6, fig. 6 (a) is a load prediction result of different algorithms, fig. 6 (b) (c) is a load prediction result amplified in the peak load time period of the rest day and the working day, and as the Adam optimization algorithm considers the second derivative of the objective function, the model has strong adaptability and good performance in processing nonlinear problems, and as can be seen from the figure, the method provided by the invention has better prediction effect compared with other two prediction algorithms, and the prediction accuracy in the uncertain time periods such as peak load underestimation is higher.
For the purpose of making the objects, technical solutions and advantages of the embodiments of the present invention more apparent, the technical solutions of the embodiments of the present invention will be clearly and completely described below with reference to the accompanying drawings in the embodiments of the present invention, and it is apparent that the described embodiments are some embodiments of the present invention, but not all embodiments of the present invention. The components of the embodiments of the present invention generally described and illustrated in the figures herein may be arranged and designed in a wide variety of different configurations. Thus, the following detailed description of the embodiments of the invention, as presented in the figures, is not intended to limit the scope of the invention, as claimed, but is merely representative of selected embodiments of the invention. All other embodiments, which can be made by those skilled in the art based on the embodiments of the invention without making any inventive effort, are intended to be within the scope of the invention.
Prediction model result comparison analysis
The case carries out predictive analysis on holiday and workday cold load data:
TABLE 2 workday prediction model result analysis
TABLE 3 analysis of holiday prediction model results
The most commonly used evaluation index root mean square relative error RMSPE and mean absolute percentage error MAPE are used for evaluation, and the difference between the predicted value and the actual value of the prediction model is compared.
Table 4 model evaluation analysis
As can be seen from Table 4, the prediction accuracy and time complexity are improved greatly compared with the conventional DBN and CRBM-DBN methods.
The invention takes the business building with the complex western security as a research object, and establishes the cold load prediction model based on the daily cold load of the business building.
In summary, the invention provides an adaptive deep confidence network-based air conditioner cold load prediction method, which aims at the problem of data loss caused by that RBM can only accept binary input, introduces a CRBM model with Gaussian distribution and the problem of slow convergence rate of the CRBM-DBN when training network model parameters, and provides an adaptive learning CRBM-DBN model method, wherein in each iteration, the convergence rate is improved by updating objective functions on corresponding parameters theta (w, a, b) through updating objective functions. Experimental results show that the method provided by the method is better than the traditional DBN and CRBM-DBN in terms of prediction accuracy and time complexity, and is an effective air conditioner cold load prediction method.
The above is only for illustrating the technical idea of the present invention, and the protection scope of the present invention is not limited by this, and any modification made on the basis of the technical scheme according to the technical idea of the present invention falls within the protection scope of the claims of the present invention.
Claims (7)
1. The air conditioner cold load prediction method based on the self-adaptive deep confidence network is characterized in that cold load data is collected firstly, missing and abnormal energy consumption data are subjected to compensation processing by adopting a Lagrange interpolation method, and the processed energy consumption data are subjected to normalization processing; the processed data is processed through independent Gaussian distribution, and the processed CRBM is input to a prediction model; training by using an RBM non-supervision mechanism, taking an RBM hidden layer of a previous layer as a visual layer input of an RBM of a next layer, and adjusting network parameters theta= { w, a, b }; then, reverse training is carried out through a supervised BP neural network, and network parameters theta= { w, a, b } are regulated again; then adopting an Adam optimization algorithm to adjust network parameters theta= { w, a, b }; finally, selecting parameters and structures of the prediction model, and carrying out structure evaluation selection on the prediction model by adopting a reconstruction error RE; the result is evaluated by adopting a root mean square relative error (RMSPE) and an average absolute percentage error (MAPE), and the air conditioner cold load prediction is completed;
in RBM non-supervision mechanism training, RBM converts input vector from visual layer to hidden layer by activating function, from hidden layer to visual layer, and extracts feature by minimizing internal energy function by training to obtain joint configuration energy function E (v, h; θ) between visual layer and hidden layer, and ensures function distribution standardization by joint distribution of visual layer and hidden layer, and calculates conditional probability distribution according to determined visual layer and hidden layer, and determines optimal model by maximizing likelihood function by using ladder-up method, when a set of training sample set S= { v is given 1 ,v 2 ,...v n ) When the objective function is calculated as a maximized lower log likelihood function Ls, the maximized lower log likelihood function Ls is:
wherein P (v) n ) The activation probability of each input sample is given, and N is the number of training samples;
the joint distribution between the visual layer and the hidden layer is:
P(v,h|θ)=exp(-E(v,h))/Z(θ)
wherein Z (θ) is a partitioning function, v is a visual layer unit, h is a hidden layer unit, and after the visual layer and the hidden layer are determined, the conditional probability distribution is calculated as follows:
wherein a is i H is the offset of the visual layer j To conceal the binary state of cell j, n is the predicted amount of cooling load for the visual layer, w ij For the weight between the two, m is the number of hidden layers corresponding to the visual layer, v i Inputting the binary state of i for the visual layer, b j Is the offset of the hidden layer.
2. The adaptive deep belief network-based air conditioner cold load prediction method of claim 1, wherein the prediction model adds continuous values of independent gaussian distributions to linear units to simulate real data, and the energy function E (v, h; θ) is:
wherein θ= { w, a, b, σ }, σ i For visual layer v i Standard deviation of corresponding gaussian noise;
the activation probabilities of the update visual layer and the hidden layer are:
wherein N (μ, σ) 2 ) Representing the mean μ and variance σ of the gaussian function 2 ,a i H is the offset of the visual layer j To hide the binary state of cell j, w ij For the weight between the two, n is the number of the predicted cold load of the visual layer, v i Inputting the binary state of i for the visual layer, b j Is the offset of the hidden layer.
3. The adaptive deep belief network-based air conditioner cooling load prediction method according to claim 1, wherein the readjusting network parameter θ= { w, a, b } is specifically:
inputting input vectors containing time, temperature, humidity and solar radiation into a network model, performing unsupervised network training, processing into continuous values through Gaussian distribution CRBM, and performing model training;
and dividing the training samples into f groups of training samples, training the f groups of training samples to adjust network parameters theta= { w, a, b } of the prediction model, and storing the network parameters after reaching the training layer number of the prediction model.
4. The adaptive deep belief network-based air conditioner cold load prediction method according to claim 1, wherein the adjusting network parameters θ= { w, a, b } by Adam optimization algorithm is specifically:
initializing the parameter vector, the first moment vector and the second moment vector, then updating each part in a loop and iteration mode to enable the parameter theta to be converged, namely, adding 1 to the time t, correspondingly updating each parameter of the deviation, and finally updating the parameter theta of the prediction model by using the calculated parameter value.
5. The adaptive deep belief network-based air conditioner cold load prediction method according to claim 1, wherein the root mean square relative error RMSPE and the mean absolute percentage error MAPE are used for evaluation, the difference between the predicted value and the actual value of the prediction model is compared, the CRBM-DBN network model is evaluated by adopting the reconstruction error, the deviation between the actual value and the predicted value constitutes the reconstruction error, the depth of the prediction model is calculated by calculating the reconstruction error, and the cold load data is predicted by training with the MAPE and the training time as evaluation criteria for the number of hidden nodes.
6. The adaptive deep belief network-based air conditioner cold load prediction method according to claim 5, wherein the root mean square relative error RMSPE and the mean absolute percentage error MAPE are specifically:
wherein y is i Andthe actual load value and the negative value at the i-th moment respectivelyA load prediction value;The average value of the air conditioner cold load true value is k, and k is the number of samples of all test sets; the reconstruction error was used for the evaluation as follows: />
Wherein RE is a reconstruction error, X represents a min-batch matrix, aV i Representing the state value of the current predicted visual element.
7. The adaptive deep belief network-based air conditioner cold load prediction method according to claim 5, wherein the depth of the prediction model is 2, the number of hidden layer nodes is 20, and the prediction step length is 1h.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010415187.7A CN111598225B (en) | 2020-05-15 | 2020-05-15 | Air conditioner cold load prediction method based on self-adaptive deep confidence network |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010415187.7A CN111598225B (en) | 2020-05-15 | 2020-05-15 | Air conditioner cold load prediction method based on self-adaptive deep confidence network |
Publications (2)
Publication Number | Publication Date |
---|---|
CN111598225A CN111598225A (en) | 2020-08-28 |
CN111598225B true CN111598225B (en) | 2023-05-02 |
Family
ID=72189761
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202010415187.7A Active CN111598225B (en) | 2020-05-15 | 2020-05-15 | Air conditioner cold load prediction method based on self-adaptive deep confidence network |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN111598225B (en) |
Families Citing this family (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112363099B (en) * | 2020-10-30 | 2023-05-09 | 天津大学 | TMR current sensor temperature drift and geomagnetic field correction device and method |
CN113177355B (en) * | 2021-04-28 | 2024-01-12 | 南方电网科学研究院有限责任公司 | Power load prediction method |
CN113449467A (en) * | 2021-06-21 | 2021-09-28 | 清华大学 | JDAN-NFN-based online security evaluation method and device for power system |
CN113776164B (en) * | 2021-09-09 | 2022-08-30 | 广州珠江新城能源有限公司 | Automatic control method and control system for centralized cooling system |
CN114814645B (en) * | 2022-06-27 | 2022-09-13 | 广东电网有限责任公司佛山供电局 | On-line analysis method and related device for capacitor voltage transformer |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2015172560A1 (en) * | 2014-05-16 | 2015-11-19 | 华南理工大学 | Central air conditioner cooling load prediction method based on bp neural network |
CN108549960A (en) * | 2018-04-20 | 2018-09-18 | 国网重庆市电力公司永川供电分公司 | A kind of 24 hours Methods of electric load forecasting |
CN108665001A (en) * | 2018-05-10 | 2018-10-16 | 河南工程学院 | It is a kind of based on depth confidence network across subject Idle state detection method |
CN110580543A (en) * | 2019-08-06 | 2019-12-17 | 天津大学 | Power load prediction method and system based on deep belief network |
CN110783964A (en) * | 2019-10-31 | 2020-02-11 | 国网河北省电力有限公司 | Risk assessment method and device for static security of power grid |
Family Cites Families (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9690898B2 (en) * | 2015-06-25 | 2017-06-27 | Globalfoundries Inc. | Generative learning for realistic and ground rule clean hot spot synthesis |
-
2020
- 2020-05-15 CN CN202010415187.7A patent/CN111598225B/en active Active
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2015172560A1 (en) * | 2014-05-16 | 2015-11-19 | 华南理工大学 | Central air conditioner cooling load prediction method based on bp neural network |
CN108549960A (en) * | 2018-04-20 | 2018-09-18 | 国网重庆市电力公司永川供电分公司 | A kind of 24 hours Methods of electric load forecasting |
CN108665001A (en) * | 2018-05-10 | 2018-10-16 | 河南工程学院 | It is a kind of based on depth confidence network across subject Idle state detection method |
CN110580543A (en) * | 2019-08-06 | 2019-12-17 | 天津大学 | Power load prediction method and system based on deep belief network |
CN110783964A (en) * | 2019-10-31 | 2020-02-11 | 国网河北省电力有限公司 | Risk assessment method and device for static security of power grid |
Non-Patent Citations (2)
Title |
---|
基于深层神经网络的电力负荷预测;何琬等;《环境与可持续发展》(第01期);全文 * |
基于深度信念网络的网络流量预测模型;任玮;《山西电子技术》(第01期);全文 * |
Also Published As
Publication number | Publication date |
---|---|
CN111598225A (en) | 2020-08-28 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN111598225B (en) | Air conditioner cold load prediction method based on self-adaptive deep confidence network | |
CN110705743B (en) | New energy consumption electric quantity prediction method based on long-term and short-term memory neural network | |
US11409347B2 (en) | Method, system and storage medium for predicting power load probability density based on deep learning | |
CN110458443B (en) | Smart home energy management method and system based on deep reinforcement learning | |
CN112733462A (en) | Ultra-short-term wind power plant power prediction method combining meteorological factors | |
CN111561732B (en) | Heat exchange station heat supply adjusting method and system based on artificial intelligence | |
CN110490385A (en) | The unified prediction of electric load and thermic load in a kind of integrated energy system | |
CN110705760A (en) | Photovoltaic power generation power prediction method based on deep belief network | |
CN109146121A (en) | The power predicating method stopped in the case of limited production based on PSO-BP model | |
CN112330487B (en) | Photovoltaic power generation short-term power prediction method | |
CN111580382A (en) | Unit-level heat supply adjusting method and system based on artificial intelligence | |
CN115965132A (en) | Distributed photovoltaic digital twin system power prediction method based on GA-BP neural network | |
CN111461466A (en) | Heating household valve adjusting method, system and equipment based on L STM time sequence | |
Li et al. | A wind power prediction method based on DE-BP neural network | |
CN113591957B (en) | Wind power output short-term rolling prediction and correction method based on LSTM and Markov chain | |
CN114298140A (en) | Wind power short-term power prediction correction method considering unit classification | |
CN115907195A (en) | Photovoltaic power generation power prediction method, system, electronic device and medium | |
CN113962454A (en) | LSTM energy consumption prediction method based on dual feature selection and particle swarm optimization | |
CN113112085A (en) | New energy station power generation load prediction method based on BP neural network | |
CN114861555A (en) | Regional comprehensive energy system short-term load prediction method based on Copula theory | |
Xu et al. | Short-term electricity consumption forecasting method for residential users based on cluster classification and backpropagation neural network | |
CN114611823B (en) | Optimized dispatching method and system for electricity-cold-heat-gas multi-energy-demand typical park | |
CN117592657A (en) | Load distribution optimization method and system for multiple water chilling units | |
CN117313351A (en) | Safe data center optimized cold prediction method and system | |
CN114234392B (en) | Air conditioner load fine prediction method based on improved PSO-LSTM |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant | ||
TR01 | Transfer of patent right |
Effective date of registration: 20241029 Address after: 518000 1002, Building A, Zhiyun Industrial Park, No. 13, Huaxing Road, Henglang Community, Longhua District, Shenzhen, Guangdong Province Patentee after: Shenzhen Wanzhida Technology Co.,Ltd. Country or region after: China Address before: 710055 Yanta Road 13, Xi'an City, Shaanxi Province Patentee before: XIAN University OF ARCHITECTURE AND TECHNOLOG Country or region before: China |