CN111209168A - Log sequence anomaly detection framework based on nLSTM-self attention - Google Patents
Log sequence anomaly detection framework based on nLSTM-self attention Download PDFInfo
- Publication number
- CN111209168A CN111209168A CN202010037427.4A CN202010037427A CN111209168A CN 111209168 A CN111209168 A CN 111209168A CN 202010037427 A CN202010037427 A CN 202010037427A CN 111209168 A CN111209168 A CN 111209168A
- Authority
- CN
- China
- Prior art keywords
- log
- sequence
- layer
- self
- anomaly detection
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 238000001514 detection method Methods 0.000 title claims abstract description 110
- 230000002159 abnormal effect Effects 0.000 claims abstract description 40
- 238000012549 training Methods 0.000 claims abstract description 39
- 238000000034 method Methods 0.000 claims abstract description 28
- 239000013598 vector Substances 0.000 claims abstract description 20
- 210000004027 cell Anatomy 0.000 claims description 20
- 238000004364 calculation method Methods 0.000 claims description 15
- 230000006870 function Effects 0.000 claims description 12
- 238000013528 artificial neural network Methods 0.000 claims description 9
- 238000004422 calculation algorithm Methods 0.000 claims description 5
- 238000012545 processing Methods 0.000 claims description 5
- 230000003044 adaptive effect Effects 0.000 claims description 3
- 230000001413 cellular effect Effects 0.000 claims description 3
- 238000011478 gradient descent method Methods 0.000 claims description 3
- 210000002569 neuron Anatomy 0.000 claims description 3
- 238000010606 normalization Methods 0.000 claims description 3
- 230000000644 propagated effect Effects 0.000 claims description 2
- 238000012360 testing method Methods 0.000 description 14
- 230000005856 abnormality Effects 0.000 description 13
- 238000002474 experimental method Methods 0.000 description 8
- 238000004458 analytical method Methods 0.000 description 7
- 230000000694 effects Effects 0.000 description 7
- 238000010586 diagram Methods 0.000 description 6
- 238000011156 evaluation Methods 0.000 description 5
- 238000010801 machine learning Methods 0.000 description 5
- 230000008569 process Effects 0.000 description 4
- 230000002547 anomalous effect Effects 0.000 description 3
- 230000001364 causal effect Effects 0.000 description 3
- 238000013135 deep learning Methods 0.000 description 3
- 238000003745 diagnosis Methods 0.000 description 3
- 230000007774 longterm Effects 0.000 description 3
- 238000005065 mining Methods 0.000 description 3
- 230000000306 recurrent effect Effects 0.000 description 3
- 238000013459 approach Methods 0.000 description 2
- 230000008859 change Effects 0.000 description 2
- 238000000605 extraction Methods 0.000 description 2
- 239000011159 matrix material Substances 0.000 description 2
- 230000007246 mechanism Effects 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 238000003058 natural language processing Methods 0.000 description 2
- 238000005457 optimization Methods 0.000 description 2
- 238000011160 research Methods 0.000 description 2
- 238000013179 statistical model Methods 0.000 description 2
- 206010000117 Abnormal behaviour Diseases 0.000 description 1
- ORILYTVJVMAKLC-UHFFFAOYSA-N Adamantane Natural products C1C(C2)CC3CC1CC2C3 ORILYTVJVMAKLC-UHFFFAOYSA-N 0.000 description 1
- 101100285899 Saccharomyces cerevisiae (strain ATCC 204508 / S288c) SSE2 gene Proteins 0.000 description 1
- FMYKJLXRRQTBOR-BZSNNMDCSA-N acetylleucyl-leucyl-norleucinal Chemical compound CCCC[C@@H](C=O)NC(=O)[C@H](CC(C)C)NC(=O)[C@H](CC(C)C)NC(C)=O FMYKJLXRRQTBOR-BZSNNMDCSA-N 0.000 description 1
- 230000006399 behavior Effects 0.000 description 1
- 238000007635 classification algorithm Methods 0.000 description 1
- 230000001427 coherent effect Effects 0.000 description 1
- 238000010219 correlation analysis Methods 0.000 description 1
- 230000007423 decrease Effects 0.000 description 1
- 230000003247 decreasing effect Effects 0.000 description 1
- 238000009795 derivation Methods 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 230000009977 dual effect Effects 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 239000000284 extract Substances 0.000 description 1
- 230000006872 improvement Effects 0.000 description 1
- 238000007689 inspection Methods 0.000 description 1
- 230000010354 integration Effects 0.000 description 1
- 238000012423 maintenance Methods 0.000 description 1
- 238000003062 neural network model Methods 0.000 description 1
- 230000009467 reduction Effects 0.000 description 1
- 230000035945 sensitivity Effects 0.000 description 1
- 230000007704 transition Effects 0.000 description 1
- 238000013519 translation Methods 0.000 description 1
- 238000012795 verification Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F11/00—Error detection; Error correction; Monitoring
- G06F11/30—Monitoring
- G06F11/3065—Monitoring arrangements determined by the means or processing involved in reporting the monitored data
- G06F11/3072—Monitoring arrangements determined by the means or processing involved in reporting the monitored data where the reporting involves data filtering, e.g. pattern matching, time or event triggered, adaptive or policy-based reporting
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
- G06N3/084—Backpropagation, e.g. using gradient descent
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- General Health & Medical Sciences (AREA)
- Computing Systems (AREA)
- Computational Linguistics (AREA)
- Data Mining & Analysis (AREA)
- Evolutionary Computation (AREA)
- Biomedical Technology (AREA)
- Molecular Biology (AREA)
- Biophysics (AREA)
- Artificial Intelligence (AREA)
- Life Sciences & Earth Sciences (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Health & Medical Sciences (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Quality & Reliability (AREA)
- Debugging And Monitoring (AREA)
Abstract
The invention relates to a log sequence anomaly detection framework based on nLSTM-self attribute, which comprises the following steps: training a model and an anomaly detection model; the training model comprises the following steps: suppose that one log file contains k log templates E ═ { E ═ E1,e2L ekThe input of the training model is a sequence of log templates, a log sequence l with the length of ht‑h,…lt‑2,lt‑1The log template l contained iniE, t-h is more than or equal to i is less than or equal to t-1, and the number | l of log templates in a sequencet‑h,...lt‑2,lt‑1I is equal to or less than h, each log template corresponds to a template number, a log template dictionary is generated, then a normal log template sequence is generated into an input sequence and target data are fed into an abnormity detection model for training(ii) a The detection stage comprises: the data input method is the same as the training stage, the model generated in the training stage is used for carrying out anomaly detection, and the model output is a probability vector P ═ (P)1,p2L pk),piRepresenting the target Log template as eiIf the actual target data is in the predicted value, the log sequence is judged to be normal, otherwise, the log sequence is judged to be abnormal.
Description
Technical Field
The invention relates to a network security technology, in particular to a log sequence anomaly detection framework based on nLSTM-self attention.
Background
The network environment is increasingly complex, attacks aiming at network application and systems are continuously emerging, and the attack is often combined and applied by various attack means, so that the existing anomaly detection method is not suitable for novel attacks any more. Once the attack is successful or the network application itself is abnormal, immeasurable loss is brought to the owner and the user of the application. The earlier attacks and false detections, the less loss will be incurred.
The network, the system and the application program can generate logs in the running process for recording the running state and important events, so that the logs contain extremely rich dynamic information, and log analysis is important for maintaining various tasks. These tasks include security tasks such as intrusion detection, internal threat detection and malware detection, and more common maintenance tasks such as detecting hardware failures and the like. By analyzing the log, we can detect abnormal behavior and mine potential security threats.
In recent years, the research of using logs as an abnormality detection data source has received more and more attention, and most of the earliest log abnormality detection methods are manual operation and rule-based methods, but as online service systems become larger, the existing methods are obviously inefficient and require manual inspection of a huge number of logs. With the development of machine learning, many researches adopt feature engineering, and various clustering methods are used for discovering abnormal points or abnormal sequence patterns for abnormality detection. Liuzhaoli et al propose an integration method using K-prototype clustering and a K-NN classification algorithm that analyzes the characteristics of system logs and applies the K-prototype clustering algorithm to divide a data set into different clusters. Obvious normal events, which typically appear as highly coherent clusters, are then filtered out, and other events are considered as anomalous candidates for further analysis. However, the clustering-based analysis method is based on the assumption that the abnormal log is an event which happens in the log file, and this assumption is not always true. He S summarizes and compares several more advanced machine learning methods for log anomaly detection in recent years. The existing machine learning algorithm firstly analyzes an original log into log templates, groups the log templates according to different grouping strategies, wherein each group corresponds to one line, and then extracts the characteristics of the grouped sequences. The feature extraction method only considers the occurrence frequency and does not consider the time sequence relation of the log, only can finish the abnormal detection, but cannot find the abnormal position, which is very unfavorable for the positioning and diagnosis of the abnormality in the later period. The clustering method has a precondition assumption that the abnormal log is a log which happens in the log. This assumption is reasonable in most cases, but there are exceptions. Because the log which occurs by chance is not necessarily abnormal, and there is a log in which abnormal information is mainly recorded according to the setting of a specific application system. In this case, the abnormality log is not a log which occurs by chance. A large number of abnormal detection log mining methods are designed for different applications, Beehive and the like perform unsupervised clustering by researching the log characteristics of network equipment and utilizing the characteristics to identify potential security threats and then manually mark abnormal logs; oprea et al use belief propagation to detect early enterprise threats to DNS logs; the PerfAugur et al system discovers anomalies in system performance by mining the characteristics of the server log. Bovenzi et al propose an operating system level anomaly detection method that is very effective for mission critical systems. Venkatakrishnan et al propose a diversity-based approach to security anomaly detection to prevent system attacks. Zhonchang et al propose a new method for analyzing DNS query behavior, respectively map the queried domain name and the host requesting query to vector space by using deep learning mechanism, apply correlation analysis and clustering, find abnormal problems in network such as botnet, etc. This method, while accurate, is only suitable for detecting anomalies in certain types of logs, is limited to specific scenario applications and requires domain experts.
The log is sequence data, and the appearance sequence of the log has a certain dependence relationship, and sometimes the dependence length is longer. For example, some new attacks are implemented, and they do not immediately damage, but reach some prerequisites, or do some normal operations before damage occurs, which is reflected in the relationship of long-term dependence in the log sequence. Routine sequence-based anomaly detection methods are also widely studied. Such methods were earlier mostly implemented based on statistical models or markov models. Although statistical model-based methods are easy to implement, their accuracy can be low if their pre-assumptions do not hold. Among the markov model-based methods, the most basic method is to model a log event sequence by using a markov chain, i.e., estimating the corresponding probability by using the frequency of events and the transition frequency, and then judging whether the test log is abnormal by calculating the occurrence probability of the test log; ye N proposes a detection method based on a first order markov chain. Although the use of higher order markov chains can improve the model description capability, the number of model parameters grows exponentially with the increase of the order, thus requiring more training logs and larger memory; bao L et al treat the traces as sequence data and use a probabilistic suffix tree based approach to organize and distinguish important statistical properties possessed by sequences. The recurrent neural network has better processing capacity for sequence data, in recent years, LSTM has good effect in sequence prediction, and researchers apply the model to a log sequence prediction task. Zhang et al uses clustering technique to generate feature sequence for original log text from multiple log sources, and inputs the feature sequence into LSTM model for hardware and software failure prediction; du Min et al is inspired by NLP to parse the original text of the system log, generate a log template sequence and input LSTM to detect denial of service attacks. One-hot coding input is adopted, in an anomaly detection part, a training and detection framework based on 2-layer stacking LSTM (2 LSTM is used for representing 2-layer stacking LSTM), although the accuracy of some data sets is greatly improved compared with that of a machine learning method, the framework has the problem of insufficient performance characteristic capability of one-hot and insufficient processing capability of LSTM on a long sequence, and does not have a good effect on all data sets. These methods are simple searches for log sequence abnormality detection using LSTM, and further improvement in detection accuracy is required.
Disclosure of Invention
The invention aims to provide a log sequence anomaly detection framework based on nLSTM-self attribute, which is used for solving the problems of the prior art.
The invention relates to a log sequence anomaly detection framework based on nLSTM-self attribute, which is characterized by comprising the following steps: training a model and an anomaly detection model; the training model comprises the following steps: suppose that one log file contains k log templates E ═ { E ═ E1,e2L ekThe input of the training model is a sequence of log templates, a log sequence l with the length of ht-h,…lt-2,lt-1The log template l contained iniE, t-h is more than or equal to i is less than or equal to t-1, and the number | l of log templates in a sequencet-h,...lt-2,lt-1If m is less than or equal to h, each log template corresponds to a template number, a log template dictionary is generated, and then a normal log template sequence is generated into an input sequence and target data are fed into an abnormal detection model for training; the detection stage comprises: the data input method is the same as the training stage, the model generated in the training stage is used for carrying out anomaly detection, and the model output is a probability vector P ═ (P)1,p2Lpk),piRepresenting the target Log template as eiIf the actual target data is in the predicted value, the log sequence is judged to be normal, otherwise, the log sequence is judged to be abnormal.
According to an embodiment of the nLSTM-self attribute-based log sequence anomaly detection framework, a trained loss function is cross entropy, and an adaptive gradient descent method is adopted for optimization of the loss function.
According to an embodiment of the nLSTM-self attribute-based log sequence anomaly detection framework of the present invention, a log file contains a plurality of event types, each event type contains a plurality of logs, the logs belonging to the same event type have a common template, and a log sequence is used as a sequence of events that occur, that is, a sequence of log templates corresponding to an original log sequence. And detecting the abnormity of the log template sequence corresponding to the original log sequence.
According to an embodiment of the nLSTM-self attribute-based log sequence anomaly detection framework of the present invention, the anomaly detection model comprises: the word embedding layer, the n layers of long-time memory neural network layer and the self-attention layer; the word embedding layer takes the log template sequence as input and is used for the front-end input of the anomaly detection framework, and the serial number of each log template in the sequence is mapped into dense word embedding; the nLSTM layer takes the distributed word embedding of each log template obtained by the word embedding layer as input; the self-attention layer firstly calculates the dependency relationship among logs in a sequence, takes the hidden states of all long-time memory neural network units at the top layer as the input of the self-attention layer, performs similarity calculation, and then performs normalization processing as the probability weight of the self-attention value: the weighted summation of the outputs of the n layers of long-time memory neural network layers is the result of the self-attention value.
According to an embodiment of the nLSTM-self attribute-based log sequence anomaly detection framework of the present invention, an LSTM unit comprises: x is the number oftWord embedding representing a Log template, CtIndicating the cellular status of the t-th LSTM unit in the current sequence, htRepresenting the hidden state of the t-th LSTM cell in the current sequence, the sigma module represents a sigmoid function, tanh represents a tanh function,indicating dot product, ⊕ indicating addition, hidden output h of an LSTM cellt:
ft=σ(Wf·[ht-1,xt]+bf) (1)
it=σ(Wi·[ht-1,xt]+bi) (2)
ot=σ(Wo·[ht-1,xt]+bo) (5)
ht=ot*tanh(Ct) (6)
Formula (1) represents a forgetting gate that determines what information to discard from the cell state; equations (2), (3), (4) represent the input gates, equations (2) and (3) determine what new information is deposited in the cell state, and equation (4) represents the new candidate; equations (5) and (6) represent the output gates.
According to an embodiment of the log sequence anomaly detection framework based on nLSTM-self attribute, a word embedding layer specifies dimensions when generating word vectors, the vectors are initialized by small random numbers, and a back propagation algorithm is adopted for training and updating.
According to an embodiment of the nLSTM-self attribute-based log sequence anomaly detection framework of the present invention, the anomaly detection model further comprises: a linear layer for converting the output result from the attention layer into a probability vector P ═ P (P) with dimension k1,p2L pk) The log file comprises k log templates, piThe next log template representing a prediction of the current sequence is ei。
According to an embodiment of the nLSTM-self association-based log sequence anomaly detection framework of the present invention, an output of an LSTM unit includes a cell state and a hidden state, the hidden state and the cell state of a previous LSTM unit are passed to a next LSTM unit, the hidden state is also passed to a stacked upper layer LSTM as an input, each LSTM unit corresponds to a word embedding of a log template in a sequence, and if h is a sequence length, each layer LSTM includes h LSTM units.
In accordance with one embodiment of the nLSTM-self attribute based log sequence anomaly detection framework of the present invention, wherein,
the self-attention layer firstly calculates the dependency relationship between logs in a sequence and expresses the dependency relationship by a similarity score, and the similarity s (h) is calculatedt,hs) Performing dot product calculation to obtain a non-normalized score:
α=s(ht,hs)=Q·QT(7)
the self-attention layer takes the hidden states of all LSTM units at the top layer as input, the size of the input data Q is batch _ size multiplied by the sequence length (h) multiplied by the number of hidden state neurons, the size of α obtained after similarity calculation is batch _ size multiplied by h, and the size represents the dependency relationship between every two logs in the sequence;
the non-normalized scores are then normalized as a probability weight for the self-attention value:
weighted summation of the LSTM outputs is a result of the calculation from attention:
the result of the self-attention value is a tensor of size bay _ size × hidden _ size × h, taking the last column new _ hidden of each bay as the final output from the attention layer [: and, -1 ].
According to an embodiment of the nLSTM-self association-based log sequence anomaly detection framework of the present invention, the nLSTM layer embeds the distributed words of each log template obtained by the embedding layer as the input of the LSTM unit at the bottom layer, the hidden state and the cell state of the last LSTM unit are transferred to the next LSTM unit, the sequence information is sequentially propagated from front to back, the hidden state is also correspondingly transferred to the stacked upper layer LSTM, and the output of each LSTM unit at the top layer is used as the output of the nLSTM layer to participate in the calculation of the self-attention layer.
The invention provides a log anomaly detection framework nLSALLog based on nLSTM-self attention, wherein n represents the number of stacked LSTM layers. The framework avoids the complicated feature extraction step of the existing machine learning method, the semantic vector representation of the log template is input into the multilayer LSTM network by means of the strong automatic feature learning capability of deep learning, the state vector of the hidden layer and the output of the multilayer LSTM are used as the input of the self-attention layer, and therefore the context information of the sequence can be better reserved and controlled, and the problem of long-term dependence of the sequence is better solved. The model obtained by normal data training can detect unknown abnormality, and the abnormality detection of the log sequence can be used for positioning the abnormal position, which has great significance for the later abnormality diagnosis.
Drawings
FIG. 1 is a schematic diagram of a log sequence anomaly detection framework based on nLSTM-self attention;
FIG. 2 is a schematic diagram of an anomaly detection model;
FIG. 3 shows a detailed view of the interior of an LSTM cell;
fig. 4a-d are schematic diagrams illustrating anomaly detection and evaluation under parameter settings in 4 models.
Detailed Description
In order to make the objects, contents, and advantages of the present invention clearer, the following detailed description of the embodiments of the present invention will be made in conjunction with the accompanying drawings and examples.
FIG. 1 is a schematic diagram of a log sequence anomaly detection framework based on nLSTM-self attribute, and as shown in FIG. 1, the log sequence anomaly detection framework includes: a training phase and a detection phase. Wherein the training phase comprises: suppose that one log file contains k log templates E ═ { E ═ E1,e2L ekOf the training phaseThe input is a sequence of log templates, a log sequence l of length ht-h,…lt-2,lt-1The log template l contained iniE, t-h is more than or equal to i is less than or equal to t-1, and the number | l of log templates in a sequencet-h,...lt-2,lt-1And h is less than or equal to m. In order to facilitate data processing, each log template corresponds to a template number, a log template dictionary is generated, and then a normal log template sequence is generated into an input sequence and target data is fed into an anomaly detection model for training. The trained loss function is cross entropy, and the optimization of the loss function is by an adaptive gradient descent method Adam. The detection stage comprises: the data input method is the same as the training stage, the model generated in the training stage is used for carrying out anomaly detection, and the model output is a probability vector P ═ (P)1,p2L pk),piRepresenting the target Log template as eiThe probability of (2) is actually understood as a multi-class problem, but the final result is a normal and abnormal two-class problem, so that further judgment is needed. According to experience, especially in the case of a small number of log templates, the target log template of one input sequence is more than one, and the log templates corresponding to the first g large probability values in P are considered to be normal. "is the target within the predicted value? The "predicted value" in "is the first g log templates with higher probability. If the actual target data is in the predicted value, the log sequence is judged to be normal, otherwise, the log sequence is judged to be abnormal. The log file comprises a plurality of event types, each event type comprises a plurality of logs, the logs belonging to the same event type have a common template, and a log sequence can be understood as a series of events, namely a sequence of the log template corresponding to an original log sequence. The invention detects the abnormity of the log template sequence corresponding to the original log sequence.
Fig. 2 is a schematic diagram of an anomaly detection model, and as shown in fig. 1 and fig. 2, the anomaly detection model in the training phase includes 3 layers: word embedding layer, nLSTM layer, self attention layer. The word embedding layer takes the log template sequence as input and is used for the front-end input of the anomaly detection framework, and the sequence number of each log template in the sequence is mapped into dense word embedding. The word embedding layer is used as a part of an anomaly detection model, the dimension needs to be specified when a word vector is generated, the vector is initialized by a small random number, and training and updating are performed by adopting a back propagation algorithm. Compared with the open source pre-training packages Word2Vec and GloVe, the Word embedding layer is a slower method, but can customize Word embedding for a specific log data set through model training. Word embedding based on neural network training contains rich context information, semantic rules of target words in the current log sequence can be well expressed, and the purpose of dimension reduction is achieved. The nLSTM layer takes distributed word embedding of each log template obtained by the word embedding layer as input, taking 2-layer LSTM as an example, the output of one LSTM unit includes a cell state and a hidden state, the hidden state and the cell state of the previous LSTM unit are transferred to the next LSTM unit, and the hidden state is also correspondingly transferred to the stacked upper layer LSTM as input. Each LSTM unit at the bottom level corresponds to a word embedding of a log template in the sequence. If h is the sequence length, each layer of LSTM contains h LSTM units.
FIG. 3 shows a detailed view of the interior of an LSTM cell, shown in FIG. 3, xtWord embedding representing a Log template, CtIndicating the cellular status of the t-th LSTM unit in the current sequence, htRepresenting the hidden state of the t-th LSTM cell in the current sequence, the sigma module represents a sigmoid function, tanh represents a tanh function,representing dot product, ⊕ representing addition hidden output h of an LSTM celltThe derivation process of (a) is shown in equations (1) to (6).
ft=σ(Wf·[ht-1,xt]+bf) (1)
it=σ(Wi·[ht-1,xt]+bi) (2)
ot=σ(Wo·[ht-1,xt]+bo)(5)
ht=ot*tanh(Ct) (6)
Formula (1) represents a forgetting gate that determines what information to discard from the cell state; formulas (2), (3) and (4) represent the input gates, (2) (3) determine what new information is stored in the cell state, and (4) represent new candidate values; (5) (6) represents the output gate, but this output will be based on the current filtered cell state. For the stacked LSTM case, this output serves as the input to the next-level LSTM unit.
The self-attention layer can learn the internal structure of a sequence, and has many successful applications in reading understanding, text classification, machine translation and the like. The invention uses this idea for the predictive analysis of log sequences. Firstly, calculating the dependency relationship between logs in a sequence, and expressing the dependency relationship by a similarity score, namely, the similarity s (h)t,hs) A non-normalized score is obtained by performing a dot product calculation:
α=s(ht,hs)=Q·QT(7)
the self-attention layer takes the hidden states of all the LSTM units at the top layer as input, the size of the input data Q is batch _ size multiplied by the sequence length (h) and the number of hidden state neurons (hidden _ size), the size of α obtained after similarity calculation is batch _ size multiplied by h, and the size represents the dependency relationship between every two logs in the sequence.
Softmax normalization was then performed on the non-normalized score as a probability weight for the attention value:
the weighted summation of the LSTM outputs is the result of calculating the attention value:
the result of the attention value is a tensor of size bay _ size × hidden _ size × h, taking the last column new _ hidden of each bay as the final output from the attention layer [: and, -1 ]. Since self-attention is the amount of attention that is calculated for each and every word, the maximum path length is only 1 regardless of how long there is between them. Long range dependencies can be captured.
Adding a linear layer to the last layer of the detection model, and converting into a probability vector P (P) with dimension k1,p2Lpk) (the log file contains k log templates), piThe next log template representing a prediction of the current sequence is ei。
The experimental platform adopts Ubuntu 16.04LT, a 64-bit system, 62.8GiB internal memory, a processorIntel Xeon (R) CPU E5-2620v4@2.1GHz x 16 processor and a Graphic Geforce GTX 1080Ti/PCIe/SSE2 dual GPU platform. The log dataset HDFS dataset contains 11,175,629 raw log records for 38.7 hours, is a log dataset collected from a cluster of 203 nodes on Amazon EC2 platform, is 1.6G in size, and has 28 templates. For the HDFS log, the block _ id identifier is used as a basic operation unit, and the log contained in each block _ id is regarded as a time domain window. Firstly, dividing the log into 575059 time domain windows according to block _ id, and generating a log template sequence corresponding to an original log sequence contained in each time domain window. Each time domain is equivalent to a sentence and each log template inside is equivalent to a word. If one word is abnormal, the sentence is considered to be abnormal. And selecting a training data set which is the same as the Deeplog, wherein all the training data sets are normal data, and the sample size accounts for less than 1% of all the data sets. Data set BlueGene/L (BGL), a public, partially labeled data set from IBM's well-known high Performance computing laboratory (Lawrence Livermore National Labs, LLNL), contains 4,747,963 raw logs of 215 days, and is 708M in size. The BGL anomaly detection is that a template sequence corresponding to an original log is divided by adopting a sliding window, logs contained in 6 or 8 hours are taken as one line, and the length of some logs is hundreds of thousands, and the number of the logs is less than 10, so that the judgment basis of' one log in one line is abnormal and one sentence is not reasonable under the division. In the invention, the BGL data set is not divided according to a fixed time window, but divided and predicted according to the sequence length specified in the model parameters, and a sliding window method is utilized to take down one sequence. 80% of the data set was selected as the training set. The information for both data sets is shown in table 1. The last two columns in the table "number of abnormal/normal (windows)" are the number of abnormal/normal logs for BGL and the number of abnormal/normal session windows for HDFS.
TABLE 1 data set information
The evaluation criteria included: for the problem of abnormality detection, the past literature often adopts Precision, Recall and F1 values for judging the problem, but these indexes do not consider the problem that most log data sets are unbalanced data sets, that is, more normal logs and fewer abnormal logs are included. The detection performance of the nLSTM-self-orientation-based anomaly detection framework provided by the invention is evaluated mainly by measuring the effect of detecting the log sequence, mainly adopting three indexes of TPR, FPR and accuracy, respectively inspecting the detection effect of the real sample condition of the anomalous log sequence and the normal log sequence, and inspecting the overall accuracy, thus being not influenced by the unbalance of the data set. The anomaly log detection confusion matrix is shown in table 2.
In table 2, TP represents the number of abnormal log sequences correctly detected as abnormal; FN indicates the number of abnormal log sequences that were erroneously detected as normal; FP represents the number of normal log sequences that were erroneously detected as anomalous; TN indicates the number of normal log sequences correctly detected as normal. The corresponding several evaluation indexes are as follows.
Table 2 log sequence anomaly detection confusion matrix
(1) The true class rate (TP), also called detection rate, represents the ratio of the number of abnormal log sequences that are correctly detected as abnormal to the total number of abnormal log sequences, with higher values yielding better performance. The calculation formula is as follows:
(2) the False Positive Rate (FPR), also called false positive rate, represents the ratio of the normal log sequence that is erroneously detected as abnormal to the total number of the actual normal log sequence, and the smaller the value, the better the performance. The calculation formula is as follows:
(3) accuracy (Accuracy), which represents the ratio of the number of correct samples to the total number of samples in the test result, the larger the value, the better the performance. The calculation formula is as follows:
the specific process is as follows: respectively loading the normal log set and the abnormal log set in the test set into trained models for testing by taking the same sequence length as that in the training process, and adding 1 to the normal log sequence set and FP if the first g log templates obtained by the current sequence prediction of the models do not contain the actual next log template; for an abnormal log sequence set, TP is incremented by 1. If one of the predictions in a sequence is abnormal and one is abnormal.
The performance analysis includes: and (4) setting experimental parameters, and determining the optimal parameters of the model by repeatedly testing different parameter combinations. And performing comparison experiments by adopting the optimal parameters in subsequent experiments. For the HDFS dataset, the sequence length h is set to 10, the word embedding dimension input _ size is set to 10, the number of hidden layer nodes in LSTM unit hidden layer _ size is set to 64, the number of LSTM layer layers is set to 2, the learning rate lr is set to 0.001, and in the detection phase, g is set to 9 as described in section 3.2. For the BGL dataset, the sequence length h is set to 10, the word embedding dimension input _ size is set to 100, the LSTM unit hidden layer node number hidden _ size is set to 64, the LSTM layer numbers layers is set to 2, the learning rate lr is set to 0.0005, and in the detection phase, g is set to 20.
The analysis of the experimental results comprises: the method is used for respectively carrying out experiments on the BGL data set and the HDFS data set, and comparing the experiments with the 2-layer recurrent neural network RNN and GRU to obtain the 2-layer LSTM which is superior to the task. Furthermore, in order to illustrate the influence of the word embedding layer and the self-attention layer on the detection model, the detection model is divided into three conditions of 2LSTM [9], word embedding layer +2LSTM and word embedding layer +2LSTM + self-attention, and the same parameter setting is adopted for testing respectively. The results of the experiment are shown in tables 3 and 4.
As can be seen from Table 3, the 2LSTM detection model is superior to the 2RNN and 2GRU models, the performance of the detection model is greatly improved after the 2LSTM is added with the word embedding layer, the TPR and the FPR are optimized after the self-attention layer is added, and the total accuracy is improved. The number of abnormal log sequences correctly detected as abnormalities is increased by 33 more, and the number of normal log sequences erroneously detected as abnormalities is decreased by 4.
As can be seen from Table 4, the FPR obtained by detecting the BGL data set by the recurrent neural network models 2RNN, 2GRU and 2LSTM is higher, and the accuracy of the 2LSTM is the highest among the three models. After the 2LSTM is added with the word embedding layer, the FPR is reduced by 57%, the TPR is only reduced by 0.06%, the total accuracy is improved by 9.8%, further, after the attention layer is added, the FPR is reduced by 18.2%, the TPR still has small change, and the total accuracy is improved.
Through experimental verification on two data sets, the word embedding layer +2LSTM + self-attention layer model greatly improves the overall detection performance, particularly, the 2LSTM plus the word embedding layer greatly improves the model detection performance, which shows the effectiveness of word embedding on hidden log mode mining, because a log sequence embodies a causal relationship, the causal relationship is a semantic relationship, semantic features model a log template sequence through word embedding, then the 2LSTM + self-attention layer automatically learns hidden semantic information of a whole sentence, and experimental results also show the effectiveness of the natural language processing thought for reference. The self-attention layer obtains corresponding dependency scores between every two log templates in the current sequence, so that the dependency relationship between every two log templates can be better represented, and the causal relationship in the sequence can be better reflected. Therefore, after the self-attention layer is added, experiments prove that the detection performance of the model is improved.
The data set in table 1 was tested on the corresponding pre-trained model using 2LSTM, word embedding layer +2LSTM and the method of the present invention, respectively, the experimental environment is as described in section 4.1, and the run time is as shown in tables 5 and 6. From the table, it is seen that the test time does not increase with the complexity of the model, but instead the test time decreases by 11% after the input vector dimensions are reduced by word embedding. After the self-attention layer is added, the test time is basically not changed, but the detection performance is improved. The test times for the three models on the BGL data set did not change significantly because the test set was small and the difference in run time was not apparent. Therefore, the detection model of the invention increases the detection effect on the basis of not increasing the time cost.
TABLE 3 comparison of the test results of different test models on HDFS dataset
TABLE 4 comparison of detection Effect of different detection models on BGL dataset
TABLE 5 run time (HDFS)
And (3) analyzing parameter sensitivity: several main parameters: the basis of setting the number of LSTM layers, word vector dimension input _ size, hidden layer node number hidden _ size, and sequence length h, taking BGL dataset as an example, fig. 4 shows an abnormality detection evaluation schematic diagram under different parameter settings, as shown in fig. 4, only the parameters to be evaluated are changed during the experiment, and the other parameters are kept the optimal settings unchanged. The evaluation indexes with data labels in the figure are FPR and Accuracy. On the basis of high accuracy, there is a balanced state between relatively low FPR and relatively high TPR, which is the basis for selecting the optimal parameters. As shown in fig. 4(c), when the word embedding dimension is set to 100, there is a relatively balanced state of FPR and TPR, because when the word embedding dimension is set to 90, although there is a higher TPR, at the expense of a high FPR (the lower the FPR, the better), the overall accuracy is rather reduced.
In order to fully utilize the dependency relationship among log sequences, the invention provides a general log sequence anomaly detection framework based on nLSTM-Self attribute, a model obtained by training normal data can detect unknown anomalies, the model can embed words of a log template into a multi-layer LSTM network by means of the strong automatic learning characteristic capacity of deep learning, and the obtained hidden layer state vector and the output of the multi-layer LSTM are used as the input of a Self-Attention layer, so that the information of all logs in the sequence can be better focused, the long-term dependence problem of the sequence can be better solved, a linear layer is added in the last layer of the detection model and converted into a probability vector, and the prediction of the current sequence is completed. Experimental results show that the model provided by the invention has certain flexibility, can well detect the abnormity in the log, has no increase of the running time due to the complexity of the model, and achieves the best detection effect in the field of log sequence abnormity detection at present. Next, we will continue to develop the study in two ways: the method has the advantages that firstly, the setting of g in the detection stage is further researched, the g is uniformly set, and the better setting method is that the possible states of the next log event of the current sequence can be automatically identified according to different sequences; and secondly, performing abnormity diagnosis on the basis of log abnormity detection, positioning an abnormity position and analyzing an abnormity reason, and providing help for network and system administrators.
The invention has the advantages that: (1) a general log sequence anomaly detection framework (nLSALLog) based on nLSTM-Self Attention is proposed, and a Self-attentive mechanism is first used for log anomaly detection. (2) Theoretical analysis of nLSALLog is provided, and the correctness and the expandability of the nLSALLog are illustrated; (3) the detection performance and the cost of the nLSALLog provided by the invention are verified through experiments, and the basis of parameter setting is discussed.
The above description is only a preferred embodiment of the present invention, and it should be noted that, for those skilled in the art, several modifications and variations can be made without departing from the technical principle of the present invention, and these modifications and variations should also be regarded as the protection scope of the present invention.
Claims (10)
1. An nLSTM-self attribute based log sequence anomaly detection framework, comprising: training a model and an anomaly detection model; the training model comprises the following steps: suppose that one log file contains k log templates E ═ { E ═ E1,e2LekThe input of the training model is a sequence of log templates, a log sequence l with the length of ht-h,…lt-2,lt-1The log template l contained iniE, t-h is more than or equal to i is less than or equal to t-1, and the number | l of log templates in a sequencet-h,...lt-2,lt-1If m is less than or equal to h, each log template corresponds to a template number, a log template dictionary is generated, and then a normal log template sequence is generated into an input sequence and target data are fed into an abnormal detection model for training; the detection stage comprises: the data input method is the same as the training stage, the model generated in the training stage is used for carrying out anomaly detection, and the model output is a probability vector P ═ (P)1,p2L pk),piRepresenting the target Log template as eiIf the actual target data is in the predicted value, the log sequence is judged to be normal, otherwise, the log sequence is judged to be abnormal.
2. The nLSTM-self attribute based log sequence anomaly detection framework of claim 1, wherein the trained loss function is cross entropy and the optimized adoption of the loss function is an adaptive gradient descent method.
3. The nLSTM-self attribute-based log sequence anomaly detection framework of claim 1, wherein a log file comprises a plurality of event types, each event type comprises a plurality of logs, the logs belonging to the same event type have a common template, and a log sequence is used as a sequence of occurring events, namely a sequence of log templates corresponding to an original log sequence. And detecting the abnormity of the log template sequence corresponding to the original log sequence.
4. The nLSTM-self attribute based log sequence anomaly detection framework of claim 1, wherein the anomaly detection model comprises in combination: the word embedding layer, the n layers of long-time memory neural network layer and the self-attention layer; the word embedding layer takes the log template sequence as input and is used for the front-end input of the anomaly detection framework, and the serial number of each log template in the sequence is mapped into dense word embedding; the nLSTM layer takes the distributed word embedding of each log template obtained by the word embedding layer as input; the self-attention layer firstly calculates the dependency relationship among logs in a sequence, takes the hidden states of all long-time memory neural network units at the top layer as the input of the self-attention layer, performs similarity calculation, and then performs normalization processing as the probability weight of the self-attention value: the weighted summation of the outputs of the n layers of long-time memory neural network layers is the result of the self-attention value.
5. The nLSTM-self attention-based log sequence exception detection of claim 1Survey frame, its characterized in that, LSTM unit includes: x is the number oftWord embedding representing a Log template, CtIndicating the cellular status of the t-th LSTM unit in the current sequence, htRepresenting the hidden state of the t-th LSTM cell in the current sequence, the sigma module represents a sigmoid function, tanh represents a tanh function,it is shown that the dot-product,indicating the addition, a hidden output h of an LSTM cellt:
ft=σ(Wf·[ht-1,xt]+bf) (1)
it=σ(Wi·[ht-1,xt]+bi) (2)
ot=σ(Wo·[ht-1,xt]+bo) (5)
ht=ot*tanh(Ct) (6)
Formula (1) represents a forgetting gate that determines what information to discard from the cell state; equations (2), (3), (4) represent the input gates, equations (2) and (3) determine what new information is deposited in the cell state, and equation (4) represents the new candidate; equations (5) and (6) represent the output gates.
6. The nLSTM-self association-based log sequence anomaly detection framework of claim 4, wherein the word embedding layer specifies dimensions when generating word vectors, the vectors are initialized with small random numbers, and a back propagation algorithm is used for training and updating.
7. The nLSTM-self attribute based log sequence anomaly detection framework of claim 4, wherein the anomaly detection model further comprises: a linear layer for converting the output result from the attention layer into a probability vector P ═ P (P) with dimension k1,p2Lpk) The log file comprises k log templates, piThe next log template representing a prediction of the current sequence is ei。
8. The nLSTM-self association-based log sequence anomaly detection framework of claim 1, wherein the output of an LSTM unit comprises cell states and hidden states, the hidden states and cell states of a previous LSTM unit are passed to a next LSTM unit, the hidden states are also passed to a stacked upper layer LSTM, and as its input, each LSTM unit corresponds to a word embedding of a log template in the sequence, and if h is the sequence length, each layer LSTM contains h LSTM units.
9. The nLSTM-self attribute based log sequence anomaly detection framework of claim 4,
the self-attention layer firstly calculates the dependency relationship between logs in a sequence and expresses the dependency relationship by a similarity score, and the similarity s (h) is calculatedt,hs) Performing dot product calculation to obtain a non-normalized score:
α=s(ht,hs)=Q·QT(7)
the self-attention layer takes the hidden states of all LSTM units at the top layer as input, the size of the input data Q is batch _ size multiplied by the sequence length (h) multiplied by the number of hidden state neurons, the size of α obtained after similarity calculation is batch _ size multiplied by h, and the size represents the dependency relationship between every two logs in the sequence;
the non-normalized scores are then normalized as a probability weight for the self-attention value:
weighted summation of the LSTM outputs is a result of the calculation from attention:
the result of the self-attention value is a tensor of size bay _ size × hidden _ size × h, taking the last column new _ hidden of each bay as the final output from the attention layer [: and, -1 ].
10. The nLSTM-self association-based log sequence anomaly detection framework of claim 4, wherein the nLSTM layer embeds the distributed words of each log template obtained by the embedding layer as the input of the LSTM unit at the bottom layer, the hidden state and the cell state of the last LSTM unit are transmitted to the next LSTM unit, the sequence information is sequentially propagated from front to back, the hidden state is also correspondingly transmitted to the upper layer LSTM of the stack, and the output of each LSTM unit at the top layer is used as the output of the nLSTM layer to participate in the calculation of the self-attention layer.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010037427.4A CN111209168A (en) | 2020-01-14 | 2020-01-14 | Log sequence anomaly detection framework based on nLSTM-self attention |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010037427.4A CN111209168A (en) | 2020-01-14 | 2020-01-14 | Log sequence anomaly detection framework based on nLSTM-self attention |
Publications (1)
Publication Number | Publication Date |
---|---|
CN111209168A true CN111209168A (en) | 2020-05-29 |
Family
ID=70786070
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202010037427.4A Pending CN111209168A (en) | 2020-01-14 | 2020-01-14 | Log sequence anomaly detection framework based on nLSTM-self attention |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN111209168A (en) |
Cited By (43)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111782460A (en) * | 2020-06-04 | 2020-10-16 | 昆山伊莱智能软件科技有限公司 | Large-scale log data anomaly detection method and device and storage medium |
CN111797978A (en) * | 2020-07-08 | 2020-10-20 | 北京天融信网络安全技术有限公司 | Internal threat detection method and device, electronic equipment and storage medium |
CN111930597A (en) * | 2020-08-13 | 2020-11-13 | 南开大学 | Log anomaly detection method based on transfer learning |
CN111930592A (en) * | 2020-07-20 | 2020-11-13 | 国网浙江省电力有限公司嘉兴供电公司 | Method and system for detecting log sequence abnormity in real time |
CN111930903A (en) * | 2020-06-30 | 2020-11-13 | 山东师范大学 | System anomaly detection method and system based on deep log sequence analysis |
CN111949480A (en) * | 2020-08-10 | 2020-11-17 | 重庆大学 | Log anomaly detection method based on component perception |
CN111967508A (en) * | 2020-07-31 | 2020-11-20 | 复旦大学 | Time series abnormal point detection method based on saliency map |
CN111984514A (en) * | 2020-09-02 | 2020-11-24 | 大连大学 | Prophet-bLSTM-DTW-based log anomaly detection method |
CN112015705A (en) * | 2020-06-05 | 2020-12-01 | 浙商银行股份有限公司 | Block chain log monitoring method based on deep learning |
CN112069787A (en) * | 2020-08-27 | 2020-12-11 | 西安交通大学 | Log parameter anomaly detection method based on word embedding |
CN112085866A (en) * | 2020-08-14 | 2020-12-15 | 陕西千山航空电子有限责任公司 | Airplane abnormal state identification method based on flight parameter data |
CN112202726A (en) * | 2020-09-10 | 2021-01-08 | 西安交通大学 | System anomaly detection method based on context sensing |
CN112363896A (en) * | 2020-09-02 | 2021-02-12 | 大连大学 | Log anomaly detection system |
CN112597704A (en) * | 2020-12-24 | 2021-04-02 | 东北大学 | Engine abnormity reason analysis method, system, equipment and medium |
CN112613032A (en) * | 2020-12-15 | 2021-04-06 | 中国科学院信息工程研究所 | Host intrusion detection method and device based on system call sequence |
CN112882899A (en) * | 2021-02-25 | 2021-06-01 | 中国烟草总公司郑州烟草研究院 | Method and device for detecting log abnormity |
CN113111908A (en) * | 2021-03-03 | 2021-07-13 | 长沙理工大学 | BERT (binary inverse transcription) anomaly detection method and equipment based on template sequence or word sequence |
CN113286128A (en) * | 2021-06-11 | 2021-08-20 | 上海兴容信息技术有限公司 | Method and system for detecting target object |
CN113312447A (en) * | 2021-03-10 | 2021-08-27 | 天津大学 | Semi-supervised log anomaly detection method based on probability label estimation |
CN113434357A (en) * | 2021-05-17 | 2021-09-24 | 中国科学院信息工程研究所 | Log abnormity detection method and device based on sequence prediction |
CN113472742A (en) * | 2021-05-28 | 2021-10-01 | 中国科学院信息工程研究所 | Internal threat detection method and device based on gated cyclic unit |
CN113553052A (en) * | 2021-06-09 | 2021-10-26 | 麒麟软件有限公司 | Method for automatically recognizing security-related code submissions using an Attention-coded representation |
CN113704201A (en) * | 2021-09-02 | 2021-11-26 | 国家电网有限公司信息通信分公司 | Log anomaly detection method and device and server |
CN113778733A (en) * | 2021-08-31 | 2021-12-10 | 大连海事大学 | Log sequence anomaly detection method based on multi-scale MASS |
CN113792820A (en) * | 2021-11-15 | 2021-12-14 | 航天宏康智能科技(北京)有限公司 | Countermeasure training method and device for user behavior log anomaly detection model |
CN113918367A (en) * | 2021-09-26 | 2022-01-11 | 南京邮电大学 | Large-scale system log anomaly detection method based on attention mechanism |
CN113988202A (en) * | 2021-11-04 | 2022-01-28 | 季华实验室 | Mechanical arm abnormal vibration detection method based on deep learning |
CN114138973A (en) * | 2021-12-03 | 2022-03-04 | 大连海事大学 | Log sequence anomaly detection method based on contrast countertraining |
CN114401135A (en) * | 2022-01-14 | 2022-04-26 | 国网河北省电力有限公司电力科学研究院 | Internal threat detection method based on LSTM-Attention user and entity behavior analysis technology |
CN114490235A (en) * | 2021-09-01 | 2022-05-13 | 北京云集智造科技有限公司 | Algorithm model for intelligently identifying quantity relation and abnormity of log data |
CN114819925A (en) * | 2022-06-29 | 2022-07-29 | 成都秦川物联网科技股份有限公司 | Industrial Internet of things based on event sequence analysis and prediction and control method thereof |
CN115017015A (en) * | 2022-08-04 | 2022-09-06 | 北京航空航天大学 | Method and system for detecting abnormal behavior of program in edge computing environment |
CN115034286A (en) * | 2022-04-24 | 2022-09-09 | 国家计算机网络与信息安全管理中心 | Abnormal user identification method and device based on adaptive loss function |
CN115277180A (en) * | 2022-07-26 | 2022-11-01 | 电子科技大学 | Block chain log anomaly detection and tracing system |
CN115270125A (en) * | 2022-08-11 | 2022-11-01 | 江苏安超云软件有限公司 | IDS log classification prediction method, device, equipment and storage medium |
CN115604003A (en) * | 2022-10-14 | 2023-01-13 | 浙江工业大学(Cn) | System anomaly detection method based on program log data |
CN115794465A (en) * | 2022-11-10 | 2023-03-14 | 上海鼎茂信息技术有限公司 | Method and system for detecting log abnormity |
CN116232770A (en) * | 2023-05-08 | 2023-06-06 | 中国石油大学(华东) | Enterprise network safety protection system and method based on SDN controller |
CN116349211A (en) * | 2020-09-14 | 2023-06-27 | 华为云计算技术有限公司 | Self-attention based deep learning distributed trace anomaly detection |
CN117493220A (en) * | 2024-01-03 | 2024-02-02 | 安徽思高智能科技有限公司 | RPA flow operation abnormity detection method, device and storage device |
CN117539739A (en) * | 2023-12-11 | 2024-02-09 | 国网河南省电力公司经济技术研究院 | User continuous behavior anomaly monitoring method based on double features |
CN117786564A (en) * | 2023-11-23 | 2024-03-29 | 重庆邮电大学 | Abnormal electricity consumption intelligent detection method |
CN117938555A (en) * | 2024-03-25 | 2024-04-26 | 衢州海易科技有限公司 | Log sequence and parameter anomaly detection method and system for cloud platform of Internet of vehicles |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN103514398A (en) * | 2013-10-18 | 2014-01-15 | 中国科学院信息工程研究所 | Real-time online log detection method and system |
WO2019060327A1 (en) * | 2017-09-20 | 2019-03-28 | University Of Utah Research Foundation | Online detection of anomalies within a log using machine learning |
CN110381079A (en) * | 2019-07-31 | 2019-10-25 | 福建师范大学 | Network log method for detecting abnormality is carried out in conjunction with GRU and SVDD |
-
2020
- 2020-01-14 CN CN202010037427.4A patent/CN111209168A/en active Pending
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN103514398A (en) * | 2013-10-18 | 2014-01-15 | 中国科学院信息工程研究所 | Real-time online log detection method and system |
WO2019060327A1 (en) * | 2017-09-20 | 2019-03-28 | University Of Utah Research Foundation | Online detection of anomalies within a log using machine learning |
CN110381079A (en) * | 2019-07-31 | 2019-10-25 | 福建师范大学 | Network log method for detecting abnormality is carried out in conjunction with GRU and SVDD |
Non-Patent Citations (2)
Title |
---|
RUIPENG YANG ET AL: "nLSALog: An Anomaly Detection Framework for Log Sequence in Security Management", 《IEEE ACCESS》 * |
杨瑞朋等: "基于改进时间卷积网络的日志序列异常检测", 《计算机工程》 * |
Cited By (63)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111782460A (en) * | 2020-06-04 | 2020-10-16 | 昆山伊莱智能软件科技有限公司 | Large-scale log data anomaly detection method and device and storage medium |
CN112015705A (en) * | 2020-06-05 | 2020-12-01 | 浙商银行股份有限公司 | Block chain log monitoring method based on deep learning |
CN111930903A (en) * | 2020-06-30 | 2020-11-13 | 山东师范大学 | System anomaly detection method and system based on deep log sequence analysis |
CN111797978A (en) * | 2020-07-08 | 2020-10-20 | 北京天融信网络安全技术有限公司 | Internal threat detection method and device, electronic equipment and storage medium |
CN111930592A (en) * | 2020-07-20 | 2020-11-13 | 国网浙江省电力有限公司嘉兴供电公司 | Method and system for detecting log sequence abnormity in real time |
CN111967508A (en) * | 2020-07-31 | 2020-11-20 | 复旦大学 | Time series abnormal point detection method based on saliency map |
CN111949480A (en) * | 2020-08-10 | 2020-11-17 | 重庆大学 | Log anomaly detection method based on component perception |
CN111949480B (en) * | 2020-08-10 | 2023-08-11 | 重庆大学 | Log anomaly detection method based on component perception |
CN111930597A (en) * | 2020-08-13 | 2020-11-13 | 南开大学 | Log anomaly detection method based on transfer learning |
CN111930597B (en) * | 2020-08-13 | 2023-09-22 | 南开大学 | Log abnormality detection method based on transfer learning |
CN112085866A (en) * | 2020-08-14 | 2020-12-15 | 陕西千山航空电子有限责任公司 | Airplane abnormal state identification method based on flight parameter data |
CN112069787A (en) * | 2020-08-27 | 2020-12-11 | 西安交通大学 | Log parameter anomaly detection method based on word embedding |
CN112363896B (en) * | 2020-09-02 | 2023-12-05 | 大连大学 | Log abnormality detection system |
CN111984514B (en) * | 2020-09-02 | 2023-05-23 | 大连大学 | Log anomaly detection method based on Prophet-bLSTM-DTW |
CN112363896A (en) * | 2020-09-02 | 2021-02-12 | 大连大学 | Log anomaly detection system |
CN111984514A (en) * | 2020-09-02 | 2020-11-24 | 大连大学 | Prophet-bLSTM-DTW-based log anomaly detection method |
CN112202726A (en) * | 2020-09-10 | 2021-01-08 | 西安交通大学 | System anomaly detection method based on context sensing |
CN116349211A (en) * | 2020-09-14 | 2023-06-27 | 华为云计算技术有限公司 | Self-attention based deep learning distributed trace anomaly detection |
CN112613032B (en) * | 2020-12-15 | 2024-03-26 | 中国科学院信息工程研究所 | Host intrusion detection method and device based on system call sequence |
CN112613032A (en) * | 2020-12-15 | 2021-04-06 | 中国科学院信息工程研究所 | Host intrusion detection method and device based on system call sequence |
CN112597704B (en) * | 2020-12-24 | 2024-02-06 | 东北大学 | Engine abnormality cause analysis method, system, equipment and medium |
CN112597704A (en) * | 2020-12-24 | 2021-04-02 | 东北大学 | Engine abnormity reason analysis method, system, equipment and medium |
CN112882899A (en) * | 2021-02-25 | 2021-06-01 | 中国烟草总公司郑州烟草研究院 | Method and device for detecting log abnormity |
CN113111908A (en) * | 2021-03-03 | 2021-07-13 | 长沙理工大学 | BERT (binary inverse transcription) anomaly detection method and equipment based on template sequence or word sequence |
CN113312447A (en) * | 2021-03-10 | 2021-08-27 | 天津大学 | Semi-supervised log anomaly detection method based on probability label estimation |
CN113312447B (en) * | 2021-03-10 | 2022-07-12 | 天津大学 | Semi-supervised log anomaly detection method based on probability label estimation |
CN113434357B (en) * | 2021-05-17 | 2023-04-11 | 中国科学院信息工程研究所 | Log anomaly detection method and device based on sequence prediction |
CN113434357A (en) * | 2021-05-17 | 2021-09-24 | 中国科学院信息工程研究所 | Log abnormity detection method and device based on sequence prediction |
CN113472742B (en) * | 2021-05-28 | 2022-09-27 | 中国科学院信息工程研究所 | Internal threat detection method and device based on gated cyclic unit |
CN113472742A (en) * | 2021-05-28 | 2021-10-01 | 中国科学院信息工程研究所 | Internal threat detection method and device based on gated cyclic unit |
CN113553052A (en) * | 2021-06-09 | 2021-10-26 | 麒麟软件有限公司 | Method for automatically recognizing security-related code submissions using an Attention-coded representation |
CN113286128A (en) * | 2021-06-11 | 2021-08-20 | 上海兴容信息技术有限公司 | Method and system for detecting target object |
CN113778733A (en) * | 2021-08-31 | 2021-12-10 | 大连海事大学 | Log sequence anomaly detection method based on multi-scale MASS |
CN113778733B (en) * | 2021-08-31 | 2024-03-15 | 大连海事大学 | Log sequence anomaly detection method based on multi-scale MASS |
CN114490235A (en) * | 2021-09-01 | 2022-05-13 | 北京云集智造科技有限公司 | Algorithm model for intelligently identifying quantity relation and abnormity of log data |
CN113704201A (en) * | 2021-09-02 | 2021-11-26 | 国家电网有限公司信息通信分公司 | Log anomaly detection method and device and server |
CN113918367B (en) * | 2021-09-26 | 2024-11-08 | 南京邮电大学 | Large-scale system log abnormality detection method based on attention mechanism |
CN113918367A (en) * | 2021-09-26 | 2022-01-11 | 南京邮电大学 | Large-scale system log anomaly detection method based on attention mechanism |
CN113988202A (en) * | 2021-11-04 | 2022-01-28 | 季华实验室 | Mechanical arm abnormal vibration detection method based on deep learning |
CN113792820A (en) * | 2021-11-15 | 2021-12-14 | 航天宏康智能科技(北京)有限公司 | Countermeasure training method and device for user behavior log anomaly detection model |
CN114138973B (en) * | 2021-12-03 | 2024-07-16 | 大连海事大学 | Log sequence anomaly detection method based on contrast countermeasure training |
CN114138973A (en) * | 2021-12-03 | 2022-03-04 | 大连海事大学 | Log sequence anomaly detection method based on contrast countertraining |
CN114401135A (en) * | 2022-01-14 | 2022-04-26 | 国网河北省电力有限公司电力科学研究院 | Internal threat detection method based on LSTM-Attention user and entity behavior analysis technology |
CN115034286A (en) * | 2022-04-24 | 2022-09-09 | 国家计算机网络与信息安全管理中心 | Abnormal user identification method and device based on adaptive loss function |
CN115034286B (en) * | 2022-04-24 | 2024-07-02 | 国家计算机网络与信息安全管理中心 | Abnormal user identification method and device based on self-adaptive loss function |
CN114819925A (en) * | 2022-06-29 | 2022-07-29 | 成都秦川物联网科技股份有限公司 | Industrial Internet of things based on event sequence analysis and prediction and control method thereof |
US11754999B1 (en) | 2022-06-29 | 2023-09-12 | Chengdu Qinchuan Iot Technology Co., Ltd. | Industrial internet of things based on event sequence analysis and prediction, prediction method, and storage medium thereof |
CN114819925B (en) * | 2022-06-29 | 2022-10-11 | 成都秦川物联网科技股份有限公司 | Industrial Internet of things system based on event sequence analysis and prediction and control method thereof |
CN115277180A (en) * | 2022-07-26 | 2022-11-01 | 电子科技大学 | Block chain log anomaly detection and tracing system |
CN115277180B (en) * | 2022-07-26 | 2023-04-28 | 电子科技大学 | Block chain log anomaly detection and tracing system |
CN115017015A (en) * | 2022-08-04 | 2022-09-06 | 北京航空航天大学 | Method and system for detecting abnormal behavior of program in edge computing environment |
CN115017015B (en) * | 2022-08-04 | 2023-01-03 | 北京航空航天大学 | Method and system for detecting abnormal behavior of program in edge computing environment |
CN115270125A (en) * | 2022-08-11 | 2022-11-01 | 江苏安超云软件有限公司 | IDS log classification prediction method, device, equipment and storage medium |
CN115604003B (en) * | 2022-10-14 | 2024-04-05 | 浙江工业大学 | System abnormality detection method based on program log data |
CN115604003A (en) * | 2022-10-14 | 2023-01-13 | 浙江工业大学(Cn) | System anomaly detection method based on program log data |
CN115794465B (en) * | 2022-11-10 | 2023-12-19 | 上海鼎茂信息技术有限公司 | Log abnormality detection method and system |
CN115794465A (en) * | 2022-11-10 | 2023-03-14 | 上海鼎茂信息技术有限公司 | Method and system for detecting log abnormity |
CN116232770A (en) * | 2023-05-08 | 2023-06-06 | 中国石油大学(华东) | Enterprise network safety protection system and method based on SDN controller |
CN117786564A (en) * | 2023-11-23 | 2024-03-29 | 重庆邮电大学 | Abnormal electricity consumption intelligent detection method |
CN117539739A (en) * | 2023-12-11 | 2024-02-09 | 国网河南省电力公司经济技术研究院 | User continuous behavior anomaly monitoring method based on double features |
CN117493220B (en) * | 2024-01-03 | 2024-03-26 | 安徽思高智能科技有限公司 | RPA flow operation abnormity detection method, device and storage device |
CN117493220A (en) * | 2024-01-03 | 2024-02-02 | 安徽思高智能科技有限公司 | RPA flow operation abnormity detection method, device and storage device |
CN117938555A (en) * | 2024-03-25 | 2024-04-26 | 衢州海易科技有限公司 | Log sequence and parameter anomaly detection method and system for cloud platform of Internet of vehicles |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN111209168A (en) | Log sequence anomaly detection framework based on nLSTM-self attention | |
Song et al. | Auditing data provenance in text-generation models | |
US10600005B2 (en) | System for automatic, simultaneous feature selection and hyperparameter tuning for a machine learning model | |
Uwagbole et al. | Applied machine learning predictive analytics to SQL injection attack detection and prevention | |
CN109408389B (en) | Code defect detection method and device based on deep learning | |
Yang et al. | NLSALog: An anomaly detection framework for log sequence in security management | |
CN111600919B (en) | Method and device for constructing intelligent network application protection system model | |
KR102457003B1 (en) | A SYSTEM AND METHOD FOR DETECTING DOMAIN GENERATION ALGORITHMS (DGAs) USING DEEP LEARNING AND SIGNAL PROCESSING TECHNIQUES | |
Mezina et al. | Obfuscated malware detection using dilated convolutional network | |
CN113672931B (en) | Software vulnerability automatic detection method and device based on pre-training | |
Nowotny | Two challenges of correct validation in pattern recognition | |
CN112016097A (en) | Method for predicting time of network security vulnerability being utilized | |
CN109656818B (en) | Fault prediction method for software intensive system | |
Zhou et al. | Discrimination of rock fracture and blast events based on signal complexity and machine learning | |
CN116232673A (en) | Hidden channel identification method, hidden channel identification device, computer equipment and storage medium | |
US20220327394A1 (en) | Learning support apparatus, learning support methods, and computer-readable recording medium | |
US11977633B2 (en) | Augmented machine learning malware detection based on static and dynamic analysis | |
Moskal et al. | Translating intrusion alerts to cyberattack stages using pseudo-active transfer learning (PATRL) | |
CN112613032B (en) | Host intrusion detection method and device based on system call sequence | |
Li et al. | Enhancing robustness of deep neural networks against adversarial malware samples: Principles, framework, and application to AICS’2019 challenge | |
Liang et al. | Automatic security classification based on incremental learning and similarity comparison | |
CN117729003A (en) | Threat information credibility analysis system and method based on machine learning | |
Lighari | Hybrid model of rule based and clustering analysis for big data security | |
Catania et al. | An analysis of convolutional neural networks for detecting DGA | |
Zhang | Clement: Machine learning methods for malware recognition based on semantic behaviours |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
RJ01 | Rejection of invention patent application after publication | ||
RJ01 | Rejection of invention patent application after publication |
Application publication date: 20200529 |