CN111209848B - A real-time fall detection method based on deep learning - Google Patents
A real-time fall detection method based on deep learning Download PDFInfo
- Publication number
- CN111209848B CN111209848B CN202010006573.0A CN202010006573A CN111209848B CN 111209848 B CN111209848 B CN 111209848B CN 202010006573 A CN202010006573 A CN 202010006573A CN 111209848 B CN111209848 B CN 111209848B
- Authority
- CN
- China
- Prior art keywords
- network
- human body
- stage
- detection
- confidence
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
- 238000001514 detection method Methods 0.000 title claims abstract description 47
- 238000013135 deep learning Methods 0.000 title claims abstract description 7
- 238000004422 calculation algorithm Methods 0.000 claims abstract description 19
- 238000000034 method Methods 0.000 claims abstract description 16
- 230000006870 function Effects 0.000 claims description 12
- 238000012549 training Methods 0.000 claims description 12
- 239000013598 vector Substances 0.000 claims description 7
- 230000006399 behavior Effects 0.000 claims description 6
- 208000009144 Pure autonomic failure Diseases 0.000 claims description 5
- 239000013312 porous aromatic framework Substances 0.000 claims description 5
- 238000013528 artificial neural network Methods 0.000 claims description 4
- 238000007635 classification algorithm Methods 0.000 claims description 4
- 230000008569 process Effects 0.000 claims description 4
- 230000001629 suppression Effects 0.000 claims description 4
- 238000004364 calculation method Methods 0.000 claims description 3
- 238000000605 extraction Methods 0.000 claims description 3
- 230000004927 fusion Effects 0.000 claims description 3
- 238000013527 convolutional neural network Methods 0.000 claims description 2
- 238000005457 optimization Methods 0.000 claims description 2
- 238000012546 transfer Methods 0.000 claims description 2
- 238000004458 analytical method Methods 0.000 abstract description 2
- 238000005516 engineering process Methods 0.000 description 6
- 238000012544 monitoring process Methods 0.000 description 6
- 230000008901 benefit Effects 0.000 description 4
- 238000010586 diagram Methods 0.000 description 4
- 230000007613 environmental effect Effects 0.000 description 3
- 230000000474 nursing effect Effects 0.000 description 3
- 238000012545 processing Methods 0.000 description 3
- 238000012706 support-vector machine Methods 0.000 description 3
- 230000002159 abnormal effect Effects 0.000 description 2
- 210000000988 bone and bone Anatomy 0.000 description 2
- 230000007547 defect Effects 0.000 description 2
- 230000036541 health Effects 0.000 description 2
- 239000000463 material Substances 0.000 description 2
- 238000013526 transfer learning Methods 0.000 description 2
- 241001465754 Metazoa Species 0.000 description 1
- 208000027418 Wounds and injury Diseases 0.000 description 1
- 230000001133 acceleration Effects 0.000 description 1
- 230000032683 aging Effects 0.000 description 1
- 230000002547 anomalous effect Effects 0.000 description 1
- 238000013473 artificial intelligence Methods 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 230000006378 damage Effects 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 230000018109 developmental process Effects 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 238000002474 experimental method Methods 0.000 description 1
- 208000014674 injury Diseases 0.000 description 1
- 238000009434 installation Methods 0.000 description 1
- 230000008520 organization Effects 0.000 description 1
- 238000003909 pattern recognition Methods 0.000 description 1
- 239000002699 waste material Substances 0.000 description 1
- 230000003313 weakening effect Effects 0.000 description 1
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/103—Static body considered as a whole, e.g. static pedestrian or occupant recognition
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/214—Generating training patterns; Bootstrap methods, e.g. bagging or boosting
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
- G06F18/241—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/25—Fusion techniques
- G06F18/253—Fusion techniques of extracted features
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/40—Extraction of image or video features
- G06V10/46—Descriptors for shape, contour or point-related descriptors, e.g. scale invariant feature transform [SIFT] or bags of words [BoW]; Salient regional features
- G06V10/462—Salient features, e.g. scale invariant feature transforms [SIFT]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V2201/00—Indexing scheme relating to image or video recognition or understanding
- G06V2201/07—Target detection
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- Data Mining & Analysis (AREA)
- General Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- Evolutionary Computation (AREA)
- Life Sciences & Earth Sciences (AREA)
- Artificial Intelligence (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Evolutionary Biology (AREA)
- Bioinformatics & Computational Biology (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Multimedia (AREA)
- Biomedical Technology (AREA)
- General Health & Medical Sciences (AREA)
- Molecular Biology (AREA)
- Computing Systems (AREA)
- Human Computer Interaction (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Computational Linguistics (AREA)
- Biophysics (AREA)
- Health & Medical Sciences (AREA)
- Image Analysis (AREA)
Abstract
Description
技术领域technical field
本发明涉及一种基于深度学习的实时跌倒检测方法。根据openpose人体姿态估计算法结合SVDD分类算法,设计出一种更加准确高效的实时跌倒检测算法,属于深度学习技术领域、人体姿态估计技术领域、和图像处理技术领域。The invention relates to a real-time fall detection method based on deep learning. According to the openpose human body posture estimation algorithm combined with the SVDD classification algorithm, a more accurate and efficient real-time fall detection algorithm is designed, which belongs to the field of deep learning technology, human body posture estimation technology, and image processing technology.
背景技术Background technique
近年来,随着我国人口老龄化的速度不断加快,老年人的安全和健康已经成为了社会广泛关注的话题。根据世界卫生组织的报告,每年65岁及以上的老年人中,有大约28-35%曾经跌倒,因此,准确并且快速识别跌倒并且通知老年人的监护人或者医护人员是非常有必要的。目前很多养老院都存在着看护人员不足,不能及时发现老年人身体安全出现状况的问题,对于老年人的监护存在着很大的缺陷。In recent years, with the acceleration of my country's population aging, the safety and health of the elderly has become a topic of widespread concern in society. According to the report of the World Health Organization, about 28-35% of the elderly aged 65 and above have fallen every year. Therefore, it is very necessary to accurately and quickly identify falls and notify the guardians or medical staff of the elderly. At present, there is a shortage of nursing staff in many nursing homes, and the problem of the physical safety of the elderly cannot be discovered in time, and there are great defects in the supervision of the elderly.
近年来,研究人员提出了多种实现自动跌倒检测的方法,分别是基于可穿戴式设备的跌倒检测方法、基于环境传感器的跌倒检测方法和基于机器视觉的跌倒检测方法。In recent years, researchers have proposed a variety of methods for automatic fall detection, including wearable device-based fall detection methods, environmental sensor-based fall detection methods, and machine vision-based fall detection methods.
可穿戴式跌倒检测系统检测不受空间的限制,但往往需要被检测者长时间佩戴可穿戴式设备,这些设备通常安装较为繁琐,长时间佩戴使用者会出现厌烦情绪,而且老年人通常记忆力都会有一定程度的减弱从而忘记佩戴可穿戴设备。基于环境传感器的跌倒检测方法具有无需穿戴、检测区域大的优点,而且具有一次安装长期多人使用的优势,可以适用于家庭、养老院等场合。缺点在于对比其他两种方法对于环境设备的要求更高,成本也比较高,同时准确率也存在不稳定的问题。Wearable fall detection system detection is not limited by space, but it often requires the person to be tested to wear wearable devices for a long time. These devices are usually cumbersome to install, and users will feel bored after wearing them for a long time. Moreover, the elderly usually have a certain degree of memory weakening and forget to wear wearable devices. The fall detection method based on the environmental sensor has the advantages of not needing to be worn, having a large detection area, and having the advantage of being used by multiple people for a long time after one installation, and can be applied to occasions such as families and nursing homes. The disadvantage is that compared with the other two methods, the requirements for environmental equipment are higher, the cost is relatively high, and the accuracy rate is also unstable.
随着人工智能、模式识别等技术的发展,越来越的基于机器视觉的跌倒检测方法被提出,监控系统作为公共安全领域的一项重要技术已经得到广泛应用。但是,目前的大多数监控系统仅仅停留在对捕获的视频信号进行人工监控和事后录像分析阶段,需要大量的物力人力。所谓的智能监控系统,要求计算机自动对捕捉到的图像数据进行实时分析和处理,对目标进行检测、识别与跟踪,并对目标的行为进行分析,当目标行为异常时,监控系统会执行报警、保存视频数据等一系列操作。把智能监控系统和跌倒检测结合在一起,可以在老年人摔倒的第一时间通知监护人或者报警,大大缩短了老年人从跌倒到得到救治的时间,减少了二次伤害的可能性,减少后期治疗等费用,对于提高老年人的晚年生活质量,减少社会公共资源的浪费具有重大的意义。With the development of technologies such as artificial intelligence and pattern recognition, more and more fall detection methods based on machine vision have been proposed, and monitoring systems have been widely used as an important technology in the field of public safety. However, most of the current monitoring systems only stay at the stage of manual monitoring of captured video signals and post-event video analysis, which requires a lot of material and manpower. The so-called intelligent monitoring system requires the computer to automatically analyze and process the captured image data in real time, detect, identify and track the target, and analyze the behavior of the target. When the target behaves abnormally, the monitoring system will perform a series of operations such as alarming and saving video data. Combining the intelligent monitoring system with fall detection, the guardian or the police can be notified as soon as the elderly fall, which greatly shortens the time for the elderly from falling to receiving treatment, reduces the possibility of secondary injury, and reduces the cost of later treatment. It is of great significance for improving the quality of life of the elderly in their later years and reducing the waste of social public resources.
发明内容Contents of the invention
本发明要解决的技术问题在于,针对现有的使用传统图像处理进行跌倒检测的技术误检率和实时性上的缺陷,提高检测的速度和准确率。。The technical problem to be solved by the present invention is to improve the speed and accuracy of detection in view of the defects of false detection rate and real-time performance of the existing technology of fall detection using traditional image processing. .
为了解决上述技术问题,本发明提供一种基于深度学习的实时跌倒检测方法。In order to solve the above technical problems, the present invention provides a real-time fall detection method based on deep learning.
本跌倒检测算法主要步骤如下:The main steps of the fall detection algorithm are as follows:
步骤1,使用摄像头等设备采集人体图像;Step 1, using a camera and other equipment to collect human body images;
步骤2,进行人体骨骼关键点识别;Step 2, carry out key point recognition of human skeleton;
步骤2.1,使用神经网络卷积VGG-19提取特征,预测每个关键点的热点图谱,通过多阶段的网络迭代,每一个阶段的输出作为下一个阶段的输入,对上一个阶段的结果优化;Step 2.1, use the neural network convolution VGG-19 to extract features, predict the heat map of each key point, through multi-stage network iteration, the output of each stage is used as the input of the next stage, and optimize the results of the previous stage;
步骤2.2,对于预测的关键点增加矢量编码,对于图像中的关键点进行组合,同一个人的不同部位进行连接;Step 2.2, add vector encoding to the predicted key points, combine the key points in the image, and connect different parts of the same person;
步骤2.3,迁移学习训练图像数据集,使模型可以更好地应用于特定场景;Step 2.3, transfer learning training image data set, so that the model can be better applied to specific scenarios;
步骤3,使用SSD-Mobilenet目标检测算法对检测的人体关键点区域进行目标检测,去掉非人体的部分;Step 3, use the SSD-Mobilenet target detection algorithm to perform target detection on the detected human body key point area, and remove the non-human body part;
步骤3.1,先使用Mobilenet网络对COCO数据集进行预训练,生成预训练模型;Step 3.1, first use the Mobilenet network to pre-train the COCO dataset to generate a pre-training model;
步骤3.2,使用步骤2.3采集的数据生成tfrecord数据用于迁移学习;Step 3.2, using the data collected in step 2.3 to generate tfrecord data for transfer learning;
步骤3.3,步骤3.3把步骤3.2生成的数据和步骤3.1生成的预训练模型融合特征传入SSD网络,使模型经过较短时间的训练就可以收敛并具有较好的检测能力;In step 3.3, step 3.3 transfers the data generated in step 3.2 and the fusion feature of the pre-training model generated in step 3.1 into the SSD network, so that the model can converge after a short period of training and has better detection ability;
步骤4,使用SVDD分类算法对采集到的人体骨骼关键点进行分类。Step 4, use the SVDD classification algorithm to classify the collected key points of human bones.
有益效果Beneficial effect
本方法的优势和创新性主要体现以下方面:The advantages and innovations of this method are mainly reflected in the following aspects:
(1)实现一种全自动的算法流程,无需人工干预或最小程度干预,解决了传统视频监控方法耗费人力物力的缺点。(1) Realize a fully automatic algorithm process without manual intervention or minimal intervention, which solves the shortcomings of traditional video surveillance methods that consume manpower and material resources.
(2)使用openpose和SSD-Mobilenet算法相结合的人体姿态估计算法(2) Human body pose estimation algorithm using the combination of openpose and SSD-Mobilenet algorithm
对于使用openpose算法检测到的关键点,通过SSD-Mobilenet目标检测算法进行人体目标检测,过滤掉非人体的区域,减小算法的误判率。For the key points detected by the openpose algorithm, the SSD-Mobilenet target detection algorithm is used to detect human targets, filter out non-human areas, and reduce the misjudgment rate of the algorithm.
(3)针对跌倒数据样本不平衡的问题,使用SVDD算法进行分类,通过(3) For the problem of unbalanced data samples of falls, use the SVDD algorithm to classify, and pass
网格搜索等方法对参数进行优化。形成正样本训练的正常域,并通过正常域检测跌倒行为。The parameters are optimized by grid search and other methods. Form the normal domain of positive sample training, and detect the fall behavior through the normal domain.
(4)将(2)和(3)结合构建一个鲁棒性较好,准确率较高的实时跌倒检测框架,使用多种目标检测,图像处理等多领域的算法,具有算法应用的创新型。(4) Combining (2) and (3) to build a robust and accurate real-time fall detection framework, using a variety of target detection, image processing and other multi-field algorithms, with innovative algorithm applications.
附图说明Description of drawings
本发明上述的和/或附加的方面和优点从下面结合附图对实施例的描述中将变得明显和容易理解,其中:The above and/or additional aspects and advantages of the present invention will become apparent and easy to understand from the following description of the embodiments in conjunction with the accompanying drawings, wherein:
图1为本发明实时例的跌倒检测整体流程图。FIG. 1 is an overall flow chart of fall detection in a real-time example of the present invention.
图2为本发明实施例的神经网络特征提取部分网络结构图。FIG. 2 is a network structure diagram of the neural network feature extraction part of the embodiment of the present invention.
图3为本发明实施例的人体关键点检测结果图。Fig. 3 is a diagram of detection results of key points of a human body according to an embodiment of the present invention.
图4为本发明实施例的SVDD分类结果图。Fig. 4 is a diagram of the SVDD classification result of the embodiment of the present invention.
图5为本发明实施例的人体跌倒检测结果图。Fig. 5 is a diagram of human fall detection results according to an embodiment of the present invention.
具体实施方式Detailed ways
下面详细描述本发明的实施例,所述实施例的示例在附图中示出,其中自始至终相同或类似的标号表示相同或类似的元件或具有相同或类似功能的元件。下面通过参考附图描述的实施例是示例性的,仅用于解释本发明,而不能解释为对本发明的限制。Embodiments of the present invention are described in detail below, examples of which are shown in the drawings, wherein the same or similar reference numerals designate the same or similar elements or elements having the same or similar functions throughout. The embodiments described below by referring to the figures are exemplary only for explaining the present invention and should not be construed as limiting the present invention.
如图1所示,本发明提出一种基于深度学习的实时跌倒检测算法,具体步骤如下所示:As shown in Figure 1, the present invention proposes a real-time fall detection algorithm based on deep learning, and the specific steps are as follows:
步骤1,使用摄像头等设备采集人体图像;Step 1, using a camera and other equipment to collect human body images;
步骤2,进行人体骨骼关键点识别;Step 2, carry out key point recognition of human skeleton;
步骤2.1,使用神经网络卷积提取特征,预测每个关键点的热点图谱,通过多阶段的学习,每个阶段对上一个阶段的结果优化;Step 2.1, use neural network convolution to extract features, predict the heat map of each key point, and through multi-stage learning, each stage optimizes the results of the previous stage;
如图2所示,本方法使用VGG-19网络进行特征提取特征图fearture-map,由VGG-19的前10层初始化并进行微调,生成一组输入到每个分支第一阶段的特征映射F,在第一阶段,网络产生一组置信图的预测S1=ρ1(F)和一组部分亲和力场其中ρ1和/>是用于阶段1推断的CNN。在随后的阶段,来自前一阶段的两个分支的预测以及原始图像F被用于精确预测。特征F被连接并用于产生精确预测As shown in Figure 2, this method uses the VGG-19 network for feature extraction. The feature map fear-map is initialized and fine-tuned by the first 10 layers of VGG-19 to generate a set of feature maps F that are input to the first stage of each branch. In the first stage, the network generates a set of confidence map predictions S 1 = ρ 1 (F) and a set of partial affinity fields where ρ 1 and /> is the CNN used for stage 1 inference. In subsequent stages, the predictions from the two branches of the previous stage together with the original image F are used for the precise prediction. Features F are concatenated and used to produce accurate predictions
公式中ρt和是阶段t的预测。In the formula, ρ t and is the forecast at stage t.
在网络早期身体部位之间可能存在混淆,但是通过后期网络迭代,人体关键点的预测越来越准确。网络的两个分支有着两个损失函数,每个分支一个,在预测和groudtruth的maps和fields之间使用L2损失防止过拟合的情况出现,对于损失函数加权计算。There may be confusion between body parts in the early stage of the network, but through later network iterations, the prediction of human body key points becomes more and more accurate. The two branches of the network have two loss functions, one for each branch, using L2 loss between the maps and fields of the prediction and groudtruth to prevent overfitting, and weighting the calculation of the loss function.
和/>分别是从1开始一直到C不同阶段t的L2损失函数,f为总体的损失函数。为每个人生成个人的confidence maps,由网络预测出各个置信度的集合。 and /> They are the L2 loss function from 1 to the different stage t of C, and f is the overall loss function. Personal confidence maps are generated for each person, and the network predicts a set of confidence levels.
在网络训练过程中,对于某一类给定的关键点j,对于某一个人k,他的置信度标志图只有一个峰值,置信度标签为每一个点到数据真实点的高斯距离。对于多个人,使用非极大值抑制(NMS)剔除网络中得分低的数据。In the process of network training, for a certain type of given key point j, for a certain person k, his confidence mark map has only one peak, and the confidence mark is the Gaussian distance from each point to the real point of the data. For multiple persons, non-maximum suppression (NMS) is used to cull low-scoring data from the network.
步骤2.2,对于预测的关键点增加矢量编码,对同一个人的不同部位进行连接;给定一组检测到的身体部位,对于每一对身体部位检测关联进行置信度度量,检测关联的一种方式是检测肢体上每对肢体之间的额外中点,并检查候选部位检测之间的发生率,当人体叠加在一起时这些中点可能会提供虚假关联,为了解决这些局限性,使用PAFs特征表示法,它保留了肢体支撑区域的位置和方向信息,PAFs是每个肢体的2D矢量场,对于属于特定肢体的区域中的每个像素,2D矢量编码从肢体的一部分指向另一部分的方向。每种类型的肢体都有一个对应的PAFs区域,用于连接两个相关的身体部位。对检测置信度图进行非极大值抑制(NMS),以获得一组离散的部位候选位置。对于每个部分,可能会有多个候选位置,由于图像中有多个人或误报,这些部分候选位置定义了一大组可能的肢体。使用PAF上的线积分计算来评分每个候选肢体,寻找最佳解析的问题对应于已知为NP-Hard的K维匹配问题,使用匈牙利算法来获得最佳匹配。Step 2.2, add vector encodings for the predicted keypoints, and connect different parts of the same person; given a set of detected body parts, perform a confidence measure for each pair of body part detection associations, one way to detect associations is to detect additional midpoints between each pair of limbs on the limbs, and check the incidence between candidate part detections, these midpoints may provide false associations when the human body is superimposed. A 2D vector field for each limb, for each pixel in the region belonging to a particular limb, the 2D vector encodes the direction from one part of the limb to another. Each type of limb has a corresponding region of PAFs that connect two related body parts. Non-maximum suppression (NMS) is performed on the detection confidence map to obtain a discrete set of part candidate locations. For each part, there may be multiple candidate positions, which define a large set of possible limbs due to multiple people in the image or false positives. Each candidate limb is scored using a line integral computation over the PAF, and the problem of finding the best resolution corresponds to a K-dimensional matching problem known as NP-Hard, using the Hungarian algorithm to obtain the best match.
步骤2.3,为了使模型更加适用于本方法应用场景,采集了1000张人体各个动作的RGB图片,包含人体日常行为和跌倒图像,使用tensorflow框架对采集图片进行训练,随着迭代次数不断增加,损失函数值不算减小,设置迭代次数为5000次,当达到5000时,损失值为0.0925,使用模型对测试集进行测试,准确率为91.2%,证明算法满足使用要求,使用此步骤实施后效果如图3所示。In step 2.3, in order to make the model more suitable for the application scenarios of this method, 1000 RGB pictures of various human actions were collected, including daily human behavior and fall images, and the collected pictures were trained using the tensorflow framework. shown.
步骤3,使用SSD-Mobilenet目标检测模型对检测的人体关键点区域进行目标检测,去掉非人体的部分。Step 3, use the SSD-Mobilenet target detection model to perform target detection on the detected human body key point area, and remove the non-human body part.
步骤3.1首先使用Mobilenet网络对COCO数据集进行预训练,这个数据集包括80类目标,包括人体,动物,交通工具,家用电器等12个大类,得到预训练的参数。Step 3.1 First, use the Mobilenet network to pre-train the COCO data set. This data set includes 80 types of targets, including 12 categories such as human body, animals, vehicles, and household appliances, to obtain pre-trained parameters.
步骤3.2使用步骤2.3中的自建人体数据集,使用LabelImg工具标注人体样本,样本被标注好后,生成一一对应的xml文件,然后调用xml_to_csv.py生成csv表格,接下来使用generate_tfrecord.py文件把csv文件转化为tensorflow框架能识别的tfrecord格式。Step 3.2 Use the self-built human body dataset in step 2.3, and use the LabelImg tool to label human samples. After the samples are labeled, generate a one-to-one corresponding xml file, and then call xml_to_csv.py to generate a csv table, and then use the generate_tfrecord.py file to convert the csv file into a tfrecord format that the tensorflow framework can recognize.
步骤3.3把步骤3.2生成的数据和步骤3.1生成的预训练模型融合特征传入SSD网络,设置迭代次10000次,当损失达到1.8时停止训练,总损失为1.8,平均精度为78.4%。此步骤实施后可以去除冗余的关键点,只保留正确的人体骨骼关键点。Step 3.3 Introduce the fusion features of the data generated in step 3.2 and the pre-trained model generated in step 3.1 to the SSD network, set the number of iterations to 10,000, stop training when the loss reaches 1.8, the total loss is 1.8, and the average accuracy is 78.4%. After this step is implemented, the redundant key points can be removed, and only the correct human skeleton key points can be kept.
步骤4,使用SVDD分类算法对采集到的人体骨骼关键点进行分类。Step 4, use the SVDD classification algorithm to classify the collected key points of human bones.
虽然获得了关节点的轨迹,但仍存在数据不平衡的问题。在日常生活中,由于跌落类型的多样性,获得足够的实际样本下降不仅需要大量的人力和资源,而且难以获得,这导致样本不平衡,因此如何使用不平衡数据来检测跌倒已成为一个基本问题。本方法采用SVDD异常检测方法,通过网格搜索等方法对参数进行优化。形成正样本训练的正常域,并通过正常域检测跌倒行为。SVDD最初主要用于异常检测和图像分类。与分类器支持向量机(SVM)不同,SVDD属于单类SVM算法。主要思想是通过训练阳性样本来构建正常的域超球面,而超球面外部的样本是异常样本。对于给定的阳性样本集X,其包含N个阳性样本(Xi=0,1,2,...,N),以便在构建超球面时减少异常数据点被包括在正常域中的影响,惩罚因子C和松弛变量ξi,当半球R的超球面中心时,让正样本完全被球体包围,相应的优化方程是:Although the trajectories of joint points are obtained, there is still the problem of data imbalance. In daily life, due to the diversity of fall types, obtaining enough actual sample falls not only requires a lot of manpower and resources, but also is difficult to obtain, which leads to sample imbalance, so how to use imbalanced data to detect falls has become a fundamental problem. This method adopts the SVDD anomaly detection method, and optimizes the parameters by grid search and other methods. Form the normal domain of positive sample training, and detect the fall behavior through the normal domain. SVDD was originally mainly used for anomaly detection and image classification. Unlike the classifier Support Vector Machine (SVM), SVDD is a one-class SVM algorithm. The main idea is to construct a normal domain hypersphere by training positive samples, while samples outside the hypersphere are abnormal samples. For a given positive sample set X, which contains N positive samples (Xi=0, 1, 2, ..., N), in order to reduce the impact of abnormal data points being included in the normal domain when constructing a hypersphere, the penalty factor C and the relaxation variable ξ i , when the hypersphere center of the hemisphere R, let the positive samples be completely surrounded by a sphere, the corresponding optimization equation is:
对于典型的二次规划问题,通过引入拉格朗日乘数并求解相应的函数。通常,在删除异常数据点之后,数据将不会以球形方式分布。因此,引入核函数K将低维空间中的非线性问题转化为高维线性问题。图4显示了二维SVDD异常检测图,红点是正常数据,蓝色交叉点是异常数据,决策边界被红点包围。决策中的边界点是支持向量。支持向量确定决策边界的形状和大小,这是模型的准确性。For typical quadratic programming problems, by introducing Lagrangian multipliers and solving the corresponding functions. Typically, after removing outlier data points, the data will not be distributed in a spherical fashion. Therefore, the kernel function K is introduced to transform the nonlinear problem in the low-dimensional space into a high-dimensional linear problem. Figure 4 shows the two-dimensional SVDD anomaly detection map, the red dots are normal data, the blue intersections are anomalous data, and the decision boundary is surrounded by red dots. The boundary points in the decision are support vectors. Support vectors determine the shape and size of the decision boundary, which is the accuracy of the model.
此步骤实施后效果如图5所示。经过实验证明可见本方法的准确性和可靠性。The effect after this step is implemented is shown in Figure 5. The experiment proves the accuracy and reliability of this method.
Claims (1)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010006573.0A CN111209848B (en) | 2020-01-03 | 2020-01-03 | A real-time fall detection method based on deep learning |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010006573.0A CN111209848B (en) | 2020-01-03 | 2020-01-03 | A real-time fall detection method based on deep learning |
Publications (2)
Publication Number | Publication Date |
---|---|
CN111209848A CN111209848A (en) | 2020-05-29 |
CN111209848B true CN111209848B (en) | 2023-07-21 |
Family
ID=70786558
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202010006573.0A Active CN111209848B (en) | 2020-01-03 | 2020-01-03 | A real-time fall detection method based on deep learning |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN111209848B (en) |
Families Citing this family (13)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111814588B (en) * | 2020-06-18 | 2023-08-01 | 浙江大华技术股份有限公司 | Behavior detection method, related equipment and device |
CN111899470B (en) * | 2020-08-26 | 2022-07-22 | 歌尔科技有限公司 | Human body falling detection method, device, equipment and storage medium |
CN112215185B (en) * | 2020-10-21 | 2022-08-05 | 成都信息工程大学 | A system and method for detecting falling behavior from surveillance video |
CN112270381B (en) * | 2020-11-16 | 2022-06-03 | 电子科技大学 | People flow detection method based on deep learning |
CN112633327B (en) * | 2020-12-02 | 2023-06-30 | 西安电子科技大学 | Staged metal surface defect detection method, system, medium, equipment and application |
CN112766091B (en) * | 2021-01-05 | 2023-09-29 | 中科院成都信息技术股份有限公司 | Video unsafe behavior recognition system and method based on human skeleton key points |
CN112861686B (en) * | 2021-02-01 | 2022-08-30 | 内蒙古大学 | SVM-based image target detection method |
CN113177468B (en) * | 2021-04-27 | 2023-10-27 | 北京百度网讯科技有限公司 | Human behavior detection method and device, electronic equipment and storage medium |
CN113368487A (en) * | 2021-06-10 | 2021-09-10 | 福州大学 | OpenPose-based 3D private fitness system and working method thereof |
CN113724853A (en) * | 2021-10-08 | 2021-11-30 | 联想新视界(南昌)人工智能工研院有限公司 | Intelligent medical system based on deep learning |
CN114220162A (en) * | 2021-11-17 | 2022-03-22 | 深圳职业技术学院 | Pig gesture recognition method and device |
CN114913597B (en) * | 2022-05-06 | 2024-08-02 | 山东光汇控股有限公司 | Fall detection method and system based on OpenPose and lightweight neural network |
CN115471826B (en) * | 2022-08-23 | 2024-03-26 | 中国航空油料集团有限公司 | Method and device for judging safe driving behavior of aviation fueller and safe operation and maintenance system |
Family Cites Families (13)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102722721A (en) * | 2012-05-25 | 2012-10-10 | 山东大学 | Human falling detection method based on machine vision |
CN107220604A (en) * | 2017-05-18 | 2017-09-29 | 清华大学深圳研究生院 | A kind of fall detection method based on video |
CN107506706A (en) * | 2017-08-14 | 2017-12-22 | 南京邮电大学 | A kind of tumble detection method for human body based on three-dimensional camera |
CN108182410A (en) * | 2017-12-28 | 2018-06-19 | 南通大学 | A kind of joint objective zone location and the tumble recognizer of depth characteristic study |
CN108805032B (en) * | 2018-05-17 | 2021-06-15 | 郑州大学 | A fall detection method based on deep convolutional network |
CN108960056B (en) * | 2018-05-30 | 2022-06-03 | 西南交通大学 | A Fall Detection Method Based on Attitude Analysis and Support Vector Data Description |
CN108875614A (en) * | 2018-06-07 | 2018-11-23 | 南京邮电大学 | It is a kind of that detection method is fallen down based on deep learning image procossing |
CN108961675A (en) * | 2018-06-14 | 2018-12-07 | 江南大学 | Fall detection method based on convolutional neural networks |
CN109492612B (en) * | 2018-11-28 | 2024-07-02 | 平安科技(深圳)有限公司 | Fall detection method and device based on bone points |
CN109800802A (en) * | 2019-01-10 | 2019-05-24 | 深圳绿米联创科技有限公司 | Visual sensor and object detecting method and device applied to visual sensor |
CN110555368B (en) * | 2019-06-28 | 2022-05-03 | 西安理工大学 | Fall behavior recognition method based on three-dimensional convolutional neural network |
CN110490080B (en) * | 2019-07-22 | 2023-05-09 | 毕昇云(武汉)信息技术有限公司 | An image-based human fall detection method |
CN110633736A (en) * | 2019-08-27 | 2019-12-31 | 电子科技大学 | A human fall detection method based on multi-source heterogeneous data fusion |
-
2020
- 2020-01-03 CN CN202010006573.0A patent/CN111209848B/en active Active
Non-Patent Citations (2)
Title |
---|
Doppler Radar Fall Activity Detection Using the Wavelet Transform;Bo Yu Su 等;IEEE Transactions on Biomedical Engineering;全文 * |
基于特征自动提取的跌倒检测算法;胡双杰 等;传感技术学报;全文 * |
Also Published As
Publication number | Publication date |
---|---|
CN111209848A (en) | 2020-05-29 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN111209848B (en) | A real-time fall detection method based on deep learning | |
CN110781838B (en) | Multi-mode track prediction method for pedestrians in complex scene | |
Chackravarthy et al. | Intelligent crime anomaly detection in smart cities using deep learning | |
CN106022229B (en) | Abnormal Behavior Recognition Method Based on Video Motion Information Feature Extraction and Adaptive Enhancement Algorithm Error Backpropagation Network | |
CN108549841A (en) | A kind of recognition methods of the Falls Among Old People behavior based on deep learning | |
CN112163564B (en) | Tumble prejudging method based on human body key point behavior identification and LSTM (least Square TM) | |
CN113378676A (en) | Method for detecting figure interaction in image based on multi-feature fusion | |
CN108509938A (en) | A kind of fall detection method based on video monitoring | |
CN112434723B (en) | Day/night image classification and object detection method based on attention network | |
CN105139029A (en) | Activity recognition method and activity recognition device for persons serving sentences | |
CN107609477A (en) | It is a kind of that detection method is fallen down with what Intelligent bracelet was combined based on deep learning | |
CN107967941A (en) | A kind of unmanned plane health monitoring method and system based on intelligent vision reconstruct | |
Du et al. | Convolutional neural network-based data anomaly detection considering class imbalance with limited data | |
CN113297972A (en) | Transformer substation equipment defect intelligent analysis method based on data fusion deep learning | |
CN111881802A (en) | Traffic police gesture recognition method based on double-branch space-time graph convolutional network | |
CN115731579A (en) | Individual recognition method of terrestrial animals based on cross-attention Transformer network | |
CN118762332A (en) | Work site safety detection methods, safety detection systems and safety fences | |
Gorodnichev et al. | Research and Development of a System for Determining Abnormal Human Behavior by Video Image Based on Deepstream Technology | |
CN117037264A (en) | Prison personnel abnormal behavior identification method based on target and key point detection | |
CN114638441B (en) | An ocean current monitoring and early warning system based on satellite remote sensing images | |
CN115393802A (en) | A method for identifying uncommon intrusion targets in railway scenes based on small sample learning | |
Rezaee et al. | Intelligent detection of the falls in the elderly using fuzzy inference system and video-based motion estimation method | |
CN111178134B (en) | A Fall Detection Method Based on Deep Learning and Network Compression | |
CN110826459A (en) | Migratable campus violent behavior video identification method based on attitude estimation | |
CN114913585A (en) | A fall detection method for the elderly at home based on facial expressions |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |